@@ -160,6 +160,11 @@ export namespace LLM {
160160 ...input . messages ,
161161 ]
162162
163+ const maxOutputTokens =
164+ isOpenaiOauth || provider . id . includes ( "github-copilot" )
165+ ? undefined
166+ : ProviderTransform . maxOutputTokens ( input . model )
167+
163168 const params = await Plugin . trigger (
164169 "chat.params" ,
165170 {
@@ -175,6 +180,7 @@ export namespace LLM {
175180 : undefined ,
176181 topP : input . agent . topP ?? ProviderTransform . topP ( input . model ) ,
177182 topK : ProviderTransform . topK ( input . model ) ,
183+ maxOutputTokens,
178184 options,
179185 } ,
180186 )
@@ -193,11 +199,6 @@ export namespace LLM {
193199 } ,
194200 )
195201
196- const maxOutputTokens =
197- isOpenaiOauth || provider . id . includes ( "github-copilot" )
198- ? undefined
199- : ProviderTransform . maxOutputTokens ( input . model )
200-
201202 const tools = await resolveTools ( input )
202203
203204 // LiteLLM and some Anthropic proxies require the tools parameter to be present
@@ -291,7 +292,7 @@ export namespace LLM {
291292 activeTools : Object . keys ( tools ) . filter ( ( x ) => x !== "invalid" ) ,
292293 tools,
293294 toolChoice : input . toolChoice ,
294- maxOutputTokens,
295+ maxOutputTokens : params . maxOutputTokens ,
295296 abortSignal : input . abort ,
296297 headers : {
297298 ...( input . model . providerID . startsWith ( "opencode" )
0 commit comments