From 29ea9fcf2545d4dbea5c51c7d02b955dd57dc350 Mon Sep 17 00:00:00 2001 From: Aiden Cline Date: Wed, 28 Jan 2026 21:55:50 -0600 Subject: [PATCH] fix: ensure variants for copilot models work w/ maxTokens being set --- packages/opencode/src/provider/transform.ts | 4 ++-- packages/opencode/src/session/llm.ts | 13 ------------- 2 files changed, 2 insertions(+), 15 deletions(-) diff --git a/packages/opencode/src/provider/transform.ts b/packages/opencode/src/provider/transform.ts index 39eef6c91..57042d817 100644 --- a/packages/opencode/src/provider/transform.ts +++ b/packages/opencode/src/provider/transform.ts @@ -428,13 +428,13 @@ export namespace ProviderTransform { high: { thinking: { type: "enabled", - budgetTokens: 16000, + budgetTokens: Math.min(16_000, Math.floor(model.limit.output / 2 - 1)), }, }, max: { thinking: { type: "enabled", - budgetTokens: 31999, + budgetTokens: Math.min(31_999, model.limit.output - 1), }, }, } diff --git a/packages/opencode/src/session/llm.ts b/packages/opencode/src/session/llm.ts index 033e4862c..1e409b03f 100644 --- a/packages/opencode/src/session/llm.ts +++ b/packages/opencode/src/session/llm.ts @@ -158,19 +158,6 @@ export namespace LLM { input.model.limit.output, OUTPUT_TOKEN_MAX, ) - log.info("max_output_tokens", { - tokens: ProviderTransform.maxOutputTokens( - input.model.api.npm, - params.options, - input.model.limit.output, - OUTPUT_TOKEN_MAX, - ), - modelOptions: params.options, - outputLimit: input.model.limit.output, - }) - // tokens = 32000 - // outputLimit = 64000 - // modelOptions={"reasoningEffort":"minimal"} const tools = await resolveTools(input)