From d98bd4bd523bd5ceb6411143020713c3ee4d7e7c Mon Sep 17 00:00:00 2001 From: Aiden Cline <63023139+rekram1-node@users.noreply.github.com> Date: Tue, 10 Feb 2026 23:30:29 -0600 Subject: [PATCH] fix: add additional context overflow cases, remove overcorrecting ones (#13077) --- packages/opencode/src/provider/error.ts | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/packages/opencode/src/provider/error.ts b/packages/opencode/src/provider/error.ts index 2693df04f..0db03576e 100644 --- a/packages/opencode/src/provider/error.ts +++ b/packages/opencode/src/provider/error.ts @@ -12,15 +12,13 @@ export namespace ProviderError { /input token count.*exceeds the maximum/i, // Google (Gemini) /maximum prompt length is \d+/i, // xAI (Grok) /reduce the length of the messages/i, // Groq - /maximum context length is \d+ tokens/i, // OpenRouter + /maximum context length is \d+ tokens/i, // OpenRouter, DeepSeek /exceeds the limit of \d+/i, // GitHub Copilot /exceeds the available context size/i, // llama.cpp server /greater than the context length/i, // LM Studio /context window exceeds limit/i, // MiniMax - /exceeded model token limit/i, // Kimi For Coding + /exceeded model token limit/i, // Kimi For Coding, Moonshot /context[_ ]length[_ ]exceeded/i, // Generic fallback - /too many tokens/i, // Generic fallback - /token limit exceeded/i, // Generic fallback ] function isOpenAiErrorRetryable(e: APICallError) {