From 39d9616ed7eac55ed2c8424afdded6f5aa36cc88 Mon Sep 17 00:00:00 2001 From: Juan Camilo Date: Wed, 1 Apr 2026 17:03:57 +0200 Subject: [PATCH] fix: update DeepSeek context window from 64k to 128k DeepSeek V3 documentation specifies 128k context window for both deepseek-chat and deepseek-reasoner. The previous 64k value caused premature compaction and underutilization of available context. Relates to #39 Co-Authored-By: Juan Camilo --- src/utils/model/openaiContextWindows.ts | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/src/utils/model/openaiContextWindows.ts b/src/utils/model/openaiContextWindows.ts index fd6fb15a..63ac5f2c 100644 --- a/src/utils/model/openaiContextWindows.ts +++ b/src/utils/model/openaiContextWindows.ts @@ -24,9 +24,9 @@ const OPENAI_CONTEXT_WINDOWS: Record = { 'o4-mini': 200_000, 'o3': 200_000, - // DeepSeek - 'deepseek-chat': 64_000, - 'deepseek-reasoner': 64_000, + // DeepSeek (V3: 128k context per official docs) + 'deepseek-chat': 128_000, + 'deepseek-reasoner': 128_000, // Groq (fast inference) 'llama-3.3-70b-versatile': 128_000,