From 29493bde1aa038b6647df9c35777327a1eddd810 Mon Sep 17 00:00:00 2001 From: Vasanthdev2004 Date: Wed, 1 Apr 2026 21:59:21 +0530 Subject: [PATCH] test: cover gpt-4o max token limits --- src/utils/context.test.ts | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) diff --git a/src/utils/context.test.ts b/src/utils/context.test.ts index 17895b88..16a6c857 100644 --- a/src/utils/context.test.ts +++ b/src/utils/context.test.ts @@ -35,3 +35,22 @@ test('deepseek-chat clamps oversized max output overrides to the provider limit' expect(getMaxOutputTokensForModel('deepseek-chat')).toBe(8_192) }) + +test('gpt-4o uses provider-specific context and output caps', () => { + process.env.CLAUDE_CODE_USE_OPENAI = '1' + delete process.env.CLAUDE_CODE_MAX_OUTPUT_TOKENS + + expect(getContextWindowForModel('gpt-4o')).toBe(128_000) + expect(getModelMaxOutputTokens('gpt-4o')).toEqual({ + default: 16_384, + upperLimit: 16_384, + }) + expect(getMaxOutputTokensForModel('gpt-4o')).toBe(16_384) +}) + +test('gpt-4o clamps oversized max output overrides to the provider limit', () => { + process.env.CLAUDE_CODE_USE_OPENAI = '1' + process.env.CLAUDE_CODE_MAX_OUTPUT_TOKENS = '32000' + + expect(getMaxOutputTokensForModel('gpt-4o')).toBe(16_384) +})