Merge pull request #85 from nusquama/fix/azure-openai-max-completion-tokens

fix: use max_completion_tokens instead of max_tokens for OpenAI-compatible APIs
This commit is contained in:
Kevin Codex
2026-04-02 08:49:23 +08:00
committed by GitHub

View File

@@ -636,9 +636,13 @@ class OpenAIShimMessages {
const body: Record<string, unknown> = {
model: request.resolvedModel,
messages: openaiMessages,
max_tokens: params.max_tokens,
stream: params.stream ?? false,
}
if (params.max_tokens !== undefined) {
body.max_completion_tokens = params.max_tokens
} else if ((params as Record<string, unknown>).max_completion_tokens !== undefined) {
body.max_completion_tokens = (params as Record<string, unknown>).max_completion_tokens
}
if (params.stream) {
body.stream_options = { include_usage: true }