* Add local OpenAI-compatible model discovery to /model * Guard local OpenAI model discovery from Codex routing * Preserve remote OpenAI Codex alias behavior
86 lines
3.1 KiB
TypeScript
86 lines
3.1 KiB
TypeScript
import { afterEach, expect, test } from 'bun:test'
|
|
|
|
import {
|
|
getAdditionalModelOptionsCacheScope,
|
|
isLocalProviderUrl,
|
|
resolveProviderRequest,
|
|
} from './providerConfig.js'
|
|
|
|
const originalEnv = {
|
|
CLAUDE_CODE_USE_OPENAI: process.env.CLAUDE_CODE_USE_OPENAI,
|
|
OPENAI_BASE_URL: process.env.OPENAI_BASE_URL,
|
|
OPENAI_MODEL: process.env.OPENAI_MODEL,
|
|
}
|
|
|
|
afterEach(() => {
|
|
process.env.CLAUDE_CODE_USE_OPENAI = originalEnv.CLAUDE_CODE_USE_OPENAI
|
|
process.env.OPENAI_BASE_URL = originalEnv.OPENAI_BASE_URL
|
|
process.env.OPENAI_MODEL = originalEnv.OPENAI_MODEL
|
|
})
|
|
|
|
test('treats localhost endpoints as local', () => {
|
|
expect(isLocalProviderUrl('http://localhost:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://127.0.0.1:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://0.0.0.0:11434/v1')).toBe(true)
|
|
// Full 127.0.0.0/8 loopback range should be treated as local
|
|
expect(isLocalProviderUrl('http://127.0.0.2:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://127.1.2.3:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://127.255.255.255:11434/v1')).toBe(true)
|
|
})
|
|
|
|
test('treats private IPv4 endpoints as local', () => {
|
|
expect(isLocalProviderUrl('http://10.0.0.1:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://172.16.0.1:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://192.168.0.1:11434/v1')).toBe(true)
|
|
})
|
|
|
|
test('treats .local hostnames as local', () => {
|
|
expect(isLocalProviderUrl('http://ollama.local:11434/v1')).toBe(true)
|
|
})
|
|
|
|
test('treats private IPv6 endpoints as local', () => {
|
|
expect(isLocalProviderUrl('http://[fd00::1]:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://[fe80::1]:11434/v1')).toBe(true)
|
|
expect(isLocalProviderUrl('http://[::1]:11434/v1')).toBe(true)
|
|
})
|
|
|
|
test('treats public hosts as remote', () => {
|
|
expect(isLocalProviderUrl('http://203.0.113.1:11434/v1')).toBe(false)
|
|
expect(isLocalProviderUrl('https://example.com/v1')).toBe(false)
|
|
expect(isLocalProviderUrl('http://[2001:4860:4860::8888]:11434/v1')).toBe(false)
|
|
})
|
|
|
|
test('creates a cache scope for local openai-compatible providers', () => {
|
|
process.env.CLAUDE_CODE_USE_OPENAI = '1'
|
|
process.env.OPENAI_BASE_URL = 'http://localhost:1234/v1'
|
|
process.env.OPENAI_MODEL = 'llama-3.2-3b-instruct'
|
|
|
|
expect(getAdditionalModelOptionsCacheScope()).toBe(
|
|
'openai:http://localhost:1234/v1',
|
|
)
|
|
})
|
|
|
|
test('keeps codex alias models on chat completions for local openai-compatible providers', () => {
|
|
process.env.CLAUDE_CODE_USE_OPENAI = '1'
|
|
process.env.OPENAI_BASE_URL = 'http://127.0.0.1:8080/v1'
|
|
process.env.OPENAI_MODEL = 'gpt-5.4'
|
|
|
|
expect(resolveProviderRequest()).toMatchObject({
|
|
transport: 'chat_completions',
|
|
requestedModel: 'gpt-5.4',
|
|
resolvedModel: 'gpt-5.4',
|
|
baseUrl: 'http://127.0.0.1:8080/v1',
|
|
})
|
|
expect(getAdditionalModelOptionsCacheScope()).toBe(
|
|
'openai:http://127.0.0.1:8080/v1',
|
|
)
|
|
})
|
|
|
|
test('skips local model cache scope for remote openai-compatible providers', () => {
|
|
process.env.CLAUDE_CODE_USE_OPENAI = '1'
|
|
process.env.OPENAI_BASE_URL = 'https://api.openai.com/v1'
|
|
process.env.OPENAI_MODEL = 'gpt-4o'
|
|
|
|
expect(getAdditionalModelOptionsCacheScope()).toBeNull()
|
|
})
|