// @ts-nocheck import { spawn } from 'node:child_process' import { existsSync, readFileSync } from 'node:fs' import { resolve } from 'node:path' type ProviderProfile = 'openai' | 'ollama' type ProfileFile = { profile: ProviderProfile env?: { OPENAI_BASE_URL?: string OPENAI_MODEL?: string OPENAI_API_KEY?: string } } type LaunchOptions = { requestedProfile: ProviderProfile | 'auto' | null passthroughArgs: string[] fast: boolean } function parseLaunchOptions(argv: string[]): LaunchOptions { let requestedProfile: ProviderProfile | 'auto' | null = 'auto' const passthroughArgs: string[] = [] let fast = false for (const arg of argv) { const lower = arg.toLowerCase() if (lower === '--fast') { fast = true continue } if ((lower === 'auto' || lower === 'openai' || lower === 'ollama') && requestedProfile === 'auto') { requestedProfile = lower as ProviderProfile | 'auto' continue } if (arg.startsWith('--')) { passthroughArgs.push(arg) continue } if (requestedProfile === 'auto') { requestedProfile = null break } passthroughArgs.push(arg) } return { requestedProfile, passthroughArgs, fast, } } function loadPersistedProfile(): ProfileFile | null { const path = resolve(process.cwd(), '.openclaude-profile.json') if (!existsSync(path)) return null try { const parsed = JSON.parse(readFileSync(path, 'utf8')) as ProfileFile if (parsed.profile === 'openai' || parsed.profile === 'ollama') { return parsed } return null } catch { return null } } async function hasLocalOllama(): Promise { const endpoint = 'http://localhost:11434/api/tags' const controller = new AbortController() const timeout = setTimeout(() => controller.abort(), 1200) try { const response = await fetch(endpoint, { signal: controller.signal }) return response.ok } catch { return false } finally { clearTimeout(timeout) } } function runCommand(command: string, env: NodeJS.ProcessEnv): Promise { return new Promise(resolve => { const child = spawn(command, { cwd: process.cwd(), env, stdio: 'inherit', shell: true, }) child.on('close', code => resolve(code ?? 1)) child.on('error', () => resolve(1)) }) } function buildEnv(profile: ProviderProfile, persisted: ProfileFile | null): NodeJS.ProcessEnv { const persistedEnv = persisted?.env ?? {} const env: NodeJS.ProcessEnv = { ...process.env, CLAUDE_CODE_USE_OPENAI: '1', } if (profile === 'ollama') { env.OPENAI_BASE_URL = persistedEnv.OPENAI_BASE_URL || process.env.OPENAI_BASE_URL || 'http://localhost:11434/v1' env.OPENAI_MODEL = persistedEnv.OPENAI_MODEL || process.env.OPENAI_MODEL || 'llama3.1:8b' if (!process.env.OPENAI_API_KEY || process.env.OPENAI_API_KEY === 'SUA_CHAVE') { delete env.OPENAI_API_KEY } return env } env.OPENAI_BASE_URL = process.env.OPENAI_BASE_URL || persistedEnv.OPENAI_BASE_URL || 'https://api.openai.com/v1' env.OPENAI_MODEL = process.env.OPENAI_MODEL || persistedEnv.OPENAI_MODEL || 'gpt-4o' env.OPENAI_API_KEY = process.env.OPENAI_API_KEY || persistedEnv.OPENAI_API_KEY return env } function applyFastFlags(env: NodeJS.ProcessEnv): NodeJS.ProcessEnv { env.CLAUDE_CODE_SIMPLE ??= '1' env.CLAUDE_CODE_DISABLE_THINKING ??= '1' env.DISABLE_INTERLEAVED_THINKING ??= '1' env.DISABLE_AUTO_COMPACT ??= '1' env.CLAUDE_CODE_DISABLE_AUTO_MEMORY ??= '1' env.CLAUDE_CODE_DISABLE_BACKGROUND_TASKS ??= '1' return env } function quoteArg(arg: string): string { if (!arg.includes(' ') && !arg.includes('"')) return arg return `"${arg.replace(/"/g, '\\"')}"` } function printSummary(profile: ProviderProfile, env: NodeJS.ProcessEnv): void { const keySet = Boolean(env.OPENAI_API_KEY) console.log(`Launching profile: ${profile}`) console.log(`OPENAI_BASE_URL=${env.OPENAI_BASE_URL}`) console.log(`OPENAI_MODEL=${env.OPENAI_MODEL}`) console.log(`OPENAI_API_KEY_SET=${keySet}`) } async function main(): Promise { const options = parseLaunchOptions(process.argv.slice(2)) const requestedProfile = options.requestedProfile if (!requestedProfile) { console.error('Usage: bun run scripts/provider-launch.ts [openai|ollama|auto] [--fast] [-- ]') process.exit(1) } const persisted = loadPersistedProfile() let profile: ProviderProfile if (requestedProfile === 'auto') { if (persisted) { profile = persisted.profile } else { profile = (await hasLocalOllama()) ? 'ollama' : 'openai' } } else { profile = requestedProfile } const env = buildEnv(profile, persisted) if (options.fast) { applyFastFlags(env) } if (profile === 'openai' && (!env.OPENAI_API_KEY || env.OPENAI_API_KEY === 'SUA_CHAVE')) { console.error('OPENAI_API_KEY is required for openai profile and cannot be SUA_CHAVE. Run: bun run profile:init -- --provider openai --api-key ') process.exit(1) } printSummary(profile, env) const doctorCode = await runCommand('bun run scripts/system-check.ts', env) if (doctorCode !== 0) { console.error('Runtime doctor failed. Fix configuration before launching.') process.exit(doctorCode) } const cliArgs = options.passthroughArgs.map(quoteArg).join(' ') const devCommand = cliArgs ? `bun run dev -- ${cliArgs}` : 'bun run dev' const devCode = await runCommand(devCommand, env) process.exit(devCode) } await main() export {}