Compare commits
8 Commits
fix/skillt
...
fix/repl-a
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
98f38d8bfc | ||
|
|
279cd1a7e1 | ||
|
|
5c13223aa4 | ||
|
|
2c8842f87c | ||
|
|
858f06d964 | ||
|
|
600c01faf7 | ||
|
|
b07bafa5bd | ||
|
|
85aa8b0985 |
@@ -8,6 +8,34 @@ import {
|
||||
validateProviderEnvOrExit,
|
||||
} from '../utils/providerValidation.js'
|
||||
|
||||
// OpenClaude: polyfill globalThis.File for Node < 20.
|
||||
// undici v7 references `File` at module evaluation time (webidl type
|
||||
// assertions). Node 18 lacks the global, causing a ReferenceError inside
|
||||
// the bundled __commonJS require chain which deadlocks the process when a
|
||||
// proxy is configured (configureGlobalAgents → require_undici).
|
||||
// eslint-disable-next-line custom-rules/no-top-level-side-effects
|
||||
if (typeof globalThis.File === 'undefined') {
|
||||
try {
|
||||
// Node 18.13+ exposes File in node:buffer but not as a global.
|
||||
// eslint-disable-next-line @typescript-eslint/no-require-imports
|
||||
const { File: NodeFile } = require('node:buffer')
|
||||
// @ts-expect-error -- polyfilling missing global
|
||||
globalThis.File = NodeFile
|
||||
} catch {
|
||||
// Absolute fallback: stub so `MakeTypeAssertion(File)` doesn't throw.
|
||||
// @ts-expect-error -- minimal polyfill
|
||||
globalThis.File = class File extends Blob {
|
||||
name: string
|
||||
lastModified: number
|
||||
constructor(parts: BlobPart[], name: string, opts?: FilePropertyBag) {
|
||||
super(parts, opts)
|
||||
this.name = name
|
||||
this.lastModified = opts?.lastModified ?? Date.now()
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// OpenClaude: disable experimental API betas by default.
|
||||
// Tool search (defer_loading), global cache scope, and context management
|
||||
// require internal API support not available to external accounts → 500.
|
||||
|
||||
@@ -201,6 +201,117 @@ describe('Codex request translation', () => {
|
||||
])
|
||||
})
|
||||
|
||||
test('preserves Grep tool pattern field in Codex strict schemas', () => {
|
||||
const tools = convertToolsToResponsesTools([
|
||||
{
|
||||
name: 'Grep',
|
||||
description: 'Search file contents',
|
||||
input_schema: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: { type: 'string', description: 'Search pattern' },
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
},
|
||||
])
|
||||
|
||||
expect(tools).toEqual([
|
||||
{
|
||||
type: 'function',
|
||||
name: 'Grep',
|
||||
description: 'Search file contents',
|
||||
parameters: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: { type: 'string', description: 'Search pattern' },
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern', 'path'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
strict: true,
|
||||
},
|
||||
])
|
||||
})
|
||||
|
||||
test('preserves Glob tool pattern field in Codex strict schemas', () => {
|
||||
const tools = convertToolsToResponsesTools([
|
||||
{
|
||||
name: 'Glob',
|
||||
description: 'Find files by pattern',
|
||||
input_schema: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: { type: 'string', description: 'Glob pattern' },
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
},
|
||||
])
|
||||
|
||||
expect(tools).toEqual([
|
||||
{
|
||||
type: 'function',
|
||||
name: 'Glob',
|
||||
description: 'Find files by pattern',
|
||||
parameters: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: { type: 'string', description: 'Glob pattern' },
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern', 'path'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
strict: true,
|
||||
},
|
||||
])
|
||||
})
|
||||
|
||||
test('strips validator pattern keyword but keeps string field named pattern in Codex schemas', () => {
|
||||
const tools = convertToolsToResponsesTools([
|
||||
{
|
||||
name: 'RegexProbe',
|
||||
description: 'Probe regex schema handling',
|
||||
input_schema: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: {
|
||||
type: 'string',
|
||||
pattern: '^[a-z]+$',
|
||||
},
|
||||
},
|
||||
required: ['pattern'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
},
|
||||
])
|
||||
|
||||
expect(tools).toEqual([
|
||||
{
|
||||
type: 'function',
|
||||
name: 'RegexProbe',
|
||||
description: 'Probe regex schema handling',
|
||||
parameters: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: {
|
||||
type: 'string',
|
||||
},
|
||||
},
|
||||
required: ['pattern'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
strict: true,
|
||||
},
|
||||
])
|
||||
})
|
||||
|
||||
test('removes unsupported uri format from strict Responses schemas', () => {
|
||||
const tools = convertToolsToResponsesTools([
|
||||
{
|
||||
|
||||
@@ -261,6 +261,73 @@ test('preserves Gemini tool call extra_content in follow-up requests', async ()
|
||||
})
|
||||
})
|
||||
|
||||
test('preserves Grep tool pattern field in OpenAI-compatible schemas', async () => {
|
||||
let requestBody: Record<string, unknown> | undefined
|
||||
|
||||
globalThis.fetch = (async (_input, init) => {
|
||||
requestBody = JSON.parse(String(init?.body))
|
||||
|
||||
return new Response(
|
||||
JSON.stringify({
|
||||
id: 'chatcmpl-grep-schema',
|
||||
model: 'qwen/qwen3.6-plus',
|
||||
choices: [
|
||||
{
|
||||
message: {
|
||||
role: 'assistant',
|
||||
content: 'done',
|
||||
},
|
||||
finish_reason: 'stop',
|
||||
},
|
||||
],
|
||||
usage: {
|
||||
prompt_tokens: 12,
|
||||
completion_tokens: 4,
|
||||
total_tokens: 16,
|
||||
},
|
||||
}),
|
||||
{
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
},
|
||||
)
|
||||
}) as FetchType
|
||||
|
||||
const client = createOpenAIShimClient({}) as OpenAIShimClient
|
||||
|
||||
await client.beta.messages.create({
|
||||
model: 'qwen/qwen3.6-plus',
|
||||
system: 'test system',
|
||||
messages: [{ role: 'user', content: 'Use Grep' }],
|
||||
tools: [
|
||||
{
|
||||
name: 'Grep',
|
||||
description: 'Search file contents',
|
||||
input_schema: {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: { type: 'string', description: 'Search pattern' },
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern'],
|
||||
additionalProperties: false,
|
||||
},
|
||||
},
|
||||
],
|
||||
max_tokens: 64,
|
||||
stream: false,
|
||||
})
|
||||
|
||||
const tools = requestBody?.tools as Array<Record<string, unknown>> | undefined
|
||||
const grepTool = tools?.find(tool => (tool.function as Record<string, unknown>)?.name === 'Grep') as
|
||||
| { function?: { parameters?: { properties?: Record<string, unknown>; required?: string[] } } }
|
||||
| undefined
|
||||
|
||||
expect(Object.keys(grepTool?.function?.parameters?.properties ?? {})).toContain('pattern')
|
||||
expect(grepTool?.function?.parameters?.required).toContain('pattern')
|
||||
})
|
||||
|
||||
test('does not infer Gemini mode from OPENAI_BASE_URL path substrings', async () => {
|
||||
let capturedAuthorization: string | null = null
|
||||
|
||||
|
||||
@@ -195,10 +195,12 @@ function convertContentBlocks(
|
||||
// handled separately
|
||||
break
|
||||
case 'thinking':
|
||||
// Append thinking as text with a marker for models that support reasoning
|
||||
if (block.thinking) {
|
||||
parts.push({ type: 'text', text: `<thinking>${block.thinking}</thinking>` })
|
||||
}
|
||||
case 'redacted_thinking':
|
||||
// Strip thinking blocks for OpenAI-compatible providers.
|
||||
// These are Anthropic-specific content types that 3P providers
|
||||
// don't understand. Serializing them as <thinking> text corrupts
|
||||
// multi-turn context: the model sees the tags as part of its
|
||||
// previous reply and may mimic or misattribute them.
|
||||
break
|
||||
default:
|
||||
if (block.text) {
|
||||
|
||||
@@ -72,16 +72,23 @@ export function getContextWindowForModel(
|
||||
return 1_000_000
|
||||
}
|
||||
|
||||
// OpenAI-compatible provider — use known context windows for the model
|
||||
if (
|
||||
// OpenAI-compatible provider — use known context windows for the model.
|
||||
// Unknown models get a conservative 8k default so auto-compact triggers
|
||||
// before hitting a hard context_window_exceeded error (issue #248 finding 3).
|
||||
const isOpenAIProvider =
|
||||
isEnvTruthy(process.env.CLAUDE_CODE_USE_OPENAI) ||
|
||||
isEnvTruthy(process.env.CLAUDE_CODE_USE_GEMINI) ||
|
||||
isEnvTruthy(process.env.CLAUDE_CODE_USE_GITHUB)
|
||||
) {
|
||||
if (isOpenAIProvider) {
|
||||
const openaiWindow = getOpenAIContextWindow(model)
|
||||
if (openaiWindow !== undefined) {
|
||||
return openaiWindow
|
||||
}
|
||||
console.error(
|
||||
`[context] Warning: model "${model}" not in context window table — using conservative 8k default. ` +
|
||||
'Add it to src/utils/model/openaiContextWindows.ts for accurate compaction.',
|
||||
)
|
||||
return 8_000
|
||||
}
|
||||
|
||||
const cap = getModelCapability(model)
|
||||
|
||||
@@ -69,3 +69,93 @@ test('loadConversationForResume rejects oversized transcripts before resume hook
|
||||
)
|
||||
expect(hookSpy).not.toHaveBeenCalled()
|
||||
})
|
||||
|
||||
test('deserializeMessagesWithInterruptDetection strips thinking blocks only for OpenAI-compatible providers', async () => {
|
||||
const serializedMessages = [
|
||||
user(id(10), 'hello'),
|
||||
{
|
||||
type: 'assistant',
|
||||
uuid: id(11),
|
||||
parentUuid: id(10),
|
||||
timestamp: ts,
|
||||
cwd: '/tmp',
|
||||
sessionId,
|
||||
version: 'test',
|
||||
message: {
|
||||
role: 'assistant',
|
||||
content: [
|
||||
{ type: 'thinking', thinking: 'secret reasoning' },
|
||||
{ type: 'text', text: 'visible reply' },
|
||||
],
|
||||
},
|
||||
},
|
||||
{
|
||||
type: 'assistant',
|
||||
uuid: id(12),
|
||||
parentUuid: id(11),
|
||||
timestamp: ts,
|
||||
cwd: '/tmp',
|
||||
sessionId,
|
||||
version: 'test',
|
||||
message: {
|
||||
role: 'assistant',
|
||||
content: [{ type: 'thinking', thinking: 'only hidden reasoning' }],
|
||||
},
|
||||
},
|
||||
user(id(13), 'follow up'),
|
||||
]
|
||||
|
||||
mock.module('./model/providers.js', () => ({
|
||||
getAPIProvider: () => 'openai',
|
||||
isOpenAICompatibleProvider: (provider: string) =>
|
||||
provider === 'openai' ||
|
||||
provider === 'gemini' ||
|
||||
provider === 'github' ||
|
||||
provider === 'codex',
|
||||
}))
|
||||
|
||||
const openaiModule = await import(`./conversationRecovery.ts?provider=openai-${Date.now()}`)
|
||||
const thirdParty = openaiModule.deserializeMessagesWithInterruptDetection(serializedMessages as never[])
|
||||
const thirdPartyAssistantMessages = thirdParty.messages.filter(
|
||||
message => message.type === 'assistant',
|
||||
)
|
||||
|
||||
expect(thirdPartyAssistantMessages).toHaveLength(2)
|
||||
expect(thirdPartyAssistantMessages[0]?.message?.content).toEqual([
|
||||
{ type: 'text', text: 'visible reply' },
|
||||
])
|
||||
expect(
|
||||
JSON.stringify(thirdPartyAssistantMessages.map(message => message.message?.content)),
|
||||
).not.toContain('secret reasoning')
|
||||
expect(
|
||||
JSON.stringify(thirdPartyAssistantMessages.map(message => message.message?.content)),
|
||||
).not.toContain('only hidden reasoning')
|
||||
|
||||
mock.restore()
|
||||
mock.module('./model/providers.js', () => ({
|
||||
getAPIProvider: () => 'bedrock',
|
||||
isOpenAICompatibleProvider: (provider: string) =>
|
||||
provider === 'openai' ||
|
||||
provider === 'gemini' ||
|
||||
provider === 'github' ||
|
||||
provider === 'codex',
|
||||
}))
|
||||
|
||||
const bedrockModule = await import(`./conversationRecovery.ts?provider=bedrock-${Date.now()}`)
|
||||
const anthropicCompatible = bedrockModule.deserializeMessagesWithInterruptDetection(serializedMessages as never[])
|
||||
const anthropicAssistantMessages = anthropicCompatible.messages.filter(
|
||||
message => message.type === 'assistant',
|
||||
)
|
||||
|
||||
expect(anthropicAssistantMessages).toHaveLength(2)
|
||||
expect(anthropicAssistantMessages[0]?.message?.content).toEqual([
|
||||
{ type: 'thinking', thinking: 'secret reasoning' },
|
||||
{ type: 'text', text: 'visible reply' },
|
||||
])
|
||||
expect(
|
||||
JSON.stringify(anthropicAssistantMessages.map(message => message.message?.content)),
|
||||
).toContain('secret reasoning')
|
||||
expect(
|
||||
JSON.stringify(anthropicAssistantMessages.map(message => message.message?.content)),
|
||||
).not.toContain('only hidden reasoning')
|
||||
})
|
||||
|
||||
@@ -13,6 +13,7 @@ const originalSimple = process.env.CLAUDE_CODE_SIMPLE
|
||||
const sessionId = '00000000-0000-4000-8000-000000001999'
|
||||
const ts = '2026-04-02T00:00:00.000Z'
|
||||
|
||||
|
||||
function id(n: number): string {
|
||||
return `00000000-0000-4000-8000-${String(n).padStart(12, '0')}`
|
||||
}
|
||||
@@ -76,4 +77,3 @@ test('loadConversationForResume rejects oversized reconstructed transcripts', as
|
||||
'Reconstructed transcript is too large to resume safely',
|
||||
)
|
||||
})
|
||||
|
||||
|
||||
@@ -24,6 +24,7 @@ import {
|
||||
type FileHistorySnapshot,
|
||||
} from './fileHistory.js'
|
||||
import { logError } from './log.js'
|
||||
import { getAPIProvider } from './model/providers.js'
|
||||
import {
|
||||
createAssistantMessage,
|
||||
createUserMessage,
|
||||
@@ -177,6 +178,25 @@ export type DeserializeResult = {
|
||||
turnInterruptionState: TurnInterruptionState
|
||||
}
|
||||
|
||||
/**
|
||||
* Remove thinking/redacted_thinking content blocks from assistant messages.
|
||||
* Messages that become empty after stripping are removed entirely.
|
||||
*/
|
||||
function stripThinkingBlocks(messages: NormalizedMessage[]): NormalizedMessage[] {
|
||||
return messages.reduce<NormalizedMessage[]>((acc, msg) => {
|
||||
if (msg.type !== 'assistant' || !Array.isArray(msg.message?.content)) {
|
||||
acc.push(msg)
|
||||
return acc
|
||||
}
|
||||
const filtered = msg.message.content.filter(
|
||||
(block: { type?: string }) => block.type !== 'thinking' && block.type !== 'redacted_thinking',
|
||||
)
|
||||
if (filtered.length === 0) return acc
|
||||
acc.push({ ...msg, message: { ...msg.message, content: filtered } })
|
||||
return acc
|
||||
}, [])
|
||||
}
|
||||
|
||||
/**
|
||||
* Deserializes messages from a log file into the format expected by the REPL.
|
||||
* Filters unresolved tool uses, orphaned thinking messages, and appends a
|
||||
@@ -227,10 +247,19 @@ export function deserializeMessagesWithInterruptDetection(
|
||||
filteredToolUses,
|
||||
) as NormalizedMessage[]
|
||||
|
||||
// Strip thinking/redacted_thinking content blocks from assistant messages
|
||||
// when resuming against a 3P provider. These Anthropic-specific blocks cause
|
||||
// 400 errors or context corruption on OpenAI-compatible providers (issue #248 finding 5).
|
||||
const provider = getAPIProvider()
|
||||
const isThirdPartyProvider = provider !== 'firstParty' && provider !== 'bedrock' && provider !== 'vertex' && provider !== 'foundry'
|
||||
const thinkingStripped = isThirdPartyProvider
|
||||
? stripThinkingBlocks(filteredThinking)
|
||||
: filteredThinking
|
||||
|
||||
// Filter out assistant messages with only whitespace text content.
|
||||
// This can happen when model outputs "\n\n" before thinking, user cancels mid-stream.
|
||||
const filteredMessages = filterWhitespaceOnlyAssistantMessages(
|
||||
filteredThinking,
|
||||
thinkingStripped,
|
||||
) as NormalizedMessage[]
|
||||
|
||||
const internalState = detectTurnInterruption(filteredMessages)
|
||||
|
||||
@@ -1,11 +1,52 @@
|
||||
import { expect, test } from 'bun:test'
|
||||
import path from 'path'
|
||||
|
||||
import { wrapRipgrepUnavailableError } from './ripgrep.ts'
|
||||
import { resolveRipgrepConfig, wrapRipgrepUnavailableError } from './ripgrep.js'
|
||||
|
||||
const MOCK_BUILTIN_PATH = path.normalize(
|
||||
process.platform === 'win32'
|
||||
? `vendor/ripgrep/${process.arch}-win32/rg.exe`
|
||||
: `vendor/ripgrep/${process.arch}-${process.platform}/rg`,
|
||||
)
|
||||
|
||||
test('ripgrepCommand falls back to system rg when builtin binary is missing', () => {
|
||||
const config = resolveRipgrepConfig({
|
||||
userWantsSystemRipgrep: false,
|
||||
bundledMode: false,
|
||||
builtinCommand: MOCK_BUILTIN_PATH,
|
||||
builtinExists: false,
|
||||
systemExecutablePath: '/usr/bin/rg',
|
||||
processExecPath: '/fake/bun',
|
||||
})
|
||||
|
||||
expect(config).toMatchObject({
|
||||
mode: 'system',
|
||||
command: 'rg',
|
||||
args: [],
|
||||
})
|
||||
})
|
||||
|
||||
test('ripgrepCommand keeps builtin mode when bundled binary exists', () => {
|
||||
const config = resolveRipgrepConfig({
|
||||
userWantsSystemRipgrep: false,
|
||||
bundledMode: false,
|
||||
builtinCommand: MOCK_BUILTIN_PATH,
|
||||
builtinExists: true,
|
||||
systemExecutablePath: '/usr/bin/rg',
|
||||
processExecPath: '/fake/bun',
|
||||
})
|
||||
|
||||
expect(config).toMatchObject({
|
||||
mode: 'builtin',
|
||||
command: MOCK_BUILTIN_PATH,
|
||||
args: [],
|
||||
})
|
||||
})
|
||||
|
||||
test('wrapRipgrepUnavailableError explains missing packaged fallback', () => {
|
||||
const error = wrapRipgrepUnavailableError(
|
||||
{ code: 'ENOENT', message: 'spawn rg ENOENT' },
|
||||
{ mode: 'builtin', command: 'C:\\fake\\vendor\\ripgrep\\rg.exe' },
|
||||
{ mode: 'builtin', command: 'C:\\fake\\vendor\\ripgrep\\rg.exe', args: [] },
|
||||
'win32',
|
||||
)
|
||||
|
||||
@@ -18,7 +59,7 @@ test('wrapRipgrepUnavailableError explains missing packaged fallback', () => {
|
||||
test('wrapRipgrepUnavailableError explains missing system ripgrep', () => {
|
||||
const error = wrapRipgrepUnavailableError(
|
||||
{ code: 'ENOENT', message: 'spawn rg ENOENT' },
|
||||
{ mode: 'system', command: 'rg' },
|
||||
{ mode: 'system', command: 'rg', args: [] },
|
||||
'linux',
|
||||
)
|
||||
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
import type { ChildProcess, ExecFileException } from 'child_process'
|
||||
import { execFile, spawn } from 'child_process'
|
||||
import { existsSync } from 'fs'
|
||||
import memoize from 'lodash-es/memoize.js'
|
||||
import { homedir } from 'os'
|
||||
import * as path from 'path'
|
||||
@@ -30,40 +31,72 @@ type RipgrepConfig = {
|
||||
|
||||
type RipgrepErrorLike = Pick<NodeJS.ErrnoException, 'code' | 'message'>
|
||||
|
||||
const getRipgrepConfig = memoize((): RipgrepConfig => {
|
||||
const userWantsSystemRipgrep = isEnvDefinedFalsy(
|
||||
process.env.USE_BUILTIN_RIPGREP,
|
||||
)
|
||||
function isErrnoException(error: unknown): error is NodeJS.ErrnoException {
|
||||
return error instanceof Error
|
||||
}
|
||||
|
||||
// Try system ripgrep if user wants it
|
||||
if (userWantsSystemRipgrep) {
|
||||
const { cmd: systemPath } = findExecutable('rg', [])
|
||||
if (systemPath !== 'rg') {
|
||||
// SECURITY: Use command name 'rg' instead of systemPath to prevent PATH hijacking
|
||||
// If we used systemPath, a malicious ./rg.exe in current directory could be executed
|
||||
// Using just 'rg' lets the OS resolve it safely with NoDefaultCurrentDirectoryInExePath protection
|
||||
return { mode: 'system', command: 'rg', args: [] }
|
||||
}
|
||||
type ResolveRipgrepConfigArgs = {
|
||||
userWantsSystemRipgrep: boolean
|
||||
bundledMode: boolean
|
||||
builtinCommand: string
|
||||
builtinExists: boolean
|
||||
systemExecutablePath: string
|
||||
processExecPath?: string
|
||||
}
|
||||
|
||||
export function resolveRipgrepConfig({
|
||||
userWantsSystemRipgrep,
|
||||
bundledMode,
|
||||
builtinCommand,
|
||||
builtinExists,
|
||||
systemExecutablePath,
|
||||
processExecPath = process.execPath,
|
||||
}: ResolveRipgrepConfigArgs): RipgrepConfig {
|
||||
if (userWantsSystemRipgrep && systemExecutablePath !== 'rg') {
|
||||
// SECURITY: Use command name 'rg' instead of systemExecutablePath to prevent PATH hijacking
|
||||
return { mode: 'system', command: 'rg', args: [] }
|
||||
}
|
||||
|
||||
// In bundled (native) mode, ripgrep is statically compiled into bun-internal
|
||||
// and dispatches based on argv[0]. We spawn ourselves with argv0='rg'.
|
||||
if (isInBundledMode()) {
|
||||
if (bundledMode) {
|
||||
return {
|
||||
mode: 'embedded',
|
||||
command: process.execPath,
|
||||
command: processExecPath,
|
||||
args: ['--no-config'],
|
||||
argv0: 'rg',
|
||||
}
|
||||
}
|
||||
|
||||
if (builtinExists) {
|
||||
return { mode: 'builtin', command: builtinCommand, args: [] }
|
||||
}
|
||||
|
||||
if (systemExecutablePath !== 'rg') {
|
||||
return { mode: 'system', command: 'rg', args: [] }
|
||||
}
|
||||
|
||||
return { mode: 'builtin', command: builtinCommand, args: [] }
|
||||
}
|
||||
|
||||
const getRipgrepConfig = memoize((): RipgrepConfig => {
|
||||
const userWantsSystemRipgrep = isEnvDefinedFalsy(
|
||||
process.env.USE_BUILTIN_RIPGREP,
|
||||
)
|
||||
const bundledMode = isInBundledMode()
|
||||
const rgRoot = path.resolve(__dirname, 'vendor', 'ripgrep')
|
||||
const command =
|
||||
const builtinCommand =
|
||||
process.platform === 'win32'
|
||||
? path.resolve(rgRoot, `${process.arch}-win32`, 'rg.exe')
|
||||
: path.resolve(rgRoot, `${process.arch}-${process.platform}`, 'rg')
|
||||
const builtinExists = existsSync(builtinCommand)
|
||||
const { cmd: systemExecutablePath } = findExecutable('rg', [])
|
||||
|
||||
return { mode: 'builtin', command, args: [] }
|
||||
return resolveRipgrepConfig({
|
||||
userWantsSystemRipgrep,
|
||||
bundledMode,
|
||||
builtinCommand,
|
||||
builtinExists,
|
||||
systemExecutablePath,
|
||||
})
|
||||
})
|
||||
|
||||
export function ripgrepCommand(): {
|
||||
@@ -324,7 +357,9 @@ async function ripGrepFileCount(
|
||||
if (settled) return
|
||||
settled = true
|
||||
reject(
|
||||
err.code === 'ENOENT' ? wrapRipgrepUnavailableError(err) : err,
|
||||
isErrnoException(err) && err.code === 'ENOENT'
|
||||
? wrapRipgrepUnavailableError(err)
|
||||
: err,
|
||||
)
|
||||
})
|
||||
})
|
||||
@@ -388,7 +423,9 @@ export async function ripGrepStream(
|
||||
if (settled) return
|
||||
settled = true
|
||||
reject(
|
||||
err.code === 'ENOENT' ? wrapRipgrepUnavailableError(err) : err,
|
||||
isErrnoException(err) && err.code === 'ENOENT'
|
||||
? wrapRipgrepUnavailableError(err)
|
||||
: err,
|
||||
)
|
||||
})
|
||||
})
|
||||
@@ -436,7 +473,9 @@ export async function ripGrep(
|
||||
const CRITICAL_ERROR_CODES = ['ENOENT', 'EACCES', 'EPERM']
|
||||
if (CRITICAL_ERROR_CODES.includes(error.code as string)) {
|
||||
reject(
|
||||
error.code === 'ENOENT' ? wrapRipgrepUnavailableError(error) : error,
|
||||
isErrnoException(error) && error.code === 'ENOENT'
|
||||
? wrapRipgrepUnavailableError(error)
|
||||
: error,
|
||||
)
|
||||
return
|
||||
}
|
||||
|
||||
68
src/utils/schemaSanitizer.test.ts
Normal file
68
src/utils/schemaSanitizer.test.ts
Normal file
@@ -0,0 +1,68 @@
|
||||
import { describe, expect, test } from 'bun:test'
|
||||
|
||||
import { sanitizeSchemaForOpenAICompat } from './schemaSanitizer'
|
||||
|
||||
describe('sanitizeSchemaForOpenAICompat', () => {
|
||||
test('preserves Grep-like properties.pattern while keeping it required', () => {
|
||||
const schema = {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: {
|
||||
type: 'string',
|
||||
description: 'The regular expression pattern to search for in file contents',
|
||||
},
|
||||
path: { type: 'string' },
|
||||
glob: { type: 'string' },
|
||||
},
|
||||
required: ['pattern'],
|
||||
}
|
||||
|
||||
const sanitized = sanitizeSchemaForOpenAICompat(schema)
|
||||
const properties = sanitized.properties as Record<string, unknown> | undefined
|
||||
|
||||
expect(Object.keys(properties ?? {})).toEqual(['pattern', 'path', 'glob'])
|
||||
expect(properties?.pattern).toEqual({
|
||||
type: 'string',
|
||||
description: 'The regular expression pattern to search for in file contents',
|
||||
})
|
||||
expect(sanitized.required).toEqual(['pattern'])
|
||||
})
|
||||
|
||||
test('preserves Glob-like properties.pattern while keeping it required', () => {
|
||||
const schema = {
|
||||
type: 'object',
|
||||
properties: {
|
||||
pattern: {
|
||||
type: 'string',
|
||||
description: 'The glob pattern to match files against',
|
||||
},
|
||||
path: { type: 'string' },
|
||||
},
|
||||
required: ['pattern'],
|
||||
}
|
||||
|
||||
const sanitized = sanitizeSchemaForOpenAICompat(schema)
|
||||
const properties = sanitized.properties as Record<string, unknown> | undefined
|
||||
|
||||
expect(Object.keys(properties ?? {})).toEqual(['pattern', 'path'])
|
||||
expect(properties?.pattern).toEqual({
|
||||
type: 'string',
|
||||
description: 'The glob pattern to match files against',
|
||||
})
|
||||
expect(sanitized.required).toEqual(['pattern'])
|
||||
})
|
||||
|
||||
test('strips JSON Schema validator pattern from string schemas', () => {
|
||||
const schema = {
|
||||
type: 'string',
|
||||
pattern: '^[a-z]+$',
|
||||
minLength: 1,
|
||||
}
|
||||
|
||||
const sanitized = sanitizeSchemaForOpenAICompat(schema)
|
||||
|
||||
expect(sanitized).toEqual({
|
||||
type: 'string',
|
||||
})
|
||||
})
|
||||
})
|
||||
@@ -33,6 +33,15 @@ function stripSchemaKeywords(schema: unknown, keywords: Set<string>): unknown {
|
||||
|
||||
const result: Record<string, unknown> = {}
|
||||
for (const [key, value] of Object.entries(schema)) {
|
||||
if (key === 'properties' && isSchemaRecord(value)) {
|
||||
const sanitizedProps: Record<string, unknown> = {}
|
||||
for (const [propName, propSchema] of Object.entries(value)) {
|
||||
sanitizedProps[propName] = stripSchemaKeywords(propSchema, keywords)
|
||||
}
|
||||
result[key] = sanitizedProps
|
||||
continue
|
||||
}
|
||||
|
||||
if (keywords.has(key)) {
|
||||
continue
|
||||
}
|
||||
@@ -215,10 +224,13 @@ export function sanitizeSchemaForOpenAICompat(
|
||||
}
|
||||
}
|
||||
|
||||
if (Array.isArray(record.required) && isSchemaRecord(record.properties)) {
|
||||
const properties = isSchemaRecord(record.properties)
|
||||
? record.properties
|
||||
: undefined
|
||||
|
||||
if (Array.isArray(record.required) && properties) {
|
||||
record.required = record.required.filter(
|
||||
(value): value is string =>
|
||||
typeof value === 'string' && value in record.properties,
|
||||
(value): value is string => typeof value === 'string' && value in properties,
|
||||
)
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user