Add DeepSeek V4 flash/pro support and DeepSeek thinking compatibility (#877)

* Add DeepSeek V4 support and thinking compatibility

* Fix DeepSeek profile persistence regression

* Align multi-model handling with openai-multi-model
This commit is contained in:
JATMN
2026-04-24 11:29:46 -07:00
committed by GitHub
parent c4cb98a4f0
commit ff2a380723
15 changed files with 356 additions and 31 deletions

View File

@@ -174,7 +174,7 @@ export function getProviderPresetDefaults(
provider: 'openai',
name: 'DeepSeek',
baseUrl: 'https://api.deepseek.com/v1',
model: 'deepseek-chat',
model: 'deepseek-v4-flash, deepseek-v4-pro, deepseek-chat, deepseek-reasoner',
apiKey: '',
requiresApiKey: true,
}
@@ -839,7 +839,7 @@ export function persistActiveProviderProfileModel(
/**
* Generate model options from a provider profile's model field.
* Each comma-separated model becomes a separate option in the picker.
* Each parsed model becomes a separate option in the picker.
*/
export function getProfileModelOptions(profile: ProviderProfile): ModelOption[] {
const models = parseModelList(profile.model)