chore: 更新codex配置部分支持gpt-5.4的长上下文
This commit is contained in:
@@ -512,12 +512,14 @@ function generateOpenAIFiles(baseUrl: string, apiKey: string): FileConfig[] {
|
|||||||
|
|
||||||
// config.toml content
|
// config.toml content
|
||||||
const configContent = `model_provider = "OpenAI"
|
const configContent = `model_provider = "OpenAI"
|
||||||
model = "gpt-5.3-codex"
|
model = "gpt-5.4"
|
||||||
review_model = "gpt-5.3-codex"
|
review_model = "gpt-5.4"
|
||||||
model_reasoning_effort = "xhigh"
|
model_reasoning_effort = "high"
|
||||||
disable_response_storage = true
|
disable_response_storage = true
|
||||||
network_access = "enabled"
|
network_access = "enabled"
|
||||||
windows_wsl_setup_acknowledged = true
|
windows_wsl_setup_acknowledged = true
|
||||||
|
model_context_window = 1000000
|
||||||
|
model_auto_compact_token_limit = 900000
|
||||||
|
|
||||||
[model_providers.OpenAI]
|
[model_providers.OpenAI]
|
||||||
name = "OpenAI"
|
name = "OpenAI"
|
||||||
@@ -549,12 +551,14 @@ function generateOpenAIWsFiles(baseUrl: string, apiKey: string): FileConfig[] {
|
|||||||
|
|
||||||
// config.toml content with WebSocket v2
|
// config.toml content with WebSocket v2
|
||||||
const configContent = `model_provider = "OpenAI"
|
const configContent = `model_provider = "OpenAI"
|
||||||
model = "gpt-5.3-codex"
|
model = "gpt-5.4"
|
||||||
review_model = "gpt-5.3-codex"
|
review_model = "gpt-5.4"
|
||||||
model_reasoning_effort = "xhigh"
|
model_reasoning_effort = "high"
|
||||||
disable_response_storage = true
|
disable_response_storage = true
|
||||||
network_access = "enabled"
|
network_access = "enabled"
|
||||||
windows_wsl_setup_acknowledged = true
|
windows_wsl_setup_acknowledged = true
|
||||||
|
model_context_window = 1000000
|
||||||
|
model_auto_compact_token_limit = 900000
|
||||||
|
|
||||||
[model_providers.OpenAI]
|
[model_providers.OpenAI]
|
||||||
name = "OpenAI"
|
name = "OpenAI"
|
||||||
|
|||||||
Reference in New Issue
Block a user