From 225fd035ae7462d8d7bf0b8d375c61166fee80a4 Mon Sep 17 00:00:00 2001 From: shaw Date: Fri, 6 Mar 2026 10:55:09 +0800 Subject: [PATCH] =?UTF-8?q?chore:=20=E6=9B=B4=E6=96=B0codex=E9=85=8D?= =?UTF-8?q?=E7=BD=AE=E9=83=A8=E5=88=86=E6=94=AF=E6=8C=81gpt-5.4=E7=9A=84?= =?UTF-8?q?=E9=95=BF=E4=B8=8A=E4=B8=8B=E6=96=87?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- frontend/src/components/keys/UseKeyModal.vue | 16 ++++++++++------ 1 file changed, 10 insertions(+), 6 deletions(-) diff --git a/frontend/src/components/keys/UseKeyModal.vue b/frontend/src/components/keys/UseKeyModal.vue index a61ce6b4..5ae2fa3f 100644 --- a/frontend/src/components/keys/UseKeyModal.vue +++ b/frontend/src/components/keys/UseKeyModal.vue @@ -512,12 +512,14 @@ function generateOpenAIFiles(baseUrl: string, apiKey: string): FileConfig[] { // config.toml content const configContent = `model_provider = "OpenAI" -model = "gpt-5.3-codex" -review_model = "gpt-5.3-codex" -model_reasoning_effort = "xhigh" +model = "gpt-5.4" +review_model = "gpt-5.4" +model_reasoning_effort = "high" disable_response_storage = true network_access = "enabled" windows_wsl_setup_acknowledged = true +model_context_window = 1000000 +model_auto_compact_token_limit = 900000 [model_providers.OpenAI] name = "OpenAI" @@ -549,12 +551,14 @@ function generateOpenAIWsFiles(baseUrl: string, apiKey: string): FileConfig[] { // config.toml content with WebSocket v2 const configContent = `model_provider = "OpenAI" -model = "gpt-5.3-codex" -review_model = "gpt-5.3-codex" -model_reasoning_effort = "xhigh" +model = "gpt-5.4" +review_model = "gpt-5.4" +model_reasoning_effort = "high" disable_response_storage = true network_access = "enabled" windows_wsl_setup_acknowledged = true +model_context_window = 1000000 +model_auto_compact_token_limit = 900000 [model_providers.OpenAI] name = "OpenAI"