Merge pull request #2684 from seefs001/fix/codex-rm-max-output-tokens

fix: codex Unsupported parameter: max_output_tokens
This commit is contained in:
Calcium-Ion
2026-01-21 23:47:17 +08:00
committed by GitHub

View File

@@ -91,6 +91,9 @@ func (a *Adaptor) ConvertOpenAIResponsesRequest(c *gin.Context, info *relaycommo
// codex: store must be false
request.Store = json.RawMessage("false")
// rm max_output_tokens
request.MaxOutputTokens = 0
request.Temperature = nil
return request, nil
}