diff --git a/relay/channel/vertex/adaptor.go b/relay/channel/vertex/adaptor.go index bd9cc97c..964c469f 100644 --- a/relay/channel/vertex/adaptor.go +++ b/relay/channel/vertex/adaptor.go @@ -24,9 +24,9 @@ import ( ) const ( - RequestModeClaude = 1 - RequestModeGemini = 2 - RequestModeLlama = 3 + RequestModeClaude = 1 + RequestModeGemini = 2 + RequestModeOpenSource = 3 ) var claudeModelMap = map[string]string{ @@ -115,7 +115,7 @@ func (a *Adaptor) Init(info *relaycommon.RelayInfo) { } else if strings.Contains(info.UpstreamModelName, "llama") || // open source models strings.Contains(info.UpstreamModelName, "-maas") { - a.RequestMode = RequestModeLlama + a.RequestMode = RequestModeOpenSource } else { a.RequestMode = RequestModeGemini } @@ -166,10 +166,9 @@ func (a *Adaptor) getRequestUrl(info *relaycommon.RelayInfo, modelName, suffix s suffix, ), nil } - } else if a.RequestMode == RequestModeLlama { + } else if a.RequestMode == RequestModeOpenSource { return fmt.Sprintf( - "https://%s-aiplatform.googleapis.com/v1beta1/projects/%s/locations/%s/endpoints/openapi/chat/completions", - region, + "https://aiplatform.googleapis.com/v1beta1/projects/%s/locations/%s/endpoints/openapi/chat/completions", adc.ProjectID, region, ), nil @@ -242,7 +241,7 @@ func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) { model = v } return a.getRequestUrl(info, model, suffix) - } else if a.RequestMode == RequestModeLlama { + } else if a.RequestMode == RequestModeOpenSource { return a.getRequestUrl(info, "", "") } return "", errors.New("unsupported request mode") @@ -340,7 +339,7 @@ func (a *Adaptor) ConvertOpenAIRequest(c *gin.Context, info *relaycommon.RelayIn } c.Set("request_model", request.Model) return geminiRequest, nil - } else if a.RequestMode == RequestModeLlama { + } else if a.RequestMode == RequestModeOpenSource { return request, nil } return nil, errors.New("unsupported request mode") @@ -375,7 +374,7 @@ func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, info *relaycom } else { return gemini.GeminiChatStreamHandler(c, info, resp) } - case RequestModeLlama: + case RequestModeOpenSource: return openai.OaiStreamHandler(c, info, resp) } } else { @@ -391,7 +390,7 @@ func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, info *relaycom } return gemini.GeminiChatHandler(c, info, resp) } - case RequestModeLlama: + case RequestModeOpenSource: return openai.OpenaiHandler(c, info, resp) } }