|
|
@@ -0,0 +1,127 @@
|
|
|
+package relay
|
|
|
+
|
|
|
+import (
|
|
|
+ "bytes"
|
|
|
+ "encoding/json"
|
|
|
+ "fmt"
|
|
|
+ "github.com/gin-gonic/gin"
|
|
|
+ "net/http"
|
|
|
+ "one-api/common"
|
|
|
+ "one-api/dto"
|
|
|
+ relaycommon "one-api/relay/common"
|
|
|
+ relayconstant "one-api/relay/constant"
|
|
|
+ "one-api/service"
|
|
|
+ "one-api/setting"
|
|
|
+)
|
|
|
+
|
|
|
+func getEmbeddingPromptToken(embeddingRequest dto.EmbeddingRequest) int {
|
|
|
+ token, _ := service.CountTokenInput(embeddingRequest.Input, embeddingRequest.Model)
|
|
|
+ return token
|
|
|
+}
|
|
|
+
|
|
|
+func EmbeddingHelper(c *gin.Context, relayMode int) (openaiErr *dto.OpenAIErrorWithStatusCode) {
|
|
|
+ relayInfo := relaycommon.GenRelayInfo(c)
|
|
|
+
|
|
|
+ var embeddingRequest *dto.EmbeddingRequest
|
|
|
+ err := common.UnmarshalBodyReusable(c, &embeddingRequest)
|
|
|
+ if err != nil {
|
|
|
+ common.LogError(c, fmt.Sprintf("getAndValidateTextRequest failed: %s", err.Error()))
|
|
|
+ return service.OpenAIErrorWrapperLocal(err, "invalid_text_request", http.StatusBadRequest)
|
|
|
+ }
|
|
|
+ if relayMode == relayconstant.RelayModeModerations && embeddingRequest.Model == "" {
|
|
|
+ embeddingRequest.Model = "m3e-base"
|
|
|
+ }
|
|
|
+ if relayMode == relayconstant.RelayModeEmbeddings && embeddingRequest.Model == "" {
|
|
|
+ embeddingRequest.Model = c.Param("model")
|
|
|
+ }
|
|
|
+ if embeddingRequest.Input == nil || len(embeddingRequest.Input) == 0 {
|
|
|
+ return service.OpenAIErrorWrapperLocal(fmt.Errorf("input is empty"), "invalid_input", http.StatusBadRequest)
|
|
|
+ }
|
|
|
+ // map model name
|
|
|
+ modelMapping := c.GetString("model_mapping")
|
|
|
+ //isModelMapped := false
|
|
|
+ if modelMapping != "" && modelMapping != "{}" {
|
|
|
+ modelMap := make(map[string]string)
|
|
|
+ err := json.Unmarshal([]byte(modelMapping), &modelMap)
|
|
|
+ if err != nil {
|
|
|
+ return service.OpenAIErrorWrapperLocal(err, "unmarshal_model_mapping_failed", http.StatusInternalServerError)
|
|
|
+ }
|
|
|
+ if modelMap[embeddingRequest.Model] != "" {
|
|
|
+ embeddingRequest.Model = modelMap[embeddingRequest.Model]
|
|
|
+ // set upstream model name
|
|
|
+ //isModelMapped = true
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ relayInfo.UpstreamModelName = embeddingRequest.Model
|
|
|
+ modelPrice, success := common.GetModelPrice(embeddingRequest.Model, false)
|
|
|
+ groupRatio := setting.GetGroupRatio(relayInfo.Group)
|
|
|
+
|
|
|
+ var preConsumedQuota int
|
|
|
+ var ratio float64
|
|
|
+ var modelRatio float64
|
|
|
+
|
|
|
+ promptToken := getEmbeddingPromptToken(*embeddingRequest)
|
|
|
+ if !success {
|
|
|
+ preConsumedTokens := promptToken
|
|
|
+ modelRatio = common.GetModelRatio(embeddingRequest.Model)
|
|
|
+ ratio = modelRatio * groupRatio
|
|
|
+ preConsumedQuota = int(float64(preConsumedTokens) * ratio)
|
|
|
+ } else {
|
|
|
+ preConsumedQuota = int(modelPrice * common.QuotaPerUnit * groupRatio)
|
|
|
+ }
|
|
|
+ relayInfo.PromptTokens = promptToken
|
|
|
+
|
|
|
+ // pre-consume quota 预消耗配额
|
|
|
+ preConsumedQuota, userQuota, openaiErr := preConsumeQuota(c, preConsumedQuota, relayInfo)
|
|
|
+ if openaiErr != nil {
|
|
|
+ return openaiErr
|
|
|
+ }
|
|
|
+ defer func() {
|
|
|
+ if openaiErr != nil {
|
|
|
+ returnPreConsumedQuota(c, relayInfo, userQuota, preConsumedQuota)
|
|
|
+ }
|
|
|
+ }()
|
|
|
+
|
|
|
+ adaptor := GetAdaptor(relayInfo.ApiType)
|
|
|
+ if adaptor == nil {
|
|
|
+ return service.OpenAIErrorWrapperLocal(fmt.Errorf("invalid api type: %d", relayInfo.ApiType), "invalid_api_type", http.StatusBadRequest)
|
|
|
+ }
|
|
|
+ adaptor.Init(relayInfo)
|
|
|
+
|
|
|
+ convertedRequest, err := adaptor.ConvertEmbeddingRequest(c,relayInfo,*embeddingRequest)
|
|
|
+
|
|
|
+ if err != nil {
|
|
|
+ return service.OpenAIErrorWrapperLocal(err, "convert_request_failed", http.StatusInternalServerError)
|
|
|
+ }
|
|
|
+ jsonData, err := json.Marshal(convertedRequest)
|
|
|
+ if err != nil {
|
|
|
+ return service.OpenAIErrorWrapperLocal(err, "json_marshal_failed", http.StatusInternalServerError)
|
|
|
+ }
|
|
|
+ requestBody := bytes.NewBuffer(jsonData)
|
|
|
+ statusCodeMappingStr := c.GetString("status_code_mapping")
|
|
|
+ resp, err := adaptor.DoRequest(c,relayInfo, requestBody)
|
|
|
+ if err != nil {
|
|
|
+ return service.OpenAIErrorWrapper(err, "do_request_failed", http.StatusInternalServerError)
|
|
|
+ }
|
|
|
+
|
|
|
+ var httpResp *http.Response
|
|
|
+ if resp != nil {
|
|
|
+ httpResp = resp.(*http.Response)
|
|
|
+ if httpResp.StatusCode != http.StatusOK {
|
|
|
+ openaiErr = service.RelayErrorHandler(httpResp)
|
|
|
+ // reset status code 重置状态码
|
|
|
+ service.ResetStatusCode(openaiErr, statusCodeMappingStr)
|
|
|
+ return openaiErr
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ usage, openaiErr := adaptor.DoResponse(c, httpResp, relayInfo)
|
|
|
+ if openaiErr != nil {
|
|
|
+ // reset status code 重置状态码
|
|
|
+ service.ResetStatusCode(openaiErr, statusCodeMappingStr)
|
|
|
+ return openaiErr
|
|
|
+ }
|
|
|
+ postConsumeQuota(c, relayInfo, embeddingRequest.Model, usage.(*dto.Usage), ratio, preConsumedQuota, userQuota, modelRatio, groupRatio, modelPrice, success, "")
|
|
|
+ return nil
|
|
|
+}
|