relay-gemini-native.go 6.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211
  1. package gemini
  2. import (
  3. "io"
  4. "net/http"
  5. "one-api/common"
  6. "one-api/dto"
  7. "one-api/logger"
  8. relaycommon "one-api/relay/common"
  9. "one-api/relay/helper"
  10. "one-api/service"
  11. "one-api/types"
  12. "strings"
  13. "github.com/pkg/errors"
  14. "github.com/gin-gonic/gin"
  15. )
  16. func GeminiTextGenerationHandler(c *gin.Context, info *relaycommon.RelayInfo, resp *http.Response) (*dto.Usage, *types.NewAPIError) {
  17. defer service.CloseResponseBodyGracefully(resp)
  18. // 读取响应体
  19. responseBody, err := io.ReadAll(resp.Body)
  20. if err != nil {
  21. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  22. }
  23. if common.DebugEnabled {
  24. println(string(responseBody))
  25. }
  26. // 解析为 Gemini 原生响应格式
  27. var geminiResponse dto.GeminiChatResponse
  28. err = common.Unmarshal(responseBody, &geminiResponse)
  29. if err != nil {
  30. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  31. }
  32. // 计算使用量(基于 UsageMetadata)
  33. usage := dto.Usage{
  34. PromptTokens: geminiResponse.UsageMetadata.PromptTokenCount,
  35. CompletionTokens: geminiResponse.UsageMetadata.CandidatesTokenCount + geminiResponse.UsageMetadata.ThoughtsTokenCount,
  36. TotalTokens: geminiResponse.UsageMetadata.TotalTokenCount,
  37. }
  38. usage.CompletionTokenDetails.ReasoningTokens = geminiResponse.UsageMetadata.ThoughtsTokenCount
  39. if strings.HasPrefix(info.UpstreamModelName, "gemini-2.5-flash-image-preview") {
  40. imageOutputCounts := 0
  41. for _, candidate := range geminiResponse.Candidates {
  42. for _, part := range candidate.Content.Parts {
  43. if part.InlineData != nil && strings.HasPrefix(part.InlineData.MimeType, "image/") {
  44. imageOutputCounts++
  45. }
  46. }
  47. }
  48. if imageOutputCounts != 0 {
  49. usage.CompletionTokens = usage.CompletionTokens - imageOutputCounts*1290
  50. usage.TotalTokens = usage.TotalTokens - imageOutputCounts*1290
  51. c.Set("gemini_image_tokens", imageOutputCounts*1290)
  52. }
  53. }
  54. // if strings.HasPrefix(info.UpstreamModelName, "gemini-2.5-flash-image-preview") {
  55. // for _, detail := range geminiResponse.UsageMetadata.CandidatesTokensDetails {
  56. // if detail.Modality == "IMAGE" {
  57. // usage.CompletionTokens = usage.CompletionTokens - detail.TokenCount
  58. // usage.TotalTokens = usage.TotalTokens - detail.TokenCount
  59. // c.Set("gemini_image_tokens", detail.TokenCount)
  60. // }
  61. // }
  62. // }
  63. for _, detail := range geminiResponse.UsageMetadata.PromptTokensDetails {
  64. if detail.Modality == "AUDIO" {
  65. usage.PromptTokensDetails.AudioTokens = detail.TokenCount
  66. } else if detail.Modality == "TEXT" {
  67. usage.PromptTokensDetails.TextTokens = detail.TokenCount
  68. }
  69. }
  70. service.IOCopyBytesGracefully(c, resp, responseBody)
  71. return &usage, nil
  72. }
  73. func NativeGeminiEmbeddingHandler(c *gin.Context, resp *http.Response, info *relaycommon.RelayInfo) (*dto.Usage, *types.NewAPIError) {
  74. defer service.CloseResponseBodyGracefully(resp)
  75. responseBody, err := io.ReadAll(resp.Body)
  76. if err != nil {
  77. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  78. }
  79. if common.DebugEnabled {
  80. println(string(responseBody))
  81. }
  82. usage := &dto.Usage{
  83. PromptTokens: info.PromptTokens,
  84. TotalTokens: info.PromptTokens,
  85. }
  86. if info.IsGeminiBatchEmbedding {
  87. var geminiResponse dto.GeminiBatchEmbeddingResponse
  88. err = common.Unmarshal(responseBody, &geminiResponse)
  89. if err != nil {
  90. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  91. }
  92. } else {
  93. var geminiResponse dto.GeminiEmbeddingResponse
  94. err = common.Unmarshal(responseBody, &geminiResponse)
  95. if err != nil {
  96. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  97. }
  98. }
  99. service.IOCopyBytesGracefully(c, resp, responseBody)
  100. return usage, nil
  101. }
  102. func GeminiTextGenerationStreamHandler(c *gin.Context, info *relaycommon.RelayInfo, resp *http.Response) (*dto.Usage, *types.NewAPIError) {
  103. var usage = &dto.Usage{}
  104. var imageCount int
  105. helper.SetEventStreamHeaders(c)
  106. responseText := strings.Builder{}
  107. helper.StreamScannerHandler(c, resp, info, func(data string) bool {
  108. var geminiResponse dto.GeminiChatResponse
  109. err := common.UnmarshalJsonStr(data, &geminiResponse)
  110. if err != nil {
  111. logger.LogError(c, "error unmarshalling stream response: "+err.Error())
  112. return false
  113. }
  114. // 统计图片数量
  115. for _, candidate := range geminiResponse.Candidates {
  116. for _, part := range candidate.Content.Parts {
  117. if part.InlineData != nil && part.InlineData.MimeType != "" {
  118. imageCount++
  119. }
  120. if part.Text != "" {
  121. responseText.WriteString(part.Text)
  122. }
  123. }
  124. }
  125. // 更新使用量统计
  126. if geminiResponse.UsageMetadata.TotalTokenCount != 0 {
  127. usage.PromptTokens = geminiResponse.UsageMetadata.PromptTokenCount
  128. usage.CompletionTokens = geminiResponse.UsageMetadata.CandidatesTokenCount + geminiResponse.UsageMetadata.ThoughtsTokenCount
  129. usage.TotalTokens = geminiResponse.UsageMetadata.TotalTokenCount
  130. usage.CompletionTokenDetails.ReasoningTokens = geminiResponse.UsageMetadata.ThoughtsTokenCount
  131. for _, detail := range geminiResponse.UsageMetadata.PromptTokensDetails {
  132. if detail.Modality == "AUDIO" {
  133. usage.PromptTokensDetails.AudioTokens = detail.TokenCount
  134. } else if detail.Modality == "TEXT" {
  135. usage.PromptTokensDetails.TextTokens = detail.TokenCount
  136. }
  137. }
  138. if strings.HasPrefix(info.UpstreamModelName, "gemini-2.5-flash-image-preview") {
  139. for _, detail := range geminiResponse.UsageMetadata.CandidatesTokensDetails {
  140. if detail.Modality == "IMAGE" {
  141. usage.CompletionTokens = usage.CompletionTokens - detail.TokenCount
  142. usage.TotalTokens = usage.TotalTokens - detail.TokenCount
  143. c.Set("gemini_image_tokens", detail.TokenCount)
  144. }
  145. }
  146. }
  147. }
  148. // 直接发送 GeminiChatResponse 响应
  149. err = helper.StringData(c, data)
  150. if err != nil {
  151. logger.LogError(c, err.Error())
  152. }
  153. info.SendResponseCount++
  154. return true
  155. })
  156. if info.SendResponseCount == 0 {
  157. return nil, types.NewOpenAIError(errors.New("no response received from Gemini API"), types.ErrorCodeEmptyResponse, http.StatusInternalServerError)
  158. }
  159. if imageCount != 0 {
  160. if usage.CompletionTokens == 0 {
  161. usage.CompletionTokens = imageCount * 258
  162. }
  163. }
  164. // 如果usage.CompletionTokens为0,则使用本地统计的completion tokens
  165. if usage.CompletionTokens == 0 {
  166. str := responseText.String()
  167. if len(str) > 0 {
  168. usage = service.ResponseText2Usage(responseText.String(), info.UpstreamModelName, info.PromptTokens)
  169. } else {
  170. // 空补全,不需要使用量
  171. usage = &dto.Usage{}
  172. }
  173. }
  174. // 移除流式响应结尾的[Done],因为Gemini API没有发送Done的行为
  175. //helper.Done(c)
  176. return usage, nil
  177. }