relay-gemini-native.go 3.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105
  1. package gemini
  2. import (
  3. "io"
  4. "net/http"
  5. "github.com/QuantumNous/new-api/common"
  6. "github.com/QuantumNous/new-api/dto"
  7. "github.com/QuantumNous/new-api/logger"
  8. relaycommon "github.com/QuantumNous/new-api/relay/common"
  9. "github.com/QuantumNous/new-api/relay/helper"
  10. "github.com/QuantumNous/new-api/service"
  11. "github.com/QuantumNous/new-api/types"
  12. "github.com/gin-gonic/gin"
  13. )
  14. func GeminiTextGenerationHandler(c *gin.Context, info *relaycommon.RelayInfo, resp *http.Response) (*dto.Usage, *types.NewAPIError) {
  15. defer service.CloseResponseBodyGracefully(resp)
  16. // 读取响应体
  17. responseBody, err := io.ReadAll(resp.Body)
  18. if err != nil {
  19. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  20. }
  21. if common.DebugEnabled {
  22. println(string(responseBody))
  23. }
  24. // 解析为 Gemini 原生响应格式
  25. var geminiResponse dto.GeminiChatResponse
  26. err = common.Unmarshal(responseBody, &geminiResponse)
  27. if err != nil {
  28. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  29. }
  30. // 计算使用量(基于 UsageMetadata)
  31. usage := dto.Usage{
  32. PromptTokens: geminiResponse.UsageMetadata.PromptTokenCount,
  33. CompletionTokens: geminiResponse.UsageMetadata.CandidatesTokenCount + geminiResponse.UsageMetadata.ThoughtsTokenCount,
  34. TotalTokens: geminiResponse.UsageMetadata.TotalTokenCount,
  35. }
  36. usage.CompletionTokenDetails.ReasoningTokens = geminiResponse.UsageMetadata.ThoughtsTokenCount
  37. for _, detail := range geminiResponse.UsageMetadata.PromptTokensDetails {
  38. if detail.Modality == "AUDIO" {
  39. usage.PromptTokensDetails.AudioTokens = detail.TokenCount
  40. } else if detail.Modality == "TEXT" {
  41. usage.PromptTokensDetails.TextTokens = detail.TokenCount
  42. }
  43. }
  44. service.IOCopyBytesGracefully(c, resp, responseBody)
  45. return &usage, nil
  46. }
  47. func NativeGeminiEmbeddingHandler(c *gin.Context, resp *http.Response, info *relaycommon.RelayInfo) (*dto.Usage, *types.NewAPIError) {
  48. defer service.CloseResponseBodyGracefully(resp)
  49. responseBody, err := io.ReadAll(resp.Body)
  50. if err != nil {
  51. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  52. }
  53. if common.DebugEnabled {
  54. println(string(responseBody))
  55. }
  56. usage := service.ResponseText2Usage(c, "", info.UpstreamModelName, info.GetEstimatePromptTokens())
  57. if info.IsGeminiBatchEmbedding {
  58. var geminiResponse dto.GeminiBatchEmbeddingResponse
  59. err = common.Unmarshal(responseBody, &geminiResponse)
  60. if err != nil {
  61. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  62. }
  63. } else {
  64. var geminiResponse dto.GeminiEmbeddingResponse
  65. err = common.Unmarshal(responseBody, &geminiResponse)
  66. if err != nil {
  67. return nil, types.NewOpenAIError(err, types.ErrorCodeBadResponseBody, http.StatusInternalServerError)
  68. }
  69. }
  70. service.IOCopyBytesGracefully(c, resp, responseBody)
  71. return usage, nil
  72. }
  73. func GeminiTextGenerationStreamHandler(c *gin.Context, info *relaycommon.RelayInfo, resp *http.Response) (*dto.Usage, *types.NewAPIError) {
  74. helper.SetEventStreamHeaders(c)
  75. return geminiStreamHandler(c, info, resp, func(data string, geminiResponse *dto.GeminiChatResponse) bool {
  76. err := helper.StringData(c, data)
  77. if err != nil {
  78. logger.LogError(c, "failed to write stream data: "+err.Error())
  79. return false
  80. }
  81. info.SendResponseCount++
  82. return true
  83. })
  84. }