adaptor.go 9.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332
  1. package vertex
  2. import (
  3. "encoding/json"
  4. "errors"
  5. "fmt"
  6. "io"
  7. "net/http"
  8. "one-api/dto"
  9. "one-api/relay/channel"
  10. "one-api/relay/channel/claude"
  11. "one-api/relay/channel/gemini"
  12. "one-api/relay/channel/openai"
  13. relaycommon "one-api/relay/common"
  14. "one-api/relay/constant"
  15. "one-api/setting/model_setting"
  16. "one-api/types"
  17. "strings"
  18. "github.com/gin-gonic/gin"
  19. )
  20. const (
  21. RequestModeClaude = 1
  22. RequestModeGemini = 2
  23. RequestModeLlama = 3
  24. )
  25. var claudeModelMap = map[string]string{
  26. "claude-3-sonnet-20240229": "claude-3-sonnet@20240229",
  27. "claude-3-opus-20240229": "claude-3-opus@20240229",
  28. "claude-3-haiku-20240307": "claude-3-haiku@20240307",
  29. "claude-3-5-sonnet-20240620": "claude-3-5-sonnet@20240620",
  30. "claude-3-5-sonnet-20241022": "claude-3-5-sonnet-v2@20241022",
  31. "claude-3-7-sonnet-20250219": "claude-3-7-sonnet@20250219",
  32. "claude-sonnet-4-20250514": "claude-sonnet-4@20250514",
  33. "claude-opus-4-20250514": "claude-opus-4@20250514",
  34. "claude-opus-4-1-20250805": "claude-opus-4-1@20250805",
  35. }
  36. const anthropicVersion = "vertex-2023-10-16"
  37. type Adaptor struct {
  38. RequestMode int
  39. AccountCredentials Credentials
  40. }
  41. func (a *Adaptor) ConvertGeminiRequest(c *gin.Context, info *relaycommon.RelayInfo, request *dto.GeminiChatRequest) (any, error) {
  42. geminiAdaptor := gemini.Adaptor{}
  43. return geminiAdaptor.ConvertGeminiRequest(c, info, request)
  44. }
  45. func (a *Adaptor) ConvertClaudeRequest(c *gin.Context, info *relaycommon.RelayInfo, request *dto.ClaudeRequest) (any, error) {
  46. if v, ok := claudeModelMap[info.UpstreamModelName]; ok {
  47. c.Set("request_model", v)
  48. } else {
  49. c.Set("request_model", request.Model)
  50. }
  51. vertexClaudeReq := copyRequest(request, anthropicVersion)
  52. return vertexClaudeReq, nil
  53. }
  54. func (a *Adaptor) ConvertAudioRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.AudioRequest) (io.Reader, error) {
  55. //TODO implement me
  56. return nil, errors.New("not implemented")
  57. }
  58. func (a *Adaptor) ConvertImageRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.ImageRequest) (any, error) {
  59. geminiAdaptor := gemini.Adaptor{}
  60. return geminiAdaptor.ConvertImageRequest(c, info, request)
  61. }
  62. func (a *Adaptor) Init(info *relaycommon.RelayInfo) {
  63. if strings.HasPrefix(info.UpstreamModelName, "claude") {
  64. a.RequestMode = RequestModeClaude
  65. } else if strings.Contains(info.UpstreamModelName, "llama") {
  66. a.RequestMode = RequestModeLlama
  67. } else {
  68. a.RequestMode = RequestModeGemini
  69. }
  70. }
  71. func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) {
  72. adc := &Credentials{}
  73. if err := json.Unmarshal([]byte(info.ApiKey), adc); err != nil {
  74. return "", fmt.Errorf("failed to decode credentials file: %w", err)
  75. }
  76. region := GetModelRegion(info.ApiVersion, info.OriginModelName)
  77. a.AccountCredentials = *adc
  78. suffix := ""
  79. if a.RequestMode == RequestModeGemini {
  80. if model_setting.GetGeminiSettings().ThinkingAdapterEnabled {
  81. // 新增逻辑:处理 -thinking-<budget> 格式
  82. if strings.Contains(info.UpstreamModelName, "-thinking-") {
  83. parts := strings.Split(info.UpstreamModelName, "-thinking-")
  84. info.UpstreamModelName = parts[0]
  85. } else if strings.HasSuffix(info.UpstreamModelName, "-thinking") { // 旧的适配
  86. info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking")
  87. } else if strings.HasSuffix(info.UpstreamModelName, "-nothinking") {
  88. info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-nothinking")
  89. }
  90. }
  91. if info.IsStream {
  92. suffix = "streamGenerateContent?alt=sse"
  93. } else {
  94. suffix = "generateContent"
  95. }
  96. if strings.HasPrefix(info.UpstreamModelName, "imagen") {
  97. suffix = "predict"
  98. }
  99. if region == "global" {
  100. return fmt.Sprintf(
  101. "https://aiplatform.googleapis.com/v1/projects/%s/locations/global/publishers/google/models/%s:%s",
  102. adc.ProjectID,
  103. info.UpstreamModelName,
  104. suffix,
  105. ), nil
  106. } else {
  107. return fmt.Sprintf(
  108. "https://%s-aiplatform.googleapis.com/v1/projects/%s/locations/%s/publishers/google/models/%s:%s",
  109. region,
  110. adc.ProjectID,
  111. region,
  112. info.UpstreamModelName,
  113. suffix,
  114. ), nil
  115. }
  116. } else if a.RequestMode == RequestModeClaude {
  117. if info.IsStream {
  118. suffix = "streamRawPredict?alt=sse"
  119. } else {
  120. suffix = "rawPredict"
  121. }
  122. model := info.UpstreamModelName
  123. if v, ok := claudeModelMap[info.UpstreamModelName]; ok {
  124. model = v
  125. }
  126. if region == "global" {
  127. return fmt.Sprintf(
  128. "https://aiplatform.googleapis.com/v1/projects/%s/locations/global/publishers/anthropic/models/%s:%s",
  129. adc.ProjectID,
  130. model,
  131. suffix,
  132. ), nil
  133. } else {
  134. return fmt.Sprintf(
  135. "https://%s-aiplatform.googleapis.com/v1/projects/%s/locations/%s/publishers/anthropic/models/%s:%s",
  136. region,
  137. adc.ProjectID,
  138. region,
  139. model,
  140. suffix,
  141. ), nil
  142. }
  143. } else if a.RequestMode == RequestModeLlama {
  144. return fmt.Sprintf(
  145. "https://%s-aiplatform.googleapis.com/v1beta1/projects/%s/locations/%s/endpoints/openapi/chat/completions",
  146. region,
  147. adc.ProjectID,
  148. region,
  149. ), nil
  150. }
  151. return "", errors.New("unsupported request mode")
  152. }
  153. func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Header, info *relaycommon.RelayInfo) error {
  154. channel.SetupApiRequestHeader(info, c, req)
  155. accessToken, err := getAccessToken(a, info)
  156. if err != nil {
  157. return err
  158. }
  159. req.Set("Authorization", "Bearer "+accessToken)
  160. req.Set("x-goog-user-project", a.AccountCredentials.ProjectID)
  161. return nil
  162. }
  163. func (a *Adaptor) ConvertOpenAIRequest(c *gin.Context, info *relaycommon.RelayInfo, request *dto.GeneralOpenAIRequest) (any, error) {
  164. if request == nil {
  165. return nil, errors.New("request is nil")
  166. }
  167. if a.RequestMode == RequestModeGemini && strings.HasPrefix(info.UpstreamModelName, "imagen") {
  168. prompt := ""
  169. for _, m := range request.Messages {
  170. if m.Role == "user" {
  171. prompt = m.StringContent()
  172. if prompt != "" {
  173. break
  174. }
  175. }
  176. }
  177. if prompt == "" {
  178. if p, ok := request.Prompt.(string); ok {
  179. prompt = p
  180. }
  181. }
  182. if prompt == "" {
  183. return nil, errors.New("prompt is required for image generation")
  184. }
  185. imgReq := dto.ImageRequest{
  186. Model: request.Model,
  187. Prompt: prompt,
  188. N: 1,
  189. Size: "1024x1024",
  190. }
  191. if request.N > 0 {
  192. imgReq.N = uint(request.N)
  193. }
  194. if request.Size != "" {
  195. imgReq.Size = request.Size
  196. }
  197. if len(request.ExtraBody) > 0 {
  198. var extra map[string]any
  199. if err := json.Unmarshal(request.ExtraBody, &extra); err == nil {
  200. if n, ok := extra["n"].(float64); ok && n > 0 {
  201. imgReq.N = uint(n)
  202. }
  203. if size, ok := extra["size"].(string); ok {
  204. imgReq.Size = size
  205. }
  206. // accept aspectRatio in extra body (top-level or under parameters)
  207. if ar, ok := extra["aspectRatio"].(string); ok && ar != "" {
  208. imgReq.Size = ar
  209. }
  210. if params, ok := extra["parameters"].(map[string]any); ok {
  211. if ar, ok := params["aspectRatio"].(string); ok && ar != "" {
  212. imgReq.Size = ar
  213. }
  214. }
  215. }
  216. }
  217. c.Set("request_model", request.Model)
  218. return a.ConvertImageRequest(c, info, imgReq)
  219. }
  220. if a.RequestMode == RequestModeClaude {
  221. claudeReq, err := claude.RequestOpenAI2ClaudeMessage(c, *request)
  222. if err != nil {
  223. return nil, err
  224. }
  225. vertexClaudeReq := copyRequest(claudeReq, anthropicVersion)
  226. c.Set("request_model", claudeReq.Model)
  227. info.UpstreamModelName = claudeReq.Model
  228. return vertexClaudeReq, nil
  229. } else if a.RequestMode == RequestModeGemini {
  230. geminiRequest, err := gemini.CovertGemini2OpenAI(c, *request, info)
  231. if err != nil {
  232. return nil, err
  233. }
  234. c.Set("request_model", request.Model)
  235. return geminiRequest, nil
  236. } else if a.RequestMode == RequestModeLlama {
  237. return request, nil
  238. }
  239. return nil, errors.New("unsupported request mode")
  240. }
  241. func (a *Adaptor) ConvertRerankRequest(c *gin.Context, relayMode int, request dto.RerankRequest) (any, error) {
  242. return nil, nil
  243. }
  244. func (a *Adaptor) ConvertEmbeddingRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.EmbeddingRequest) (any, error) {
  245. //TODO implement me
  246. return nil, errors.New("not implemented")
  247. }
  248. func (a *Adaptor) ConvertOpenAIResponsesRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.OpenAIResponsesRequest) (any, error) {
  249. // TODO implement me
  250. return nil, errors.New("not implemented")
  251. }
  252. func (a *Adaptor) DoRequest(c *gin.Context, info *relaycommon.RelayInfo, requestBody io.Reader) (any, error) {
  253. return channel.DoApiRequest(a, c, info, requestBody)
  254. }
  255. func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, info *relaycommon.RelayInfo) (usage any, err *types.NewAPIError) {
  256. if info.IsStream {
  257. switch a.RequestMode {
  258. case RequestModeClaude:
  259. return claude.ClaudeStreamHandler(c, resp, info, claude.RequestModeMessage)
  260. case RequestModeGemini:
  261. if info.RelayMode == constant.RelayModeGemini {
  262. return gemini.GeminiTextGenerationStreamHandler(c, info, resp)
  263. } else {
  264. return gemini.GeminiChatStreamHandler(c, info, resp)
  265. }
  266. case RequestModeLlama:
  267. return openai.OaiStreamHandler(c, info, resp)
  268. }
  269. } else {
  270. switch a.RequestMode {
  271. case RequestModeClaude:
  272. return claude.ClaudeHandler(c, resp, info, claude.RequestModeMessage)
  273. case RequestModeGemini:
  274. if info.RelayMode == constant.RelayModeGemini {
  275. return gemini.GeminiTextGenerationHandler(c, info, resp)
  276. } else {
  277. if strings.HasPrefix(info.UpstreamModelName, "imagen") {
  278. return gemini.GeminiImageHandler(c, info, resp)
  279. }
  280. return gemini.GeminiChatHandler(c, info, resp)
  281. }
  282. case RequestModeLlama:
  283. return openai.OpenaiHandler(c, info, resp)
  284. }
  285. }
  286. return
  287. }
  288. func (a *Adaptor) GetModelList() []string {
  289. var modelList []string
  290. for i, s := range ModelList {
  291. modelList = append(modelList, s)
  292. ModelList[i] = s
  293. }
  294. for i, s := range claude.ModelList {
  295. modelList = append(modelList, s)
  296. claude.ModelList[i] = s
  297. }
  298. for i, s := range gemini.ModelList {
  299. modelList = append(modelList, s)
  300. gemini.ModelList[i] = s
  301. }
  302. return modelList
  303. }
  304. func (a *Adaptor) GetChannelName() string {
  305. return ChannelName
  306. }