convert.go 31 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962
  1. package service
  2. import (
  3. "encoding/json"
  4. "fmt"
  5. "strings"
  6. "github.com/QuantumNous/new-api/common"
  7. "github.com/QuantumNous/new-api/constant"
  8. "github.com/QuantumNous/new-api/dto"
  9. "github.com/QuantumNous/new-api/relay/channel/openrouter"
  10. relaycommon "github.com/QuantumNous/new-api/relay/common"
  11. "github.com/QuantumNous/new-api/relay/reasonmap"
  12. )
  13. func ClaudeToOpenAIRequest(claudeRequest dto.ClaudeRequest, info *relaycommon.RelayInfo) (*dto.GeneralOpenAIRequest, error) {
  14. openAIRequest := dto.GeneralOpenAIRequest{
  15. Model: claudeRequest.Model,
  16. MaxTokens: claudeRequest.MaxTokens,
  17. Temperature: claudeRequest.Temperature,
  18. TopP: claudeRequest.TopP,
  19. Stream: claudeRequest.Stream,
  20. }
  21. isOpenRouter := info.ChannelType == constant.ChannelTypeOpenRouter
  22. if claudeRequest.Thinking != nil && claudeRequest.Thinking.Type == "enabled" {
  23. if isOpenRouter {
  24. reasoning := openrouter.RequestReasoning{
  25. MaxTokens: claudeRequest.Thinking.GetBudgetTokens(),
  26. }
  27. reasoningJSON, err := json.Marshal(reasoning)
  28. if err != nil {
  29. return nil, fmt.Errorf("failed to marshal reasoning: %w", err)
  30. }
  31. openAIRequest.Reasoning = reasoningJSON
  32. } else {
  33. thinkingSuffix := "-thinking"
  34. if strings.HasSuffix(info.OriginModelName, thinkingSuffix) &&
  35. !strings.HasSuffix(openAIRequest.Model, thinkingSuffix) {
  36. openAIRequest.Model = openAIRequest.Model + thinkingSuffix
  37. }
  38. }
  39. }
  40. // Convert stop sequences
  41. if len(claudeRequest.StopSequences) == 1 {
  42. openAIRequest.Stop = claudeRequest.StopSequences[0]
  43. } else if len(claudeRequest.StopSequences) > 1 {
  44. openAIRequest.Stop = claudeRequest.StopSequences
  45. }
  46. // Convert tools
  47. tools, _ := common.Any2Type[[]dto.Tool](claudeRequest.Tools)
  48. openAITools := make([]dto.ToolCallRequest, 0)
  49. for _, claudeTool := range tools {
  50. openAITool := dto.ToolCallRequest{
  51. Type: "function",
  52. Function: dto.FunctionRequest{
  53. Name: claudeTool.Name,
  54. Description: claudeTool.Description,
  55. Parameters: claudeTool.InputSchema,
  56. },
  57. }
  58. openAITools = append(openAITools, openAITool)
  59. }
  60. openAIRequest.Tools = openAITools
  61. // Convert messages
  62. openAIMessages := make([]dto.Message, 0)
  63. // Add system message if present
  64. if claudeRequest.System != nil {
  65. if claudeRequest.IsStringSystem() && claudeRequest.GetStringSystem() != "" {
  66. openAIMessage := dto.Message{
  67. Role: "system",
  68. }
  69. openAIMessage.SetStringContent(claudeRequest.GetStringSystem())
  70. openAIMessages = append(openAIMessages, openAIMessage)
  71. } else {
  72. systems := claudeRequest.ParseSystem()
  73. if len(systems) > 0 {
  74. openAIMessage := dto.Message{
  75. Role: "system",
  76. }
  77. isOpenRouterClaude := isOpenRouter && strings.HasPrefix(info.UpstreamModelName, "anthropic/claude")
  78. if isOpenRouterClaude {
  79. systemMediaMessages := make([]dto.MediaContent, 0, len(systems))
  80. for _, system := range systems {
  81. message := dto.MediaContent{
  82. Type: "text",
  83. Text: system.GetText(),
  84. CacheControl: system.CacheControl,
  85. }
  86. systemMediaMessages = append(systemMediaMessages, message)
  87. }
  88. openAIMessage.SetMediaContent(systemMediaMessages)
  89. } else {
  90. systemStr := ""
  91. for _, system := range systems {
  92. if system.Text != nil {
  93. systemStr += *system.Text
  94. }
  95. }
  96. openAIMessage.SetStringContent(systemStr)
  97. }
  98. openAIMessages = append(openAIMessages, openAIMessage)
  99. }
  100. }
  101. }
  102. for _, claudeMessage := range claudeRequest.Messages {
  103. openAIMessage := dto.Message{
  104. Role: claudeMessage.Role,
  105. }
  106. //log.Printf("claudeMessage.Content: %v", claudeMessage.Content)
  107. if claudeMessage.IsStringContent() {
  108. openAIMessage.SetStringContent(claudeMessage.GetStringContent())
  109. } else {
  110. content, err := claudeMessage.ParseContent()
  111. if err != nil {
  112. return nil, err
  113. }
  114. contents := content
  115. var toolCalls []dto.ToolCallRequest
  116. mediaMessages := make([]dto.MediaContent, 0, len(contents))
  117. for _, mediaMsg := range contents {
  118. switch mediaMsg.Type {
  119. case "text", "input_text":
  120. message := dto.MediaContent{
  121. Type: "text",
  122. Text: mediaMsg.GetText(),
  123. CacheControl: mediaMsg.CacheControl,
  124. }
  125. mediaMessages = append(mediaMessages, message)
  126. case "image":
  127. // Handle image conversion (base64 to URL or keep as is)
  128. imageData := fmt.Sprintf("data:%s;base64,%s", mediaMsg.Source.MediaType, mediaMsg.Source.Data)
  129. //textContent += fmt.Sprintf("[Image: %s]", imageData)
  130. mediaMessage := dto.MediaContent{
  131. Type: "image_url",
  132. ImageUrl: &dto.MessageImageUrl{Url: imageData},
  133. }
  134. mediaMessages = append(mediaMessages, mediaMessage)
  135. case "tool_use":
  136. toolCall := dto.ToolCallRequest{
  137. ID: mediaMsg.Id,
  138. Type: "function",
  139. Function: dto.FunctionRequest{
  140. Name: mediaMsg.Name,
  141. Arguments: toJSONString(mediaMsg.Input),
  142. },
  143. }
  144. toolCalls = append(toolCalls, toolCall)
  145. case "tool_result":
  146. // Add tool result as a separate message
  147. toolName := mediaMsg.Name
  148. if toolName == "" {
  149. toolName = claudeRequest.SearchToolNameByToolCallId(mediaMsg.ToolUseId)
  150. }
  151. oaiToolMessage := dto.Message{
  152. Role: "tool",
  153. Name: &toolName,
  154. ToolCallId: mediaMsg.ToolUseId,
  155. }
  156. //oaiToolMessage.SetStringContent(*mediaMsg.GetMediaContent().Text)
  157. if mediaMsg.IsStringContent() {
  158. oaiToolMessage.SetStringContent(mediaMsg.GetStringContent())
  159. } else {
  160. mediaContents := mediaMsg.ParseMediaContent()
  161. encodeJson, _ := common.Marshal(mediaContents)
  162. oaiToolMessage.SetStringContent(string(encodeJson))
  163. }
  164. openAIMessages = append(openAIMessages, oaiToolMessage)
  165. }
  166. }
  167. if len(toolCalls) > 0 {
  168. openAIMessage.SetToolCalls(toolCalls)
  169. }
  170. if len(mediaMessages) > 0 && len(toolCalls) == 0 {
  171. openAIMessage.SetMediaContent(mediaMessages)
  172. }
  173. }
  174. if len(openAIMessage.ParseContent()) > 0 || len(openAIMessage.ToolCalls) > 0 {
  175. openAIMessages = append(openAIMessages, openAIMessage)
  176. }
  177. }
  178. openAIRequest.Messages = openAIMessages
  179. return &openAIRequest, nil
  180. }
  181. func generateStopBlock(index int) *dto.ClaudeResponse {
  182. return &dto.ClaudeResponse{
  183. Type: "content_block_stop",
  184. Index: common.GetPointer[int](index),
  185. }
  186. }
  187. func StreamResponseOpenAI2Claude(openAIResponse *dto.ChatCompletionsStreamResponse, info *relaycommon.RelayInfo) []*dto.ClaudeResponse {
  188. if info.ClaudeConvertInfo.Done {
  189. return nil
  190. }
  191. var claudeResponses []*dto.ClaudeResponse
  192. // stopOpenBlocks emits the required content_block_stop event(s) for the currently open block(s)
  193. // according to Anthropic's SSE streaming state machine:
  194. // content_block_start -> content_block_delta* -> content_block_stop (per index).
  195. //
  196. // For text/thinking, there is at most one open block at info.ClaudeConvertInfo.Index.
  197. // For tools, OpenAI tool_calls can stream multiple parallel tool_use blocks (indexed from 0),
  198. // so we may have multiple open blocks and must stop each one explicitly.
  199. stopOpenBlocks := func() {
  200. switch info.ClaudeConvertInfo.LastMessagesType {
  201. case relaycommon.LastMessageTypeText, relaycommon.LastMessageTypeThinking:
  202. claudeResponses = append(claudeResponses, generateStopBlock(info.ClaudeConvertInfo.Index))
  203. case relaycommon.LastMessageTypeTools:
  204. base := info.ClaudeConvertInfo.ToolCallBaseIndex
  205. for offset := 0; offset <= info.ClaudeConvertInfo.ToolCallMaxIndexOffset; offset++ {
  206. claudeResponses = append(claudeResponses, generateStopBlock(base+offset))
  207. }
  208. }
  209. }
  210. // stopOpenBlocksAndAdvance closes the currently open block(s) and advances the content block index
  211. // to the next available slot for subsequent content_block_start events.
  212. //
  213. // This prevents invalid streams where a content_block_delta (e.g. thinking_delta) is emitted for an
  214. // index whose active content_block type is different (the typical cause of "Mismatched content block type").
  215. stopOpenBlocksAndAdvance := func() {
  216. if info.ClaudeConvertInfo.LastMessagesType == relaycommon.LastMessageTypeNone {
  217. return
  218. }
  219. stopOpenBlocks()
  220. switch info.ClaudeConvertInfo.LastMessagesType {
  221. case relaycommon.LastMessageTypeTools:
  222. info.ClaudeConvertInfo.Index = info.ClaudeConvertInfo.ToolCallBaseIndex + info.ClaudeConvertInfo.ToolCallMaxIndexOffset + 1
  223. info.ClaudeConvertInfo.ToolCallBaseIndex = 0
  224. info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
  225. default:
  226. info.ClaudeConvertInfo.Index++
  227. }
  228. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeNone
  229. }
  230. if info.SendResponseCount == 1 {
  231. msg := &dto.ClaudeMediaMessage{
  232. Id: openAIResponse.Id,
  233. Model: openAIResponse.Model,
  234. Type: "message",
  235. Role: "assistant",
  236. Usage: &dto.ClaudeUsage{
  237. InputTokens: info.GetEstimatePromptTokens(),
  238. OutputTokens: 0,
  239. },
  240. }
  241. msg.SetContent(make([]any, 0))
  242. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  243. Type: "message_start",
  244. Message: msg,
  245. })
  246. //claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  247. // Type: "ping",
  248. //})
  249. if openAIResponse.IsToolCall() {
  250. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeTools
  251. info.ClaudeConvertInfo.ToolCallBaseIndex = 0
  252. info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
  253. var toolCall dto.ToolCallResponse
  254. if len(openAIResponse.Choices) > 0 && len(openAIResponse.Choices[0].Delta.ToolCalls) > 0 {
  255. toolCall = openAIResponse.Choices[0].Delta.ToolCalls[0]
  256. } else {
  257. first := openAIResponse.GetFirstToolCall()
  258. if first != nil {
  259. toolCall = *first
  260. } else {
  261. toolCall = dto.ToolCallResponse{}
  262. }
  263. }
  264. resp := &dto.ClaudeResponse{
  265. Type: "content_block_start",
  266. ContentBlock: &dto.ClaudeMediaMessage{
  267. Id: toolCall.ID,
  268. Type: "tool_use",
  269. Name: toolCall.Function.Name,
  270. Input: map[string]interface{}{},
  271. },
  272. }
  273. resp.SetIndex(0)
  274. claudeResponses = append(claudeResponses, resp)
  275. // 首块包含工具 delta,则追加 input_json_delta
  276. if toolCall.Function.Arguments != "" {
  277. idx := 0
  278. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  279. Index: &idx,
  280. Type: "content_block_delta",
  281. Delta: &dto.ClaudeMediaMessage{
  282. Type: "input_json_delta",
  283. PartialJson: &toolCall.Function.Arguments,
  284. },
  285. })
  286. }
  287. } else {
  288. }
  289. // 判断首个响应是否存在内容(非标准的 OpenAI 响应)
  290. if len(openAIResponse.Choices) > 0 {
  291. reasoning := openAIResponse.Choices[0].Delta.GetReasoningContent()
  292. content := openAIResponse.Choices[0].Delta.GetContentString()
  293. if reasoning != "" {
  294. if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeThinking {
  295. stopOpenBlocksAndAdvance()
  296. }
  297. idx := info.ClaudeConvertInfo.Index
  298. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  299. Index: &idx,
  300. Type: "content_block_start",
  301. ContentBlock: &dto.ClaudeMediaMessage{
  302. Type: "thinking",
  303. Thinking: common.GetPointer[string](""),
  304. },
  305. })
  306. idx2 := idx
  307. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  308. Index: &idx2,
  309. Type: "content_block_delta",
  310. Delta: &dto.ClaudeMediaMessage{
  311. Type: "thinking_delta",
  312. Thinking: &reasoning,
  313. },
  314. })
  315. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeThinking
  316. } else if content != "" {
  317. if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeText {
  318. stopOpenBlocksAndAdvance()
  319. }
  320. idx := info.ClaudeConvertInfo.Index
  321. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  322. Index: &idx,
  323. Type: "content_block_start",
  324. ContentBlock: &dto.ClaudeMediaMessage{
  325. Type: "text",
  326. Text: common.GetPointer[string](""),
  327. },
  328. })
  329. idx2 := idx
  330. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  331. Index: &idx2,
  332. Type: "content_block_delta",
  333. Delta: &dto.ClaudeMediaMessage{
  334. Type: "text_delta",
  335. Text: common.GetPointer[string](content),
  336. },
  337. })
  338. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeText
  339. }
  340. }
  341. // 如果首块就带 finish_reason,需要立即发送停止块
  342. if len(openAIResponse.Choices) > 0 && openAIResponse.Choices[0].FinishReason != nil && *openAIResponse.Choices[0].FinishReason != "" {
  343. info.FinishReason = *openAIResponse.Choices[0].FinishReason
  344. stopOpenBlocks()
  345. oaiUsage := openAIResponse.Usage
  346. if oaiUsage == nil {
  347. oaiUsage = info.ClaudeConvertInfo.Usage
  348. }
  349. if oaiUsage != nil {
  350. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  351. Type: "message_delta",
  352. Usage: &dto.ClaudeUsage{
  353. InputTokens: oaiUsage.PromptTokens,
  354. OutputTokens: oaiUsage.CompletionTokens,
  355. CacheCreationInputTokens: oaiUsage.PromptTokensDetails.CachedCreationTokens,
  356. CacheReadInputTokens: oaiUsage.PromptTokensDetails.CachedTokens,
  357. },
  358. Delta: &dto.ClaudeMediaMessage{
  359. StopReason: common.GetPointer[string](stopReasonOpenAI2Claude(info.FinishReason)),
  360. },
  361. })
  362. }
  363. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  364. Type: "message_stop",
  365. })
  366. info.ClaudeConvertInfo.Done = true
  367. }
  368. return claudeResponses
  369. }
  370. if len(openAIResponse.Choices) == 0 {
  371. // no choices
  372. // 可能为非标准的 OpenAI 响应,判断是否已经完成
  373. if info.ClaudeConvertInfo.Done {
  374. stopOpenBlocks()
  375. oaiUsage := info.ClaudeConvertInfo.Usage
  376. if oaiUsage != nil {
  377. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  378. Type: "message_delta",
  379. Usage: &dto.ClaudeUsage{
  380. InputTokens: oaiUsage.PromptTokens,
  381. OutputTokens: oaiUsage.CompletionTokens,
  382. CacheCreationInputTokens: oaiUsage.PromptTokensDetails.CachedCreationTokens,
  383. CacheReadInputTokens: oaiUsage.PromptTokensDetails.CachedTokens,
  384. },
  385. Delta: &dto.ClaudeMediaMessage{
  386. StopReason: common.GetPointer[string](stopReasonOpenAI2Claude(info.FinishReason)),
  387. },
  388. })
  389. }
  390. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  391. Type: "message_stop",
  392. })
  393. }
  394. return claudeResponses
  395. } else {
  396. chosenChoice := openAIResponse.Choices[0]
  397. doneChunk := chosenChoice.FinishReason != nil && *chosenChoice.FinishReason != ""
  398. if doneChunk {
  399. info.FinishReason = *chosenChoice.FinishReason
  400. }
  401. var claudeResponse dto.ClaudeResponse
  402. var isEmpty bool
  403. claudeResponse.Type = "content_block_delta"
  404. if len(chosenChoice.Delta.ToolCalls) > 0 {
  405. toolCalls := chosenChoice.Delta.ToolCalls
  406. if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeTools {
  407. stopOpenBlocksAndAdvance()
  408. info.ClaudeConvertInfo.ToolCallBaseIndex = info.ClaudeConvertInfo.Index
  409. info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
  410. }
  411. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeTools
  412. base := info.ClaudeConvertInfo.ToolCallBaseIndex
  413. maxOffset := info.ClaudeConvertInfo.ToolCallMaxIndexOffset
  414. for i, toolCall := range toolCalls {
  415. offset := 0
  416. if toolCall.Index != nil {
  417. offset = *toolCall.Index
  418. } else {
  419. offset = i
  420. }
  421. if offset > maxOffset {
  422. maxOffset = offset
  423. }
  424. blockIndex := base + offset
  425. idx := blockIndex
  426. if toolCall.Function.Name != "" {
  427. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  428. Index: &idx,
  429. Type: "content_block_start",
  430. ContentBlock: &dto.ClaudeMediaMessage{
  431. Id: toolCall.ID,
  432. Type: "tool_use",
  433. Name: toolCall.Function.Name,
  434. Input: map[string]interface{}{},
  435. },
  436. })
  437. }
  438. if len(toolCall.Function.Arguments) > 0 {
  439. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  440. Index: &idx,
  441. Type: "content_block_delta",
  442. Delta: &dto.ClaudeMediaMessage{
  443. Type: "input_json_delta",
  444. PartialJson: &toolCall.Function.Arguments,
  445. },
  446. })
  447. }
  448. }
  449. info.ClaudeConvertInfo.ToolCallMaxIndexOffset = maxOffset
  450. info.ClaudeConvertInfo.Index = base + maxOffset
  451. } else {
  452. reasoning := chosenChoice.Delta.GetReasoningContent()
  453. textContent := chosenChoice.Delta.GetContentString()
  454. if reasoning != "" || textContent != "" {
  455. if reasoning != "" {
  456. if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeThinking {
  457. stopOpenBlocksAndAdvance()
  458. idx := info.ClaudeConvertInfo.Index
  459. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  460. Index: &idx,
  461. Type: "content_block_start",
  462. ContentBlock: &dto.ClaudeMediaMessage{
  463. Type: "thinking",
  464. Thinking: common.GetPointer[string](""),
  465. },
  466. })
  467. }
  468. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeThinking
  469. claudeResponse.Delta = &dto.ClaudeMediaMessage{
  470. Type: "thinking_delta",
  471. Thinking: &reasoning,
  472. }
  473. } else {
  474. if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeText {
  475. stopOpenBlocksAndAdvance()
  476. idx := info.ClaudeConvertInfo.Index
  477. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  478. Index: &idx,
  479. Type: "content_block_start",
  480. ContentBlock: &dto.ClaudeMediaMessage{
  481. Type: "text",
  482. Text: common.GetPointer[string](""),
  483. },
  484. })
  485. }
  486. info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeText
  487. claudeResponse.Delta = &dto.ClaudeMediaMessage{
  488. Type: "text_delta",
  489. Text: common.GetPointer[string](textContent),
  490. }
  491. }
  492. } else {
  493. isEmpty = true
  494. }
  495. }
  496. claudeResponse.Index = common.GetPointer[int](info.ClaudeConvertInfo.Index)
  497. if !isEmpty && claudeResponse.Delta != nil {
  498. claudeResponses = append(claudeResponses, &claudeResponse)
  499. }
  500. if doneChunk || info.ClaudeConvertInfo.Done {
  501. stopOpenBlocks()
  502. oaiUsage := openAIResponse.Usage
  503. if oaiUsage == nil {
  504. oaiUsage = info.ClaudeConvertInfo.Usage
  505. }
  506. if oaiUsage != nil {
  507. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  508. Type: "message_delta",
  509. Usage: &dto.ClaudeUsage{
  510. InputTokens: oaiUsage.PromptTokens,
  511. OutputTokens: oaiUsage.CompletionTokens,
  512. CacheCreationInputTokens: oaiUsage.PromptTokensDetails.CachedCreationTokens,
  513. CacheReadInputTokens: oaiUsage.PromptTokensDetails.CachedTokens,
  514. },
  515. Delta: &dto.ClaudeMediaMessage{
  516. StopReason: common.GetPointer[string](stopReasonOpenAI2Claude(info.FinishReason)),
  517. },
  518. })
  519. }
  520. claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
  521. Type: "message_stop",
  522. })
  523. info.ClaudeConvertInfo.Done = true
  524. return claudeResponses
  525. }
  526. }
  527. return claudeResponses
  528. }
  529. func ResponseOpenAI2Claude(openAIResponse *dto.OpenAITextResponse, info *relaycommon.RelayInfo) *dto.ClaudeResponse {
  530. var stopReason string
  531. contents := make([]dto.ClaudeMediaMessage, 0)
  532. claudeResponse := &dto.ClaudeResponse{
  533. Id: openAIResponse.Id,
  534. Type: "message",
  535. Role: "assistant",
  536. Model: openAIResponse.Model,
  537. }
  538. for _, choice := range openAIResponse.Choices {
  539. stopReason = stopReasonOpenAI2Claude(choice.FinishReason)
  540. if choice.FinishReason == "tool_calls" {
  541. for _, toolUse := range choice.Message.ParseToolCalls() {
  542. claudeContent := dto.ClaudeMediaMessage{}
  543. claudeContent.Type = "tool_use"
  544. claudeContent.Id = toolUse.ID
  545. claudeContent.Name = toolUse.Function.Name
  546. var mapParams map[string]interface{}
  547. if err := common.Unmarshal([]byte(toolUse.Function.Arguments), &mapParams); err == nil {
  548. claudeContent.Input = mapParams
  549. } else {
  550. claudeContent.Input = toolUse.Function.Arguments
  551. }
  552. contents = append(contents, claudeContent)
  553. }
  554. } else {
  555. claudeContent := dto.ClaudeMediaMessage{}
  556. claudeContent.Type = "text"
  557. claudeContent.SetText(choice.Message.StringContent())
  558. contents = append(contents, claudeContent)
  559. }
  560. }
  561. claudeResponse.Content = contents
  562. claudeResponse.StopReason = stopReason
  563. claudeResponse.Usage = &dto.ClaudeUsage{
  564. InputTokens: openAIResponse.PromptTokens,
  565. OutputTokens: openAIResponse.CompletionTokens,
  566. }
  567. return claudeResponse
  568. }
  569. func stopReasonOpenAI2Claude(reason string) string {
  570. return reasonmap.OpenAIFinishReasonToClaudeStopReason(reason)
  571. }
  572. func toJSONString(v interface{}) string {
  573. b, err := json.Marshal(v)
  574. if err != nil {
  575. return "{}"
  576. }
  577. return string(b)
  578. }
  579. func GeminiToOpenAIRequest(geminiRequest *dto.GeminiChatRequest, info *relaycommon.RelayInfo) (*dto.GeneralOpenAIRequest, error) {
  580. openaiRequest := &dto.GeneralOpenAIRequest{
  581. Model: info.UpstreamModelName,
  582. Stream: info.IsStream,
  583. }
  584. // 转换 messages
  585. var messages []dto.Message
  586. for _, content := range geminiRequest.Contents {
  587. message := dto.Message{
  588. Role: convertGeminiRoleToOpenAI(content.Role),
  589. }
  590. // 处理 parts
  591. var mediaContents []dto.MediaContent
  592. var toolCalls []dto.ToolCallRequest
  593. for _, part := range content.Parts {
  594. if part.Text != "" {
  595. mediaContent := dto.MediaContent{
  596. Type: "text",
  597. Text: part.Text,
  598. }
  599. mediaContents = append(mediaContents, mediaContent)
  600. } else if part.InlineData != nil {
  601. mediaContent := dto.MediaContent{
  602. Type: "image_url",
  603. ImageUrl: &dto.MessageImageUrl{
  604. Url: fmt.Sprintf("data:%s;base64,%s", part.InlineData.MimeType, part.InlineData.Data),
  605. Detail: "auto",
  606. MimeType: part.InlineData.MimeType,
  607. },
  608. }
  609. mediaContents = append(mediaContents, mediaContent)
  610. } else if part.FileData != nil {
  611. mediaContent := dto.MediaContent{
  612. Type: "image_url",
  613. ImageUrl: &dto.MessageImageUrl{
  614. Url: part.FileData.FileUri,
  615. Detail: "auto",
  616. MimeType: part.FileData.MimeType,
  617. },
  618. }
  619. mediaContents = append(mediaContents, mediaContent)
  620. } else if part.FunctionCall != nil {
  621. // 处理 Gemini 的工具调用
  622. toolCall := dto.ToolCallRequest{
  623. ID: fmt.Sprintf("call_%d", len(toolCalls)+1), // 生成唯一ID
  624. Type: "function",
  625. Function: dto.FunctionRequest{
  626. Name: part.FunctionCall.FunctionName,
  627. Arguments: toJSONString(part.FunctionCall.Arguments),
  628. },
  629. }
  630. toolCalls = append(toolCalls, toolCall)
  631. } else if part.FunctionResponse != nil {
  632. // 处理 Gemini 的工具响应,创建单独的 tool 消息
  633. toolMessage := dto.Message{
  634. Role: "tool",
  635. ToolCallId: fmt.Sprintf("call_%d", len(toolCalls)), // 使用对应的调用ID
  636. }
  637. toolMessage.SetStringContent(toJSONString(part.FunctionResponse.Response))
  638. messages = append(messages, toolMessage)
  639. }
  640. }
  641. // 设置消息内容
  642. if len(toolCalls) > 0 {
  643. // 如果有工具调用,设置工具调用
  644. message.SetToolCalls(toolCalls)
  645. } else if len(mediaContents) == 1 && mediaContents[0].Type == "text" {
  646. // 如果只有一个文本内容,直接设置字符串
  647. message.Content = mediaContents[0].Text
  648. } else if len(mediaContents) > 0 {
  649. // 如果有多个内容或包含媒体,设置为数组
  650. message.SetMediaContent(mediaContents)
  651. }
  652. // 只有当消息有内容或工具调用时才添加
  653. if len(message.ParseContent()) > 0 || len(message.ToolCalls) > 0 {
  654. messages = append(messages, message)
  655. }
  656. }
  657. openaiRequest.Messages = messages
  658. if geminiRequest.GenerationConfig.Temperature != nil {
  659. openaiRequest.Temperature = geminiRequest.GenerationConfig.Temperature
  660. }
  661. if geminiRequest.GenerationConfig.TopP > 0 {
  662. openaiRequest.TopP = geminiRequest.GenerationConfig.TopP
  663. }
  664. if geminiRequest.GenerationConfig.TopK > 0 {
  665. openaiRequest.TopK = int(geminiRequest.GenerationConfig.TopK)
  666. }
  667. if geminiRequest.GenerationConfig.MaxOutputTokens > 0 {
  668. openaiRequest.MaxTokens = geminiRequest.GenerationConfig.MaxOutputTokens
  669. }
  670. // gemini stop sequences 最多 5 个,openai stop 最多 4 个
  671. if len(geminiRequest.GenerationConfig.StopSequences) > 0 {
  672. openaiRequest.Stop = geminiRequest.GenerationConfig.StopSequences[:4]
  673. }
  674. if geminiRequest.GenerationConfig.CandidateCount > 0 {
  675. openaiRequest.N = geminiRequest.GenerationConfig.CandidateCount
  676. }
  677. // 转换工具调用
  678. if len(geminiRequest.GetTools()) > 0 {
  679. var tools []dto.ToolCallRequest
  680. for _, tool := range geminiRequest.GetTools() {
  681. if tool.FunctionDeclarations != nil {
  682. functionDeclarations, err := common.Any2Type[[]dto.FunctionRequest](tool.FunctionDeclarations)
  683. if err != nil {
  684. common.SysError(fmt.Sprintf("failed to parse gemini function declarations: %v (type=%T)", err, tool.FunctionDeclarations))
  685. continue
  686. }
  687. for _, function := range functionDeclarations {
  688. openAITool := dto.ToolCallRequest{
  689. Type: "function",
  690. Function: dto.FunctionRequest{
  691. Name: function.Name,
  692. Description: function.Description,
  693. Parameters: function.Parameters,
  694. },
  695. }
  696. tools = append(tools, openAITool)
  697. }
  698. }
  699. }
  700. if len(tools) > 0 {
  701. openaiRequest.Tools = tools
  702. }
  703. }
  704. // gemini system instructions
  705. if geminiRequest.SystemInstructions != nil {
  706. // 将系统指令作为第一条消息插入
  707. systemMessage := dto.Message{
  708. Role: "system",
  709. Content: extractTextFromGeminiParts(geminiRequest.SystemInstructions.Parts),
  710. }
  711. openaiRequest.Messages = append([]dto.Message{systemMessage}, openaiRequest.Messages...)
  712. }
  713. return openaiRequest, nil
  714. }
  715. func convertGeminiRoleToOpenAI(geminiRole string) string {
  716. switch geminiRole {
  717. case "user":
  718. return "user"
  719. case "model":
  720. return "assistant"
  721. case "function":
  722. return "function"
  723. default:
  724. return "user"
  725. }
  726. }
  727. func extractTextFromGeminiParts(parts []dto.GeminiPart) string {
  728. var texts []string
  729. for _, part := range parts {
  730. if part.Text != "" {
  731. texts = append(texts, part.Text)
  732. }
  733. }
  734. return strings.Join(texts, "\n")
  735. }
  736. // ResponseOpenAI2Gemini 将 OpenAI 响应转换为 Gemini 格式
  737. func ResponseOpenAI2Gemini(openAIResponse *dto.OpenAITextResponse, info *relaycommon.RelayInfo) *dto.GeminiChatResponse {
  738. geminiResponse := &dto.GeminiChatResponse{
  739. Candidates: make([]dto.GeminiChatCandidate, 0, len(openAIResponse.Choices)),
  740. UsageMetadata: dto.GeminiUsageMetadata{
  741. PromptTokenCount: openAIResponse.PromptTokens,
  742. CandidatesTokenCount: openAIResponse.CompletionTokens,
  743. TotalTokenCount: openAIResponse.PromptTokens + openAIResponse.CompletionTokens,
  744. },
  745. }
  746. for _, choice := range openAIResponse.Choices {
  747. candidate := dto.GeminiChatCandidate{
  748. Index: int64(choice.Index),
  749. SafetyRatings: []dto.GeminiChatSafetyRating{},
  750. }
  751. // 设置结束原因
  752. var finishReason string
  753. switch choice.FinishReason {
  754. case "stop":
  755. finishReason = "STOP"
  756. case "length":
  757. finishReason = "MAX_TOKENS"
  758. case "content_filter":
  759. finishReason = "SAFETY"
  760. case "tool_calls":
  761. finishReason = "STOP"
  762. default:
  763. finishReason = "STOP"
  764. }
  765. candidate.FinishReason = &finishReason
  766. // 转换消息内容
  767. content := dto.GeminiChatContent{
  768. Role: "model",
  769. Parts: make([]dto.GeminiPart, 0),
  770. }
  771. // 处理工具调用
  772. toolCalls := choice.Message.ParseToolCalls()
  773. if len(toolCalls) > 0 {
  774. for _, toolCall := range toolCalls {
  775. // 解析参数
  776. var args map[string]interface{}
  777. if toolCall.Function.Arguments != "" {
  778. if err := json.Unmarshal([]byte(toolCall.Function.Arguments), &args); err != nil {
  779. args = map[string]interface{}{"arguments": toolCall.Function.Arguments}
  780. }
  781. } else {
  782. args = make(map[string]interface{})
  783. }
  784. part := dto.GeminiPart{
  785. FunctionCall: &dto.FunctionCall{
  786. FunctionName: toolCall.Function.Name,
  787. Arguments: args,
  788. },
  789. }
  790. content.Parts = append(content.Parts, part)
  791. }
  792. } else {
  793. // 处理文本内容
  794. textContent := choice.Message.StringContent()
  795. if textContent != "" {
  796. part := dto.GeminiPart{
  797. Text: textContent,
  798. }
  799. content.Parts = append(content.Parts, part)
  800. }
  801. }
  802. candidate.Content = content
  803. geminiResponse.Candidates = append(geminiResponse.Candidates, candidate)
  804. }
  805. return geminiResponse
  806. }
  807. // StreamResponseOpenAI2Gemini 将 OpenAI 流式响应转换为 Gemini 格式
  808. func StreamResponseOpenAI2Gemini(openAIResponse *dto.ChatCompletionsStreamResponse, info *relaycommon.RelayInfo) *dto.GeminiChatResponse {
  809. // 检查是否有实际内容或结束标志
  810. hasContent := false
  811. hasFinishReason := false
  812. for _, choice := range openAIResponse.Choices {
  813. if len(choice.Delta.GetContentString()) > 0 || (choice.Delta.ToolCalls != nil && len(choice.Delta.ToolCalls) > 0) {
  814. hasContent = true
  815. }
  816. if choice.FinishReason != nil {
  817. hasFinishReason = true
  818. }
  819. }
  820. // 如果没有实际内容且没有结束标志,跳过。主要针对 openai 流响应开头的空数据
  821. if !hasContent && !hasFinishReason {
  822. return nil
  823. }
  824. geminiResponse := &dto.GeminiChatResponse{
  825. Candidates: make([]dto.GeminiChatCandidate, 0, len(openAIResponse.Choices)),
  826. UsageMetadata: dto.GeminiUsageMetadata{
  827. PromptTokenCount: info.GetEstimatePromptTokens(),
  828. CandidatesTokenCount: 0, // 流式响应中可能没有完整的 usage 信息
  829. TotalTokenCount: info.GetEstimatePromptTokens(),
  830. },
  831. }
  832. if openAIResponse.Usage != nil {
  833. geminiResponse.UsageMetadata.PromptTokenCount = openAIResponse.Usage.PromptTokens
  834. geminiResponse.UsageMetadata.CandidatesTokenCount = openAIResponse.Usage.CompletionTokens
  835. geminiResponse.UsageMetadata.TotalTokenCount = openAIResponse.Usage.TotalTokens
  836. }
  837. for _, choice := range openAIResponse.Choices {
  838. candidate := dto.GeminiChatCandidate{
  839. Index: int64(choice.Index),
  840. SafetyRatings: []dto.GeminiChatSafetyRating{},
  841. }
  842. // 设置结束原因
  843. if choice.FinishReason != nil {
  844. var finishReason string
  845. switch *choice.FinishReason {
  846. case "stop":
  847. finishReason = "STOP"
  848. case "length":
  849. finishReason = "MAX_TOKENS"
  850. case "content_filter":
  851. finishReason = "SAFETY"
  852. case "tool_calls":
  853. finishReason = "STOP"
  854. default:
  855. finishReason = "STOP"
  856. }
  857. candidate.FinishReason = &finishReason
  858. }
  859. // 转换消息内容
  860. content := dto.GeminiChatContent{
  861. Role: "model",
  862. Parts: make([]dto.GeminiPart, 0),
  863. }
  864. // 处理工具调用
  865. if choice.Delta.ToolCalls != nil {
  866. for _, toolCall := range choice.Delta.ToolCalls {
  867. // 解析参数
  868. var args map[string]interface{}
  869. if toolCall.Function.Arguments != "" {
  870. if err := json.Unmarshal([]byte(toolCall.Function.Arguments), &args); err != nil {
  871. args = map[string]interface{}{"arguments": toolCall.Function.Arguments}
  872. }
  873. } else {
  874. args = make(map[string]interface{})
  875. }
  876. part := dto.GeminiPart{
  877. FunctionCall: &dto.FunctionCall{
  878. FunctionName: toolCall.Function.Name,
  879. Arguments: args,
  880. },
  881. }
  882. content.Parts = append(content.Parts, part)
  883. }
  884. } else {
  885. // 处理文本内容
  886. textContent := choice.Delta.GetContentString()
  887. if textContent != "" {
  888. part := dto.GeminiPart{
  889. Text: textContent,
  890. }
  891. content.Parts = append(content.Parts, part)
  892. }
  893. }
  894. candidate.Content = content
  895. geminiResponse.Candidates = append(geminiResponse.Candidates, candidate)
  896. }
  897. return geminiResponse
  898. }