relay-openai.go 6.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203
  1. package openai
  2. import (
  3. "bufio"
  4. "bytes"
  5. "encoding/json"
  6. "github.com/gin-gonic/gin"
  7. "io"
  8. "net/http"
  9. "one-api/common"
  10. "one-api/dto"
  11. relayconstant "one-api/relay/constant"
  12. "one-api/service"
  13. "strings"
  14. "sync"
  15. "time"
  16. )
  17. func OpenaiStreamHandler(c *gin.Context, resp *http.Response, relayMode int) (*dto.OpenAIErrorWithStatusCode, string, int) {
  18. //checkSensitive := constant.ShouldCheckCompletionSensitive()
  19. var responseTextBuilder strings.Builder
  20. toolCount := 0
  21. scanner := bufio.NewScanner(resp.Body)
  22. scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
  23. if atEOF && len(data) == 0 {
  24. return 0, nil, nil
  25. }
  26. if i := strings.Index(string(data), "\n"); i >= 0 {
  27. return i + 1, data[0:i], nil
  28. }
  29. if atEOF {
  30. return len(data), data, nil
  31. }
  32. return 0, nil, nil
  33. })
  34. dataChan := make(chan string, 5)
  35. stopChan := make(chan bool, 2)
  36. defer close(stopChan)
  37. defer close(dataChan)
  38. var wg sync.WaitGroup
  39. go func() {
  40. wg.Add(1)
  41. defer wg.Done()
  42. var streamItems []string // store stream items
  43. for scanner.Scan() {
  44. data := scanner.Text()
  45. if len(data) < 6 { // ignore blank line or wrong format
  46. continue
  47. }
  48. if data[:6] != "data: " && data[:6] != "[DONE]" {
  49. continue
  50. }
  51. common.SafeSendString(dataChan, data)
  52. data = data[6:]
  53. if !strings.HasPrefix(data, "[DONE]") {
  54. streamItems = append(streamItems, data)
  55. }
  56. }
  57. streamResp := "[" + strings.Join(streamItems, ",") + "]"
  58. switch relayMode {
  59. case relayconstant.RelayModeChatCompletions:
  60. var streamResponses []dto.ChatCompletionsStreamResponseSimple
  61. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  62. if err != nil {
  63. common.SysError("error unmarshalling stream response: " + err.Error())
  64. for _, item := range streamItems {
  65. var streamResponse dto.ChatCompletionsStreamResponseSimple
  66. err := json.Unmarshal(common.StringToByteSlice(item), &streamResponse)
  67. if err == nil {
  68. for _, choice := range streamResponse.Choices {
  69. responseTextBuilder.WriteString(choice.Delta.GetContentString())
  70. if choice.Delta.ToolCalls != nil {
  71. if len(choice.Delta.ToolCalls) > toolCount {
  72. toolCount = len(choice.Delta.ToolCalls)
  73. }
  74. for _, tool := range choice.Delta.ToolCalls {
  75. responseTextBuilder.WriteString(tool.Function.Name)
  76. responseTextBuilder.WriteString(tool.Function.Arguments)
  77. }
  78. }
  79. }
  80. }
  81. }
  82. } else {
  83. for _, streamResponse := range streamResponses {
  84. for _, choice := range streamResponse.Choices {
  85. responseTextBuilder.WriteString(choice.Delta.GetContentString())
  86. if choice.Delta.ToolCalls != nil {
  87. if len(choice.Delta.ToolCalls) > toolCount {
  88. toolCount = len(choice.Delta.ToolCalls)
  89. }
  90. for _, tool := range choice.Delta.ToolCalls {
  91. responseTextBuilder.WriteString(tool.Function.Name)
  92. responseTextBuilder.WriteString(tool.Function.Arguments)
  93. }
  94. }
  95. }
  96. }
  97. }
  98. case relayconstant.RelayModeCompletions:
  99. var streamResponses []dto.CompletionsStreamResponse
  100. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  101. if err != nil {
  102. common.SysError("error unmarshalling stream response: " + err.Error())
  103. for _, item := range streamItems {
  104. var streamResponse dto.CompletionsStreamResponse
  105. err := json.Unmarshal(common.StringToByteSlice(item), &streamResponse)
  106. if err == nil {
  107. for _, choice := range streamResponse.Choices {
  108. responseTextBuilder.WriteString(choice.Text)
  109. }
  110. }
  111. }
  112. } else {
  113. for _, streamResponse := range streamResponses {
  114. for _, choice := range streamResponse.Choices {
  115. responseTextBuilder.WriteString(choice.Text)
  116. }
  117. }
  118. }
  119. }
  120. if len(dataChan) > 0 {
  121. // wait data out
  122. time.Sleep(2 * time.Second)
  123. }
  124. common.SafeSendBool(stopChan, true)
  125. }()
  126. service.SetEventStreamHeaders(c)
  127. c.Stream(func(w io.Writer) bool {
  128. select {
  129. case data := <-dataChan:
  130. if strings.HasPrefix(data, "data: [DONE]") {
  131. data = data[:12]
  132. }
  133. // some implementations may add \r at the end of data
  134. data = strings.TrimSuffix(data, "\r")
  135. c.Render(-1, common.CustomEvent{Data: data})
  136. return true
  137. case <-stopChan:
  138. return false
  139. }
  140. })
  141. err := resp.Body.Close()
  142. if err != nil {
  143. return service.OpenAIErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), "", toolCount
  144. }
  145. wg.Wait()
  146. return nil, responseTextBuilder.String(), toolCount
  147. }
  148. func OpenaiHandler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*dto.OpenAIErrorWithStatusCode, *dto.Usage) {
  149. var simpleResponse dto.SimpleResponse
  150. responseBody, err := io.ReadAll(resp.Body)
  151. if err != nil {
  152. return service.OpenAIErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
  153. }
  154. err = resp.Body.Close()
  155. if err != nil {
  156. return service.OpenAIErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  157. }
  158. err = json.Unmarshal(responseBody, &simpleResponse)
  159. if err != nil {
  160. return service.OpenAIErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
  161. }
  162. if simpleResponse.Error.Type != "" {
  163. return &dto.OpenAIErrorWithStatusCode{
  164. Error: simpleResponse.Error,
  165. StatusCode: resp.StatusCode,
  166. }, nil
  167. }
  168. // Reset response body
  169. resp.Body = io.NopCloser(bytes.NewBuffer(responseBody))
  170. // We shouldn't set the header before we parse the response body, because the parse part may fail.
  171. // And then we will have to send an error response, but in this case, the header has already been set.
  172. // So the httpClient will be confused by the response.
  173. // For example, Postman will report error, and we cannot check the response at all.
  174. for k, v := range resp.Header {
  175. c.Writer.Header().Set(k, v[0])
  176. }
  177. c.Writer.WriteHeader(resp.StatusCode)
  178. _, err = io.Copy(c.Writer, resp.Body)
  179. if err != nil {
  180. return service.OpenAIErrorWrapper(err, "copy_response_body_failed", http.StatusInternalServerError), nil
  181. }
  182. err = resp.Body.Close()
  183. if err != nil {
  184. return service.OpenAIErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  185. }
  186. if simpleResponse.Usage.TotalTokens == 0 {
  187. completionTokens := 0
  188. for _, choice := range simpleResponse.Choices {
  189. ctkm, _, _ := service.CountTokenText(string(choice.Message.Content), model, false)
  190. completionTokens += ctkm
  191. }
  192. simpleResponse.Usage = dto.Usage{
  193. PromptTokens: promptTokens,
  194. CompletionTokens: completionTokens,
  195. TotalTokens: promptTokens + completionTokens,
  196. }
  197. }
  198. return nil, &simpleResponse.Usage
  199. }