relay-openai.go 4.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156
  1. package controller
  2. import (
  3. "bufio"
  4. "bytes"
  5. "encoding/json"
  6. "github.com/gin-gonic/gin"
  7. "io"
  8. "net/http"
  9. "one-api/common"
  10. "strings"
  11. "sync"
  12. )
  13. func openaiStreamHandler(c *gin.Context, resp *http.Response, relayMode int) (*OpenAIErrorWithStatusCode, string) {
  14. var responseTextBuilder strings.Builder
  15. scanner := bufio.NewScanner(resp.Body)
  16. scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
  17. if atEOF && len(data) == 0 {
  18. return 0, nil, nil
  19. }
  20. if i := strings.Index(string(data), "\n"); i >= 0 {
  21. return i + 1, data[0:i], nil
  22. }
  23. if atEOF {
  24. return len(data), data, nil
  25. }
  26. return 0, nil, nil
  27. })
  28. dataChan := make(chan string)
  29. stopChan := make(chan bool)
  30. var wg sync.WaitGroup
  31. go func() {
  32. wg.Add(1)
  33. var streamItems []string
  34. for scanner.Scan() {
  35. data := scanner.Text()
  36. if len(data) < 6 { // ignore blank line or wrong format
  37. continue
  38. }
  39. if data[:6] != "data: " && data[:6] != "[DONE]" {
  40. continue
  41. }
  42. dataChan <- data
  43. data = data[6:]
  44. if !strings.HasPrefix(data, "[DONE]") {
  45. streamItems = append(streamItems, data)
  46. }
  47. }
  48. streamResp := "[" + strings.Join(streamItems, ",") + "]"
  49. switch relayMode {
  50. case RelayModeChatCompletions:
  51. var streamResponses []ChatCompletionsStreamResponseSimple
  52. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  53. if err != nil {
  54. common.SysError("error unmarshalling stream response: " + err.Error())
  55. wg.Done()
  56. return // just ignore the error
  57. }
  58. for _, streamResponse := range streamResponses {
  59. for _, choice := range streamResponse.Choices {
  60. responseTextBuilder.WriteString(choice.Delta.Content)
  61. }
  62. }
  63. case RelayModeCompletions:
  64. var streamResponses []CompletionsStreamResponse
  65. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  66. if err != nil {
  67. common.SysError("error unmarshalling stream response: " + err.Error())
  68. wg.Done()
  69. return // just ignore the error
  70. }
  71. for _, streamResponse := range streamResponses {
  72. for _, choice := range streamResponse.Choices {
  73. responseTextBuilder.WriteString(choice.Text)
  74. }
  75. }
  76. }
  77. wg.Done()
  78. stopChan <- true
  79. }()
  80. setEventStreamHeaders(c)
  81. c.Stream(func(w io.Writer) bool {
  82. select {
  83. case data := <-dataChan:
  84. if strings.HasPrefix(data, "data: [DONE]") {
  85. data = data[:12]
  86. }
  87. // some implementations may add \r at the end of data
  88. data = strings.TrimSuffix(data, "\r")
  89. c.Render(-1, common.CustomEvent{Data: data})
  90. return true
  91. case <-stopChan:
  92. return false
  93. }
  94. })
  95. err := resp.Body.Close()
  96. if err != nil {
  97. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), ""
  98. }
  99. wg.Wait()
  100. return nil, responseTextBuilder.String()
  101. }
  102. func openaiHandler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*OpenAIErrorWithStatusCode, *Usage) {
  103. var textResponse TextResponse
  104. responseBody, err := io.ReadAll(resp.Body)
  105. if err != nil {
  106. return errorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
  107. }
  108. err = resp.Body.Close()
  109. if err != nil {
  110. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  111. }
  112. err = json.Unmarshal(responseBody, &textResponse)
  113. if err != nil {
  114. return errorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
  115. }
  116. if textResponse.Error.Type != "" {
  117. return &OpenAIErrorWithStatusCode{
  118. OpenAIError: textResponse.Error,
  119. StatusCode: resp.StatusCode,
  120. }, nil
  121. }
  122. // Reset response body
  123. resp.Body = io.NopCloser(bytes.NewBuffer(responseBody))
  124. // We shouldn't set the header before we parse the response body, because the parse part may fail.
  125. // And then we will have to send an error response, but in this case, the header has already been set.
  126. // So the httpClient will be confused by the response.
  127. // For example, Postman will report error, and we cannot check the response at all.
  128. for k, v := range resp.Header {
  129. c.Writer.Header().Set(k, v[0])
  130. }
  131. c.Writer.WriteHeader(resp.StatusCode)
  132. _, err = io.Copy(c.Writer, resp.Body)
  133. if err != nil {
  134. return errorWrapper(err, "copy_response_body_failed", http.StatusInternalServerError), nil
  135. }
  136. err = resp.Body.Close()
  137. if err != nil {
  138. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  139. }
  140. if textResponse.Usage.TotalTokens == 0 {
  141. completionTokens := 0
  142. for _, choice := range textResponse.Choices {
  143. completionTokens += countTokenText(string(choice.Message.Content), model)
  144. }
  145. textResponse.Usage = Usage{
  146. PromptTokens: promptTokens,
  147. CompletionTokens: completionTokens,
  148. TotalTokens: promptTokens + completionTokens,
  149. }
  150. }
  151. return nil, &textResponse.Usage
  152. }