relay-openai.go 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154
  1. package controller
  2. import (
  3. "bufio"
  4. "bytes"
  5. "encoding/json"
  6. "github.com/gin-gonic/gin"
  7. "io"
  8. "net/http"
  9. "one-api/common"
  10. "strings"
  11. "sync"
  12. )
  13. func openaiStreamHandler(c *gin.Context, resp *http.Response, relayMode int) (*OpenAIErrorWithStatusCode, string) {
  14. var responseTextBuilder strings.Builder
  15. scanner := bufio.NewScanner(resp.Body)
  16. scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
  17. if atEOF && len(data) == 0 {
  18. return 0, nil, nil
  19. }
  20. if i := strings.Index(string(data), "\n"); i >= 0 {
  21. return i + 1, data[0:i], nil
  22. }
  23. if atEOF {
  24. return len(data), data, nil
  25. }
  26. return 0, nil, nil
  27. })
  28. dataChan := make(chan string)
  29. stopChan := make(chan bool)
  30. var wg sync.WaitGroup
  31. go func() {
  32. wg.Add(1)
  33. var streamItems []string
  34. for scanner.Scan() {
  35. data := scanner.Text()
  36. if len(data) < 6 { // ignore blank line or wrong format
  37. continue
  38. }
  39. if data[:6] != "data: " && data[:6] != "[DONE]" {
  40. continue
  41. }
  42. dataChan <- data
  43. data = data[6:]
  44. if !strings.HasPrefix(data, "[DONE]") {
  45. streamItems = append(streamItems, data)
  46. }
  47. }
  48. streamResp := "[" + strings.Join(streamItems, ",") + "]"
  49. switch relayMode {
  50. case RelayModeChatCompletions:
  51. var streamResponses []ChatCompletionsStreamResponseSimple
  52. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  53. if err != nil {
  54. common.SysError("error unmarshalling stream response: " + err.Error())
  55. return // just ignore the error
  56. }
  57. for _, streamResponse := range streamResponses {
  58. for _, choice := range streamResponse.Choices {
  59. responseTextBuilder.WriteString(choice.Delta.Content)
  60. }
  61. }
  62. case RelayModeCompletions:
  63. var streamResponses []CompletionsStreamResponse
  64. err := json.Unmarshal(common.StringToByteSlice(streamResp), &streamResponses)
  65. if err != nil {
  66. common.SysError("error unmarshalling stream response: " + err.Error())
  67. return // just ignore the error
  68. }
  69. for _, streamResponse := range streamResponses {
  70. for _, choice := range streamResponse.Choices {
  71. responseTextBuilder.WriteString(choice.Text)
  72. }
  73. }
  74. }
  75. wg.Done()
  76. stopChan <- true
  77. }()
  78. setEventStreamHeaders(c)
  79. c.Stream(func(w io.Writer) bool {
  80. select {
  81. case data := <-dataChan:
  82. if strings.HasPrefix(data, "data: [DONE]") {
  83. data = data[:12]
  84. }
  85. // some implementations may add \r at the end of data
  86. data = strings.TrimSuffix(data, "\r")
  87. c.Render(-1, common.CustomEvent{Data: data})
  88. return true
  89. case <-stopChan:
  90. return false
  91. }
  92. })
  93. err := resp.Body.Close()
  94. if err != nil {
  95. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), ""
  96. }
  97. wg.Wait()
  98. return nil, responseTextBuilder.String()
  99. }
  100. func openaiHandler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*OpenAIErrorWithStatusCode, *Usage) {
  101. var textResponse TextResponse
  102. responseBody, err := io.ReadAll(resp.Body)
  103. if err != nil {
  104. return errorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
  105. }
  106. err = resp.Body.Close()
  107. if err != nil {
  108. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  109. }
  110. err = json.Unmarshal(responseBody, &textResponse)
  111. if err != nil {
  112. return errorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
  113. }
  114. if textResponse.Error.Type != "" {
  115. return &OpenAIErrorWithStatusCode{
  116. OpenAIError: textResponse.Error,
  117. StatusCode: resp.StatusCode,
  118. }, nil
  119. }
  120. // Reset response body
  121. resp.Body = io.NopCloser(bytes.NewBuffer(responseBody))
  122. // We shouldn't set the header before we parse the response body, because the parse part may fail.
  123. // And then we will have to send an error response, but in this case, the header has already been set.
  124. // So the httpClient will be confused by the response.
  125. // For example, Postman will report error, and we cannot check the response at all.
  126. for k, v := range resp.Header {
  127. c.Writer.Header().Set(k, v[0])
  128. }
  129. c.Writer.WriteHeader(resp.StatusCode)
  130. _, err = io.Copy(c.Writer, resp.Body)
  131. if err != nil {
  132. return errorWrapper(err, "copy_response_body_failed", http.StatusInternalServerError), nil
  133. }
  134. err = resp.Body.Close()
  135. if err != nil {
  136. return errorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
  137. }
  138. if textResponse.Usage.TotalTokens == 0 {
  139. completionTokens := 0
  140. for _, choice := range textResponse.Choices {
  141. completionTokens += countTokenText(string(choice.Message.Content), model)
  142. }
  143. textResponse.Usage = Usage{
  144. PromptTokens: promptTokens,
  145. CompletionTokens: completionTokens,
  146. TotalTokens: promptTokens + completionTokens,
  147. }
  148. }
  149. return nil, &textResponse.Usage
  150. }