You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

openai.go 4.6 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169
  1. // Copyright 2023 The casbin Authors. All Rights Reserved.
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. package model
  15. import (
  16. "context"
  17. "fmt"
  18. "io"
  19. "net/http"
  20. "strings"
  21. "github.com/casibase/casibase/proxy"
  22. "github.com/sashabaranov/go-openai"
  23. )
  24. // https://pkg.go.dev/github.com/sashabaranov/go-openai@v1.12.0#pkg-constants
  25. // https://platform.openai.com/docs/models/overview
  26. var __maxTokens = map[string]int{
  27. openai.GPT4: 8192,
  28. openai.GPT40613: 8192,
  29. openai.GPT432K: 32768,
  30. openai.GPT432K0613: 32768,
  31. openai.GPT40314: 8192,
  32. openai.GPT432K0314: 32768,
  33. openai.GPT3Dot5Turbo: 4097,
  34. openai.GPT3Dot5Turbo16K: 16385,
  35. openai.GPT3Dot5Turbo0613: 4097,
  36. openai.GPT3Dot5Turbo16K0613: 16385,
  37. openai.GPT3Dot5Turbo0301: 4097,
  38. openai.GPT3TextDavinci003: 4097,
  39. openai.GPT3TextDavinci002: 4097,
  40. openai.GPT3TextCurie001: 2049,
  41. openai.GPT3TextBabbage001: 2049,
  42. openai.GPT3TextAda001: 2049,
  43. openai.GPT3Davinci: 2049,
  44. openai.GPT3Curie: 2049,
  45. openai.GPT3Ada: 2049,
  46. openai.GPT3Babbage: 2049,
  47. }
  48. type OpenAiModelProvider struct {
  49. subType string
  50. secretKey string
  51. temperature float32
  52. topP float32
  53. frequencyPenalty float32
  54. presencePenalty float32
  55. }
  56. func NewOpenAiModelProvider(subType string, secretKey string, temperature float32, topP float32, frequencyPenalty float32, presencePenalty float32) (*OpenAiModelProvider, error) {
  57. p := &OpenAiModelProvider{
  58. subType: subType,
  59. secretKey: secretKey,
  60. temperature: temperature,
  61. topP: topP,
  62. frequencyPenalty: frequencyPenalty,
  63. presencePenalty: presencePenalty,
  64. }
  65. return p, nil
  66. }
  67. func getProxyClientFromToken(authToken string) *openai.Client {
  68. config := openai.DefaultConfig(authToken)
  69. config.HTTPClient = proxy.ProxyHttpClient
  70. c := openai.NewClientWithConfig(config)
  71. return c
  72. }
  73. // GetMaxTokens returns the max tokens for a given openai model.
  74. func (p *OpenAiModelProvider) GetMaxTokens() int {
  75. res, ok := __maxTokens[p.subType]
  76. if !ok {
  77. return 4097
  78. }
  79. return res
  80. }
  81. func (p *OpenAiModelProvider) QueryText(question string, writer io.Writer, builder *strings.Builder) error {
  82. client := getProxyClientFromToken(p.secretKey)
  83. ctx := context.Background()
  84. flusher, ok := writer.(http.Flusher)
  85. if !ok {
  86. return fmt.Errorf("writer does not implement http.Flusher")
  87. }
  88. model := p.subType
  89. if model == "" {
  90. model = openai.GPT3TextDavinci003
  91. }
  92. // https://platform.openai.com/tokenizer
  93. // https://github.com/pkoukk/tiktoken-go#available-encodings
  94. tokenCount, err := GetTokenSize(model, question)
  95. if err != nil {
  96. return err
  97. }
  98. maxTokens := p.GetMaxTokens() - tokenCount
  99. if maxTokens < 0 {
  100. return fmt.Errorf("The token count: [%d] exceeds the model: [%s]'s maximum token count: [%d]", tokenCount, model, p.GetMaxTokens())
  101. }
  102. temperature := p.temperature
  103. topP := p.topP
  104. frequencyPenalty := p.frequencyPenalty
  105. presencePenalty := p.presencePenalty
  106. respStream, err := client.CreateCompletionStream(
  107. ctx,
  108. openai.CompletionRequest{
  109. Model: model,
  110. Prompt: question,
  111. MaxTokens: maxTokens,
  112. Stream: true,
  113. Temperature: temperature,
  114. TopP: topP,
  115. FrequencyPenalty: frequencyPenalty,
  116. PresencePenalty: presencePenalty,
  117. },
  118. )
  119. if err != nil {
  120. return err
  121. }
  122. defer respStream.Close()
  123. isLeadingReturn := true
  124. for {
  125. completion, streamErr := respStream.Recv()
  126. if streamErr != nil {
  127. if streamErr == io.EOF {
  128. break
  129. }
  130. return streamErr
  131. }
  132. data := completion.Choices[0].Text
  133. if isLeadingReturn && len(data) != 0 {
  134. if strings.Count(data, "\n") == len(data) {
  135. continue
  136. } else {
  137. isLeadingReturn = false
  138. }
  139. }
  140. // Write the streamed data as Server-Sent Events
  141. if _, err = fmt.Fprintf(writer, "event: message\ndata: %s\n\n", data); err != nil {
  142. return err
  143. }
  144. flusher.Flush()
  145. // Append the response to the strings.Builder
  146. builder.WriteString(data)
  147. }
  148. return nil
  149. }