openai.go 3.3 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132
  1. package openai
  2. import (
  3. "context"
  4. "crypto/tls"
  5. "fmt"
  6. "github.com/0xJacky/Nginx-UI/api"
  7. "github.com/0xJacky/Nginx-UI/internal/chatbot"
  8. "github.com/0xJacky/Nginx-UI/settings"
  9. "github.com/gin-gonic/gin"
  10. "github.com/pkg/errors"
  11. "github.com/sashabaranov/go-openai"
  12. "io"
  13. "net/http"
  14. "net/url"
  15. )
  16. const ChatGPTInitPrompt = `You are a assistant who can help users write and optimise the configurations of Nginx,
  17. the first user message contains the content of the configuration file which is currently opened by the user and
  18. the current language code(CLC). You suppose to use the language corresponding to the CLC to give the first reply.
  19. Later the language environment depends on the user message.
  20. The first reply should involve the key information of the file and ask user what can you help them.`
  21. func MakeChatCompletionRequest(c *gin.Context) {
  22. var json struct {
  23. Filepath string `json:"filepath"`
  24. Messages []openai.ChatCompletionMessage `json:"messages"`
  25. }
  26. if !api.BindAndValid(c, &json) {
  27. return
  28. }
  29. messages := []openai.ChatCompletionMessage{
  30. {
  31. Role: openai.ChatMessageRoleSystem,
  32. Content: ChatGPTInitPrompt,
  33. },
  34. }
  35. messages = append(messages, json.Messages...)
  36. if json.Filepath != "" {
  37. messages = chatbot.ChatCompletionWithContext(json.Filepath, messages)
  38. }
  39. // SSE server
  40. c.Writer.Header().Set("Content-Type", "text/event-stream; charset=utf-8")
  41. c.Writer.Header().Set("Cache-Control", "no-cache")
  42. c.Writer.Header().Set("Connection", "keep-alive")
  43. c.Writer.Header().Set("Access-Control-Allow-Origin", "*")
  44. config := openai.DefaultConfig(settings.OpenAISettings.Token)
  45. if settings.OpenAISettings.Proxy != "" {
  46. proxyUrl, err := url.Parse(settings.OpenAISettings.Proxy)
  47. if err != nil {
  48. c.Stream(func(w io.Writer) bool {
  49. c.SSEvent("message", gin.H{
  50. "type": "error",
  51. "content": err.Error(),
  52. })
  53. return false
  54. })
  55. return
  56. }
  57. transport := &http.Transport{
  58. Proxy: http.ProxyURL(proxyUrl),
  59. TLSClientConfig: &tls.Config{InsecureSkipVerify: settings.ServerSettings.InsecureSkipVerify},
  60. }
  61. config.HTTPClient = &http.Client{
  62. Transport: transport,
  63. }
  64. }
  65. if settings.OpenAISettings.BaseUrl != "" {
  66. config.BaseURL = settings.OpenAISettings.BaseUrl
  67. }
  68. openaiClient := openai.NewClientWithConfig(config)
  69. ctx := context.Background()
  70. req := openai.ChatCompletionRequest{
  71. Model: settings.OpenAISettings.Model,
  72. Messages: messages,
  73. Stream: true,
  74. }
  75. stream, err := openaiClient.CreateChatCompletionStream(ctx, req)
  76. if err != nil {
  77. fmt.Printf("CompletionStream error: %v\n", err)
  78. c.Stream(func(w io.Writer) bool {
  79. c.SSEvent("message", gin.H{
  80. "type": "error",
  81. "content": err.Error(),
  82. })
  83. return false
  84. })
  85. return
  86. }
  87. defer stream.Close()
  88. msgChan := make(chan string)
  89. go func() {
  90. defer close(msgChan)
  91. for {
  92. response, err := stream.Recv()
  93. if errors.Is(err, io.EOF) {
  94. fmt.Println()
  95. return
  96. }
  97. if err != nil {
  98. fmt.Printf("Stream error: %v\n", err)
  99. return
  100. }
  101. message := fmt.Sprintf("%s", response.Choices[0].Delta.Content)
  102. msgChan <- message
  103. }
  104. }()
  105. c.Stream(func(w io.Writer) bool {
  106. if m, ok := <-msgChan; ok {
  107. c.SSEvent("message", gin.H{
  108. "type": "message",
  109. "content": m,
  110. })
  111. return true
  112. }
  113. return false
  114. })
  115. }