123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132 |
- package openai
- import (
- "context"
- "crypto/tls"
- "fmt"
- "github.com/0xJacky/Nginx-UI/api"
- "github.com/0xJacky/Nginx-UI/internal/chatbot"
- "github.com/0xJacky/Nginx-UI/settings"
- "github.com/gin-gonic/gin"
- "github.com/pkg/errors"
- "github.com/sashabaranov/go-openai"
- "io"
- "net/http"
- "net/url"
- )
- const ChatGPTInitPrompt = `You are a assistant who can help users write and optimise the configurations of Nginx,
- the first user message contains the content of the configuration file which is currently opened by the user and
- the current language code(CLC). You suppose to use the language corresponding to the CLC to give the first reply.
- Later the language environment depends on the user message.
- The first reply should involve the key information of the file and ask user what can you help them.`
- func MakeChatCompletionRequest(c *gin.Context) {
- var json struct {
- Filepath string `json:"filepath"`
- Messages []openai.ChatCompletionMessage `json:"messages"`
- }
- if !api.BindAndValid(c, &json) {
- return
- }
- messages := []openai.ChatCompletionMessage{
- {
- Role: openai.ChatMessageRoleSystem,
- Content: ChatGPTInitPrompt,
- },
- }
- messages = append(messages, json.Messages...)
- if json.Filepath != "" {
- messages = chatbot.ChatCompletionWithContext(json.Filepath, messages)
- }
- // SSE server
- c.Writer.Header().Set("Content-Type", "text/event-stream; charset=utf-8")
- c.Writer.Header().Set("Cache-Control", "no-cache")
- c.Writer.Header().Set("Connection", "keep-alive")
- c.Writer.Header().Set("Access-Control-Allow-Origin", "*")
- config := openai.DefaultConfig(settings.OpenAISettings.Token)
- if settings.OpenAISettings.Proxy != "" {
- proxyUrl, err := url.Parse(settings.OpenAISettings.Proxy)
- if err != nil {
- c.Stream(func(w io.Writer) bool {
- c.SSEvent("message", gin.H{
- "type": "error",
- "content": err.Error(),
- })
- return false
- })
- return
- }
- transport := &http.Transport{
- Proxy: http.ProxyURL(proxyUrl),
- TLSClientConfig: &tls.Config{InsecureSkipVerify: settings.ServerSettings.InsecureSkipVerify},
- }
- config.HTTPClient = &http.Client{
- Transport: transport,
- }
- }
- if settings.OpenAISettings.BaseUrl != "" {
- config.BaseURL = settings.OpenAISettings.BaseUrl
- }
- openaiClient := openai.NewClientWithConfig(config)
- ctx := context.Background()
- req := openai.ChatCompletionRequest{
- Model: settings.OpenAISettings.Model,
- Messages: messages,
- Stream: true,
- }
- stream, err := openaiClient.CreateChatCompletionStream(ctx, req)
- if err != nil {
- fmt.Printf("CompletionStream error: %v\n", err)
- c.Stream(func(w io.Writer) bool {
- c.SSEvent("message", gin.H{
- "type": "error",
- "content": err.Error(),
- })
- return false
- })
- return
- }
- defer stream.Close()
- msgChan := make(chan string)
- go func() {
- defer close(msgChan)
- for {
- response, err := stream.Recv()
- if errors.Is(err, io.EOF) {
- fmt.Println()
- return
- }
- if err != nil {
- fmt.Printf("Stream error: %v\n", err)
- return
- }
- message := fmt.Sprintf("%s", response.Choices[0].Delta.Content)
- msgChan <- message
- }
- }()
- c.Stream(func(w io.Writer) bool {
- if m, ok := <-msgChan; ok {
- c.SSEvent("message", gin.H{
- "type": "message",
- "content": m,
- })
- return true
- }
- return false
- })
- }
|