mirror of https://github.com/velour/catbase.git
gpt: use llama as a backend
This commit is contained in:
parent
efc4605f39
commit
d6bb334a14
|
@ -0,0 +1,106 @@
|
|||
package gpt
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"github.com/rs/zerolog/log"
|
||||
"github.com/velour/catbase/bot"
|
||||
"github.com/velour/catbase/config"
|
||||
"regexp"
|
||||
"slices"
|
||||
)
|
||||
|
||||
const gpt3URL = "https://api.openai.com/v1/engines/%s/completions"
|
||||
const gpt3ModURL = "https://api.openai.com/v1/moderations"
|
||||
|
||||
type GPTPlugin struct {
|
||||
b bot.Bot
|
||||
c *config.Config
|
||||
h bot.HandlerTable
|
||||
|
||||
chatCount int
|
||||
chatHistory []chatEntry
|
||||
}
|
||||
|
||||
type chatEntry struct {
|
||||
Role string `json:"role"`
|
||||
Content string `json:"content"`
|
||||
}
|
||||
|
||||
func New(b bot.Bot) *GPTPlugin {
|
||||
p := &GPTPlugin{
|
||||
b: b,
|
||||
c: b.Config(),
|
||||
}
|
||||
p.register()
|
||||
return p
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) register() {
|
||||
p.h = bot.HandlerTable{
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^gpt (?P<text>.*)`),
|
||||
HelpText: "chat completion",
|
||||
Handler: p.chatMessageForce,
|
||||
},
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^got (?P<text>.*)`),
|
||||
HelpText: "chat completion",
|
||||
Handler: p.chatMessageForce,
|
||||
},
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^gpt-prompt: (?P<text>.*)`),
|
||||
HelpText: "set the ChatGPT prompt",
|
||||
Handler: p.setPromptMessage,
|
||||
},
|
||||
}
|
||||
p.b.RegisterTable(p, p.h)
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) setPromptMessage(r bot.Request) bool {
|
||||
prompt := r.Values["text"]
|
||||
if err := p.setPrompt(prompt); err != nil {
|
||||
resp := fmt.Sprintf("Error: %s", err)
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, resp)
|
||||
}
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, fmt.Sprintf(`Okay. I set the prompt to: "%s"`, prompt))
|
||||
return true
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) chatMessage(r bot.Request) bool {
|
||||
if slices.Contains(p.c.GetArray("gpt.silence", []string{}), r.Msg.Channel) {
|
||||
log.Debug().Msgf("%s silenced", r.Msg.Channel)
|
||||
return true
|
||||
}
|
||||
return p.chatMessageForce(r)
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) chatMessageForce(r bot.Request) bool {
|
||||
p.chatHistory = append(p.chatHistory, chatEntry{
|
||||
Role: "user",
|
||||
Content: r.Values["text"],
|
||||
})
|
||||
chatResp, err := p.llama()
|
||||
if err == nil {
|
||||
p.chatHistory = append(p.chatHistory, chatResp)
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, chatResp.Content)
|
||||
return true
|
||||
} else if !errors.Is(err, InstanceNotFoundError) {
|
||||
log.Error().Err(err).Msgf("error contacting llama")
|
||||
} else {
|
||||
log.Info().Msgf("Llama is currently down")
|
||||
}
|
||||
resp, err := p.chatGPT(r.Values["text"])
|
||||
if err != nil {
|
||||
resp = fmt.Sprintf("Error: %s", err)
|
||||
}
|
||||
p.chatHistory = append(p.chatHistory, chatEntry{
|
||||
Role: "assistant",
|
||||
Content: resp,
|
||||
})
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, resp)
|
||||
return true
|
||||
}
|
|
@ -1,260 +0,0 @@
|
|||
package gpt
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io"
|
||||
"math/rand"
|
||||
"net/http"
|
||||
"reflect"
|
||||
"regexp"
|
||||
"slices"
|
||||
"strings"
|
||||
|
||||
"github.com/rs/zerolog/log"
|
||||
"github.com/velour/catbase/bot"
|
||||
"github.com/velour/catbase/config"
|
||||
)
|
||||
|
||||
const gpt3URL = "https://api.openai.com/v1/engines/%s/completions"
|
||||
const gpt3ModURL = "https://api.openai.com/v1/moderations"
|
||||
|
||||
type GPTPlugin struct {
|
||||
b bot.Bot
|
||||
c *config.Config
|
||||
h bot.HandlerTable
|
||||
|
||||
chatCount int
|
||||
}
|
||||
|
||||
func New(b bot.Bot) *GPTPlugin {
|
||||
p := &GPTPlugin{
|
||||
b: b,
|
||||
c: b.Config(),
|
||||
}
|
||||
p.register()
|
||||
return p
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) register() {
|
||||
p.h = bot.HandlerTable{
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^gpt3 (?P<text>.*)`),
|
||||
HelpText: "request text completion",
|
||||
Handler: p.message,
|
||||
},
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^gpt (?P<text>.*)`),
|
||||
HelpText: "chat completion",
|
||||
Handler: p.chatMessageForce,
|
||||
},
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^got (?P<text>.*)`),
|
||||
HelpText: "chat completion",
|
||||
Handler: p.chatMessageForce,
|
||||
},
|
||||
{
|
||||
Kind: bot.Message, IsCmd: true,
|
||||
Regex: regexp.MustCompile(`(?is)^gpt-prompt: (?P<text>.*)`),
|
||||
HelpText: "set the ChatGPT prompt",
|
||||
Handler: p.setPromptMessage,
|
||||
},
|
||||
}
|
||||
p.b.RegisterTable(p, p.h)
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) setPromptMessage(r bot.Request) bool {
|
||||
prompt := r.Values["text"]
|
||||
if err := p.setPrompt(prompt); err != nil {
|
||||
resp := fmt.Sprintf("Error: %s", err)
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, resp)
|
||||
}
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, fmt.Sprintf(`Okay. I set the prompt to: "%s"`, prompt))
|
||||
return true
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) chatMessage(r bot.Request) bool {
|
||||
if slices.Contains(p.c.GetArray("gpt.silence", []string{}), r.Msg.Channel) {
|
||||
log.Debug().Msgf("%s silenced", r.Msg.Channel)
|
||||
return true
|
||||
}
|
||||
return p.chatMessageForce(r)
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) chatMessageForce(r bot.Request) bool {
|
||||
resp, err := p.chatGPT(r.Values["text"])
|
||||
if err != nil {
|
||||
resp = fmt.Sprintf("Error: %s", err)
|
||||
}
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, resp)
|
||||
return true
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) message(r bot.Request) bool {
|
||||
stem := r.Values["text"]
|
||||
p.b.Send(r.Conn, bot.Message, r.Msg.Channel, p.gpt3(stem))
|
||||
return true
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) gpt3(stem string) string {
|
||||
log.Debug().Msgf("Got GPT3 request: %s", stem)
|
||||
if err := p.checkStem(stem); err != nil {
|
||||
return "GPT3 moderation " + err.Error()
|
||||
}
|
||||
postStruct := gpt3Request{
|
||||
Prompt: stem,
|
||||
MaxTokens: p.c.GetInt("gpt3.tokens", 16),
|
||||
Temperature: p.c.GetFloat64("gpt3.temperature", 1),
|
||||
TopP: p.c.GetFloat64("gpt3.top_p", 1),
|
||||
N: p.c.GetInt("gpt3.n", 1),
|
||||
Stop: p.c.GetArray("gpt3.stop", []string{"\n"}),
|
||||
Echo: p.c.GetBool("gpt3.echo", false),
|
||||
}
|
||||
val, err := p.mkRequest(gpt3URL, postStruct)
|
||||
if err != nil {
|
||||
return err.Error()
|
||||
}
|
||||
choices := val.(gpt3Response).Choices
|
||||
if len(choices) > 0 {
|
||||
return choices[rand.Intn(len(choices))].Text
|
||||
}
|
||||
return "OpenAI is too shitty to respond to that."
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) mkRequest(endPoint string, postStruct interface{}) (interface{}, error) {
|
||||
postBody, _ := json.Marshal(postStruct)
|
||||
client := &http.Client{}
|
||||
u := fmt.Sprintf(endPoint, p.c.Get("gpt3.engine", "ada"))
|
||||
req, err := http.NewRequest("POST", u, bytes.NewBuffer(postBody))
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("could not make gpt3 request")
|
||||
return nil, err
|
||||
}
|
||||
gpt3Key := p.c.Get("gpt3.bearer", "")
|
||||
if gpt3Key == "" {
|
||||
log.Error().Msgf("no GPT3 key given")
|
||||
return nil, err
|
||||
}
|
||||
req.Header.Add("Content-Type", "application/json")
|
||||
req.Header.Add("Authorization", fmt.Sprintf("Bearer %s", gpt3Key))
|
||||
res, err := client.Do(req)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
resBody, _ := io.ReadAll(res.Body)
|
||||
gpt3Resp := gpt3Response{}
|
||||
err = json.Unmarshal(resBody, &gpt3Resp)
|
||||
|
||||
log.Debug().
|
||||
Str("body", string(resBody)).
|
||||
Interface("resp", gpt3Resp).
|
||||
Msg("OpenAI Response")
|
||||
|
||||
return gpt3Resp, nil
|
||||
}
|
||||
|
||||
func (p *GPTPlugin) checkStem(stem string) error {
|
||||
if !p.c.GetBool("gpt3.moderation", true) {
|
||||
return nil
|
||||
}
|
||||
postBody, _ := json.Marshal(gpt3ModRequest{Input: stem})
|
||||
client := &http.Client{}
|
||||
req, err := http.NewRequest("POST", gpt3ModURL, bytes.NewBuffer(postBody))
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
gpt3Key := p.c.Get("gpt3.bearer", "")
|
||||
if gpt3Key == "" {
|
||||
return fmt.Errorf("no GPT3 API key")
|
||||
}
|
||||
req.Header.Add("Content-Type", "application/json")
|
||||
req.Header.Add("Authorization", fmt.Sprintf("Bearer %s", gpt3Key))
|
||||
res, err := client.Do(req)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
resBody, _ := io.ReadAll(res.Body)
|
||||
log.Debug().Str("resBody", string(resBody)).Msg("res")
|
||||
gpt3Resp := gpt3Moderation{}
|
||||
err = json.Unmarshal(resBody, &gpt3Resp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
log.Debug().Interface("GPT3 Moderation", gpt3Resp).Msg("Moderation result")
|
||||
for _, res := range gpt3Resp.Results {
|
||||
if res.Flagged {
|
||||
list := ""
|
||||
categories := reflect.ValueOf(res.Categories)
|
||||
fields := reflect.VisibleFields(reflect.TypeOf(res.Categories))
|
||||
for i := 0; i < categories.NumField(); i++ {
|
||||
if categories.Field(i).Bool() {
|
||||
list += fields[i].Name + ", "
|
||||
}
|
||||
}
|
||||
list = strings.TrimSuffix(list, ", ")
|
||||
return fmt.Errorf("flagged: %s", list)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
type gpt3Request struct {
|
||||
Prompt string `json:"prompt"`
|
||||
MaxTokens int `json:"max_tokens"`
|
||||
Temperature float64 `json:"temperature"`
|
||||
TopP float64 `json:"top_p"`
|
||||
N int `json:"n"`
|
||||
Stream bool `json:"stream"`
|
||||
Logprobs any `json:"logprobs"`
|
||||
Stop []string `json:"stop"`
|
||||
Echo bool `json:"echo"`
|
||||
}
|
||||
|
||||
type gpt3ModRequest struct {
|
||||
Input string `json:"input"`
|
||||
}
|
||||
|
||||
type gpt3Response struct {
|
||||
ID string `json:"id"`
|
||||
Object string `json:"object"`
|
||||
Created int `json:"created"`
|
||||
Model string `json:"model"`
|
||||
Choices []struct {
|
||||
Text string `json:"text"`
|
||||
Index int `json:"index"`
|
||||
Logprobs any `json:"logprobs"`
|
||||
FinishReason string `json:"finish_reason"`
|
||||
} `json:"choices"`
|
||||
}
|
||||
|
||||
type gpt3Moderation struct {
|
||||
ID string `json:"id"`
|
||||
Model string `json:"model"`
|
||||
Results []struct {
|
||||
Categories struct {
|
||||
Hate bool `json:"hate"`
|
||||
HateThreatening bool `json:"hate/threatening"`
|
||||
SelfHarm bool `json:"self-harm"`
|
||||
Sexual bool `json:"sexual"`
|
||||
SexualMinors bool `json:"sexual/minors"`
|
||||
Violence bool `json:"violence"`
|
||||
ViolenceGraphic bool `json:"violence/graphic"`
|
||||
} `json:"categories"`
|
||||
CategoryScores struct {
|
||||
Hate float64 `json:"hate"`
|
||||
HateThreatening float64 `json:"hate/threatening"`
|
||||
SelfHarm float64 `json:"self-harm"`
|
||||
Sexual float64 `json:"sexual"`
|
||||
SexualMinors float64 `json:"sexual/minors"`
|
||||
Violence float64 `json:"violence"`
|
||||
ViolenceGraphic float64 `json:"violence/graphic"`
|
||||
} `json:"category_scores"`
|
||||
Flagged bool `json:"flagged"`
|
||||
} `json:"results"`
|
||||
}
|
|
@ -0,0 +1,73 @@
|
|||
package gpt
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"encoding/json"
|
||||
"errors"
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"time"
|
||||
)
|
||||
|
||||
var InstanceNotFoundError = errors.New("instance not found")
|
||||
var empty = chatEntry{}
|
||||
|
||||
func (g *GPTPlugin) llama() (chatEntry, error) {
|
||||
llamaURL := g.c.Get("gpt.llamaurl", "")
|
||||
if llamaURL == "" {
|
||||
return empty, fmt.Errorf("could not find llama url")
|
||||
}
|
||||
llamaModel := g.c.Get("gpt.llamamodel", "")
|
||||
if llamaModel == "" {
|
||||
return empty, fmt.Errorf("could not find llama model")
|
||||
}
|
||||
|
||||
req := llamaRequest{
|
||||
Model: llamaModel,
|
||||
Messages: g.chatHistory,
|
||||
Stream: false,
|
||||
}
|
||||
|
||||
body, err := json.Marshal(req)
|
||||
if err != nil {
|
||||
return empty, fmt.Errorf("could not marshal llama request: %w", err)
|
||||
}
|
||||
|
||||
resp, err := http.Post(llamaURL, "application/json", bytes.NewBuffer(body))
|
||||
if err != nil {
|
||||
return empty, fmt.Errorf("could not post llama request: %w", err)
|
||||
}
|
||||
|
||||
if resp.StatusCode == 503 {
|
||||
return empty, InstanceNotFoundError
|
||||
}
|
||||
body, _ = io.ReadAll(resp.Body)
|
||||
|
||||
llamaResp := llamaResponse{}
|
||||
err = json.Unmarshal(body, &llamaResp)
|
||||
if err != nil {
|
||||
return empty, fmt.Errorf("could not unmarshal llama response: %w, raw: %s", err, string(body))
|
||||
}
|
||||
|
||||
return llamaResp.Message, nil
|
||||
}
|
||||
|
||||
type llamaRequest struct {
|
||||
Model string `json:"model"`
|
||||
Stream bool `json:"stream"`
|
||||
Messages []chatEntry `json:"messages"`
|
||||
}
|
||||
|
||||
type llamaResponse struct {
|
||||
Model string `json:"model"`
|
||||
CreatedAt time.Time `json:"created_at"`
|
||||
Message chatEntry `json:"message"`
|
||||
DoneReason string `json:"done_reason"`
|
||||
Done bool `json:"done"`
|
||||
TotalDuration int64 `json:"total_duration"`
|
||||
LoadDuration int `json:"load_duration"`
|
||||
PromptEvalDuration int `json:"prompt_eval_duration"`
|
||||
EvalCount int `json:"eval_count"`
|
||||
EvalDuration int64 `json:"eval_duration"`
|
||||
}
|
Loading…
Reference in New Issue