diff --git a/plugins/deadend/deadend.go b/plugins/deadend/deadend.go index 2524069..4c8e74d 100644 --- a/plugins/deadend/deadend.go +++ b/plugins/deadend/deadend.go @@ -6,7 +6,7 @@ import ( "regexp" ) -const defaultMessage = "I don't know how to respond to that. If you'd like to ask GPT, use the `gpt` command." +const defaultMessage = "I don't know how to respond to that. If you'd like to ask an LLM, use the `llm` command." type DeadEndPlugin struct { b bot.Bot diff --git a/plugins/llm/gpt.go b/plugins/llm/gpt.go index b0ad39b..bab3d7d 100644 --- a/plugins/llm/gpt.go +++ b/plugins/llm/gpt.go @@ -7,7 +7,6 @@ import ( "github.com/velour/catbase/bot" "github.com/velour/catbase/config" "regexp" - "slices" ) const gpt3URL = "https://api.openai.com/v1/engines/%s/completions" @@ -47,20 +46,20 @@ func (p *LLMPlugin) register() { { Kind: bot.Message, IsCmd: true, Regex: regexp.MustCompile(`(?is)^llm (?P.*)`), - HelpText: "chat completion", - Handler: p.chatMessageForce, + HelpText: "chat completion using first-available AI", + Handler: p.chatMessage, }, { Kind: bot.Message, IsCmd: true, - Regex: regexp.MustCompile(`(?is)^gpt (?P.*)`), - HelpText: "chat completion", - Handler: p.chatMessageForce, + Regex: regexp.MustCompile(`(?is)^gpt4 (?P.*)`), + HelpText: "chat completion using OpenAI", + Handler: p.gptMessage, }, { Kind: bot.Message, IsCmd: true, Regex: regexp.MustCompile(`(?is)^got (?P.*)`), HelpText: "chat completion", - Handler: p.chatMessageForce, + Handler: p.chatMessage, }, } p.b.RegisterTable(p, p.h) @@ -77,14 +76,6 @@ func (p *LLMPlugin) setPromptMessage(r bot.Request) bool { } func (p *LLMPlugin) chatMessage(r bot.Request) bool { - if slices.Contains(p.c.GetArray("gpt.silence", []string{}), r.Msg.Channel) { - log.Debug().Msgf("%s silenced", r.Msg.Channel) - return true - } - return p.chatMessageForce(r) -} - -func (p *LLMPlugin) chatMessageForce(r bot.Request) bool { p.chatHistory = append(p.chatHistory, chatEntry{ Role: "user", Content: r.Values["text"], @@ -103,6 +94,10 @@ func (p *LLMPlugin) chatMessageForce(r bot.Request) bool { } else { log.Info().Msgf("Llama is currently down") } + return p.gptMessage(r) +} + +func (p *LLMPlugin) gptMessage(r bot.Request) bool { resp, err := p.chatGPT(r.Values["text"]) if err != nil { resp = fmt.Sprintf("Error: %s", err)