Fixing user request representation in the LLM request context (#40) #41
|
@ -36,7 +36,7 @@ func (l *LlmConnector) HandleChatMessage(userMessage ChatMessage, model string,
|
||||||
historyLength := len(requestContext.Chat.History)
|
historyLength := len(requestContext.Chat.History)
|
||||||
|
|
||||||
if historyLength > 0 {
|
if historyLength > 0 {
|
||||||
systemPrompt += "\nYou have access to last " + strconv.Itoa(historyLength) + "messages in this chat."
|
systemPrompt += "\nYou have access to last " + strconv.Itoa(historyLength) + " messages in this chat."
|
||||||
}
|
}
|
||||||
|
|
||||||
req := openai.ChatCompletionRequest{
|
req := openai.ChatCompletionRequest{
|
||||||
|
|
|
@ -56,7 +56,7 @@ func (c RequestContext) Prompt() string {
|
||||||
prompt += "Chat description is \"" + c.Chat.Description + "\". "
|
prompt += "Chat description is \"" + c.Chat.Description + "\". "
|
||||||
}
|
}
|
||||||
|
|
||||||
prompt += "\nProfile of the user who mentioned you in the chat:" +
|
prompt += "\nProfile of the user who mentioned you in the chat:\n" +
|
||||||
"First name: \"" + c.User.FirstName + "\"\n"
|
"First name: \"" + c.User.FirstName + "\"\n"
|
||||||
if c.User.Username != "" {
|
if c.User.Username != "" {
|
||||||
prompt += "Username: @" + c.User.Username + ".\n"
|
prompt += "Username: @" + c.User.Username + ".\n"
|
||||||
|
|
Loading…
Reference in a new issue