mirror of
https://git.femboyfinancial.jp/james/FemScoreboard.git
synced 2024-11-24 11:22:00 -08:00
Change chat cmd max tokens, replies
This commit is contained in:
parent
19346fb2c3
commit
8346f52f23
@ -3,6 +3,7 @@ REACTIONS="💀,💯,😭"
|
||||
CLIENT="123456789012345678"
|
||||
GUILD="123456789012345678"
|
||||
|
||||
LLAMACPP_HOST=127.0.0.1
|
||||
LLAMACPP_PORT=9999
|
||||
|
||||
ENABLE_MOTD=1
|
||||
|
@ -1,17 +1,13 @@
|
||||
import {
|
||||
ChatInputCommandInteraction,
|
||||
Interaction,
|
||||
InteractionResponse,
|
||||
Message,
|
||||
MessageComponentInteraction,
|
||||
SlashCommandBuilder
|
||||
} from 'discord.js';
|
||||
import { ChatMessage, llamacpp, streamText } from 'modelfusion';
|
||||
import { llamacpp, streamText } from 'modelfusion';
|
||||
import { logInfo, logError } from '../../../logging';
|
||||
|
||||
const llamaCppServer = llamacpp.Api({
|
||||
baseUrl: {
|
||||
host: "localhost",
|
||||
host: process.env.LLAMACPP_HOST,
|
||||
port: process.env.LLAMACPP_PORT,
|
||||
}
|
||||
});
|
||||
@ -25,34 +21,32 @@ async function llamaChat(interaction: ChatInputCommandInteraction)
|
||||
model: llamacpp
|
||||
.CompletionTextGenerator({
|
||||
promptTemplate: llamacpp.prompt.Llama2,
|
||||
maxGenerationTokens: 400,
|
||||
maxGenerationTokens: 16,
|
||||
temperature: 0.7,
|
||||
api: llamaCppServer
|
||||
})
|
||||
.withChatPrompt(),
|
||||
prompt: {
|
||||
system: "You are Hatsune Miku, the famous 16-year-old Japanese virtual singer from Crypton Future Media. You have a penchant for politically incorrect humor, and are making banter with your like-minded friends.",
|
||||
system: 'You are Hatsune Miku, the famous 16-year-old Japanese virtual singer from Crypton Future Media. You have a penchant for politically incorrect humor, and are making banter with your like-minded friends.',
|
||||
messages: [{
|
||||
"role": "user",
|
||||
"content": interaction.options.getString('prompt')
|
||||
'role': 'user',
|
||||
'content': interaction.options.getString('prompt')
|
||||
}]
|
||||
}
|
||||
});
|
||||
|
||||
let outMsg: InteractionResponse;
|
||||
let allText = '';
|
||||
await interaction.deferReply();
|
||||
try {
|
||||
for await (const textPart of textStream) {
|
||||
logInfo(`[chat] Added to LLaMA response: ${textPart}`);
|
||||
allText += textPart;
|
||||
if (!outMsg) {
|
||||
outMsg = await interaction.reply(allText);
|
||||
} else {
|
||||
await outMsg.edit(allText);
|
||||
}
|
||||
await interaction.editReply(allText);
|
||||
}
|
||||
logInfo(`[chat] Final LLaMA response: ${allText}`);
|
||||
} catch (err) {
|
||||
logError(err);
|
||||
await interaction.reply(err.toString());
|
||||
await interaction.editReply(err.toString());
|
||||
}
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user