diff --git a/public/scripts/extensions/memory/index.js b/public/scripts/extensions/memory/index.js index dd3178f9c..862131474 100644 --- a/public/scripts/extensions/memory/index.js +++ b/public/scripts/extensions/memory/index.js @@ -28,6 +28,7 @@ import { ARGUMENT_TYPE, SlashCommandArgument, SlashCommandNamedArgument } from ' import { MacrosParser } from '../../macros.js'; import { countWebLlmTokens, generateWebLlmChatPrompt, getWebLlmContextSize, isWebLlmSupported } from '../shared.js'; import { commonEnumProviders } from '../../slash-commands/SlashCommandCommonEnumsProvider.js'; +import { removeReasoningFromString } from '../../reasoning.js'; export { MODULE_NAME }; const MODULE_NAME = '1_memory'; @@ -504,7 +505,7 @@ async function summarizeCallback(args, text) { case summary_sources.extras: return await callExtrasSummarizeAPI(text); case summary_sources.main: - return await generateRaw(text, '', false, false, prompt, extension_settings.memory.overrideResponseLength); + return removeReasoningFromString(await generateRaw(text, '', false, false, prompt, extension_settings.memory.overrideResponseLength)); case summary_sources.webllm: { const messages = [{ role: 'system', content: prompt }, { role: 'user', content: text }].filter(m => m.content); const params = extension_settings.memory.overrideResponseLength > 0 ? { max_tokens: extension_settings.memory.overrideResponseLength } : {}; @@ -699,7 +700,8 @@ async function summarizeChatMain(context, force, skipWIAN) { return null; } - summary = await generateRaw(rawPrompt, '', false, false, prompt, extension_settings.memory.overrideResponseLength); + const rawSummary = await generateRaw(rawPrompt, '', false, false, prompt, extension_settings.memory.overrideResponseLength); + summary = removeReasoningFromString(rawSummary); index = lastUsedIndex; } finally { inApiCall = false; diff --git a/public/scripts/extensions/vectors/index.js b/public/scripts/extensions/vectors/index.js index 4d238ff4e..81c40b3b9 100644 --- a/public/scripts/extensions/vectors/index.js +++ b/public/scripts/extensions/vectors/index.js @@ -36,6 +36,7 @@ import { slashCommandReturnHelper } from '../../slash-commands/SlashCommandRetur import { callGenericPopup, POPUP_RESULT, POPUP_TYPE } from '../../popup.js'; import { generateWebLlmChatPrompt, isWebLlmSupported } from '../shared.js'; import { WebLlmVectorProvider } from './webllm.js'; +import { removeReasoningFromString } from '../../reasoning.js'; /** * @typedef {object} HashedMessage @@ -260,7 +261,7 @@ async function summarizeExtra(element) { * @returns {Promise} Sucess */ async function summarizeMain(element) { - element.text = await generateRaw(element.text, '', false, false, settings.summary_prompt); + element.text = removeReasoningFromString(await generateRaw(element.text, '', false, false, settings.summary_prompt)); return true; } diff --git a/public/scripts/openai.js b/public/scripts/openai.js index 3d269fd07..a8a8cdb3f 100644 --- a/public/scripts/openai.js +++ b/public/scripts/openai.js @@ -765,7 +765,10 @@ async function populationInjectionPrompts(prompts, messages) { const wrap = false; // Group prompts by priority - const orderGroups = {}; + const extensionPromptsOrder = '0'; + const orderGroups = { + [extensionPromptsOrder]: [], + }; for (const prompt of depthPrompts) { const order = prompt.injection_order || 0; if (!orderGroups[order]) { @@ -788,7 +791,9 @@ async function populationInjectionPrompts(prompts, messages) { .join(separator); // Get extension prompt - const extensionPrompt = await getExtensionPrompt(extension_prompt_types.IN_CHAT, i, separator, roleTypes[role], wrap); + const extensionPrompt = order === extensionPromptsOrder + ? await getExtensionPrompt(extension_prompt_types.IN_CHAT, i, separator, roleTypes[role], wrap) + : ''; const jointPrompt = [rolePrompts, extensionPrompt].filter(x => x).map(x => x.trim()).join(separator); if (jointPrompt && jointPrompt.length) { diff --git a/src/endpoints/stable-diffusion.js b/src/endpoints/stable-diffusion.js index 74883ea58..7b25d81da 100644 --- a/src/endpoints/stable-diffusion.js +++ b/src/endpoints/stable-diffusion.js @@ -824,7 +824,8 @@ pollinations.post('/generate', async (request, response) => { height: String(request.body.height ?? 1024), nologo: String(true), nofeed: String(true), - referer: 'sillytavern', + private: String(true), + referrer: 'sillytavern', }); promptUrl.search = params.toString();