From 47cb017a45bac9c52e6978fa30f88d0bd67ba8c7 Mon Sep 17 00:00:00 2001 From: Cohee <18619528+Cohee1207@users.noreply.github.com> Date: Mon, 25 Dec 2023 02:42:03 +0200 Subject: [PATCH] #1589 Add 'cache_prompt' for l.cpp --- public/scripts/textgen-settings.js | 1 + 1 file changed, 1 insertion(+) diff --git a/public/scripts/textgen-settings.js b/public/scripts/textgen-settings.js index 475aec70f..c5af0b583 100644 --- a/public/scripts/textgen-settings.js +++ b/public/scripts/textgen-settings.js @@ -769,6 +769,7 @@ export function getTextGenGenerationData(finalPrompt, maxTokens, isImpersonate, 'logit_bias': logitBiasArray, // Conflicts with ooba's grammar_string 'grammar': settings.grammar_string, + 'cache_prompt': true, }; params = Object.assign(params, llamaCppParams); }