From 8bf679cba278ac7e4337b3ac420441492f96db14 Mon Sep 17 00:00:00 2001 From: Cohee <18619528+Cohee1207@users.noreply.github.com> Date: Sun, 27 Aug 2023 19:24:28 +0300 Subject: [PATCH] #1033 Promt => prompt --- public/script.js | 72 +++++++++++++++++----------------- public/scripts/nai-settings.js | 4 +- 2 files changed, 38 insertions(+), 38 deletions(-) diff --git a/public/script.js b/public/script.js index 52dfcbbb8..66a8f0778 100644 --- a/public/script.js +++ b/public/script.js @@ -327,7 +327,7 @@ let firstRun = false; const default_ch_mes = "Hello"; let count_view_mes = 0; -let generatedPromtCache = ""; +let generatedPromptCache = ""; let generation_started = new Date(); let characters = []; let this_chid; @@ -1445,7 +1445,7 @@ function addOneMessage(mes, { type = "normal", insertAfter = null, scroll = true var avatarImg = getUserAvatar(user_avatar); const isSystem = mes.is_system; const title = mes.title; - generatedPromtCache = ""; + generatedPromptCache = ""; //for non-user mesages if (!mes["is_user"]) { @@ -2119,7 +2119,7 @@ class StreamingProcessor { activateSendButtons(); showSwipeButtons(); setGenerationProgress(0); - generatedPromtCache = ''; + generatedPromptCache = ''; //console.log("Generated text size:", text.length, text) @@ -2621,13 +2621,13 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, const originalType = type; runGenerate(cyclePrompt); - async function runGenerate(cycleGenerationPromt = '') { + async function runGenerate(cycleGenerationPrompt = '') { if (!dryRun) { is_send_press = true; } - generatedPromtCache += cycleGenerationPromt; - if (generatedPromtCache.length == 0 || type === 'continue') { + generatedPromptCache += cycleGenerationPrompt; + if (generatedPromptCache.length == 0 || type === 'continue') { if (main_api === 'openai') { generateOpenAIPromptCache(); } @@ -2657,7 +2657,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, let mesExmString = ''; - function setPromtString() { + function setPromptString() { if (main_api == 'openai') { return; } @@ -2728,26 +2728,26 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, return promptCache; } - function checkPromtSize() { + function checkPromptSize() { console.debug('---checking Prompt size'); - setPromtString(); + setPromptString(); const prompt = [ storyString, mesExmString, mesSend.join(''), - generatedPromtCache, + generatedPromptCache, allAnchors, quiet_prompt, ].join('').replace(/\r/gm, ''); - let thisPromtContextSize = getTokenCount(prompt, power_user.token_padding); + let thisPromptContextSize = getTokenCount(prompt, power_user.token_padding); - if (thisPromtContextSize > this_max_context) { //if the prepared prompt is larger than the max context size... + if (thisPromptContextSize > this_max_context) { //if the prepared prompt is larger than the max context size... if (count_exm_add > 0) { // ..and we have example mesages.. count_exm_add--; // remove the example messages... - checkPromtSize(); // and try agin... + checkPromptSize(); // and try agin... } else if (mesSend.length > 0) { // if the chat history is longer than 0 mesSend.shift(); // remove the first (oldest) chat entry.. - checkPromtSize(); // and check size again.. + checkPromptSize(); // and check size again.. } else { //end console.debug(`---mesSend.length = ${mesSend.length}`); @@ -2755,12 +2755,12 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, } } - if (generatedPromtCache.length > 0 && main_api !== 'openai') { - console.debug('---Generated Prompt Cache length: ' + generatedPromtCache.length); - checkPromtSize(); + if (generatedPromptCache.length > 0 && main_api !== 'openai') { + console.debug('---Generated Prompt Cache length: ' + generatedPromptCache.length); + checkPromptSize(); } else { - console.debug('---calling setPromtString ' + generatedPromtCache.length) - setPromtString(); + console.debug('---calling setPromptString ' + generatedPromptCache.length) + setPromptString(); } // Fetches the combined prompt for both negative and positive prompts @@ -2849,8 +2849,8 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, } // Prune from prompt cache if it exists - if (generatedPromtCache.length !== 0) { - generatedPromtCache = cleanupPromptCache(generatedPromtCache); + if (generatedPromptCache.length !== 0) { + generatedPromptCache = cleanupPromptCache(generatedPromptCache); } // Right now, everything is suffixed with a newline @@ -2867,7 +2867,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, afterScenarioAnchor + mesExmString + mesSendString + - generatedPromtCache; + generatedPromptCache; combinedPrompt = combinedPrompt.replace(/\r/gm, ''); @@ -2980,9 +2980,9 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, afterScenarioAnchor: afterScenarioAnchor, examplesString: examplesString, mesSendString: mesSendString, - generatedPromtCache: generatedPromtCache, + generatedPromptCache: generatedPromptCache, promptBias: promptBias, - finalPromt: finalPrompt, + finalPrompt: finalPrompt, charDescription: charDescription, charPersonality: charPersonality, scenarioText: scenarioText, @@ -3058,7 +3058,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, async function onSuccess(data) { if (data.error == 'dryRun') { - generatedPromtCache = ''; + generatedPromptCache = ''; resolve(); return; } @@ -3108,7 +3108,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, } tokens_already_generated = 0; - generatedPromtCache = ""; + generatedPromptCache = ""; const substringStart = originalType !== 'continue' ? magFirst.length : 0; getMessage = message_already_generated.substring(substringStart); } @@ -3126,7 +3126,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, if (getMessage.length > 0) { if (isImpersonate) { $('#send_textarea').val(getMessage).trigger('input'); - generatedPromtCache = ""; + generatedPromptCache = ""; await eventSource.emit(event_types.IMPERSONATE_READY, getMessage); } else if (type == 'quiet') { @@ -3189,7 +3189,7 @@ async function Generate(type, { automatic_trigger, force_name2, resolve, reject, } } } else { - generatedPromtCache = ''; + generatedPromptCache = ''; activateSendButtons(); //console.log('runGenerate calling showSwipeBtns'); showSwipeButtons(); @@ -3402,21 +3402,21 @@ function addChatsSeparator(mesSendString) { } } -function appendZeroDepthAnchor(force_name2, zeroDepthAnchor, finalPromt) { +function appendZeroDepthAnchor(force_name2, zeroDepthAnchor, finalPrompt) { const trimBothEnds = !force_name2 && !is_pygmalion; let trimmedPrompt = (trimBothEnds ? zeroDepthAnchor.trim() : zeroDepthAnchor.trimEnd()); - if (trimBothEnds && !finalPromt.endsWith('\n')) { - finalPromt += '\n'; + if (trimBothEnds && !finalPrompt.endsWith('\n')) { + finalPrompt += '\n'; } - finalPromt += trimmedPrompt; + finalPrompt += trimmedPrompt; if (force_name2 || is_pygmalion) { - finalPromt += ' '; + finalPrompt += ' '; } - return finalPromt; + return finalPrompt; } function getMultigenAmount() { @@ -3553,7 +3553,7 @@ function promptItemize(itemizedPrompts, requestedMesId) { } else { //for non-OAI APIs //console.log('-- Counting non-OAI Tokens'); - var finalPromptTokens = getTokenCount(itemizedPrompts[thisPromptSet].finalPromt); + var finalPromptTokens = getTokenCount(itemizedPrompts[thisPromptSet].finalPrompt); var storyStringTokens = getTokenCount(itemizedPrompts[thisPromptSet].storyString) - worldInfoStringTokens; var examplesStringTokens = getTokenCount(itemizedPrompts[thisPromptSet].examplesString); var mesSendStringTokens = getTokenCount(itemizedPrompts[thisPromptSet].mesSendString) @@ -3570,7 +3570,7 @@ function promptItemize(itemizedPrompts, requestedMesId) { //afterScenarioAnchorTokens + //only counts if AN is set to 'after scenario' //zeroDepthAnchorTokens + //same as above, even if AN not on 0 depth promptBiasTokens; //{{}} - //- thisPrompt_padding; //not sure this way of calculating is correct, but the math results in same value as 'finalPromt' + //- thisPrompt_padding; //not sure this way of calculating is correct, but the math results in same value as 'finalPrompt' } if (this_main_api == 'openai') { diff --git a/public/scripts/nai-settings.js b/public/scripts/nai-settings.js index 71c45aadc..fb7328c10 100644 --- a/public/scripts/nai-settings.js +++ b/public/scripts/nai-settings.js @@ -453,7 +453,7 @@ export function getNovelGenerationData(finalPrompt, this_settings, this_amount_g } // Check if the prefix needs to be overriden to use instruct mode -function selectPrefix(selected_prefix, finalPromt) { +function selectPrefix(selected_prefix, finalPrompt) { let useInstruct = false; const clio = nai_settings.model_novel.includes('clio'); const kayra = nai_settings.model_novel.includes('kayra'); @@ -461,7 +461,7 @@ function selectPrefix(selected_prefix, finalPromt) { if (isNewModel) { // NovelAI claims they scan backwards 1000 characters (not tokens!) to look for instruct brackets. That's really short. - const tail = finalPromt.slice(-1500); + const tail = finalPrompt.slice(-1500); useInstruct = tail.includes("}"); return useInstruct ? "special_instruct" : selected_prefix; }