Merge branch 'staging' into connection-profile-omit

This commit is contained in:
Cohee 2024-09-26 23:27:54 +03:00
commit 8083431202
5 changed files with 35 additions and 10 deletions

View File

@ -1310,7 +1310,7 @@
</div>
</div>
<div data-tg-type="koboldcpp, aphrodite" id="xtc_block" class="wide100p">
<div data-tg-type="koboldcpp, aphrodite, tabby" id="xtc_block" class="wide100p">
<h4 class="wide100p textAlignCenter">
<label data-i18n="Exclude Top Choices (XTC)">Exclude Top Choices (XTC)</label>
<a href="https://github.com/oobabooga/text-generation-webui/pull/6335" target="_blank">

View File

@ -4863,7 +4863,7 @@ export function getMaxContextSize(overrideResponseLength = null) {
this_max_context = Math.min(max_context, 8192);
// Added special tokens and whatnot
this_max_context -= 1;
this_max_context -= 10;
}
this_max_context = this_max_context - (overrideResponseLength || amount_gen);

View File

@ -492,11 +492,37 @@ function getBadWordPermutations(text) {
export function getNovelGenerationData(finalPrompt, settings, maxLength, isImpersonate, isContinue, _cfgValues, type) {
console.debug('NovelAI generation data for', type);
const isKayra = nai_settings.model_novel.includes('kayra');
const isErato = nai_settings.model_novel.includes('erato');
const tokenizerType = getTokenizerTypeForModel(nai_settings.model_novel);
const stoppingStrings = getStoppingStrings(isImpersonate, isContinue);
// Llama 3 tokenizer, huh?
if (isErato) {
const additionalStopStrings = [];
for (const stoppingString of stoppingStrings) {
if (stoppingString.startsWith('\n')) {
additionalStopStrings.push('.' + stoppingString);
additionalStopStrings.push('!' + stoppingString);
additionalStopStrings.push('?' + stoppingString);
additionalStopStrings.push('*' + stoppingString);
additionalStopStrings.push('"' + stoppingString);
additionalStopStrings.push('_' + stoppingString);
additionalStopStrings.push('...' + stoppingString);
additionalStopStrings.push('."' + stoppingString);
additionalStopStrings.push('?"' + stoppingString);
additionalStopStrings.push('!"' + stoppingString);
additionalStopStrings.push('.*' + stoppingString);
additionalStopStrings.push(')' + stoppingString);
}
}
stoppingStrings.push(...additionalStopStrings);
}
const MAX_STOP_SEQUENCES = 1024;
const stopSequences = (tokenizerType !== tokenizers.NONE)
? getStoppingStrings(isImpersonate, isContinue)
.map(t => getTextTokens(tokenizerType, t))
? stoppingStrings.slice(0, MAX_STOP_SEQUENCES).map(t => getTextTokens(tokenizerType, t))
: undefined;
const badWordIds = (tokenizerType !== tokenizers.NONE)
@ -515,11 +541,9 @@ export function getNovelGenerationData(finalPrompt, settings, maxLength, isImper
console.log(finalPrompt);
}
const isKayra = nai_settings.model_novel.includes('kayra');
const isErato = nai_settings.model_novel.includes('erato');
if (isErato) {
finalPrompt = '<|startoftext|>' + finalPrompt;
finalPrompt = '<|startoftext|><|reserved_special_token81|>' + finalPrompt;
}
const adjustedMaxLength = (isKayra || isErato) ? getKayraMaxResponseTokens() : maximum_output_length;

View File

@ -143,7 +143,7 @@ function selectSystemPromptCallback(args, name) {
foundName = result[0].item;
}
$select.val(foundName).trigger('input');
$select.val(foundName).trigger('change');
!quiet && toastr.success(`System prompt "${foundName}" selected`);
return foundName;
}

View File

@ -3245,8 +3245,9 @@ grammarly-extension {
}
.wide_dialogue_popup {
aspect-ratio: 1 / 1;
width: unset !important;
/* FIXME: Chrome 129 broke max-height for aspect-ratio sized elements */
/* aspect-ratio: 1 / 1; */
/* width: unset !important; */
min-width: var(--sheldWidth);
}