NovelAI: Increase tier token limits

This commit is contained in:
Cohee 2024-10-17 23:47:47 +03:00
parent cb987706eb
commit 027d710716
1 changed files with 8 additions and 8 deletions

View File

@ -83,9 +83,9 @@ export function setNovelData(data) {
export function getKayraMaxContextTokens() { export function getKayraMaxContextTokens() {
switch (novel_data?.tier) { switch (novel_data?.tier) {
case 1: case 1:
return 3072; return 4096;
case 2: case 2:
return 6144; return 8192;
case 3: case 3:
return 8192; return 8192;
} }
@ -93,14 +93,14 @@ export function getKayraMaxContextTokens() {
return null; return null;
} }
export function getKayraMaxResponseTokens() { export function getNovelMaxResponseTokens() {
switch (novel_data?.tier) { switch (novel_data?.tier) {
case 1: case 1:
return 100;
case 2:
return 100;
case 3:
return 150; return 150;
case 2:
return 150;
case 3:
return 250;
} }
return maximum_output_length; return maximum_output_length;
@ -546,7 +546,7 @@ export function getNovelGenerationData(finalPrompt, settings, maxLength, isImper
finalPrompt = '<|startoftext|><|reserved_special_token81|>' + finalPrompt; finalPrompt = '<|startoftext|><|reserved_special_token81|>' + finalPrompt;
} }
const adjustedMaxLength = (isKayra || isErato) ? getKayraMaxResponseTokens() : maximum_output_length; const adjustedMaxLength = (isKayra || isErato) ? getNovelMaxResponseTokens() : maximum_output_length;
return { return {
'input': finalPrompt, 'input': finalPrompt,