mirror of
https://github.com/SillyTavern/SillyTavern.git
synced 2025-06-05 21:59:27 +02:00
Enable no-case-declarations lint
This commit is contained in:
@ -61,7 +61,6 @@ module.exports = {
|
|||||||
'no-prototype-builtins': 'off',
|
'no-prototype-builtins': 'off',
|
||||||
'no-extra-boolean-cast': 'off',
|
'no-extra-boolean-cast': 'off',
|
||||||
'require-yield': 'off',
|
'require-yield': 'off',
|
||||||
'no-case-declarations': 'off',
|
|
||||||
'no-constant-condition': ['error', {checkLoops: false}]
|
'no-constant-condition': ['error', {checkLoops: false}]
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -2755,10 +2755,11 @@ export async function generateRaw(prompt, api, instructOverride) {
|
|||||||
generateData = getKoboldGenerationData(prompt, koboldSettings, amount_gen, max_context, isHorde, 'quiet');
|
generateData = getKoboldGenerationData(prompt, koboldSettings, amount_gen, max_context, isHorde, 'quiet');
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case 'novel':
|
case 'novel': {
|
||||||
const novelSettings = novelai_settings[novelai_setting_names[nai_settings.preset_settings_novel]];
|
const novelSettings = novelai_settings[novelai_setting_names[nai_settings.preset_settings_novel]];
|
||||||
generateData = getNovelGenerationData(prompt, novelSettings, amount_gen, false, false, null);
|
generateData = getNovelGenerationData(prompt, novelSettings, amount_gen, false, false, null);
|
||||||
break;
|
break;
|
||||||
|
}
|
||||||
case 'textgenerationwebui':
|
case 'textgenerationwebui':
|
||||||
generateData = getTextGenGenerationData(prompt, amount_gen, false, false, null);
|
generateData = getTextGenGenerationData(prompt, amount_gen, false, false, null);
|
||||||
break;
|
break;
|
||||||
|
@ -236,14 +236,16 @@ class PresetManager {
|
|||||||
return nai_settings;
|
return nai_settings;
|
||||||
case "textgenerationwebui":
|
case "textgenerationwebui":
|
||||||
return textgenerationwebui_settings;
|
return textgenerationwebui_settings;
|
||||||
case "context":
|
case "context": {
|
||||||
const context_preset = getContextSettings();
|
const context_preset = getContextSettings();
|
||||||
context_preset['name'] = name || power_user.context.preset;
|
context_preset['name'] = name || power_user.context.preset;
|
||||||
return context_preset;
|
return context_preset;
|
||||||
case "instruct":
|
}
|
||||||
|
case "instruct": {
|
||||||
const instruct_preset = structuredClone(power_user.instruct);
|
const instruct_preset = structuredClone(power_user.instruct);
|
||||||
instruct_preset['name'] = name || power_user.instruct.preset;
|
instruct_preset['name'] = name || power_user.instruct.preset;
|
||||||
return instruct_preset;
|
return instruct_preset;
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
console.warn(`Unknown API ID ${apiId}`);
|
console.warn(`Unknown API ID ${apiId}`);
|
||||||
return {};
|
return {};
|
||||||
|
@ -507,9 +507,10 @@ export function getTextTokens(tokenizerType, str) {
|
|||||||
return getTextTokensRemote('/api/tokenize/mistral', str);
|
return getTextTokensRemote('/api/tokenize/mistral', str);
|
||||||
case tokenizers.YI:
|
case tokenizers.YI:
|
||||||
return getTextTokensRemote('/api/tokenize/yi', str);
|
return getTextTokensRemote('/api/tokenize/yi', str);
|
||||||
case tokenizers.OPENAI:
|
case tokenizers.OPENAI: {
|
||||||
const model = getTokenizerModel();
|
const model = getTokenizerModel();
|
||||||
return getTextTokensRemote('/api/tokenize/openai-encode', str, model);
|
return getTextTokensRemote('/api/tokenize/openai-encode', str, model);
|
||||||
|
}
|
||||||
case tokenizers.API:
|
case tokenizers.API:
|
||||||
return getTextTokensRemote('/tokenize_via_api', str);
|
return getTextTokensRemote('/tokenize_via_api', str);
|
||||||
default:
|
default:
|
||||||
@ -537,9 +538,10 @@ export function decodeTextTokens(tokenizerType, ids) {
|
|||||||
return decodeTextTokensRemote('/api/decode/mistral', ids);
|
return decodeTextTokensRemote('/api/decode/mistral', ids);
|
||||||
case tokenizers.YI:
|
case tokenizers.YI:
|
||||||
return decodeTextTokensRemote('/api/decode/yi', ids);
|
return decodeTextTokensRemote('/api/decode/yi', ids);
|
||||||
case tokenizers.OPENAI:
|
case tokenizers.OPENAI: {
|
||||||
const model = getTokenizerModel();
|
const model = getTokenizerModel();
|
||||||
return decodeTextTokensRemote('/api/decode/openai', ids, model);
|
return decodeTextTokensRemote('/api/decode/openai', ids, model);
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
console.warn("Calling decodeTextTokens with unsupported tokenizer type", tokenizerType);
|
console.warn("Calling decodeTextTokens with unsupported tokenizer type", tokenizerType);
|
||||||
return '';
|
return '';
|
||||||
|
@ -1898,7 +1898,7 @@ async function checkWorldInfo(chat, maxContext) {
|
|||||||
case world_info_position.ANBottom:
|
case world_info_position.ANBottom:
|
||||||
ANBottomEntries.unshift(entry.content);
|
ANBottomEntries.unshift(entry.content);
|
||||||
break;
|
break;
|
||||||
case world_info_position.atDepth:
|
case world_info_position.atDepth: {
|
||||||
const existingDepthIndex = WIDepthEntries.findIndex((e) => e.depth === entry.depth ?? DEFAULT_DEPTH);
|
const existingDepthIndex = WIDepthEntries.findIndex((e) => e.depth === entry.depth ?? DEFAULT_DEPTH);
|
||||||
if (existingDepthIndex !== -1) {
|
if (existingDepthIndex !== -1) {
|
||||||
WIDepthEntries[existingDepthIndex].entries.unshift(entry.content);
|
WIDepthEntries[existingDepthIndex].entries.unshift(entry.content);
|
||||||
@ -1909,6 +1909,7 @@ async function checkWorldInfo(chat, maxContext) {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -7,7 +7,7 @@ const parse = async (cardUrl, format) => {
|
|||||||
let fileFormat = format === undefined ? 'png' : format;
|
let fileFormat = format === undefined ? 'png' : format;
|
||||||
|
|
||||||
switch (fileFormat) {
|
switch (fileFormat) {
|
||||||
case 'png':
|
case 'png': {
|
||||||
const buffer = fs.readFileSync(cardUrl);
|
const buffer = fs.readFileSync(cardUrl);
|
||||||
const chunks = extract(buffer);
|
const chunks = extract(buffer);
|
||||||
|
|
||||||
@ -23,6 +23,7 @@ const parse = async (cardUrl, format) => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
return Buffer.from(textChunks[0].text, 'base64').toString('utf8');
|
return Buffer.from(textChunks[0].text, 'base64').toString('utf8');
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user