Enable no-case-declarations lint

This commit is contained in:
valadaptive
2023-12-02 10:14:06 -05:00
parent a665a4897d
commit 27e63a7a77
6 changed files with 14 additions and 8 deletions

View File

@ -61,7 +61,6 @@ module.exports = {
'no-prototype-builtins': 'off', 'no-prototype-builtins': 'off',
'no-extra-boolean-cast': 'off', 'no-extra-boolean-cast': 'off',
'require-yield': 'off', 'require-yield': 'off',
'no-case-declarations': 'off',
'no-constant-condition': ['error', {checkLoops: false}] 'no-constant-condition': ['error', {checkLoops: false}]
} }
}; };

View File

@ -2755,10 +2755,11 @@ export async function generateRaw(prompt, api, instructOverride) {
generateData = getKoboldGenerationData(prompt, koboldSettings, amount_gen, max_context, isHorde, 'quiet'); generateData = getKoboldGenerationData(prompt, koboldSettings, amount_gen, max_context, isHorde, 'quiet');
} }
break; break;
case 'novel': case 'novel': {
const novelSettings = novelai_settings[novelai_setting_names[nai_settings.preset_settings_novel]]; const novelSettings = novelai_settings[novelai_setting_names[nai_settings.preset_settings_novel]];
generateData = getNovelGenerationData(prompt, novelSettings, amount_gen, false, false, null); generateData = getNovelGenerationData(prompt, novelSettings, amount_gen, false, false, null);
break; break;
}
case 'textgenerationwebui': case 'textgenerationwebui':
generateData = getTextGenGenerationData(prompt, amount_gen, false, false, null); generateData = getTextGenGenerationData(prompt, amount_gen, false, false, null);
break; break;

View File

@ -236,14 +236,16 @@ class PresetManager {
return nai_settings; return nai_settings;
case "textgenerationwebui": case "textgenerationwebui":
return textgenerationwebui_settings; return textgenerationwebui_settings;
case "context": case "context": {
const context_preset = getContextSettings(); const context_preset = getContextSettings();
context_preset['name'] = name || power_user.context.preset; context_preset['name'] = name || power_user.context.preset;
return context_preset; return context_preset;
case "instruct": }
case "instruct": {
const instruct_preset = structuredClone(power_user.instruct); const instruct_preset = structuredClone(power_user.instruct);
instruct_preset['name'] = name || power_user.instruct.preset; instruct_preset['name'] = name || power_user.instruct.preset;
return instruct_preset; return instruct_preset;
}
default: default:
console.warn(`Unknown API ID ${apiId}`); console.warn(`Unknown API ID ${apiId}`);
return {}; return {};

View File

@ -507,9 +507,10 @@ export function getTextTokens(tokenizerType, str) {
return getTextTokensRemote('/api/tokenize/mistral', str); return getTextTokensRemote('/api/tokenize/mistral', str);
case tokenizers.YI: case tokenizers.YI:
return getTextTokensRemote('/api/tokenize/yi', str); return getTextTokensRemote('/api/tokenize/yi', str);
case tokenizers.OPENAI: case tokenizers.OPENAI: {
const model = getTokenizerModel(); const model = getTokenizerModel();
return getTextTokensRemote('/api/tokenize/openai-encode', str, model); return getTextTokensRemote('/api/tokenize/openai-encode', str, model);
}
case tokenizers.API: case tokenizers.API:
return getTextTokensRemote('/tokenize_via_api', str); return getTextTokensRemote('/tokenize_via_api', str);
default: default:
@ -537,9 +538,10 @@ export function decodeTextTokens(tokenizerType, ids) {
return decodeTextTokensRemote('/api/decode/mistral', ids); return decodeTextTokensRemote('/api/decode/mistral', ids);
case tokenizers.YI: case tokenizers.YI:
return decodeTextTokensRemote('/api/decode/yi', ids); return decodeTextTokensRemote('/api/decode/yi', ids);
case tokenizers.OPENAI: case tokenizers.OPENAI: {
const model = getTokenizerModel(); const model = getTokenizerModel();
return decodeTextTokensRemote('/api/decode/openai', ids, model); return decodeTextTokensRemote('/api/decode/openai', ids, model);
}
default: default:
console.warn("Calling decodeTextTokens with unsupported tokenizer type", tokenizerType); console.warn("Calling decodeTextTokens with unsupported tokenizer type", tokenizerType);
return ''; return '';

View File

@ -1898,7 +1898,7 @@ async function checkWorldInfo(chat, maxContext) {
case world_info_position.ANBottom: case world_info_position.ANBottom:
ANBottomEntries.unshift(entry.content); ANBottomEntries.unshift(entry.content);
break; break;
case world_info_position.atDepth: case world_info_position.atDepth: {
const existingDepthIndex = WIDepthEntries.findIndex((e) => e.depth === entry.depth ?? DEFAULT_DEPTH); const existingDepthIndex = WIDepthEntries.findIndex((e) => e.depth === entry.depth ?? DEFAULT_DEPTH);
if (existingDepthIndex !== -1) { if (existingDepthIndex !== -1) {
WIDepthEntries[existingDepthIndex].entries.unshift(entry.content); WIDepthEntries[existingDepthIndex].entries.unshift(entry.content);
@ -1909,6 +1909,7 @@ async function checkWorldInfo(chat, maxContext) {
}); });
} }
break; break;
}
default: default:
break; break;
} }

View File

@ -7,7 +7,7 @@ const parse = async (cardUrl, format) => {
let fileFormat = format === undefined ? 'png' : format; let fileFormat = format === undefined ? 'png' : format;
switch (fileFormat) { switch (fileFormat) {
case 'png': case 'png': {
const buffer = fs.readFileSync(cardUrl); const buffer = fs.readFileSync(cardUrl);
const chunks = extract(buffer); const chunks = extract(buffer);
@ -23,6 +23,7 @@ const parse = async (cardUrl, format) => {
} }
return Buffer.from(textChunks[0].text, 'base64').toString('utf8'); return Buffer.from(textChunks[0].text, 'base64').toString('utf8');
}
default: default:
break; break;
} }