Don't set negative prompt from CFG extension to ooba at guidance scale 1.0
This commit is contained in:
parent
b751643364
commit
fad6c164cb
|
@ -233,9 +233,13 @@ async function generateTextGenWithStreaming(generate_data, signal) {
|
|||
}
|
||||
}
|
||||
|
||||
export function getTextGenGenerationData(finalPromt, this_amount_gen, isImpersonate, cfgValues) {
|
||||
export function getTextGenGenerationData(finalPrompt, this_amount_gen, isImpersonate, cfgValues) {
|
||||
if (cfgValues?.guidanceScale?.value === 1) {
|
||||
cfgValues = null;
|
||||
}
|
||||
|
||||
return {
|
||||
'prompt': finalPromt,
|
||||
'prompt': finalPrompt,
|
||||
'max_new_tokens': this_amount_gen,
|
||||
'do_sample': textgenerationwebui_settings.do_sample,
|
||||
'temperature': textgenerationwebui_settings.temp,
|
||||
|
|
Loading…
Reference in New Issue