mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Oops
This commit is contained in:
@@ -4848,12 +4848,13 @@ def core_generate(text: list, min: int, max: int, found_entries: set):
|
|||||||
# This generation function is tangled with koboldai_vars intentionally. It
|
# This generation function is tangled with koboldai_vars intentionally. It
|
||||||
# is meant for the story and nothing else.
|
# is meant for the story and nothing else.
|
||||||
|
|
||||||
|
gen_in = torch.tensor(text, dtype=torch.long)[None]
|
||||||
|
|
||||||
if koboldai_vars.is_model_torch():
|
if koboldai_vars.is_model_torch():
|
||||||
# Torch stuff
|
# Torch stuff
|
||||||
if koboldai_vars.full_determinism:
|
if koboldai_vars.full_determinism:
|
||||||
torch.manual_seed(koboldai_vars.seed)
|
torch.manual_seed(koboldai_vars.seed)
|
||||||
|
|
||||||
gen_in = torch.tensor(text, dtype=torch.long)[None]
|
|
||||||
if koboldai_vars.sp is not None:
|
if koboldai_vars.sp is not None:
|
||||||
soft_tokens = torch.arange(
|
soft_tokens = torch.arange(
|
||||||
model.config.vocab_size,
|
model.config.vocab_size,
|
||||||
|
Reference in New Issue
Block a user