mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
More debug
This commit is contained in:
@@ -4782,6 +4782,8 @@ def calcsubmit(txt):
|
|||||||
sendtocluster(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
sendtocluster(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
||||||
# elif koboldai_vars.use_colab_tpu or koboldai_vars.model in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX"):
|
# elif koboldai_vars.use_colab_tpu or koboldai_vars.model in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX"):
|
||||||
# tpumtjgenerate(subtxt, min, max, found_entries=found_entries)
|
# tpumtjgenerate(subtxt, min, max, found_entries=found_entries)
|
||||||
|
else:
|
||||||
|
print(":(", koboldai_vars.model)
|
||||||
|
|
||||||
# For InferKit web API
|
# For InferKit web API
|
||||||
else:
|
else:
|
||||||
@@ -5043,6 +5045,7 @@ def tpu_raw_generate(
|
|||||||
batch_count: int,
|
batch_count: int,
|
||||||
):
|
):
|
||||||
# Mostly lifted from apiactionsubmit_tpumtjgenerate
|
# Mostly lifted from apiactionsubmit_tpumtjgenerate
|
||||||
|
print("we are generating")
|
||||||
soft_tokens = tpumtjgetsofttokens()
|
soft_tokens = tpumtjgetsofttokens()
|
||||||
genout = tpool.execute(
|
genout = tpool.execute(
|
||||||
tpu_mtj_backend.infer_static,
|
tpu_mtj_backend.infer_static,
|
||||||
@@ -5191,6 +5194,7 @@ def oai_raw_generate(
|
|||||||
#==================================================================#
|
#==================================================================#
|
||||||
|
|
||||||
def generate(txt, minimum, maximum, found_entries=None):
|
def generate(txt, minimum, maximum, found_entries=None):
|
||||||
|
print("ring ring", txt, minimum, maximum, found_entries)
|
||||||
koboldai_vars.generated_tkns = 0
|
koboldai_vars.generated_tkns = 0
|
||||||
|
|
||||||
if(found_entries is None):
|
if(found_entries is None):
|
||||||
|
Reference in New Issue
Block a user