mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Push it on raw_generate!
This commit is contained in:
11
aiserver.py
11
aiserver.py
@@ -4770,9 +4770,8 @@ def calcsubmit(txt):
|
|||||||
"Colab",
|
"Colab",
|
||||||
"API",
|
"API",
|
||||||
"CLUSTER",
|
"CLUSTER",
|
||||||
"OAI",
|
# "TPUMeshTransformerGPTJ",
|
||||||
"TPUMeshTransformerGPTJ",
|
# "TPUMeshTransformerGPTNeoX"
|
||||||
"TPUMeshTransformerGPTNeoX"
|
|
||||||
):
|
):
|
||||||
generate(subtxt, min, max, found_entries)
|
generate(subtxt, min, max, found_entries)
|
||||||
elif koboldai_vars.model == "Colab":
|
elif koboldai_vars.model == "Colab":
|
||||||
@@ -4781,10 +4780,8 @@ def calcsubmit(txt):
|
|||||||
sendtoapi(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
sendtoapi(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
||||||
elif koboldai_vars.model == "CLUSTER":
|
elif koboldai_vars.model == "CLUSTER":
|
||||||
sendtocluster(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
sendtocluster(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
||||||
elif koboldai_vars.model == "OAI":
|
# elif koboldai_vars.use_colab_tpu or koboldai_vars.model in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX"):
|
||||||
oairequest(utils.decodenewlines(tokenizer.decode(subtxt)), min, max)
|
# tpumtjgenerate(subtxt, min, max, found_entries=found_entries)
|
||||||
elif koboldai_vars.use_colab_tpu or koboldai_vars.model in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX"):
|
|
||||||
tpumtjgenerate(subtxt, min, max, found_entries=found_entries)
|
|
||||||
|
|
||||||
# For InferKit web API
|
# For InferKit web API
|
||||||
else:
|
else:
|
||||||
|
Reference in New Issue
Block a user