mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
TPU Fix?
This commit is contained in:
@@ -1548,7 +1548,7 @@ def general_startup(override_args=None):
|
|||||||
koboldai_vars.custmodpth = modpath
|
koboldai_vars.custmodpth = modpath
|
||||||
elif args.model:
|
elif args.model:
|
||||||
logger.message(f"Welcome to KoboldAI!")
|
logger.message(f"Welcome to KoboldAI!")
|
||||||
logger.message(f"You have selected the following Model: {koboldai_vars.model}")
|
logger.message(f"You have selected the following Model: {args.model}")
|
||||||
if args.path:
|
if args.path:
|
||||||
logger.message(f"You have selected the following path for your Model: {args.path}")
|
logger.message(f"You have selected the following path for your Model: {args.path}")
|
||||||
koboldai_vars.custmodpth = args.path;
|
koboldai_vars.custmodpth = args.path;
|
||||||
|
@@ -187,7 +187,6 @@ class model_backend(HFInferenceModel):
|
|||||||
logger.info(self.model_type)
|
logger.info(self.model_type)
|
||||||
tpu_mtj_backend.load_model(
|
tpu_mtj_backend.load_model(
|
||||||
self.model,
|
self.model,
|
||||||
self.model_type,
|
|
||||||
hf_checkpoint=self.model
|
hf_checkpoint=self.model
|
||||||
not in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX")
|
not in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX")
|
||||||
and utils.koboldai_vars.use_colab_tpu,
|
and utils.koboldai_vars.use_colab_tpu,
|
||||||
|
Reference in New Issue
Block a user