From 6a627265754ef9b2cb2cfb20a476a8af1d383398 Mon Sep 17 00:00:00 2001 From: ebolam Date: Wed, 24 May 2023 19:30:23 -0400 Subject: [PATCH] TPU Fix? --- aiserver.py | 2 +- modeling/inference_models/hf_mtj/class.py | 1 - 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/aiserver.py b/aiserver.py index b06aaa83..998441c8 100644 --- a/aiserver.py +++ b/aiserver.py @@ -1548,7 +1548,7 @@ def general_startup(override_args=None): koboldai_vars.custmodpth = modpath elif args.model: logger.message(f"Welcome to KoboldAI!") - logger.message(f"You have selected the following Model: {koboldai_vars.model}") + logger.message(f"You have selected the following Model: {args.model}") if args.path: logger.message(f"You have selected the following path for your Model: {args.path}") koboldai_vars.custmodpth = args.path; diff --git a/modeling/inference_models/hf_mtj/class.py b/modeling/inference_models/hf_mtj/class.py index 00fbfec3..bc31b3fa 100644 --- a/modeling/inference_models/hf_mtj/class.py +++ b/modeling/inference_models/hf_mtj/class.py @@ -187,7 +187,6 @@ class model_backend(HFInferenceModel): logger.info(self.model_type) tpu_mtj_backend.load_model( self.model, - self.model_type, hf_checkpoint=self.model not in ("TPUMeshTransformerGPTJ", "TPUMeshTransformerGPTNeoX") and utils.koboldai_vars.use_colab_tpu,