From 70cddc46e263f700dee355c6e4d65e90ccdb5ac5 Mon Sep 17 00:00:00 2001 From: somebody Date: Sat, 4 Mar 2023 17:51:37 -0600 Subject: [PATCH] Model: Small cleanup --- modeling/inference_model.py | 2 -- modeling/inference_models/hf_mtj.py | 2 +- 2 files changed, 1 insertion(+), 3 deletions(-) diff --git a/modeling/inference_model.py b/modeling/inference_model.py index 3f054f79..1bf35146 100644 --- a/modeling/inference_model.py +++ b/modeling/inference_model.py @@ -176,8 +176,6 @@ class InferenceModel: global current_model current_model = self - print(self.raw_generate("Hi guys,", 20).__dict__) - def _post_load(self) -> None: """Post load hook. Called after `_load()`.""" diff --git a/modeling/inference_models/hf_mtj.py b/modeling/inference_models/hf_mtj.py index 984123b2..3a16f6f7 100644 --- a/modeling/inference_models/hf_mtj.py +++ b/modeling/inference_models/hf_mtj.py @@ -266,7 +266,7 @@ class HFMTJInferenceModel(HFInferenceModel): soft_tokens = self.get_soft_tokens() dynamic_inference = kwargs.get("tpu_dynamic_inference", False) - print(f"DYNAMIC_INFERENCE={dynamic_inference} KWARGS={kwargs}") + logger.info(f"dynamic_inference={dynamic_inference}") if not dynamic_inference: genout = tpool.execute(