From ce3d4659729b619c9c658babaf02240eb45e1f50 Mon Sep 17 00:00:00 2001 From: somebody Date: Mon, 1 May 2023 16:59:30 -0500 Subject: [PATCH] Remove some debug --- modeling/inference_models/hf_mtj.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/modeling/inference_models/hf_mtj.py b/modeling/inference_models/hf_mtj.py index b848bea3..7661a67f 100644 --- a/modeling/inference_models/hf_mtj.py +++ b/modeling/inference_models/hf_mtj.py @@ -111,10 +111,8 @@ class HFMTJInferenceModel(HFInferenceModel): _, _, _, used_world_info = utils.koboldai_vars.calc_ai_text( submitted_text=decoded ) - print(utils.koboldai_vars.calc_ai_text()) # found -= excluded_world_info[i] if used_world_info: - print("lets regen") regeneration_required = True break return regeneration_required, halt @@ -321,8 +319,6 @@ class HFMTJInferenceModel(HFInferenceModel): if utils.koboldai_vars.abort or halt or not regeneration_required: break - print("(regeneration triggered)") - encoded = [] for i in range(utils.koboldai_vars.numseqs): txt = utils.decodenewlines(self.tokenizer.decode(past[i])) @@ -365,7 +361,6 @@ class HFMTJInferenceModel(HFInferenceModel): # ) # print(genout) # print(type(genout)) - print(context) genout = np.array(genout) return GenerationResult(