From 1da4580e8bf91b8497bc5a022a7818f6aefe1113 Mon Sep 17 00:00:00 2001 From: Henk Date: Thu, 22 Jun 2023 07:07:02 +0200 Subject: [PATCH] Remove wrong usegpu behavior --- modeling/inference_models/hf_torch.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/modeling/inference_models/hf_torch.py b/modeling/inference_models/hf_torch.py index 514a1e5b..7719f022 100644 --- a/modeling/inference_models/hf_torch.py +++ b/modeling/inference_models/hf_torch.py @@ -114,9 +114,6 @@ class HFTorchInferenceModel(HFInferenceModel): self.breakmodel_config.gpu_blocks = self.layers self.breakmodel_config.disk_blocks = self.disk_layers - # HACK: Prevent get_auxiliary_device from returning cuda - utils.koboldai_vars.hascuda = self.usegpu - return ret def _get_target_dtype(self) -> Union[torch.float16, torch.float32]: @@ -124,8 +121,6 @@ class HFTorchInferenceModel(HFInferenceModel): return torch.float32 elif utils.args.cpu: return torch.float32 - elif not self.usegpu: - return torch.float32 return torch.float16 def _apply_warpers(