mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix for loading model multiple times loosing the gpu/cpu splits
This commit is contained in:
@@ -197,12 +197,6 @@ class HFInferenceModel(InferenceModel):
|
||||
torch.cuda.empty_cache()
|
||||
except:
|
||||
pass
|
||||
if self.hf_torch:
|
||||
if 'breakmodel' in sys.modules:
|
||||
import breakmodel
|
||||
breakmodel.breakmodel = True
|
||||
breakmodel.gpu_blocks = []
|
||||
breakmodel.disk_blocks = 0
|
||||
|
||||
def _post_load(self) -> None:
|
||||
# These are model specific tokenizer overrides if a model has bad defaults
|
||||
|
Reference in New Issue
Block a user