mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Ensure we don't save 8bit models
This commit is contained in:
@@ -2962,7 +2962,7 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
|
|
||||||
torch._utils._rebuild_tensor = old_rebuild_tensor
|
torch._utils._rebuild_tensor = old_rebuild_tensor
|
||||||
|
|
||||||
if not args.colab or args.savemodel:
|
if (not args.colab or args.savemodel) and not use_8_bit:
|
||||||
import shutil
|
import shutil
|
||||||
tokenizer.save_pretrained("models/{}".format(koboldai_vars.model.replace('/', '_')))
|
tokenizer.save_pretrained("models/{}".format(koboldai_vars.model.replace('/', '_')))
|
||||||
if koboldai_vars.fp32_model: # Use save_pretrained to convert fp32 models to fp16
|
if koboldai_vars.fp32_model: # Use save_pretrained to convert fp32 models to fp16
|
||||||
|
Reference in New Issue
Block a user