mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
GenericHFTorch: Change use_4_bit to quantization in __init__
This commit is contained in:
@@ -29,7 +29,7 @@ model_backend_type = "Huggingface" #This should be a generic name in case multip
|
|||||||
class model_backend(HFTorchInferenceModel):
|
class model_backend(HFTorchInferenceModel):
|
||||||
def __init__(self) -> None:
|
def __init__(self) -> None:
|
||||||
super().__init__()
|
super().__init__()
|
||||||
self.use_4_bit = False
|
self.quantization = False
|
||||||
|
|
||||||
def is_valid(self, model_name, model_path, menu_path):
|
def is_valid(self, model_name, model_path, menu_path):
|
||||||
base_is_valid = super().is_valid(model_name, model_path, menu_path)
|
base_is_valid = super().is_valid(model_name, model_path, menu_path)
|
||||||
|
Reference in New Issue
Block a user