Removed extra load_in_4bit.

This commit is contained in:
Alephrin
2023-07-17 04:53:47 -06:00
committed by GitHub
parent e9913d657a
commit 145a43a000

View File

@@ -80,7 +80,6 @@ class model_backend(HFTorchInferenceModel):
if self.use_4_bit:
self.lazy_load = False
tf_kwargs.update({
"load_in_4bit": True,
"quantization_config":BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_compute_dtype=torch.float16,