HF's workaround breaks stuff

This commit is contained in:
Henk
2023-07-22 16:29:55 +02:00
parent fa9d17b3d3
commit 8dd7b93a6c

View File

@@ -81,7 +81,6 @@ class model_backend(HFTorchInferenceModel):
tf_kwargs = {
"low_cpu_mem_usage": True,
"pretraining_tp": 1,
}
if self.quantization == "8bit":