4-bit on Colab

This commit is contained in:
Henk
2023-07-18 01:48:01 +02:00
parent da9226fba5
commit 5bbcdc47da
3 changed files with 6 additions and 4 deletions

View File

@@ -82,7 +82,7 @@ class model_backend(HFTorchInferenceModel):
"low_cpu_mem_usage": True,
}
if self.use_4_bit:
if self.use_4_bit or utils.koboldai_vars.colab_arg:
tf_kwargs.update({
"quantization_config":BitsAndBytesConfig(
load_in_4bit=True,