[gptq_hf_torch] Fix typo in model type check

`model_tseype` -> `model_type`
This commit is contained in:
Nick Perez
2023-09-23 00:53:48 -04:00
committed by GitHub
parent db5ff9e748
commit d8877b642d

View File

@@ -362,7 +362,7 @@ class model_backend(HFTorchInferenceModel):
model = load_quant_offload_device_map(llama_load_quant, location, gptq_file, gptq_bits, gptq_groupsize, device_map, force_bias=v2_bias)
elif model_type == "opt":
model = load_quant_offload_device_map(opt_load_quant, location, gptq_file, gptq_bits, gptq_groupsize, device_map, force_bias=v2_bias)
elif model_tseype == "mpt":
elif model_type == "mpt":
model = load_quant_offload_device_map(mpt_load_quant, location, gptq_file, gptq_bits, gptq_groupsize, device_map, force_bias=v2_bias)
elif model_type == "gpt_bigcode":
model = load_quant_offload_device_map(bigcode_load_quant, location, gptq_file, gptq_bits, gptq_groupsize, device_map, force_bias=v2_bias).half()