mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Reimplement HF workaround only for llama
This commit is contained in:
@@ -110,6 +110,11 @@ class model_backend(HFTorchInferenceModel):
|
|||||||
# Also, lazy loader doesn't support GPT-2 models
|
# Also, lazy loader doesn't support GPT-2 models
|
||||||
self.lazy_load = False
|
self.lazy_load = False
|
||||||
|
|
||||||
|
if self.model_type == "llama":
|
||||||
|
tf_kwargs.update({
|
||||||
|
"pretraining_tp": 1 # Workaround recommended by HF to fix their mistake on the config.json tuners adopted
|
||||||
|
})
|
||||||
|
|
||||||
logger.debug(
|
logger.debug(
|
||||||
"lazy_load: {} hascuda: {} breakmodel: {} nobreakmode: {}".format(
|
"lazy_load: {} hascuda: {} breakmodel: {} nobreakmode: {}".format(
|
||||||
self.lazy_load,
|
self.lazy_load,
|
||||||
|
Reference in New Issue
Block a user