mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-02-20 05:30:57 +01:00
Fix some typos in XGLM breakmodel
This commit is contained in:
parent
8e1169ea61
commit
d847d04605
@ -384,7 +384,7 @@ def device_config(model):
|
|||||||
else:
|
else:
|
||||||
model.model.embed_tokens.to(breakmodel.primary_device)
|
model.model.embed_tokens.to(breakmodel.primary_device)
|
||||||
model.model.layer_norm.to(breakmodel.primary_device)
|
model.model.layer_norm.to(breakmodel.primary_device)
|
||||||
model.model.lm_head.to(breakmodel.primary_device)
|
model.lm_head.to(breakmodel.primary_device)
|
||||||
model.model.embed_positions.to(breakmodel.primary_device)
|
model.model.embed_positions.to(breakmodel.primary_device)
|
||||||
gc.collect()
|
gc.collect()
|
||||||
GPTNeoModel.forward = breakmodel.new_forward_neo
|
GPTNeoModel.forward = breakmodel.new_forward_neo
|
||||||
@ -736,7 +736,7 @@ if(not vars.model in ["InferKit", "Colab", "OAI", "ReadOnly", "TPUMeshTransforme
|
|||||||
from transformers import StoppingCriteria, GPT2TokenizerFast, GPT2LMHeadModel, GPTNeoForCausalLM, GPTNeoModel, AutoModelForCausalLM, AutoTokenizer
|
from transformers import StoppingCriteria, GPT2TokenizerFast, GPT2LMHeadModel, GPTNeoForCausalLM, GPTNeoModel, AutoModelForCausalLM, AutoTokenizer
|
||||||
for m in ("GPTJModel", "XGLMModel"):
|
for m in ("GPTJModel", "XGLMModel"):
|
||||||
try:
|
try:
|
||||||
globals()[m] = __import__("transformers." + m, fromlist=[...])
|
globals()[m] = getattr(__import__("transformers"), m)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
import transformers.generation_utils
|
import transformers.generation_utils
|
||||||
|
Loading…
x
Reference in New Issue
Block a user