mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fallback to transformers if hf_bleeding_edge not available
This commit is contained in:
@@ -7,7 +7,10 @@ import shutil
|
|||||||
from typing import Union
|
from typing import Union
|
||||||
|
|
||||||
from transformers import GPTNeoForCausalLM, GPT2LMHeadModel, BitsAndBytesConfig
|
from transformers import GPTNeoForCausalLM, GPT2LMHeadModel, BitsAndBytesConfig
|
||||||
|
try:
|
||||||
from hf_bleeding_edge import AutoModelForCausalLM
|
from hf_bleeding_edge import AutoModelForCausalLM
|
||||||
|
except ImportError:
|
||||||
|
from transformers import AutoModelForCausalLM
|
||||||
|
|
||||||
from transformers.utils import WEIGHTS_NAME, WEIGHTS_INDEX_NAME, TF2_WEIGHTS_NAME, TF2_WEIGHTS_INDEX_NAME, TF_WEIGHTS_NAME, FLAX_WEIGHTS_NAME, FLAX_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SAFE_WEIGHTS_INDEX_NAME
|
from transformers.utils import WEIGHTS_NAME, WEIGHTS_INDEX_NAME, TF2_WEIGHTS_NAME, TF2_WEIGHTS_INDEX_NAME, TF_WEIGHTS_NAME, FLAX_WEIGHTS_NAME, FLAX_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SAFE_WEIGHTS_INDEX_NAME
|
||||||
|
|
||||||
|
@@ -10,8 +10,11 @@ import sys
|
|||||||
from typing import Union
|
from typing import Union
|
||||||
|
|
||||||
from transformers import GPTNeoForCausalLM, AutoTokenizer, LlamaTokenizer
|
from transformers import GPTNeoForCausalLM, AutoTokenizer, LlamaTokenizer
|
||||||
|
try:
|
||||||
import hf_bleeding_edge
|
import hf_bleeding_edge
|
||||||
from hf_bleeding_edge import AutoModelForCausalLM
|
from hf_bleeding_edge import AutoModelForCausalLM
|
||||||
|
except ImportError:
|
||||||
|
from transformers import AutoModelForCausalLM
|
||||||
|
|
||||||
import utils
|
import utils
|
||||||
import modeling.lazy_loader as lazy_loader
|
import modeling.lazy_loader as lazy_loader
|
||||||
|
@@ -1,6 +1,9 @@
|
|||||||
import os, sys
|
import os, sys
|
||||||
from typing import Optional
|
from typing import Optional
|
||||||
|
try:
|
||||||
from hf_bleeding_edge import AutoConfig
|
from hf_bleeding_edge import AutoConfig
|
||||||
|
except ImportError:
|
||||||
|
from transformers import AutoConfig
|
||||||
|
|
||||||
import warnings
|
import warnings
|
||||||
import utils
|
import utils
|
||||||
|
@@ -19,7 +19,10 @@ from transformers import (
|
|||||||
GPT2LMHeadModel,
|
GPT2LMHeadModel,
|
||||||
LogitsProcessorList,
|
LogitsProcessorList,
|
||||||
)
|
)
|
||||||
|
try:
|
||||||
from hf_bleeding_edge import AutoModelForCausalLM
|
from hf_bleeding_edge import AutoModelForCausalLM
|
||||||
|
except ImportError:
|
||||||
|
from transformers import AutoModelForCausalLM
|
||||||
|
|
||||||
import utils
|
import utils
|
||||||
import modeling.lazy_loader as lazy_loader
|
import modeling.lazy_loader as lazy_loader
|
||||||
|
Reference in New Issue
Block a user