mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix tokenizer fallback for llama
This commit is contained in:
@@ -223,7 +223,8 @@ class InferenceModel:
|
||||
for i, try_get_tokenizer in enumerate(suppliers):
|
||||
try:
|
||||
return GenericTokenizer(try_get_tokenizer())
|
||||
except:
|
||||
except Exception as e:
|
||||
logger.warn(f"Tokenizer falling back due to {e}")
|
||||
# If we error on each attempt, raise the last one
|
||||
if i == len(suppliers) - 1:
|
||||
raise
|
||||
|
@@ -9,7 +9,10 @@ class GenericTokenizer:
|
||||
|
||||
def __init__(self, tokenizer: Union[Tokenizer, PreTrainedTokenizer]) -> None:
|
||||
self.tokenizer = tokenizer
|
||||
self.valid_tokens = set(self.tokenizer.vocab.values())
|
||||
try:
|
||||
self.valid_tokens = set(self.tokenizer.vocab.values())
|
||||
except AttributeError:
|
||||
self.valid_tokens = set(self.tokenizer.get_vocab().values())
|
||||
|
||||
def __getattr__(self, name: str) -> Any:
|
||||
# Fall back to tokenizer for non-generic stuff
|
||||
|
Reference in New Issue
Block a user