Remove HF's llama workaround

This commit is contained in:
Henk
2023-05-03 20:18:40 +02:00
parent 7f5242db17
commit a87d5d6f23

View File

@@ -21,7 +21,6 @@ class HFInferenceModel(InferenceModel):
# These are model specific tokenizer overrides if a model has bad defaults
if utils.koboldai_vars.model_type == "llama":
# Note: self.tokenizer is a GenericTokenizer, and self.tokenizer.tokenizer is the actual LlamaTokenizer
self.tokenizer.decode_with_prefix_space = True # Note, not supported anymore, hence the workaround below.
self.tokenizer.add_bos_token = False
# HF transformers no longer supports decode_with_prefix_space