From 6d1bf76ef1d502ace9542d6f4740e6355c46c86c Mon Sep 17 00:00:00 2001 From: henk717 Date: Mon, 27 Dec 2021 01:56:59 +0100 Subject: [PATCH] Path Fixes Fixes the tokenizer cache being hit when we already have a local model --- aiserver.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/aiserver.py b/aiserver.py index 463d8959..922e5255 100644 --- a/aiserver.py +++ b/aiserver.py @@ -902,9 +902,9 @@ if(not vars.model in ["InferKit", "Colab", "OAI", "ReadOnly", "TPUMeshTransforme elif(os.path.isdir(vars.model.replace('/', '_'))): with(maybe_use_float16()): try: - tokenizer = AutoTokenizer.from_pretrained(vars.custmodpth, cache_dir="cache/") + tokenizer = AutoTokenizer.from_pretrained(vars.model.replace('/', '_'), cache_dir="cache/") except ValueError as e: - tokenizer = GPT2TokenizerFast.from_pretrained(vars.custmodpth, cache_dir="cache/") + tokenizer = GPT2TokenizerFast.from_pretrained(vars.model.replace('/', '_'), cache_dir="cache/") try: model = AutoModelForCausalLM.from_pretrained(vars.model.replace('/', '_'), cache_dir="cache/", **lowmem) except ValueError as e: @@ -912,9 +912,9 @@ if(not vars.model in ["InferKit", "Colab", "OAI", "ReadOnly", "TPUMeshTransforme else: print("Model does not exist locally, attempting to download from Huggingface...") try: - tokenizer = AutoTokenizer.from_pretrained(vars.custmodpth, cache_dir="cache/") + tokenizer = AutoTokenizer.from_pretrained(vars.model, cache_dir="cache/") except ValueError as e: - tokenizer = GPT2TokenizerFast.from_pretrained(vars.custmodpth, cache_dir="cache/") + tokenizer = GPT2TokenizerFast.from_pretrained(vars.model, cache_dir="cache/") with(maybe_use_float16()): try: model = AutoModelForCausalLM.from_pretrained(vars.model, cache_dir="cache/", **lowmem)