From 14e5fcd355c7cf1614fa6da06caf79dc9009d278 Mon Sep 17 00:00:00 2001 From: henk717 Date: Sat, 25 Dec 2021 00:48:12 +0100 Subject: [PATCH] AutoTokenizer --- aiserver.py | 1 - 1 file changed, 1 deletion(-) diff --git a/aiserver.py b/aiserver.py index b2fe91eb..69c45b93 100644 --- a/aiserver.py +++ b/aiserver.py @@ -910,7 +910,6 @@ if(not vars.model in ["InferKit", "Colab", "OAI", "ReadOnly", "TPUMeshTransforme except ValueError as e: tokenizer = GPT2TokenizerFast.from_pretrained(vars.custmodpth, cache_dir="cache/") with(maybe_use_float16()): - tokenizer = GPT2TokenizerFast.from_pretrained(vars.model, cache_dir="cache/") try: model = AutoModelForCausalLM.from_pretrained(vars.model, cache_dir="cache/", **lowmem) except ValueError as e: