Fix memory leak in OAI tokenizer

This commit is contained in:
SillyLossy
2023-03-31 21:56:24 +03:00
parent 3acb43a7a4
commit d73ae6d0f7

View File

@ -1854,6 +1854,8 @@ app.post("/tokenize_openai", jsonParser, function (request, response_tokenize_op
}
num_tokens += 2;
tokenizer.free();
response_tokenize_openai.send({ "token_count": num_tokens });
});