Commit Graph

31 Commits

Author SHA1 Message Date
Henk
ec745d8b80 Dont accidentally block pad tokens 2023-07-21 16:25:32 +02:00
somebody
20b4b4bcef Add basic hf backend 2023-07-08 17:12:16 -05:00
somebody
1546b9efaa Hello its breaking breakmodel time 2023-05-27 16:31:53 -05:00
ebolam
47276c3424 Bug Fix 2023-05-27 08:49:21 -04:00
ebolam
cce5c1932c Fix for custom model names 2023-05-26 21:40:39 -04:00
ebolam
9723154bed Fix for --path 2023-05-26 20:10:11 -04:00
ebolam
51cea7eb9d Added ability to add labels that are based on validation data in model loading settings 2023-05-26 12:43:17 -04:00
ebolam
1a1b79a16d Change default for HF on non-breakmodel models to use GPU instead of CPU 2023-05-24 12:44:26 -04:00
ebolam
068173b24a Potential BadWords fix 2023-05-24 09:08:34 -04:00
ebolam
9bd445c2a8 gpt2 fixed 2023-05-23 20:33:55 -04:00
ebolam
839d56ebf2 Potential fix for gpt-neo and gpt-j 2023-05-23 19:25:01 -04:00
ebolam
4c25d6fbbb Fix for loading model multiple times loosing the gpu/cpu splits 2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676 Fix for breakmodel loading to CPU when set to GPU 2023-05-22 20:24:57 -04:00
ebolam
dc20e6dde9 Fix for unloading models 2023-05-22 15:04:33 -04:00
ebolam
925cad2e2f Better compatibility with hf model backend 2023-05-22 14:50:13 -04:00
ebolam
513b8575e7 Fix for missing import
Fix for model name being a path which caused save issues
2023-05-20 11:01:49 -04:00
ebolam
9df1f03b12 Fix for custom huggingface model menu entry 2023-05-19 14:28:36 -04:00
ebolam
99cffd4755 Colab GPU edition fixes 2023-05-19 09:11:08 -04:00
ebolam
06f59a7b7b Moved model backends to separate folders
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
ebolam
71aee4dbd8 First concept of model plugins with a conceptual UI.
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
Henk
a87d5d6f23 Remove HF's llama workaround 2023-05-03 20:18:40 +02:00
Llama
35d344b951 Remove torch dependency and more generic dim0 workaround
Remove torch dependency from hf.py
Make workaround for dimension zero values of token_ids
more generic to handle every token, not just newlines.
2023-05-03 09:48:16 -07:00
Llama
3768848548 Fix tokenization and whitespace issues with llama-derived models
Work around the 'soft' prefix space behavior of sentencepiece.
Override encode to restore the deleted HF support for decode_with_prefix_space.
Override decode to skip the soft space and return true decoded tokens.
Allow submitting chat messages with embedded newlines.
Split sentences between punctuation and whitespace, rather than after whitespace.
Also include trailing quotes and brackets after sentence stoppers.
This avoids splitting ." and .) into two tokens, for instance.
Insert whitespace at the beginning of the author's note, since sentences are
split with leading whitespace.
Remove spurious newlines at the end of chat responses.
2023-05-03 01:27:11 -07:00
somebody
a0f4ab5c6a Move bad token grabber until after newlinemode has been deduced 2023-05-02 20:23:36 -05:00
somebody
efe268df60 Move overrides to better places 2023-05-02 20:18:33 -05:00
somebody
933dbd634a HFInferenceModel: Make badwordsids not unique to torch 2023-05-01 17:13:33 -05:00
onesome
d4f7b60dc9 Fix for multiple paths 2023-04-26 16:49:12 -05:00
onesome
6776a71532 Add more info to custom model error 2023-04-26 16:36:52 -05:00
onesome
bbf4963d6e Fix custmodpth stuff for hf loading 2023-04-26 16:18:45 -05:00
somebody
8c8bdfaf6a Model: Fix assorted bugs
and ignore warnings in pytest
2023-03-10 20:00:39 -06:00
somebody
54cecd4d5d Model: And another refactor 2023-03-01 19:16:35 -06:00