ebolam
9bd445c2a8
gpt2 fixed
2023-05-23 20:33:55 -04:00
ebolam
839d56ebf2
Potential fix for gpt-neo and gpt-j
2023-05-23 19:25:01 -04:00
ebolam
4c25d6fbbb
Fix for loading model multiple times loosing the gpu/cpu splits
2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676
Fix for breakmodel loading to CPU when set to GPU
2023-05-22 20:24:57 -04:00
ebolam
dc20e6dde9
Fix for unloading models
2023-05-22 15:04:33 -04:00
ebolam
925cad2e2f
Better compatibility with hf model backend
2023-05-22 14:50:13 -04:00
ebolam
513b8575e7
Fix for missing import
...
Fix for model name being a path which caused save issues
2023-05-20 11:01:49 -04:00
ebolam
9df1f03b12
Fix for custom huggingface model menu entry
2023-05-19 14:28:36 -04:00
ebolam
99cffd4755
Colab GPU edition fixes
2023-05-19 09:11:08 -04:00
ebolam
06f59a7b7b
Moved model backends to separate folders
...
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
ebolam
71aee4dbd8
First concept of model plugins with a conceptual UI.
...
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
Henk
a87d5d6f23
Remove HF's llama workaround
2023-05-03 20:18:40 +02:00
Llama
35d344b951
Remove torch dependency and more generic dim0 workaround
...
Remove torch dependency from hf.py
Make workaround for dimension zero values of token_ids
more generic to handle every token, not just newlines.
2023-05-03 09:48:16 -07:00
Llama
3768848548
Fix tokenization and whitespace issues with llama-derived models
...
Work around the 'soft' prefix space behavior of sentencepiece.
Override encode to restore the deleted HF support for decode_with_prefix_space.
Override decode to skip the soft space and return true decoded tokens.
Allow submitting chat messages with embedded newlines.
Split sentences between punctuation and whitespace, rather than after whitespace.
Also include trailing quotes and brackets after sentence stoppers.
This avoids splitting ." and .) into two tokens, for instance.
Insert whitespace at the beginning of the author's note, since sentences are
split with leading whitespace.
Remove spurious newlines at the end of chat responses.
2023-05-03 01:27:11 -07:00
somebody
a0f4ab5c6a
Move bad token grabber until after newlinemode has been deduced
2023-05-02 20:23:36 -05:00
somebody
efe268df60
Move overrides to better places
2023-05-02 20:18:33 -05:00
somebody
933dbd634a
HFInferenceModel: Make badwordsids not unique to torch
2023-05-01 17:13:33 -05:00
onesome
d4f7b60dc9
Fix for multiple paths
2023-04-26 16:49:12 -05:00
onesome
6776a71532
Add more info to custom model error
2023-04-26 16:36:52 -05:00
onesome
bbf4963d6e
Fix custmodpth stuff for hf loading
2023-04-26 16:18:45 -05:00
somebody
8c8bdfaf6a
Model: Fix assorted bugs
...
and ignore warnings in pytest
2023-03-10 20:00:39 -06:00
somebody
54cecd4d5d
Model: And another refactor
2023-03-01 19:16:35 -06:00