Commit Graph

50 Commits

Author SHA1 Message Date
Henk
9e51a50bcd Llama fixes for Mistral 2023-09-28 01:08:53 +02:00
Henk
0d0a671bb9 Better use_cache implementation 2023-09-07 04:29:28 +02:00
Henk
dfb63b2340 HF 4.33.1 2023-09-06 20:21:23 +02:00
Henk
f40236c04a Modern llama tokenizer 2023-08-25 14:27:44 +02:00
Henk
d93631c889 GPTQ improvements 2023-08-19 14:45:45 +02:00
somebody
906d1f2522 Merge branch 'united' of https://github.com/henk717/KoboldAI into fixing-time 2023-08-07 16:22:04 -05:00
0cc4m
a9aa04fd1b Merge remote-tracking branch 'upstream/united' into 4bit-plugin 2023-07-23 07:18:58 +02:00
0cc4m
09bb1021dd Fallback to transformers if hf_bleeding_edge not available 2023-07-23 07:16:52 +02:00
Henk
432cdc9a08 Fix models with good pad tokens 2023-07-21 16:39:58 +02:00
Henk
ec745d8b80 Dont accidentally block pad tokens 2023-07-21 16:25:32 +02:00
somebody
fef42a6273 API: Fix loading 2023-07-19 11:52:39 -05:00
0cc4m
e78361fc8f Pull upstream changes, fix conflicts 2023-07-15 23:01:52 +02:00
somebody
20b4b4bcef Add basic hf backend 2023-07-08 17:12:16 -05:00
0cc4m
e49d35afc9 Add 4bit plugin 2023-05-28 22:54:36 +02:00
0cc4m
d71a63fa49 Merge ebolam's model-plugins branch 2023-05-28 09:26:13 +02:00
somebody
1546b9efaa Hello its breaking breakmodel time 2023-05-27 16:31:53 -05:00
ebolam
47276c3424 Bug Fix 2023-05-27 08:49:21 -04:00
ebolam
cce5c1932c Fix for custom model names 2023-05-26 21:40:39 -04:00
ebolam
9723154bed Fix for --path 2023-05-26 20:10:11 -04:00
ebolam
51cea7eb9d Added ability to add labels that are based on validation data in model loading settings 2023-05-26 12:43:17 -04:00
ebolam
1a1b79a16d Change default for HF on non-breakmodel models to use GPU instead of CPU 2023-05-24 12:44:26 -04:00
ebolam
068173b24a Potential BadWords fix 2023-05-24 09:08:34 -04:00
ebolam
9bd445c2a8 gpt2 fixed 2023-05-23 20:33:55 -04:00
ebolam
839d56ebf2 Potential fix for gpt-neo and gpt-j 2023-05-23 19:25:01 -04:00
ebolam
4c25d6fbbb Fix for loading model multiple times loosing the gpu/cpu splits 2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676 Fix for breakmodel loading to CPU when set to GPU 2023-05-22 20:24:57 -04:00
ebolam
dc20e6dde9 Fix for unloading models 2023-05-22 15:04:33 -04:00
ebolam
925cad2e2f Better compatibility with hf model backend 2023-05-22 14:50:13 -04:00
ebolam
513b8575e7 Fix for missing import
Fix for model name being a path which caused save issues
2023-05-20 11:01:49 -04:00
ebolam
9df1f03b12 Fix for custom huggingface model menu entry 2023-05-19 14:28:36 -04:00
ebolam
99cffd4755 Colab GPU edition fixes 2023-05-19 09:11:08 -04:00
ebolam
06f59a7b7b Moved model backends to separate folders
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
ebolam
71aee4dbd8 First concept of model plugins with a conceptual UI.
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
0cc4m
a2d01bb9e4 Update to GPTQ module 0.0.2, add support for upstream cuda quantizations, automatic detection 2023-05-09 22:20:35 +02:00
0cc4m
2f7856f0d1 Use GPTQ python module, add MPT quantized support 2023-05-06 20:52:42 +02:00
0cc4m
43b0afc7a8 Add safe MPT support 2023-05-05 20:07:10 +02:00
0cc4m
ef358fdf5a Merge remote-tracking branch 'origin/united' into model-structure-update 2023-05-04 07:31:13 +02:00
Henk
a87d5d6f23 Remove HF's llama workaround 2023-05-03 20:18:40 +02:00
Llama
35d344b951 Remove torch dependency and more generic dim0 workaround
Remove torch dependency from hf.py
Make workaround for dimension zero values of token_ids
more generic to handle every token, not just newlines.
2023-05-03 09:48:16 -07:00
0cc4m
58f0a336cb Merge upstream changes, fix conflict 2023-05-03 18:33:11 +02:00
Llama
3768848548 Fix tokenization and whitespace issues with llama-derived models
Work around the 'soft' prefix space behavior of sentencepiece.
Override encode to restore the deleted HF support for decode_with_prefix_space.
Override decode to skip the soft space and return true decoded tokens.
Allow submitting chat messages with embedded newlines.
Split sentences between punctuation and whitespace, rather than after whitespace.
Also include trailing quotes and brackets after sentence stoppers.
This avoids splitting ." and .) into two tokens, for instance.
Insert whitespace at the beginning of the author's note, since sentences are
split with leading whitespace.
Remove spurious newlines at the end of chat responses.
2023-05-03 01:27:11 -07:00
somebody
a0f4ab5c6a Move bad token grabber until after newlinemode has been deduced 2023-05-02 20:23:36 -05:00
somebody
efe268df60 Move overrides to better places 2023-05-02 20:18:33 -05:00
0cc4m
9c3d578d6c Work on model download support 2023-05-02 21:32:20 +02:00
somebody
933dbd634a HFInferenceModel: Make badwordsids not unique to torch 2023-05-01 17:13:33 -05:00
onesome
d4f7b60dc9 Fix for multiple paths 2023-04-26 16:49:12 -05:00
onesome
6776a71532 Add more info to custom model error 2023-04-26 16:36:52 -05:00
onesome
bbf4963d6e Fix custmodpth stuff for hf loading 2023-04-26 16:18:45 -05:00
somebody
8c8bdfaf6a Model: Fix assorted bugs
and ignore warnings in pytest
2023-03-10 20:00:39 -06:00
somebody
54cecd4d5d Model: And another refactor 2023-03-01 19:16:35 -06:00