Commit Graph

1617 Commits

Author SHA1 Message Date
ebolam
182ecff202 Added in model backend to the command line arguments 2023-05-18 16:01:17 -04:00
ebolam
f027d8b6e5 Better working valid detection and named model backends for UI 2023-05-17 21:15:31 -04:00
ebolam
c6b17889d0 Updated to latest united 2023-05-12 07:53:27 -04:00
Henk
67df9b917f Reintroduce 4.29 Transformers 2023-05-12 09:08:07 +02:00
ebolam
aaa9133899 Disk Cache working
UI valid marker broken for disk cache
2023-05-11 21:22:33 -04:00
ebolam
69d942c00c Kind of working breakmodel 2023-05-11 20:22:30 -04:00
Henk
20b54eb9ff Revert 4.29 due to unforseen consequences 2023-05-11 19:06:39 +02:00
ebolam
4605d10c37 Next iteration. Model Loading is broken completely now :) 2023-05-11 12:08:35 -04:00
ebolam
77dd5aa725 Minor update 2023-05-11 09:09:09 -04:00
Henk
e932364a1e RWKV support 2023-05-11 14:56:12 +02:00
ebolam
71aee4dbd8 First concept of model plugins with a conceptual UI.
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
somebody
a9e342ca64 Fix TPU API errors 2023-05-08 17:34:59 -05:00
somebody
f02ddab7c7 Merge branch 'united' of https://github.com/henk717/KoboldAI into peft 2023-05-06 10:47:14 -05:00
Henk
2730879c61 Better warning until something more robust is in 2023-05-05 21:28:06 +02:00
Henk
d508b4a319 More max_context_length flexibility 2023-05-05 19:50:56 +02:00
Henk
33969b5845 Basic HF code execution support 2023-05-05 17:23:01 +02:00
somebody
35b56117e6 Basic PEFT support 2023-05-03 18:51:01 -05:00
henk717
7f5242db17 Merge pull request #344 from pi6am/fix/llama-tokens
Fix/llama tokens
2023-05-03 19:07:47 +02:00
ebolam
fa3611b994 Update to United
Update to United
2023-05-03 10:54:17 -04:00
Llama
3768848548 Fix tokenization and whitespace issues with llama-derived models
Work around the 'soft' prefix space behavior of sentencepiece.
Override encode to restore the deleted HF support for decode_with_prefix_space.
Override decode to skip the soft space and return true decoded tokens.
Allow submitting chat messages with embedded newlines.
Split sentences between punctuation and whitespace, rather than after whitespace.
Also include trailing quotes and brackets after sentence stoppers.
This avoids splitting ." and .) into two tokens, for instance.
Insert whitespace at the beginning of the author's note, since sentences are
split with leading whitespace.
Remove spurious newlines at the end of chat responses.
2023-05-03 01:27:11 -07:00
Henk
5d1ee39250 Fix loadmodelsettings 2023-05-03 04:21:37 +02:00
henk717
724ba43dc1 Merge pull request #342 from one-some/model-structure-and-maybe-rwkv
Move overrides to better places
2023-05-03 03:34:17 +02:00
somebody
4b3b240bce Move loadmodelsettings 2023-05-02 20:33:37 -05:00
Henk
480919a2a7 Nicer way of serving lite 2023-05-03 01:16:02 +02:00
Henk
03e10bed82 /lite (Not functional yet) 2023-05-03 01:04:51 +02:00
somebody
c95be636a4 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-05-01 17:08:20 -05:00
ebolam
5a32159e58 Remove debug prints 2023-05-01 10:53:02 -04:00
Henk
545f79086d Ban EOS token in N mode 2023-04-30 18:48:22 +02:00
one-some
19817a271b More colab 2023-04-28 10:16:15 -05:00
one-some
b3614b64b1 Hello Colab 2023-04-28 10:10:26 -05:00
one-some
f9d162c001 Cut out things until it works 2023-04-27 10:10:17 -05:00
henk717
b0bbdc0c29 Merge pull request #333 from ebolam/united
New Editor
2023-04-27 15:52:13 +02:00
onesome
9579298df7 Better fallback 2023-04-25 22:28:07 -05:00
onesome
98cd6aa246 Make RWKV experimental 2023-04-25 17:13:06 -05:00
onesome
b8bef641ff Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-04-25 16:54:53 -05:00
Henk
9eaa2aba47 Isolate OPT Tokenizer Fix to OPT models 2023-04-25 22:49:56 +02:00
SmolBleat
692fe2e5ee Add Nerybus Models 2023-04-24 21:01:29 +02:00
ebolam
d7c46f668c fix tab vs space error 2023-04-21 14:11:51 -04:00
ebolam
3990e67b81 Merge commit 'refs/pull/378/head' of https://github.com/ebolam/KoboldAI into united 2023-04-21 08:27:36 -04:00
ebolam
a34ce85d21 Fix for UI1 remote mode file loading from browser 2023-04-16 20:43:03 -04:00
Henk
2926dac72b Don't print allowed_ips if unused 2023-04-16 23:59:13 +02:00
somebody
5dd67d027a Workaround for socketio context errors for loading 2023-04-15 18:54:21 -05:00
somebody
08b4e317ff Fix double slashing 2023-04-15 13:30:05 -05:00
somebody
d3a73aaeba Fix api 2023-04-15 13:17:20 -05:00
somebody
4dcf570407 Fix legacy model loading 2023-04-15 12:57:35 -05:00
Henk
b68860b3de Workaround to make --host work again 2023-04-15 18:31:39 +02:00
one-some
1b500c7179 Merge pull request #5 from LostRuins/concedo_api
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
b2e6fcfe3a Remove line that sets disk_layers to None always
whoops
2023-04-15 10:41:10 -05:00
Henk
3eda7269f7 Fix incorrect host merge 2023-04-15 14:58:24 +02:00
Concedo
9705b7b79c increase API version (+1 squashed commits)
Squashed commits:

[c168c08] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00