Commit Graph

1609 Commits

Author SHA1 Message Date
somebody
b93c339145 Model: Lazyload backends 2023-03-13 20:29:29 -05:00
somebody
3adc67c7a4 RWKV: Move import right before usage
So we don't needlessly compile the cuda kernel
2023-03-13 19:37:45 -05:00
somebody
b10b201701 Model: Add basic RWKV implementation 2023-03-13 19:34:38 -05:00
somebody
60793eb121 Make modellist easier to work with 2023-03-13 15:40:24 -05:00
jojorne
c87ef60db1 revert more unrelated code 2023-03-11 13:48:41 -03:00
jojorne
423c022648 Fix save loading between v1 and v2 to v3 with wi features 2023-03-11 13:14:46 -03:00
somebody
38c4edac40 Model: Fix eos/bos padding issue
Weird config None assignments
2023-03-10 20:01:01 -06:00
SillyLossy
3ac3e41dd8 Fix setting softprompt via V1 API 2023-03-10 20:00:39 -06:00
somebody
a472bdf6c3 Cleanup 2023-03-10 20:00:39 -06:00
whjms
b5000b35d9 fix 404 handler 2023-03-09 18:17:30 -05:00
whjms
ad2c2b6722 move aetherroom import to separate module 2023-03-09 18:17:30 -05:00
SillyLossy
0d1771501f Fix setting softprompt via V1 API 2023-03-09 13:15:50 +02:00
somebody
dc3dd38c4e Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-03-07 17:02:28 -06:00
somebody
3b05359e4b Model: Refuse to serve certain models over the API 2023-03-07 16:54:08 -06:00
henk717
b307cdff57 Merge pull request #301 from jojorne/jojorne-patch-use-str-for-wi-keys
Access WI uid keys with str
2023-03-06 21:15:44 +01:00
Henk
099c818c06 Experimental unbanning of <> 2023-03-06 20:56:15 +01:00
Henk
b7f30a41c5 LLama tokenizer workaround 2023-03-06 19:32:28 +01:00
jojorne
4f17bf0c87 Access WI uid keys with str. 2023-03-05 13:22:33 -03:00
Henk
9ee6adb35c Enable BMS API again under a new name 2023-03-04 19:02:19 -06:00
Henk
55349d84e1 Remote no longer unblocks the port by default 2023-03-04 19:02:15 -06:00
Henk
59e1f74f7c Enable BMS API again under a new name 2023-03-04 14:17:47 +01:00
henk717
aa558da49c Merge pull request #297 from aibosan/united-singleline-stopper
Created SinglelineStopper...
2023-03-02 20:51:54 +01:00
Henk
f03b538cab Remote no longer unblocks the port by default 2023-03-02 17:43:23 +01:00
Ondřej Benda
aa124b65db Fix: incorrect newline evaluation 2023-03-02 08:22:31 +01:00
somebody
54cecd4d5d Model: And another refactor 2023-03-01 19:16:35 -06:00
Ondřej Benda
0ba7ac96d3 Created SinglelineStopper, which interrupts token generation when a newline is reached if singleline mode is enabled 2023-03-02 00:04:17 +01:00
Henk
93bf023bd7 Use our own horde URL 2023-03-01 17:54:39 +01:00
somebody
225dcf1a0a Model: Documentation part 1 2023-02-28 20:10:09 -06:00
somebody
ef1155291f Model: TPU Fixes 2023-02-28 18:05:34 -06:00
somebody
bd3bbdaad8 Model: More tpu fixes 2023-02-27 19:29:03 -06:00
somebody
b99c16f562 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-02-26 13:06:30 -06:00
somebody
56bcd32f6d Model: Fix ReadOnly
whoops
2023-02-26 12:40:20 -06:00
somebody
a73804ca1e Accelerate: Remove HAS_ACCELERATE
Accelerate has been a dependency for a while, and as such we probably
shouldn't be lugging around code that assumes it isn't present.
2023-02-26 12:18:06 -06:00
somebody
5e3b0062ee Model: Tokenizer fix 2023-02-26 12:17:49 -06:00
somebody
8d49b5cce1 Model: More TPU stuff 2023-02-26 12:09:26 -06:00
Henk
cdfc7326e6 Horde API Setting Consolidation 2023-02-26 17:59:36 +01:00
henk717
642d173579 Merge pull request #294 from db0/kaimerge
Changes to work with the merged hordes
2023-02-26 16:00:48 +01:00
somebody
ffe4f25349 Model: Work on stoppers and stuff 2023-02-25 17:12:16 -06:00
somebody
6b4905de30 Model: Port rest of models over
Generation's still broke but it's a start
2023-02-25 16:05:56 -06:00
Henk
165c7219eb Frequency Penalty for OAI 2023-02-25 22:15:05 +01:00
somebody
f8c4158ebc Model: Successful load implementation
The goal of this series of commits is to have an implementation-agnostic
interface for models, thus being less reliant on HF Transformers for model
support. A model object will have a method for generation, a list of callbacks
to be run on every token generation, a list of samplers that will modify
probabilities, etc. Basically anything HF can do should be easily
implementable with the new interface :^)

Currently I've tested the loading of pre-downloaded models with
breakmodel between GPUs and that works, though essentially no testing
has been done in the larger scheme of things. Currently this is about
the only supported configuration, and generation isn't very functional.
2023-02-24 21:41:44 -06:00
Divided by Zer0
6426e3ca24 changes 2023-02-23 18:34:46 +01:00
Divided by Zer0
2de9672b95 attempt1 2023-02-23 18:27:11 +01:00
Divided by Zer0
d459bdb1c0 adjust 2023-02-22 23:12:34 +01:00
jojorne
d3bedfcbda Include koboldai_vars.save_story(). 2023-02-22 15:42:56 -03:00
jojorne
d6c9f5f1f5 Save/Load Story with UTF-8 encoding. 2023-02-22 14:40:42 -03:00
Henk
9e6a5db745 UI1 Botname 2023-02-19 16:22:26 +01:00
henk717
93f313d6e3 Merge pull request #291 from pi6am/fix/save-as
Fix an exception using Save As from the classic UI
2023-02-19 13:21:10 +01:00
Llama
117f0659c3 Fix exception using Save As from the classic UI
The `saveas` method was modified to take a data dict but one of the
else blocks still referred to the previous `name` parameter. Assign
to `name` to fix the `NameError: name 'name' is not defined` exception.
2023-02-18 23:41:17 -08:00
Henk
cd566caf20 Revision Fixes (Removes the workaround) 2023-02-19 00:51:50 +01:00