Commit Graph

4438 Commits

Author SHA1 Message Date
henk717
04a5e05692 Merge pull request #422 from one-some/fix-prioritization
Fix prioritization (probably)
2023-07-24 17:40:36 +02:00
somebody
2fb877db40 Backends: Probably fix sorting 2023-07-24 10:28:22 -05:00
somebody
81e4c8a807 Backends: Fix GPTQ priority 2023-07-24 10:25:44 -05:00
Henk
30495cf8d8 Fix GPT2 2023-07-24 02:05:07 +02:00
Henk
9fc9cb92f7 Fancy streaming by default 2023-07-24 01:39:16 +02:00
henk717
c60a29e544 Merge pull request #419 from one-some/streaming-fix-2
Rework single-gen streaming
2023-07-24 01:31:51 +02:00
Henk
70d2da55e5 Readme changes 2023-07-24 01:03:46 +02:00
somebody
f4593ed04b Streaming: Fix bad streamingwindow sync
sometimes the good solution is the best solution
2023-07-23 17:50:53 -05:00
somebody
8de610df8c Streaming: Rework single-gen streaming
Now has its own packet and element seperate from actions. Fixes several
bugs related to desyncing. Also adds smooth typing effect
(fixes https://github.com/henk717/KoboldAI/issues/263)
2023-07-23 17:32:52 -05:00
Henk
a963c97acb Make 4-bit the default part 2 2023-07-24 00:06:20 +02:00
Henk
3409853dfc Remove GPTQ for Colab 2023-07-23 23:26:35 +02:00
Henk
70dddf9fdc Prioritize GPTQ 2023-07-23 23:22:02 +02:00
henk717
ae9ec38ae2 Merge pull request #418 from one-some/united
index prioritization
2023-07-23 23:09:42 +02:00
Henk
0f913275a9 4-bit as Default 2023-07-23 23:08:11 +02:00
somebody
3aa677ce11 Indexed prioritization
not a great way to get that data
2023-07-23 16:04:23 -05:00
Henk
89637ae9d7 GPTQ Requirements 2023-07-23 22:51:47 +02:00
henk717
1facc73b66 Merge pull request #367 from 0cc4m/4bit-plugin
GPTQ module
2023-07-23 22:32:20 +02:00
0cc4m
73953068c0 Remove exllama backend, pending further fixes 2023-07-23 22:12:31 +02:00
0cc4m
973aea12ea Only import big python modules for GPTQ once they get used 2023-07-23 22:07:34 +02:00
0cc4m
49740aa5ab Fix ntk alpha 2023-07-23 21:56:48 +02:00
Henk
e33a58b74a Adventure stoppers = regex 2023-07-23 20:45:48 +02:00
henk717
d70481874c Merge pull request #334 from YellowRoseCx/YellowRoseCx-advpatch-1
Add AdventureMode stopper
2023-07-23 20:38:51 +02:00
henk717
66fb8b8937 Merge pull request #415 from one-some/fixes-forever
Fixes forever
2023-07-23 20:05:38 +02:00
henk717
b3b67bf50d Merge pull request #417 from LostRuins/concedo_united
updated lite to v50
2023-07-23 16:11:23 +02:00
Concedo
dd8e5f5d05 updated lite to v50 2023-07-23 21:40:08 +08:00
0cc4m
31a984aa3d Automatically install exllama module 2023-07-23 07:33:51 +02:00
0cc4m
a9aa04fd1b Merge remote-tracking branch 'upstream/united' into 4bit-plugin 2023-07-23 07:18:58 +02:00
0cc4m
09bb1021dd Fallback to transformers if hf_bleeding_edge not available 2023-07-23 07:16:52 +02:00
0cc4m
748e5ef318 Add sliders for exllama context size and related methods 2023-07-23 07:11:28 +02:00
Henk
7a5d813b92 Reimplement HF workaround only for llama 2023-07-22 16:59:49 +02:00
Henk
8dd7b93a6c HF's workaround breaks stuff 2023-07-22 16:29:55 +02:00
Henk
fa9d17b3d3 HF 4.31 2023-07-22 15:25:14 +02:00
somebody
bc8ba91429 Private Mode improvements
- cleaned up ui a bit
- disallow setting empty password (user could press the hotkey by
  mistake and be confused), giving the user a notification if they try
  to.
- gives the login attempter a message if the password is incorrect, as
  well as screaming in the logs
- obscures story name in page title
2023-07-21 21:44:10 -05:00
Henk
7823da564e Link to Lite 2023-07-22 04:04:17 +02:00
somebody
a93c9d20b1 Don't let logo container gobble up clicks 2023-07-21 19:01:51 -05:00
somebody
79b1ef1aac Fix "hide welcome logo" tweak 2023-07-21 19:01:40 -05:00
somebody
9188323331 Biases: Don't crash on empty token seq 2023-07-21 18:56:29 -05:00
somebody
5f4216730e Make logit bias work correctly(?) when prob is -inf
samplers'll do that to you

though now i am curious: what kind of effect would running the bias
before the samplers have? maybe a future option
2023-07-21 18:33:35 -05:00
somebody
418f341560 Fix a/n depth being visually apart from a/n 2023-07-21 18:13:57 -05:00
somebody
560fb3bd2d Fix occasional action highlight issue 2023-07-21 18:08:21 -05:00
henk717
83e5c29260 Merge pull request #413 from one-some/bug-hunt
Fix WI comment editing
2023-07-22 00:34:46 +02:00
somebody
e68972a270 Fix WI comments 2023-07-21 16:14:13 -05:00
Henk
a17d7aae60 Easier english 2023-07-21 19:42:49 +02:00
Henk
da9b54ec1c Don't show API link during load 2023-07-21 19:31:38 +02:00
Henk
432cdc9a08 Fix models with good pad tokens 2023-07-21 16:39:58 +02:00
Henk
ec745d8b80 Dont accidentally block pad tokens 2023-07-21 16:25:32 +02:00
henk717
dc4404f29c Merge pull request #409 from nkpz/bnb8bit
Configurable quantization level, fix for broken toggles in model settings
2023-07-19 14:22:44 +02:00
Nick Perez
9581e51476 feature(load model): select control for quantization level 2023-07-19 07:58:12 -04:00
0cc4m
58908ab846 Revert aiserver.py changes 2023-07-19 07:14:03 +02:00
0cc4m
19f511dc9f Load GPTQ module from GPTQ repo docs 2023-07-19 07:12:37 +02:00