somebody
3aa677ce11
Indexed prioritization
...
not a great way to get that data
2023-07-23 16:04:23 -05:00
Henk
89637ae9d7
GPTQ Requirements
2023-07-23 22:51:47 +02:00
henk717
1facc73b66
Merge pull request #367 from 0cc4m/4bit-plugin
...
GPTQ module
2023-07-23 22:32:20 +02:00
0cc4m
73953068c0
Remove exllama backend, pending further fixes
2023-07-23 22:12:31 +02:00
0cc4m
973aea12ea
Only import big python modules for GPTQ once they get used
2023-07-23 22:07:34 +02:00
0cc4m
49740aa5ab
Fix ntk alpha
2023-07-23 21:56:48 +02:00
Henk
e33a58b74a
Adventure stoppers = regex
2023-07-23 20:45:48 +02:00
henk717
d70481874c
Merge pull request #334 from YellowRoseCx/YellowRoseCx-advpatch-1
...
Add AdventureMode stopper
2023-07-23 20:38:51 +02:00
henk717
66fb8b8937
Merge pull request #415 from one-some/fixes-forever
...
Fixes forever
2023-07-23 20:05:38 +02:00
henk717
b3b67bf50d
Merge pull request #417 from LostRuins/concedo_united
...
updated lite to v50
2023-07-23 16:11:23 +02:00
Concedo
dd8e5f5d05
updated lite to v50
2023-07-23 21:40:08 +08:00
0cc4m
31a984aa3d
Automatically install exllama module
2023-07-23 07:33:51 +02:00
0cc4m
a9aa04fd1b
Merge remote-tracking branch 'upstream/united' into 4bit-plugin
2023-07-23 07:18:58 +02:00
0cc4m
09bb1021dd
Fallback to transformers if hf_bleeding_edge not available
2023-07-23 07:16:52 +02:00
0cc4m
748e5ef318
Add sliders for exllama context size and related methods
2023-07-23 07:11:28 +02:00
Henk
7a5d813b92
Reimplement HF workaround only for llama
2023-07-22 16:59:49 +02:00
Henk
8dd7b93a6c
HF's workaround breaks stuff
2023-07-22 16:29:55 +02:00
Henk
fa9d17b3d3
HF 4.31
2023-07-22 15:25:14 +02:00
somebody
bc8ba91429
Private Mode improvements
...
- cleaned up ui a bit
- disallow setting empty password (user could press the hotkey by
mistake and be confused), giving the user a notification if they try
to.
- gives the login attempter a message if the password is incorrect, as
well as screaming in the logs
- obscures story name in page title
2023-07-21 21:44:10 -05:00
Henk
7823da564e
Link to Lite
2023-07-22 04:04:17 +02:00
somebody
a93c9d20b1
Don't let logo container gobble up clicks
2023-07-21 19:01:51 -05:00
somebody
79b1ef1aac
Fix "hide welcome logo" tweak
2023-07-21 19:01:40 -05:00
somebody
9188323331
Biases: Don't crash on empty token seq
2023-07-21 18:56:29 -05:00
somebody
5f4216730e
Make logit bias work correctly(?) when prob is -inf
...
samplers'll do that to you
though now i am curious: what kind of effect would running the bias
before the samplers have? maybe a future option
2023-07-21 18:33:35 -05:00
somebody
418f341560
Fix a/n depth being visually apart from a/n
2023-07-21 18:13:57 -05:00
somebody
560fb3bd2d
Fix occasional action highlight issue
2023-07-21 18:08:21 -05:00
henk717
83e5c29260
Merge pull request #413 from one-some/bug-hunt
...
Fix WI comment editing
2023-07-22 00:34:46 +02:00
somebody
e68972a270
Fix WI comments
2023-07-21 16:14:13 -05:00
Henk
a17d7aae60
Easier english
2023-07-21 19:42:49 +02:00
Henk
da9b54ec1c
Don't show API link during load
2023-07-21 19:31:38 +02:00
Henk
432cdc9a08
Fix models with good pad tokens
2023-07-21 16:39:58 +02:00
Henk
ec745d8b80
Dont accidentally block pad tokens
2023-07-21 16:25:32 +02:00
henk717
dc4404f29c
Merge pull request #409 from nkpz/bnb8bit
...
Configurable quantization level, fix for broken toggles in model settings
2023-07-19 14:22:44 +02:00
Nick Perez
9581e51476
feature(load model): select control for quantization level
2023-07-19 07:58:12 -04:00
0cc4m
58908ab846
Revert aiserver.py changes
2023-07-19 07:14:03 +02:00
0cc4m
19f511dc9f
Load GPTQ module from GPTQ repo docs
2023-07-19 07:12:37 +02:00
0cc4m
1c5da2bbf3
Move pip docs from KoboldAI into GPTQ repo
2023-07-19 07:08:39 +02:00
0cc4m
7516ecf00d
Merge upstream changes, fix conflict
2023-07-19 07:02:29 +02:00
0cc4m
c84d063be8
Revert settings changes
2023-07-19 07:01:11 +02:00
0cc4m
9aa6c5fbbf
Merge upstream changes, fix conflict, adapt backends to changes
2023-07-19 06:56:09 +02:00
Nick Perez
0142913060
8 bit toggle, fix for broken toggle values
2023-07-18 23:29:38 -04:00
Henk
22e7baec52
Permit CPU layers on 4-bit (Worse than GGML)
2023-07-18 21:44:34 +02:00
henk717
5f2600d338
Merge pull request #406 from ebolam/Model_Plugins
...
Clarified message on what's required for model backend parameters
2023-07-18 02:42:23 +02:00
ebolam
66192efdb7
Clarified message on what's required for model backend parameters in the command line
2023-07-17 20:30:41 -04:00
Henk
5bbcdc47da
4-bit on Colab
2023-07-18 01:48:01 +02:00
henk717
da9226fba5
Merge pull request #401 from ebolam/Model_Plugins
...
Save the 4-bit flag to the model settings.
2023-07-18 01:19:43 +02:00
henk717
fee79928c8
Merge pull request #404 from one-some/united
...
Delete basic 4bit
2023-07-18 01:19:14 +02:00
somebody
1637760fa1
Delete basic 4bit
...
And add code to handle dangling __pycache__s
2023-07-17 18:16:03 -05:00
henk717
5c3a8e295a
Merge pull request #402 from one-some/united
...
Patches: Make lazyload work with quantization
2023-07-17 23:53:14 +02:00
somebody
23b95343bd
Patches: Make lazyload work on quantized
...
i wanna watch youtube while my model is loading without locking up my
system >:(
2023-07-17 16:47:31 -05:00