Commit Graph

4427 Commits

Author SHA1 Message Date
ebolam
03a0542f71 Fix for model backends that use toggles always returning true 2023-06-21 19:19:12 -04:00
somebody
d4b923a054 Remove debug 2023-06-21 17:41:15 -05:00
somebody
5278174a62 Materialize on cpu 2023-06-21 17:40:47 -05:00
somebody
947bcc58e4 Experiments 2023-06-21 17:33:14 -05:00
somebody
0012158eac Remove old 2023-06-21 16:58:59 -05:00
somebody
6bdcf2645e Merge branch 'united' of https://github.com/henk717/KoboldAI into accelerate-offloading 2023-06-21 16:58:39 -05:00
somebody
c40649a74e Probably fix f32 2023-06-21 16:54:41 -05:00
somebody
70f113141c Fix Transformers 4.30 2023-06-21 16:40:12 -05:00
somebody
c56214c275 Fix loading bar 2023-06-21 16:27:22 -05:00
Henk
adf108ecd6 TPU link fix 2023-06-21 22:21:22 +02:00
Henk
b41b868528 Remove duplicate links 2023-06-21 22:13:55 +02:00
Henk
a13c7d0f40 New link messages 2023-06-21 21:55:18 +02:00
Henk
fc4d659e13 Merge branch 'main' into united 2023-06-21 21:32:55 +02:00
somebody
aca2b532d7 Remove debug 2023-06-21 14:15:38 -05:00
somebody
5f224e1366 Restore choice of lazyload or not 2023-06-21 14:13:14 -05:00
somebody
0052ad401a Basic breakmodel ui support
Seems to work
2023-06-21 13:57:32 -05:00
Henk
0c19855587 HF_Hub bump 2023-06-21 19:40:07 +02:00
Henk
bbecdaeedb Silently disable MTJ when Jax is not installed 2023-06-21 17:08:45 +02:00
0cc4m
adad81639d Remove rocm gptq install from environments file 2023-06-21 15:47:46 +02:00
0cc4m
e8741a1b57 Disable scaled_dot_product_attention if torch version < 2 2023-06-20 09:19:43 +02:00
0cc4m
a191855b37 Track token generation progress 2023-06-19 19:14:26 +02:00
0cc4m
e874f0c1c2 Add token streaming support for exllama 2023-06-19 19:14:26 +02:00
henk717
d46663ac0d Merge pull request #376 from LostRuins/concedo_united
Updated Kobold Lite to v41
2023-06-18 14:30:08 +02:00
LostRuins
e7f1f47d94 Merge branch 'henk717:united' into concedo_united 2023-06-18 15:40:19 +08:00
Concedo
f42d5a4b10 Updated Kobold Lite to v41 2023-06-18 15:37:38 +08:00
YellowRoseCx
8b742b2bd4 add missing @staticmethod 2023-06-15 17:20:38 -05:00
YellowRoseCx
83493dff2e modify adv stopper 2023-06-15 17:15:33 -05:00
YellowRoseCx
877028ec7f Update hf_torch.py with adv mode stopper 2023-06-15 16:07:54 -05:00
YellowRoseCx
73c06bf0a5 add adventuremode stopper
adds a stopper token for adventure mode when it detects the bot generating impersonating text after " > You"
2023-06-15 16:02:20 -05:00
henk717
f863d5db2d Merge pull request #373 from ebolam/Model_Plugins
Making model backends respond to a specific type in the aiserver menu for now
2023-06-14 02:13:44 +02:00
ebolam
abe07a2e95 Fix for model loading from paths 2023-06-13 20:05:50 -04:00
0cc4m
0c7eaefb1a Fix AMD ROCm exllama inference 2023-06-13 10:11:29 +02:00
ebolam
e2801fb5c1 Merge branch 'henk717:united' into Model_Plugins 2023-06-12 17:36:06 -04:00
ebolam
dfb097d048 Moving basic hf to a new branch 2023-06-12 17:35:34 -04:00
0cc4m
ebf7e2cf57 Update GPTQ module to 0.0.6 2023-06-12 08:27:30 +02:00
0cc4m
0001ae00ab Add v2 with bias support (e.g. for Tulu-30b) 2023-06-12 07:18:22 +02:00
0cc4m
12df8220fb Add gpt_bigcode support, fix 8-bit GPTQ incoherence 2023-06-12 07:14:36 +02:00
0cc4m
47b371b9d3 Fix multigpu 2023-06-06 19:51:38 +02:00
0cc4m
39dfb18455 Replace exllama samplers with kobold's inbuilt ones 2023-06-06 19:21:34 +02:00
0cc4m
94520d5c80 Fix exllama model unload 2023-06-05 18:43:57 +02:00
henk717
22b2a3f327 Merge pull request #371 from LostRuins/concedo_united
updated kobold lite to v37
2023-06-04 17:01:51 +02:00
Concedo
49a64fb655 updated kobold lite to v37 2023-06-04 22:27:52 +08:00
0cc4m
b35f61e987 Basic exllama plugin 2023-06-04 15:40:12 +02:00
Henk
4d213ddc51 Update drive disconnect fix 2023-06-04 14:12:00 +02:00
0cc4m
c82625490a Rename gptq backend folder 2023-06-04 12:31:24 +02:00
0cc4m
05a0bfe6c4 Don't show HF support if no HF model files are found 2023-06-04 09:44:28 +02:00
0cc4m
eace95cc72 Pull upstream changes, fix conflict 2023-06-04 09:06:31 +02:00
0cc4m
974328ed22 Add 4bit requirements to requirements.txt 2023-06-04 09:02:15 +02:00
0cc4m
b7838c7dde Fall back to autogptq if available and model not supported by gptq-koboldai 2023-06-04 08:06:48 +02:00
ebolam
ac265723e4 Merge branch 'henk717:united' into Model_Plugins 2023-06-03 11:08:00 -04:00