Commit Graph

4539 Commits

Author SHA1 Message Date
Henk
5d1ee39250 Fix loadmodelsettings 2023-05-03 04:21:37 +02:00
henk717
724ba43dc1 Merge pull request #342 from one-some/model-structure-and-maybe-rwkv
Move overrides to better places
2023-05-03 03:34:17 +02:00
somebody
4b3b240bce Move loadmodelsettings 2023-05-02 20:33:37 -05:00
somebody
a0f4ab5c6a Move bad token grabber until after newlinemode has been deduced 2023-05-02 20:23:36 -05:00
somebody
efe268df60 Move overrides to better places 2023-05-02 20:18:33 -05:00
Henk
480919a2a7 Nicer way of serving lite 2023-05-03 01:16:02 +02:00
Henk
03e10bed82 /lite (Not functional yet) 2023-05-03 01:04:51 +02:00
Henk
de7b760048 Typo Fix 2023-05-03 01:02:50 +02:00
0cc4m
dd6644aaf0 Pytorch 2.0 (#18)
* Update huggingface.yml to Pytorch 2.0 and CUDA 11.8

* Update github docs pip wheel hub

Update ROCm requirements

* Add rocm wheel
2023-05-02 22:11:28 +02:00
0cc4m
9c3d578d6c Work on model download support 2023-05-02 21:32:20 +02:00
henk717
50c9ed3af1 Merge pull request #299 from one-some/model-structure-and-maybe-rwkv
Structure changes
2023-05-02 18:07:09 +02:00
somebody
111028642e Fix tokenizer fallback for llama 2023-05-01 19:42:52 -05:00
somebody
f6b5548131 Support safetensors in get_sharded_checkpoint_num_tensors 2023-05-01 19:15:27 -05:00
somebody
97e84928ba Download all shards correctly on aria2 and raise on bad load key 2023-05-01 18:53:36 -05:00
somebody
933dbd634a HFInferenceModel: Make badwordsids not unique to torch 2023-05-01 17:13:33 -05:00
somebody
c95be636a4 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-05-01 17:08:20 -05:00
somebody
ce3d465972 Remove some debug 2023-05-01 17:03:34 -05:00
ebolam
5a32159e58 Remove debug prints 2023-05-01 10:53:02 -04:00
ebolam
137d056cb3 Fix for pasting text in the middle of an action 2023-05-01 10:48:45 -04:00
0cc4m
f83a0aa122 Merge latest changes, fix conflict 2023-05-01 08:01:54 +02:00
Llama
eb4e89c2fa Merge pull request #29 from henk717/united
Merge united
2023-04-30 14:20:12 -07:00
0cc4m
aa67135d42 Implement new model format
Remove 4bit toggle
2023-04-30 21:59:22 +02:00
Henk
545f79086d Ban EOS token in N mode 2023-04-30 18:48:22 +02:00
0cc4m
20a5587d66 Always use offloader script, because it speeds up multi gpu 2023-04-30 18:17:43 +02:00
henk717
61511a5714 Merge pull request #341 from TinkerTankAI/united
Update KoboldAI-Horde-Bridge to the latest version
2023-04-29 15:09:15 +02:00
Tijs Zwinkels
2ad66ebcc0 Update KoboldAI-Horde-Bridge to the latest version
This version contains a timeout on http requests,
preventing a hang in my worker.
2023-04-29 15:07:32 +02:00
0cc4m
2859c67c67 Merge remote-tracking branch 'origin/united' into latestgptq 2023-04-29 13:57:34 +02:00
henk717
b19bd9c89e Merge branch 'KoboldAI:main' into united 2023-04-29 02:45:45 +02:00
Henk
1499763472 Flask fix 2023-04-29 02:44:41 +02:00
somebody
bfef79d2b8 Dynamic Fix 2023-04-28 18:55:38 -05:00
henk717
32ca5a716b Merge branch 'KoboldAI:main' into united 2023-04-28 22:29:22 +02:00
0cc4m
852005fef4 Always use GPU offloader if splitting across GPUs, this increases speed considerably 2023-04-28 18:32:34 +02:00
one-some
455b8257a9 Implement softprompt hack 2023-04-28 10:26:59 -05:00
one-some
fa6bb4b956 Potential fix for tpu dynamic 2023-04-28 10:17:48 -05:00
one-some
19817a271b More colab 2023-04-28 10:16:15 -05:00
one-some
b3614b64b1 Hello Colab 2023-04-28 10:10:26 -05:00
0cc4m
81f92ec402 Fix missing 4bit setting 2023-04-28 10:55:22 +02:00
0cc4m
5ac7527ade Merge upstream changes, fix conflict 2023-04-28 07:07:21 +02:00
somebody
ace4364339 Two more time 2023-04-27 21:13:26 -05:00
somebody
446f38ee9d One more time 2023-04-27 21:07:34 -05:00
somebody
2eee535540 Actually fix decoding with soft prompts
it really wants a tensor
2023-04-27 21:01:12 -05:00
somebody
ffa7b22734 Experiment 2023-04-27 20:28:04 -05:00
somebody
cd1eb97c2a Debuuuug 2023-04-27 20:12:29 -05:00
somebody
4559112551 Potential fix 2023-04-27 19:51:10 -05:00
somebody
b256a8fbc7 Debug 2023-04-27 19:33:03 -05:00
henk717
e9e09c4b45 Merge pull request #338 from one-some/united
Fix phantom 8 bit variable
2023-04-28 02:09:07 +02:00
somebody
720b2ec7f0 Fix phantom variable 2023-04-27 19:05:38 -05:00
0cc4m
d8949042d4 Merge latestgptq branch changes 2023-04-27 19:12:29 +02:00
one-some
f9d162c001 Cut out things until it works 2023-04-27 10:10:17 -05:00
henk717
945aec2320 Merge pull request #288 from vanderh0ff/dev
updated micro mamba user directory to match username
2023-04-27 16:12:29 +02:00