Henk
1da4580e8b
Remove wrong usegpu behavior
2023-06-22 07:07:02 +02:00
somebody
5ee20bd7d6
Fix for CPU loading
2023-06-21 21:18:43 -05:00
somebody
b81f61b820
Clean debug
2023-06-21 18:35:56 -05:00
somebody
947bcc58e4
Experiments
2023-06-21 17:33:14 -05:00
somebody
c40649a74e
Probably fix f32
2023-06-21 16:54:41 -05:00
somebody
aca2b532d7
Remove debug
2023-06-21 14:15:38 -05:00
somebody
5f224e1366
Restore choice of lazyload or not
2023-06-21 14:13:14 -05:00
somebody
0052ad401a
Basic breakmodel ui support
...
Seems to work
2023-06-21 13:57:32 -05:00
somebody
f326fc07e8
Seems to work
2023-05-31 14:42:05 -05:00
somebody
24b0b32829
Maybe works now...?
2023-05-31 14:31:08 -05:00
somebody
ac4384ef75
Auto _no_split_modules
2023-05-31 10:55:46 -05:00
somebody
58ffad237b
OPT hack
2023-05-29 13:34:11 -05:00
somebody
ceaefa9f5e
Not quite
2023-05-28 14:57:45 -05:00
somebody
1546b9efaa
Hello its breaking breakmodel time
2023-05-27 16:31:53 -05:00
ebolam
5561cc1f22
Fix for GPU generation
2023-05-23 08:33:19 -04:00
ebolam
4c25d6fbbb
Fix for loading model multiple times loosing the gpu/cpu splits
2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676
Fix for breakmodel loading to CPU when set to GPU
2023-05-22 20:24:57 -04:00
ebolam
f1a16f260f
Potential breakmodel fix
2023-05-22 16:10:41 -04:00
ebolam
ca770844b0
Fix for breakmodel
2023-05-22 15:07:59 -04:00
ebolam
3db231562f
Merge pull request #382 from henk717/united
...
Update to united
2023-05-19 06:05:25 -04:00
ebolam
56d2705f4b
removed breakmodel command line arguments (except nobreakmodel)
2023-05-18 20:19:33 -04:00
ebolam
06f59a7b7b
Moved model backends to separate folders
...
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
Henk
205c64f1ea
More universal pytorch folder detection
2023-05-13 20:26:55 +02:00
somebody
3065c1b40e
Ignore missing keys in get_original_key
2023-05-11 17:10:43 -05:00
somebody
c16336f646
Add traceback to debug log on fallback
2023-05-11 17:10:19 -05:00
ebolam
71aee4dbd8
First concept of model plugins with a conceptual UI.
...
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
Henk
bb206f598e
Don't load peft when unused
2023-05-06 18:55:26 +02:00
somebody
b7db709c47
PEFT: Change directory structure to be inside model
2023-05-06 11:16:09 -05:00
somebody
f02ddab7c7
Merge branch 'united' of https://github.com/henk717/KoboldAI into peft
2023-05-06 10:47:14 -05:00
Henk
33969b5845
Basic HF code execution support
2023-05-05 17:23:01 +02:00
somebody
35b56117e6
Basic PEFT support
2023-05-03 18:51:01 -05:00
somebody
efe268df60
Move overrides to better places
2023-05-02 20:18:33 -05:00
somebody
f6b5548131
Support safetensors in get_sharded_checkpoint_num_tensors
2023-05-01 19:15:27 -05:00
somebody
97e84928ba
Download all shards correctly on aria2 and raise on bad load key
2023-05-01 18:53:36 -05:00
somebody
933dbd634a
HFInferenceModel: Make badwordsids not unique to torch
2023-05-01 17:13:33 -05:00
onesome
467f2f25eb
More loading fixes
2023-04-26 16:58:33 -05:00
onesome
9579298df7
Better fallback
2023-04-25 22:28:07 -05:00
onesome
b8bef641ff
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-25 16:54:53 -05:00
somebody
f9fb5eba89
Remove debug
2023-04-15 18:56:49 -05:00
somebody
5dd67d027a
Workaround for socketio context errors for loading
2023-04-15 18:54:21 -05:00
one-some
1b500c7179
Merge pull request #5 from LostRuins/concedo_api
...
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3
Remove legacy no accelerate fallback code
...
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Concedo
9705b7b79c
increase API version (+1 squashed commits)
...
Squashed commits:
[c168c08
] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
somebody
ea8df4c0d3
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-14 20:38:56 -05:00
somebody
77f0797b1a
Model fix
2023-04-02 15:47:52 -05:00
somebody
9d70646e4d
Lazyload: Safetensors
2023-04-02 15:40:34 -05:00
somebody
ffe85ce8a1
Modeling: Fix logits processors (probs, biasing, lua)
2023-03-17 16:56:47 -05:00
somebody
692dbfeb37
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-03-17 16:20:13 -05:00
somebody
0320678b27
Model: WIP horde and API tests
2023-03-13 14:11:06 -05:00
somebody
cd8ccf0a5e
Modeling: Add seed parameter to raw_generate
...
Yahooo, decoupling from koboldai_vars. This makes the generation test
pass in `test_generation.py`, and makes full determinism outside of
core_generate work.
2023-03-12 21:49:10 -05:00