Commit Graph

60 Commits

Author SHA1 Message Date
Henk
81e72329af CPU fixes 2023-07-02 21:50:23 +02:00
Henk
1da4580e8b Remove wrong usegpu behavior 2023-06-22 07:07:02 +02:00
somebody
5ee20bd7d6 Fix for CPU loading 2023-06-21 21:18:43 -05:00
somebody
b81f61b820 Clean debug 2023-06-21 18:35:56 -05:00
somebody
947bcc58e4 Experiments 2023-06-21 17:33:14 -05:00
somebody
c40649a74e Probably fix f32 2023-06-21 16:54:41 -05:00
somebody
aca2b532d7 Remove debug 2023-06-21 14:15:38 -05:00
somebody
5f224e1366 Restore choice of lazyload or not 2023-06-21 14:13:14 -05:00
somebody
0052ad401a Basic breakmodel ui support
Seems to work
2023-06-21 13:57:32 -05:00
somebody
f326fc07e8 Seems to work 2023-05-31 14:42:05 -05:00
somebody
24b0b32829 Maybe works now...? 2023-05-31 14:31:08 -05:00
somebody
ac4384ef75 Auto _no_split_modules 2023-05-31 10:55:46 -05:00
somebody
58ffad237b OPT hack 2023-05-29 13:34:11 -05:00
somebody
ceaefa9f5e Not quite 2023-05-28 14:57:45 -05:00
somebody
1546b9efaa Hello its breaking breakmodel time 2023-05-27 16:31:53 -05:00
ebolam
5561cc1f22 Fix for GPU generation 2023-05-23 08:33:19 -04:00
ebolam
4c25d6fbbb Fix for loading model multiple times loosing the gpu/cpu splits 2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676 Fix for breakmodel loading to CPU when set to GPU 2023-05-22 20:24:57 -04:00
ebolam
f1a16f260f Potential breakmodel fix 2023-05-22 16:10:41 -04:00
ebolam
ca770844b0 Fix for breakmodel 2023-05-22 15:07:59 -04:00
ebolam
3db231562f Merge pull request #382 from henk717/united
Update to united
2023-05-19 06:05:25 -04:00
ebolam
56d2705f4b removed breakmodel command line arguments (except nobreakmodel) 2023-05-18 20:19:33 -04:00
ebolam
06f59a7b7b Moved model backends to separate folders
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
Henk
205c64f1ea More universal pytorch folder detection 2023-05-13 20:26:55 +02:00
somebody
3065c1b40e Ignore missing keys in get_original_key 2023-05-11 17:10:43 -05:00
somebody
c16336f646 Add traceback to debug log on fallback 2023-05-11 17:10:19 -05:00
ebolam
71aee4dbd8 First concept of model plugins with a conceptual UI.
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
Henk
bb206f598e Don't load peft when unused 2023-05-06 18:55:26 +02:00
somebody
b7db709c47 PEFT: Change directory structure to be inside model 2023-05-06 11:16:09 -05:00
somebody
f02ddab7c7 Merge branch 'united' of https://github.com/henk717/KoboldAI into peft 2023-05-06 10:47:14 -05:00
Henk
33969b5845 Basic HF code execution support 2023-05-05 17:23:01 +02:00
somebody
35b56117e6 Basic PEFT support 2023-05-03 18:51:01 -05:00
somebody
efe268df60 Move overrides to better places 2023-05-02 20:18:33 -05:00
somebody
f6b5548131 Support safetensors in get_sharded_checkpoint_num_tensors 2023-05-01 19:15:27 -05:00
somebody
97e84928ba Download all shards correctly on aria2 and raise on bad load key 2023-05-01 18:53:36 -05:00
somebody
933dbd634a HFInferenceModel: Make badwordsids not unique to torch 2023-05-01 17:13:33 -05:00
onesome
467f2f25eb More loading fixes 2023-04-26 16:58:33 -05:00
onesome
9579298df7 Better fallback 2023-04-25 22:28:07 -05:00
onesome
b8bef641ff Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-04-25 16:54:53 -05:00
somebody
f9fb5eba89 Remove debug 2023-04-15 18:56:49 -05:00
somebody
5dd67d027a Workaround for socketio context errors for loading 2023-04-15 18:54:21 -05:00
one-some
1b500c7179 Merge pull request #5 from LostRuins/concedo_api
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3 Remove legacy no accelerate fallback code
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Concedo
9705b7b79c increase API version (+1 squashed commits)
Squashed commits:

[c168c08] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
somebody
ea8df4c0d3 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-04-14 20:38:56 -05:00
somebody
77f0797b1a Model fix 2023-04-02 15:47:52 -05:00
somebody
9d70646e4d Lazyload: Safetensors 2023-04-02 15:40:34 -05:00
somebody
ffe85ce8a1 Modeling: Fix logits processors (probs, biasing, lua) 2023-03-17 16:56:47 -05:00
somebody
692dbfeb37 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-03-17 16:20:13 -05:00
somebody
0320678b27 Model: WIP horde and API tests 2023-03-13 14:11:06 -05:00