Commit Graph

892 Commits

Author SHA1 Message Date
SmolBleat 692fe2e5ee Add Nerybus Models 2023-04-24 21:01:29 +02:00
Henk 93bf023bd7 Use our own horde URL 2023-03-01 17:54:39 +01:00
Divided by Zer0 6426e3ca24 changes 2023-02-23 18:34:46 +01:00
Divided by Zer0 2de9672b95 attempt1 2023-02-23 18:27:11 +01:00
Henk 739cccd8ed Revision Fixes 2023-01-31 04:48:46 +01:00
Henk eeb1774d42 Cleaner implementation of zipfolder 2022-12-10 19:23:08 +01:00
Henk 9a8e8a0005 New pytorch zipfile support 2022-12-10 19:11:07 +01:00
somebody e6656d68a1 Move probability visualization to after logitwarpers 2022-12-09 13:47:38 -06:00
Henk 2603f1fd5d Version bump 2022-11-20 16:22:33 +01:00
Henk 3084552c05 Sampler Order Fix for Models 2022-11-14 17:15:39 +01:00
Henk 13dff68de8 Sampler Order Loading Fix 2022-11-14 16:59:53 +01:00
Henk a66e1443fd New Models 2022-11-12 16:54:40 +01:00
henk717 2e3a80b8ea
Merge branch 'KoboldAI:main' into united 2022-10-26 23:11:26 +02:00
vfbd 3233e78c56 Fix "is on the meta device" error when loading model with disk cache 2022-10-26 16:00:45 -04:00
henk717 351fb3c80b
Merge pull request #232 from VE-FORBRYDERNE/mkultra
Universal mkultra-based soft prompt tuner
2022-10-22 14:13:42 +02:00
henk717 f8be854e09
Merge branch 'KoboldAI:main' into united 2022-10-17 21:06:10 +02:00
vfbd 9ff50d81fd Fix regex for the prompt parameter of the POST /story/end endpoint 2022-10-17 14:36:23 -04:00
Llama e5d0cc7b49 Fix exception thrown by kobold.modeltype in Lua
Fixes this exception:
  File "aiserver.py", line 3389, in lua_get_modeltype
    hidden_size = get_hidden_size_from_model(model)
NameError: name 'get_hidden_size_from_model' is not defined

The kobold.modeltype method eventually attempts to call
get_hidden_size_from_model in Python, but this method
was previously defined only within a local scope and so
is not visible from within lua_get_modeltype.  Since
get_hidden_size_from_model only accesses its model argument,
there is no reason not to make it a module-level method.

Also change the severity of several more Lua error logs to error.
2022-10-14 09:20:33 -07:00
Llama 8357c3e485 Merge branch 'united' into feature/anote-kwarg 2022-10-12 23:37:45 -07:00
Llama 4a01f345de Add include_anote kwarg to lua_compute_context.
Add an optional keyword argument to lua_compute_context to control
whether the author's note should be included in the context.  The
default value is true, so if the include_anote kwarg is not specified
then the author's note will be included, which was the default
behavior prior to this change.

Also update the Lua API documentation to describe this kwarg.
2022-10-12 23:18:19 -07:00
Henk 64715b18d6 Version bump 2022-10-12 14:54:11 +02:00
Henk d5143eeb80 LUA Error as Error 2022-10-12 01:23:00 +02:00
vfbd bdfa6d86b7 Seed has to be a 64-bit unsigned int or PyTorch will throw an error
tpu_mtj_backend's seed can be an integer of arbitrary size but we will
limit it to a 64-bit unsigned integer anyways for consistency.
2022-10-02 17:50:32 -04:00
vfbd dd1c25241d Allow sampler seed and full determinism to be read/written in /config 2022-10-02 17:43:54 -04:00
vfbd 1a59a4acea Allow changing sampler seed and sampler order from API 2022-10-02 16:25:51 -04:00
scythe000 a482ec16d8
Update aiserver.py - typo fix
Changed 'beakmodel' to 'breakmodel' in the example comment.
2022-09-30 10:29:32 -07:00
Divided by Zer0 90022d05c8 fix endpoint for get_cluster_models 2022-09-30 00:26:55 +02:00
vfbd 6758d5b538 Merge branch 'united' into mkultra 2022-09-28 14:30:34 -04:00
ebolam e7973e13ac Fix for GPT models downloading even when present in model folder 2022-09-28 12:47:50 -04:00
ebolam 72fc68c6e4 Fix for lazy loading on models after a non-lazy load model 2022-09-27 19:52:35 -04:00
ebolam 4aa842eada Merge commit 'refs/pull/180/head' of https://github.com/ebolam/KoboldAI into united 2022-09-27 19:29:05 -04:00
ebolam be719a7e5e Fix for loading models that don't support breakmodel (GPU/CPU support in UI) 2022-09-27 19:02:37 -04:00
Henk 52e120c706 Disable breakmodel if we error on the check 2022-09-28 01:00:06 +02:00
henk717 d2ff32be32
Merge pull request #220 from ebolam/united
Fix for loading models on CPU only that don't support breakmodel
2022-09-28 00:46:37 +02:00
Henk 057ddb4fb2 Better --cpu handling 2022-09-28 00:45:17 +02:00
ebolam 168ae8083c Remove debug print 2022-09-27 18:30:20 -04:00
ebolam 0311cc215e Fix for loading models on CPU only that don't support breakmodel 2022-09-27 18:29:32 -04:00
ebolam edd50fc809 Fix for GPT2 breakmodel in the UI 2022-09-27 17:58:51 -04:00
Henk f1d63f61f3 Syntax fix 2022-09-27 22:43:09 +02:00
ebolam 908dc8ea60 Fix for older model loading 2022-09-27 15:59:56 -04:00
Henk 62921c4896 getmodelname for configname 2022-09-27 21:11:31 +02:00
Henk 60d09899ea Don't use Fast tokenizers when we don't have to 2022-09-27 18:26:13 +02:00
Henk 11455697ef Tokenizer Fixes (Slow first to keep coherency) 2022-09-27 17:57:18 +02:00
Henk 07896867b2 Revert Tokenizer Change 2022-09-27 15:36:08 +02:00
Henk 82a250aa1b Revert "Fix tokenizer selection code"
This reverts commit 7fba1fd28a.
2022-09-27 15:33:08 +02:00
vfbd 79ae0f17ec Merge branch 'main' into merge 2022-09-26 16:10:10 -04:00
vfbd 7fba1fd28a Fix tokenizer selection code 2022-09-26 14:37:25 -04:00
henk717 c88f88f54d
Merge pull request #215 from ebolam/united
Update for horde selection to pull models automatically
2022-09-25 19:50:44 +02:00
ebolam 1f6861d55c Update for horde selection to pull models automatically (or on typing with a 1 second delay 2022-09-25 13:49:02 -04:00
Henk d7ed577bf7 Don't stream in chatmode 2022-09-25 17:26:52 +02:00