Commit Graph

1825 Commits

Author SHA1 Message Date
henk717 0a926e41e4
Merge pull request #235 from VE-FORBRYDERNE/patch
Fix materialize function for galactica models
2022-12-12 20:15:54 +01:00
vfbd 33ba3e7e27 Fix materialize function for galactica models 2022-12-12 14:11:08 -05:00
Henk eeb1774d42 Cleaner implementation of zipfolder 2022-12-10 19:23:08 +01:00
Henk 9a8e8a0005 New pytorch zipfile support 2022-12-10 19:11:07 +01:00
henk717 dd7363548c
Merge pull request #191 from henk717/united
Probability Viewer Fix
2022-12-09 21:56:41 +01:00
henk717 686845cd21
Merge pull request #234 from one-some/united
Move probability visualization to after logitwarpers
2022-12-09 21:22:33 +01:00
somebody e6656d68a1 Move probability visualization to after logitwarpers 2022-12-09 13:47:38 -06:00
henk717 55ef53f39b
Typo fix 2022-12-08 15:17:10 +01:00
henk717 0b3e22ee13
Merge pull request #185 from henk717/united
Pin transformers version
2022-12-02 02:03:23 +01:00
Henk d0cb463c53 Pin transformers version
To avoid breaking changes lets force the exact transformers version we code against. This will be automatically picked up by all the automatic updaters.
2022-12-02 01:48:12 +01:00
henk717 e8245478d6
Merge pull request #184 from henk717/united
Cap transformers version
2022-12-02 01:27:18 +01:00
henk717 f72ceeadd0
Cap transformers version
Since MTJ is low level, we force a fixed transformers version to have more controlled updates when needed
2022-12-02 01:10:59 +01:00
henk717 04d9172fcd
Merge pull request #180 from VE-FORBRYDERNE/patch
Only enable TPU transpose optimization if loading from HF model
2022-11-21 20:02:14 +01:00
vfbd 9a3f0eaab2 Only enable TPU transpose optimization if loading from HF model 2022-11-21 13:47:18 -05:00
henk717 f2077b8e58
Merge pull request #179 from henk717/united
1.19.2
2022-11-20 16:26:03 +01:00
Henk 2603f1fd5d Version bump 2022-11-20 16:22:33 +01:00
Henk 3084552c05 Sampler Order Fix for Models 2022-11-14 17:15:39 +01:00
Henk 13dff68de8 Sampler Order Loading Fix 2022-11-14 16:59:53 +01:00
Henk a66e1443fd New Models 2022-11-12 16:54:40 +01:00
Henk 440c5c333e Clear flask_session on launch
Can help with version switching bugs
2022-11-12 15:43:06 +01:00
Henk f1e4664d56 Dependency improvements
Adding psutil from conda to avoid the need for a compiler, finetuneanon should no longer be used. If people really want to use it they are on their own.
2022-11-11 21:13:51 +01:00
Henk eb52ebd082 Merge branch 'main' into united 2022-11-03 00:22:30 +01:00
henk717 09b5ffc09d
Merge pull request #175 from VE-FORBRYDERNE/gptj-patch
Fix GPT-J model loading in TPU Colab when `vocab_size` is not divisible by 8
2022-11-03 00:13:50 +01:00
vfbd b20d80ca2a Add vocab padding to embedding bias in gptj.json 2022-11-02 19:02:09 -04:00
henk717 2e3a80b8ea
Merge branch 'KoboldAI:main' into united 2022-10-26 23:11:26 +02:00
henk717 7b5a766b4a
Merge pull request #172 from VE-FORBRYDERNE/accelerate-patch
Fix "is on the meta device" error when loading model with disk cache
2022-10-26 22:42:05 +02:00
vfbd 3233e78c56 Fix "is on the meta device" error when loading model with disk cache 2022-10-26 16:00:45 -04:00
Henk 442a9760b8 Hide V2 Saves 2022-10-23 19:03:18 +02:00
henk717 2300fb46ff
Merge branch 'KoboldAI:main' into united 2022-10-23 18:29:28 +02:00
Henk 8ee795055c Force compatible HF Hub 2022-10-23 18:28:50 +02:00
Henk ea8b50d31e Conda fix for update script 2022-10-23 16:00:18 +02:00
Henk 0da404d4f8 Conda conflict fix 2022-10-23 14:10:44 +02:00
Henk 4699ded3ce Tuner Dependencies 2022-10-22 19:00:06 +02:00
henk717 351fb3c80b
Merge pull request #232 from VE-FORBRYDERNE/mkultra
Universal mkultra-based soft prompt tuner
2022-10-22 14:13:42 +02:00
henk717 10a779d8c1
Merge pull request #231 from ebolam/united
Add parameter to Colab to use google drive
2022-10-22 14:13:32 +02:00
vfbd f7b799be56 Apply tokenizer fixes to prompt_tuner.py 2022-10-21 17:06:17 -04:00
ebolam d588dc0096 Check if dir exists before creating 2022-10-19 11:19:04 -04:00
ebolam 73865ba066 Add parameter to Colab for not using google drive (data would be ephemeral) 2022-10-19 11:05:17 -04:00
henk717 f8be854e09
Merge branch 'KoboldAI:main' into united 2022-10-17 21:06:10 +02:00
henk717 2795ced3a4
Merge pull request #168 from VE-FORBRYDERNE/api-patch
Fix regex for the prompt parameter of the POST /story/end endpoint
2022-10-17 20:38:34 +02:00
vfbd 9ff50d81fd Fix regex for the prompt parameter of the POST /story/end endpoint 2022-10-17 14:36:23 -04:00
henk717 c6ed656a76
Merge pull request #230 from pi6am/fix/lua_kobold_modeltype
Fix/lua kobold modeltype
2022-10-14 19:50:19 +02:00
Llama e5d0cc7b49 Fix exception thrown by kobold.modeltype in Lua
Fixes this exception:
  File "aiserver.py", line 3389, in lua_get_modeltype
    hidden_size = get_hidden_size_from_model(model)
NameError: name 'get_hidden_size_from_model' is not defined

The kobold.modeltype method eventually attempts to call
get_hidden_size_from_model in Python, but this method
was previously defined only within a local scope and so
is not visible from within lua_get_modeltype.  Since
get_hidden_size_from_model only accesses its model argument,
there is no reason not to make it a module-level method.

Also change the severity of several more Lua error logs to error.
2022-10-14 09:20:33 -07:00
Llama 6eb3abbdb8
Merge pull request #2 from henk717/united
Merging henk717/united
2022-10-13 20:33:34 -07:00
henk717 fff7837a4a
Merge pull request #229 from pi6am/feature/anote-kwarg
Feature/anote kwarg
2022-10-13 23:04:46 +02:00
henk717 be5ffe763c
Merge pull request #228 from VE-FORBRYDERNE/transpose
Slightly decrease TPU loading times
2022-10-13 15:35:28 +02:00
Llama 8357c3e485 Merge branch 'united' into feature/anote-kwarg 2022-10-12 23:37:45 -07:00
Llama 05bcd3af11
Merge pull request #1 from henk717/united
Version bump
2022-10-12 23:32:25 -07:00
Llama 4a01f345de Add include_anote kwarg to lua_compute_context.
Add an optional keyword argument to lua_compute_context to control
whether the author's note should be included in the context.  The
default value is true, so if the include_anote kwarg is not specified
then the author's note will be included, which was the default
behavior prior to this change.

Also update the Lua API documentation to describe this kwarg.
2022-10-12 23:18:19 -07:00
vfbd bdc73ef393 Decrease TPU loading times by eliminating a transpose operation 2022-10-12 14:31:18 -04:00