Commit Graph

1881 Commits

Author SHA1 Message Date
henk717 a04f99891f
Merge pull request #194 from Gouvernathor/patch-1
Update usage instructions for git-clone use
2022-12-21 15:36:15 +01:00
henk717 75fecb86cc
Merge pull request #196 from henk717/united
Improved model support & Shortcut Fixes
2022-12-18 20:18:13 +01:00
Gouvernathor a4f49c097a
Add git clone command and Linux case 2022-12-17 12:13:51 +01:00
Gouvernathor 55cf5f2f67
Update usage instructions for git-clone use 2022-12-17 04:10:56 +01:00
henk717 23b2d3a99e
Merge pull request #236 from one-some/united
Move shortcuts to Alt
2022-12-17 00:37:18 +01:00
somebody 9efbe381cf Move shortcuts to Alt from Ctrl 2022-12-16 16:47:22 -06:00
henk717 0a926e41e4
Merge pull request #235 from VE-FORBRYDERNE/patch
Fix materialize function for galactica models
2022-12-12 20:15:54 +01:00
vfbd 33ba3e7e27 Fix materialize function for galactica models 2022-12-12 14:11:08 -05:00
Henk eeb1774d42 Cleaner implementation of zipfolder 2022-12-10 19:23:08 +01:00
Henk 9a8e8a0005 New pytorch zipfile support 2022-12-10 19:11:07 +01:00
henk717 dd7363548c
Merge pull request #191 from henk717/united
Probability Viewer Fix
2022-12-09 21:56:41 +01:00
henk717 686845cd21
Merge pull request #234 from one-some/united
Move probability visualization to after logitwarpers
2022-12-09 21:22:33 +01:00
somebody e6656d68a1 Move probability visualization to after logitwarpers 2022-12-09 13:47:38 -06:00
henk717 55ef53f39b
Typo fix 2022-12-08 15:17:10 +01:00
henk717 0b3e22ee13
Merge pull request #185 from henk717/united
Pin transformers version
2022-12-02 02:03:23 +01:00
Henk d0cb463c53 Pin transformers version
To avoid breaking changes lets force the exact transformers version we code against. This will be automatically picked up by all the automatic updaters.
2022-12-02 01:48:12 +01:00
henk717 e8245478d6
Merge pull request #184 from henk717/united
Cap transformers version
2022-12-02 01:27:18 +01:00
henk717 f72ceeadd0
Cap transformers version
Since MTJ is low level, we force a fixed transformers version to have more controlled updates when needed
2022-12-02 01:10:59 +01:00
henk717 04d9172fcd
Merge pull request #180 from VE-FORBRYDERNE/patch
Only enable TPU transpose optimization if loading from HF model
2022-11-21 20:02:14 +01:00
vfbd 9a3f0eaab2 Only enable TPU transpose optimization if loading from HF model 2022-11-21 13:47:18 -05:00
henk717 f2077b8e58
Merge pull request #179 from henk717/united
1.19.2
2022-11-20 16:26:03 +01:00
Henk 2603f1fd5d Version bump 2022-11-20 16:22:33 +01:00
Henk 3084552c05 Sampler Order Fix for Models 2022-11-14 17:15:39 +01:00
Henk 13dff68de8 Sampler Order Loading Fix 2022-11-14 16:59:53 +01:00
Henk a66e1443fd New Models 2022-11-12 16:54:40 +01:00
Henk 440c5c333e Clear flask_session on launch
Can help with version switching bugs
2022-11-12 15:43:06 +01:00
Henk f1e4664d56 Dependency improvements
Adding psutil from conda to avoid the need for a compiler, finetuneanon should no longer be used. If people really want to use it they are on their own.
2022-11-11 21:13:51 +01:00
Henk eb52ebd082 Merge branch 'main' into united 2022-11-03 00:22:30 +01:00
henk717 09b5ffc09d
Merge pull request #175 from VE-FORBRYDERNE/gptj-patch
Fix GPT-J model loading in TPU Colab when `vocab_size` is not divisible by 8
2022-11-03 00:13:50 +01:00
vfbd b20d80ca2a Add vocab padding to embedding bias in gptj.json 2022-11-02 19:02:09 -04:00
henk717 2e3a80b8ea
Merge branch 'KoboldAI:main' into united 2022-10-26 23:11:26 +02:00
henk717 7b5a766b4a
Merge pull request #172 from VE-FORBRYDERNE/accelerate-patch
Fix "is on the meta device" error when loading model with disk cache
2022-10-26 22:42:05 +02:00
vfbd 3233e78c56 Fix "is on the meta device" error when loading model with disk cache 2022-10-26 16:00:45 -04:00
Henk 442a9760b8 Hide V2 Saves 2022-10-23 19:03:18 +02:00
henk717 2300fb46ff
Merge branch 'KoboldAI:main' into united 2022-10-23 18:29:28 +02:00
Henk 8ee795055c Force compatible HF Hub 2022-10-23 18:28:50 +02:00
Henk ea8b50d31e Conda fix for update script 2022-10-23 16:00:18 +02:00
Henk 0da404d4f8 Conda conflict fix 2022-10-23 14:10:44 +02:00
Henk 4699ded3ce Tuner Dependencies 2022-10-22 19:00:06 +02:00
henk717 351fb3c80b
Merge pull request #232 from VE-FORBRYDERNE/mkultra
Universal mkultra-based soft prompt tuner
2022-10-22 14:13:42 +02:00
henk717 10a779d8c1
Merge pull request #231 from ebolam/united
Add parameter to Colab to use google drive
2022-10-22 14:13:32 +02:00
vfbd f7b799be56 Apply tokenizer fixes to prompt_tuner.py 2022-10-21 17:06:17 -04:00
ebolam d588dc0096 Check if dir exists before creating 2022-10-19 11:19:04 -04:00
ebolam 73865ba066 Add parameter to Colab for not using google drive (data would be ephemeral) 2022-10-19 11:05:17 -04:00
henk717 f8be854e09
Merge branch 'KoboldAI:main' into united 2022-10-17 21:06:10 +02:00
henk717 2795ced3a4
Merge pull request #168 from VE-FORBRYDERNE/api-patch
Fix regex for the prompt parameter of the POST /story/end endpoint
2022-10-17 20:38:34 +02:00
vfbd 9ff50d81fd Fix regex for the prompt parameter of the POST /story/end endpoint 2022-10-17 14:36:23 -04:00
henk717 c6ed656a76
Merge pull request #230 from pi6am/fix/lua_kobold_modeltype
Fix/lua kobold modeltype
2022-10-14 19:50:19 +02:00
Llama e5d0cc7b49 Fix exception thrown by kobold.modeltype in Lua
Fixes this exception:
  File "aiserver.py", line 3389, in lua_get_modeltype
    hidden_size = get_hidden_size_from_model(model)
NameError: name 'get_hidden_size_from_model' is not defined

The kobold.modeltype method eventually attempts to call
get_hidden_size_from_model in Python, but this method
was previously defined only within a local scope and so
is not visible from within lua_get_modeltype.  Since
get_hidden_size_from_model only accesses its model argument,
there is no reason not to make it a module-level method.

Also change the severity of several more Lua error logs to error.
2022-10-14 09:20:33 -07:00
Llama 6eb3abbdb8
Merge pull request #2 from henk717/united
Merging henk717/united
2022-10-13 20:33:34 -07:00