Commit Graph

4791 Commits

Author SHA1 Message Date
Henk
cf4adfc7b4 No longer mess with quotes 2023-09-06 04:50:58 +02:00
Henk
ac61cd8e59 Slow tokenizer first 2023-09-06 04:35:25 +02:00
henk717
0e7f27247f Merge pull request #453 from AlpinDale/united
Add new models
2023-09-06 03:50:10 +02:00
henk717
256f633f11 Merge pull request #454 from ebolam/Model_Plugins
Fix for options not showing up on refresh
2023-09-06 03:43:46 +02:00
ebolam
9f1e87fb49 Remove debug text 2023-09-05 19:22:25 -04:00
ebolam
fb77d8ebd5 Merge branch 'Model_Plugins' of https://github.com/ebolam/KoboldAI into Model_Plugins 2023-09-05 19:21:44 -04:00
ebolam
bcfe5c7fdd Fix for text options not showing after multi-gen when refreshing web page. 2023-09-05 19:21:31 -04:00
ebolam
2d0040871d Fix for text options not showing after multi-gen when refreshing web page. 2023-09-05 19:20:35 -04:00
AlpinDale
9209a9fdab Add new models 2023-09-05 20:07:08 +00:00
henk717
4260fa60ad Merge pull request #451 from ebolam/Model_Plugins
Fix for mobile touch on setting sliders
2023-09-04 22:31:37 +02:00
ebolam
0a55d9c735 Merge branch 'Model_Plugins' of https://github.com/ebolam/KoboldAI into Model_Plugins 2023-09-04 16:03:04 -04:00
ebolam
45520684df Fix for mobile (Tested on Chrome Android Samsung phone) 2023-09-04 16:03:00 -04:00
ebolam
e2eb96f443 Merge branch 'henk717:united' into Model_Plugins 2023-09-04 15:32:12 -04:00
ebolam
5d23f10bf3 Fix for mobile touch 2023-09-04 15:31:30 -04:00
henk717
5c6ff8faa5 Merge pull request #450 from Disty0/united
Initial Intel ARC support with IPEX
2023-09-04 20:56:34 +02:00
Disty0
7cd23ed1a8 Fix return_xpu 2023-09-04 11:43:59 +03:00
Disty0
8151948698 Initial Intel ARC support with IPEX 2023-09-04 04:06:41 +03:00
Henk
af330f0445 Remove legnth cap 2023-09-04 01:23:31 +02:00
Henk
cc213e028d Ignore unknown generation parameters 2023-09-04 00:41:39 +02:00
Henk
0ab11e1322 Allow unlimited stop sequences 2023-09-03 23:25:35 +02:00
henk717
ce32dddef3 Merge pull request #449 from LostRuins/concedo_united_PR
Updated Kobold Lite to v60
2023-09-03 16:09:06 +02:00
Concedo
966b6f8a2e Updated Kobold Lite to v60 2023-09-03 11:40:28 +08:00
Henk
4c03402e33 API Example Polish and Fixes 2023-09-02 03:41:01 +02:00
henk717
d1430d703d Merge pull request #448 from YellowRoseCx/patch-3
Create api_example.py
2023-09-02 03:39:25 +02:00
YellowRoseCx
cb577bd491 Create api_example.py 2023-09-01 20:38:58 -05:00
Henk
565ab8a38f badwordids -> badwordsids typofix 2023-08-30 17:00:51 +02:00
henk717
1d32d16ba4 Merge branch 'KoboldAI:main' into united 2023-08-30 13:13:20 +02:00
Llama
fe53cb2f7e Merge pull request #70 from pi6am/feat/exllama-unban-eos
Hook up use_default_badwordids in exllama
2023-08-29 23:17:37 -07:00
Llama
d6ed75f993 Hook up use_default_badwordids in exllama
Use the value of the use_default_badwordids setting to configure
bad_words_ids. Also add square brackets to bad_words_ids if the
use_default_badwordids setting is True. Fix an issue with
attempting to use the tokenizer too early, and fix an exception
populating Lua bridge data when zero tokens are generated, which
can now happen if use_default_badwordids is False and the first
token generated is EOS.
2023-08-29 23:08:51 -07:00
Llama
36f53cc915 Merge pull request #69 from pi6am/merge/united-exllama
Merge branch henk717/united into exllama
2023-08-29 19:54:14 -07:00
Llama
b5b0e3faea Merge branch 'henk717/united' into merge/united-exllama 2023-08-29 19:48:44 -07:00
Henk
49fa63052f Allow EOS unbanning 2023-08-29 20:51:09 +02:00
Henk
d77acf17eb Transformers 4.32.1 2023-08-28 22:55:33 +02:00
Henk
ec4cecf2f8 Restore Horde UI for local users 2023-08-28 21:46:30 +02:00
Henk
c29a5019b3 Disable Horde UI due to lockups 2023-08-28 20:02:29 +02:00
Llama
6e64763b63 Merge pull request #68 from pi6am/merge/united-exllama
Merge henk717/united into exllama
2023-08-28 09:57:35 -07:00
Llama
2c48e05f7c Add exllama dependency back to requirements. 2023-08-28 09:52:31 -07:00
Llama
6151cbd053 Merge branch 'united' into merge/united-exllama 2023-08-28 09:32:19 -07:00
henk717
030ca26aa3 Merge pull request #445 from db0/scribe
fix: Avoid conflicting args
2023-08-28 18:04:15 +02:00
db0
b20b910568 remove debug 2023-08-28 17:49:34 +02:00
db0
31161409f6 fix argsv placement 2023-08-28 17:47:52 +02:00
db0
25b78abcd2 Merge branch 'united' into scribe 2023-08-28 17:45:49 +02:00
Llama
5229987ab7 Merge pull request #66 from pi6am/feat/exllama-config
Modify exllama to load unrenamed gptq quantized models
2023-08-28 00:09:50 -07:00
Llama
554af7b175 Modify exllama to load unrenamed gptq quantized models
Read config.json and enable exllama loading if the model has a
`quantization_config` with `quant_methdod` of `gptq`. Note that this
implementation is limited and only supports model.safetensors.
That said, this supports loading popular gptq quantized models
without renaming or symlinking the model file.
2023-08-27 23:56:02 -07:00
Llama
812df5ea56 Merge pull request #65 from pi6am/feat/exllama-badwords
Add the eos token to exllama bad words.
2023-08-27 17:03:25 -07:00
Llama
08ff7c138c Add the eos token to exllama bad words.
The bos token was already hardcoded as a bad word id.
Store badwords in a list and iterate over them during generation.
Add the Llama eos token to the list of bad words.
Also support "single line mode", which adds newline (13) to badwords.
2023-08-27 16:34:52 -07:00
Henk
3e0b8279f2 Rename GPTQ loading 2023-08-27 20:51:14 +02:00
Llama
0d150e412e Merge pull request #64 from pi6am/fix/multinomial-workaround
Resample to work around a bug in torch.multinomial
2023-08-26 22:42:21 -07:00
Llama
b7e38b4757 Resample to work around a bug in torch.multinomial
There is a bug in PyTorch 2.0.1 that allows torch.multinomial to
sometimes choose elements that have zero probability. Since
this is uncommon we can continue to use torch.multinomial as
long as we verify that the results are valid. If they aren't,
try again until the probability of each selected token is positive.
2023-08-26 22:26:26 -07:00
Henk
290f2ce05e CPU only warning 2023-08-26 00:03:28 +02:00