|  | 6c20d0d657 | Nevermind, dividing by 4 is actually correct... | 2022-03-19 00:55:04 -04:00 |  | 
			
				
					|  | f16b61ec77 | Should divide NeoX replicated parameters by 8 (not by 4) Also, suppresses the PyTorch 1.11 warning about transposing tensors with
ndim != 2 in the new code | 2022-03-19 00:48:33 -04:00 |  | 
			
				
					|  | c2c139e940 | Change default PE type for NeoX to neox_rotary | 2022-03-19 00:26:04 -04:00 |  | 
			
				
					|  | 85a4959efa | Merge branch 'united' into neox | 2022-03-18 11:19:03 -04:00 |  | 
			
				
					| 
							
							
								 henk717 | f581fe89cb | Torch version changes | 2022-03-17 21:11:36 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 9e9c1c3fe0 | Merge pull request #100 from VE-FORBRYDERNE/patch Add PyTorch 1.11 support for lazy loader | 2022-03-17 21:06:38 +01:00 |  | 
			
				
					|  | c444260eac | Silence PyTorch warning about transposing tensors with dimension != 2 | 2022-03-17 15:16:56 -04:00 |  | 
			
				
					|  | ef21ab9c91 | PyTorch 1.9 lazy loader compatibility bugfix | 2022-03-17 14:10:51 -04:00 |  | 
			
				
					|  | eaf190469d | Add PyTorch 1.11 support for lazy loader | 2022-03-17 12:51:41 -04:00 |  | 
			
				
					| 
							
							
								 henk717 | 9235754eb9 | Dependency Fixes | 2022-03-17 00:35:59 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | a3e5e052b3 | Newer umamba + slope tweak | 2022-03-16 18:34:02 +01:00 |  | 
			
				
					|  | 95c4251db9 | Print two newlines before loading HF models | 2022-03-15 13:58:53 -04:00 |  | 
			
				
					|  | 9e2848e48f | Show parameter count when loading GPT-NeoX in Colab TPU instance | 2022-03-15 13:55:27 -04:00 |  | 
			
				
					|  | 9dc48b15f0 | Add custom badwords and pad token ID for GPT-NeoX | 2022-03-14 23:31:49 -04:00 |  | 
			
				
					|  | 88f247d535 | GPT-NeoX-20B support in Colab TPU instances | 2022-03-14 23:14:20 -04:00 |  | 
			
				
					| 
							
							
								 henk717 | 4892556059 | Model saving for colab mode | 2022-03-13 11:22:44 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | ccadeabbde | Merge pull request #99 from VE-FORBRYDERNE/model-patch Model loading fixes | 2022-03-13 11:10:15 +01:00 |  | 
			
				
					|  | 2b8c46338e | Change current working directory to KoboldAI folder | 2022-03-13 01:22:11 -05:00 |  | 
			
				
					|  | 48d07adb54 | Also fallback to generic GPT2 tokenizer in Colab TPU instances | 2022-03-12 23:19:35 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | d29a629320 | Merge pull request #98 from ebolam/united Fix for retry | 2022-03-12 16:52:07 +01:00 |  | 
			
				
					| 
							
							
								 ebolam | 45eed78d21 | Merge branch 'united' of https://github.com/ebolam/KoboldAI into united | 2022-03-12 10:33:01 -05:00 |  | 
			
				
					| 
							
							
								 ebolam | b55e5a8e0b | Retry Bug Fix | 2022-03-12 10:32:27 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | 2e1b3c82f9 | Merge pull request #97 from ebolam/united Fix for retry causing issues for future redo actions | 2022-03-11 17:41:49 +01:00 |  | 
			
				
					| 
							
							
								 ebolam | ae854bab3d | Fix for retry causing issues for future redo actions | 2022-03-11 11:40:55 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | 2c66461c14 | Merge pull request #96 from VE-FORBRYDERNE/dlpack Use DLPack to convert PyTorch tensors to JAX arrays | 2022-03-10 22:00:38 +01:00 |  | 
			
				
					|  | a99eb8724d | Use DLPack to convert PyTorch tensors to JAX arrays | 2022-03-10 15:12:42 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | b02d5e8696 | Allows missing model_config again | 2022-03-10 19:59:10 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 172a548fa1 | Fallback to generic GPT2 Tokenizer | 2022-03-10 19:52:15 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 68281184bf | Remove Lowmem from TPU | 2022-03-09 19:21:15 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 9dee9b5c6d | Ignore incorrect problems | 2022-03-09 12:03:37 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | a28e553412 | Remove unused gettokenids | 2022-03-09 11:59:33 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 7434c9221b | Expand OAI Setting Compatibility | 2022-03-07 08:56:47 +01:00 |  | 
			
				
					| 
							
							
								 ebolam | f6c95f18fa | Fix for Redo (#94) * Corrected redo to skip blank steps (blank from "deleting" the chunk with the edit function)
* Removed debug code | 2022-03-06 23:18:14 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | f857696224 | OAI ConfigName Bugfix | 2022-03-06 20:18:42 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 3ddc9647eb | Basic GooseAI Support | 2022-03-06 20:10:30 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | f1b0ea711e | Merge branch 'KoboldAI:main' into united | 2022-03-06 19:02:59 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 932aabc2f3 | Merge pull request #103 from henk717/main Modern ROCm Docker | 2022-03-06 19:02:38 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 4332074c89 | Modern ROCm Docker Brings the ROCm container up to a modern standard in line with the CUDA docker. | 2022-03-06 19:01:25 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 4835192041 | Load TK on demand | 2022-03-06 14:12:01 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | daea4b8d15 | Fix Breakmodel RAM Regression | 2022-03-06 08:26:50 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 105d3831b5 | Lazy Load Float32 for CPU | 2022-03-06 07:56:04 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 77cc2ee789 | Merge pull request #93 from VE-FORBRYDERNE/lazy-loader Lazy loader | 2022-03-05 20:32:31 +01:00 |  | 
			
				
					|  | 373f7b9bd5 | Don't convert tensors to float16 if using CPU-only mode | 2022-03-05 14:30:26 -05:00 |  | 
			
				
					|  | 579e85820c | Resolve merge conflict | 2022-03-05 14:13:56 -05:00 |  | 
			
				
					|  | 2e19ea1bb6 | Auto detect if we're in a Colab TPU instance | 2022-03-05 14:07:23 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | 3a5793c815 | No longer uses --colab_tpu | 2022-03-05 19:58:24 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 935c7e5786 | Improved TPU support | 2022-03-05 19:47:51 +01:00 |  | 
			
				
					| 
							
							
								 henk717 | 6f2febb142 | Merge pull request #92 from ebolam/united Hopefully Last Redo Fix | 2022-03-05 19:26:15 +01:00 |  | 
			
				
					| 
							
							
								 ebolam | 4a8d7f5e0b | Merge branch 'henk717:united' into united | 2022-03-05 13:25:10 -05:00 |  | 
			
				
					| 
							
							
								 henk717 | c20435855b | Merge pull request #91 from VE-FORBRYDERNE/transformers-version-check Put the XGLM embedding patch behind a version check | 2022-03-05 19:03:00 +01:00 |  |