Very Large GGUFs
					Collection
				
GGUF quantized versions of very large models - over 100B parameters
					β’ 
				51 items
				β’ 
				Updated
					
				β’
					
					5
'Make knowledge free for everyone'
Experimental, based on: https://github.com/ggml-org/llama.cpp/pull/16831
Quantized version of: MiniMaxAI/MiniMax-M2
Hexagon test 0 Shot with Q4_K_M
| Model | Perplexity (PPL) | Β± Error | 
|---|---|---|
| Minimax IQ1_M | 11.8447 | 0.21162 | 
| Minimax IQ2_XXS | 9.1211 | 0.15936 | 
| Minimax Q2_K | 7.6598 | 0.13421 | 
| Minimax Q3_K | 6.7349 | 0.11651 | 
| Minimax Q4_K_M | 6.5625 | 0.11302 | 
1-bit
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
MiniMaxAI/MiniMax-M2