Magistral-Small-2509-IQ4_KS GGUF
A high-quality quantized version of mistralai/Magistral-Small-2509 using ik_llama.cpp with importance matrix optimization and custom tensor-specific quantization.
Model Details
- Base Model: mistralai/Magistral-Small-2509
- Quantization Method: Custom IQ4_KS recipe with importance matrix
- Target VRAM: 12GB for full offload
- Quantization Tool: ik_llama.cpp with imatrix calibration
Benchmark Results
Perplexity (wiki.test.raw, ctx=512)
| Metric | Value |
|---|---|
| Final PPL | 5.6245 ยฑ 0.03312 |
| Evaluation Chunks | 594 |
| Evaluation Tokens | 304,128 |
| Context Size | 512 |
- Downloads last month
- 32
Hardware compatibility
Log In
to view the estimation
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for nimishchaudhari/Magistral-Small-2509-IQ4_KS-GGUF
Base model
mistralai/Mistral-Small-3.1-24B-Base-2503
Finetuned
mistralai/Magistral-Small-2509