Qwen3-8B-4bit-SINQ / quantization_config.json
avinashhm's picture
Push quantized Qwen3-8B 4-bit SINQ model with README
d2627ab verified
{
"model_name": "Qwen/Qwen3-8B",
"nbits": 4,
"group_size": 128,
"tiling_mode": "1D",
"method": "sinq",
"compute_dtype": "torch.bfloat16",
"timestamp": "2025-10-06T17:31:58.512995"
}