This model is a quantized version of Qwen3-32B converted to AWQ Q4 format using the mlx library for efficient inference. It retains the core capabilities of Qwen3 while optimizing for resource constraints.
Chat template
Files info
Base model