MiMo-VL-7B Quantize
					Collection
				
XiaomiMiMo/MiMo-VL-7B Quantized Model
					• 
				5 items
				• 
				Updated
					
				
This is a 4-bit quantized version of XiaomiMiMo/MiMo-VL-7B-SFT,
using the BitsAndBytes library.
Quantization reduces memory usage and makes it possible to run this model on consumer GPUs
(≤ 12 GB VRAM), at the cost of a small reduction in generation quality.
fp4)  safetensorsUnable to build the model tree, the base model loops to the model itself. Learn more.