Custom GGUF quants of Metaβs Llama-3.2-Instruct's finetunes, where the Output Tensors are quantized to Q8_0 or F32 and the Embeddings are kept @F32
Joseph
Joseph717171
AI & ML interests
None yet
Recent Activity
liked
a model
about 13 hours ago
bkhmsi/micro-olmo-1b
liked
a model
about 14 hours ago
bkhmsi/micro-llama-3b
liked
a model
3 days ago
nvidia/DLER-Llama-Nemotron-8B-Merge-Research