Zen Embedding 0.6B GGUF
High-performance text embedding model based on Qwen3-Embedding, optimized for efficient inference.
Downloads
| Source | URL |
|---|---|
| HuggingFace | hf download zenlm/zen-embedding-0.6B-GGUF |
| Direct | https://download.hanzo.ai/llm-models/zen-embedding-0.6B-Q8_0.gguf |
Features
- 100+ language support
- Optimized for semantic search and retrieval
- GGUF format for efficient CPU/GPU inference
- Q8_0 quantization (639 MB)
Usage
Works with llama.cpp and compatible inference engines.
License
Apache 2.0 (inherited from Qwen3-Embedding)
- Downloads last month
- 44
Hardware compatibility
Log In
to view the estimation
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support