Zen Embedding 0.6B GGUF
High-performance text embedding model from the Zen model family optimized for efficient inference.
Downloads
| Source | URL |
|---|---|
| HuggingFace | hf download zenlm/zen-embedding-0.6B-GGUF |
| Direct | https://download.hanzo.ai/llm-models/zen-embedding-0.6B-Q8_0.gguf |
Features
- 100+ language support
- Optimized for semantic search and retrieval
- GGUF format for efficient CPU/GPU inference
- Q8_0 quantization (639 MB)
Usage
Works with llama.cpp and compatible inference engines.
License
Apache 2.0
- Downloads last month
- 103
Hardware compatibility
Log In to add your hardware
8-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for zenlm/zen-embedding-0.6B-GGUF
Unable to build the model tree, the base model loops to the model itself. Learn more.