Zen Embedding 0.6B GGUF

High-performance text embedding model from the Zen model family optimized for efficient inference.

Downloads

Source URL
HuggingFace hf download zenlm/zen-embedding-0.6B-GGUF
Direct https://download.hanzo.ai/llm-models/zen-embedding-0.6B-Q8_0.gguf

Features

  • 100+ language support
  • Optimized for semantic search and retrieval
  • GGUF format for efficient CPU/GPU inference
  • Q8_0 quantization (639 MB)

Usage

Works with llama.cpp and compatible inference engines.

License

Apache 2.0

Downloads last month
103
GGUF
Model size
0.6B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for zenlm/zen-embedding-0.6B-GGUF

Unable to build the model tree, the base model loops to the model itself. Learn more.