Qwen3 Embedding 0.6B
Qwen3 Embedding 0.6B is a compact 0.6-billion-parameter text embedding model with context of 32.8K tokens and 1024-dimensional vectors, built for cost-efficient semantic search and multilingual retrieval across more than 100 languages.
import { embed } from 'ai';
const result = await embed({ model: 'alibaba/qwen3-embedding-0.6b', value: 'Sunny day at the beach',})Frequently Asked Questions
What vector dimensions does Qwen3 Embedding 0.6B produce, and can I reduce them?
The model outputs 1024-dimensional vectors by default. Via Matryoshka Representation Learning (MRL), you can truncate these to a shorter prefix to reduce storage and query cost, though very short truncations may reduce retrieval quality.
How many languages does Qwen3 Embedding 0.6B cover?
The model supports over 100 natural languages as well as multiple programming languages, enabling cross-lingual and code-retrieval tasks within a single embedding space.
What is the maximum input length for a single embedding call?
The context window is 32.8K tokens. Inputs longer than this must be chunked before embedding.
How does this model compare to the 4B and 8B variants?
All three variants share the same context of 32.8K tokens and MRL support. The 0.6B model uses a 1024-dimensional output and 28 layers, making it the fastest and least expensive option; the larger variants produce higher-dimensional vectors that tend to perform better on precision-sensitive benchmarks.
Can I use custom task instructions with this model?
Yes. The model supports user-defined instruction prefixes on queries, which shift the embedding space to match specific retrieval intents, for example, distinguishing document-retrieval queries from code-search queries.
Is this model suitable for production RAG pipelines?
Yes. The compact vector size and multilingual coverage make it a natural fit for RAG pipelines where you embed a large knowledge base once and query it repeatedly, especially when cost per embedded token is a primary concern.