Skip to content

Qwen3 Embedding 8B

alibaba/qwen3-embedding-8b

Qwen3 Embedding 8B is Alibaba's 8B-tier text embedding model in the Qwen3 Embedding line, producing 4096-dimensional vectors and ranking first on the MTEB multilingual leaderboard at release, built for demanding cross-lingual retrieval and RAG workloads.

index.ts
import { embed } from 'ai';
const result = await embed({
model: 'alibaba/qwen3-embedding-8b',
value: 'Sunny day at the beach',
})

What To Consider When Choosing a Provider

  • Zero Data Retention

    AI Gateway supports Zero Data Retention for this model via direct gateway requests (BYOK is not included). To configure this, check the documentation.

    Authentication

    AI Gateway authenticates requests using an API key or OIDC token. You do not need to manage provider credentials directly.

For workloads indexing sensitive documents, confirm that your chosen provider's data-residency region aligns with your compliance requirements before routing production traffic.

When to Use Qwen3 Embedding 8B

Best For

  • MTEB-driven production retrieval:

    Systems where MTEB multilingual scores from the model's release evaluations are the primary criterion

  • Long-document RAG:

    Pipelines that benefit from context of 32.8K tokens and 4096-dimensional representations to preserve semantic detail

  • Cross-lingual knowledge bases:

    Indexes spanning many languages and programming environments

  • Research and evaluation workloads:

    MTEB-adjacent benchmarks serve as a proxy for real retrieval performance

Consider Alternatives When

  • Tight embedding cost budgets:

    Per-token cost dominates and slightly lower accuracy is acceptable, the 0.6B or 4B variants may provide sufficient quality

  • Memory-constrained deployments:

    Environments with strict memory limits make a fully loaded 8B model impractical

  • Generative output required:

    This model produces embeddings only; use a generative model when you need text output

Conclusion

Qwen3 Embedding 8B is the right tool when retrieval accuracy across languages and domains can't be compromised. Its first-place standing on the MTEB multilingual leaderboard at release and its 4096-dimensional output make it a strong foundation for enterprise-grade semantic search and RAG systems willing to invest in embedding quality.

FAQ

As of June 5, 2025, the model scored 70.58 on the MTEB multilingual leaderboard, placing it first among publicly evaluated embedding models at that date.

The default output is 4096-dimensional. Using Matryoshka Representation Learning (MRL), you can truncate these vectors to shorter prefix lengths for use cases where storage or query latency is constrained.

Both the 8B and 4B models use 36 transformer layers, but the 8B model has wider layers with more parameters per layer. It produces 4096-dimensional vectors compared to 2560 for the 4B. This additional resolution typically improves performance on dense retrieval and clustering tasks, particularly for technical and multilingual corpora.

Yes. Each individual text input can be up to 32.8K tokens. If a document exceeds this limit it must be split into chunks before embedding.

You can prepend a task-specific instruction to your query (e.g., describing the retrieval goal) to shift the embedding space toward that intent. This is particularly effective for asymmetric retrieval where query phrasing differs significantly from document phrasing.

Yes. The Qwen3 Embedding training explicitly covers multiple programming languages, so a unified vector index mixing code files and documentation is a supported pattern.