Codestral Embed
Codestral Embed is Mistral AI's first embedding model specialized for code, outperforming general-purpose and competing code embedding models on real-world retrieval benchmarks.
import { embed } from 'ai';
const result = await embed({ model: 'mistral/codestral-embed', value: 'Sunny day at the beach',})What To Consider When Choosing a Provider
- Configuration: Codestral Embed supports variable embedding dimensions, letting you tune the size-versus-quality tradeoff to match your vector store's cost and latency constraints.
- Zero Data Retention: AI Gateway supports Zero Data Retention for this model via direct gateway requests (BYOK is not included). To configure this, check the documentation.
- Authentication: AI Gateway authenticates requests using an API key or OIDC token. You do not need to manage provider credentials directly.
When to Use Codestral Embed
Best For
- RAG pipelines for coding agents: Building pipelines that retrieve relevant code snippets
- Semantic code search: Large repositories where keyword search is insufficient
- Duplicate code detection: Similarity analysis across codebases using Codestral Embed's 85% average on code retrieval benchmarks
- Code clustering: For analytics, refactoring identification, or repository organization
- Large-scale indexing pipelines: Workloads where embedding cost is a primary concern at millions of documents
Consider Alternatives When
- General text documentation: You need to embed prose rather than source code (consider Mistral AI Embed)
- Natural language retrieval: Your workload is primarily over descriptions of code rather than code itself
- Generation alongside embedding: You require both capabilities in a single model call
Conclusion
Codestral Embed fills a gap that general-purpose embedding models leave open. Code has structural patterns, syntax, and semantics that differ from prose, and a model trained on real-world code data retrieves it more accurately. For teams building coding agents, code search, or repository analytics, Codestral Embed is more precise than adapting a text embedding model.
Frequently Asked Questions
What makes Codestral Embed different from a general text embedding model?
Codestral Embed was trained on real-world code data and optimized for code retrieval tasks. These tasks involve matching function signatures, logic patterns, and structural similarities that general text models don't capture well.
What is the context window for Codestral Embed?
0 tokens. For files larger than this, chunk at 3,000 characters with 1,000-character overlap.
What embedding dimensions does Codestral Embed support?
Variable dimensions with ordered relevance: you can keep the first n dimensions for a quality-versus-cost tradeoff. Codestral Embed scores well on retrieval at 256 dimensions with int8 precision in published benchmarks.
What is the pricing for Codestral Embed?
Check the pricing panel on this page for today's numbers. AI Gateway tracks rates across every provider that serves Codestral Embed.
Can Codestral Embed be used for duplicate code detection?
Yes. Semantic similarity via embeddings is an effective approach for identifying duplicate or near-duplicate code patterns that differ syntactically but are logically equivalent.
How does Codestral Embed compare to Mistral AI Embed?
Mistral AI Embed is a general-purpose text embedding model. Codestral Embed was trained specifically for code and outperforms general models on code retrieval benchmarks. Use Codestral Embed when your corpus is source code; use Mistral AI Embed for natural language documents.