Skip to content
Vercel April 2026 security incident

Mistral Small

mistral/mistral-small

Mistral Small (September 1, 2024) is a 22B mid-tier model with improved reasoning, alignment, and code at about 80% lower list pricing than its predecessor in Mistral AI's September 2024 table.

Tool UseVision (Image)
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'mistral/mistral-small',
prompt: 'Why is the sky blue?'
})

What To Consider When Choosing a Provider

  • Zero Data Retention

    AI Gateway supports Zero Data Retention for this model via direct gateway requests (BYOK is not included). To configure this, check the documentation.

    Authentication

    AI Gateway authenticates requests using an API key or OIDC token. You do not need to manage provider credentials directly.

At 22B parameters, Mistral Small sits between Mistral AI NeMo 12B and Mistral AI Large 2 as a mid-point in the lineup, per Mistral AI's September 2024 post.

When to Use Mistral Small

Best For

  • Translation and summarization at scale:

    High-volume translation, summarization, and sentiment analysis workloads

  • 22B mid-tier production:

    Workloads that need competitive per-token pricing

  • Mid-point between NeMo and Large:

    Applications where this capability fit is right

  • Cost optimization:

    For workloads previously using larger Mistral AI models

Consider Alternatives When

  • Higher reasoning capability:

    For workloads that need more depth (consider Mistral AI Large 3 or Magistral)

  • Tekken tokenizer compression:

    You need multilingual compression advantages (consider Mistral AI NeMo)

  • Vision input:

    Your workload needs image input (consider Pixtral models)

Conclusion

Mistral Small's September 1, 2024 update cut list pricing and improved alignment, reasoning, and code quality in Mistral AI's post. An 80% cost reduction shipped alongside those capability changes. Mistral Small remains a fit for teams that want a 22B mid-tier model through AI Gateway.

FAQ

This page lists the current rates. Multiple providers can serve Mistral Small, so AI Gateway surfaces live pricing rather than a single fixed figure.

22 billion parameters.

Between Mistral AI NeMo (12B) and Mistral AI Large 2, described as a convenient mid-point for enterprise use cases.

Improved human alignment, stronger reasoning, and better code generation compared to the prior Mistral Small version.

Translation, summarization, and sentiment analysis. Mistral AI positions Mistral Small for tasks that don't need Mistral AI Large-scale breadth.

Mistral Small is larger (22B vs 12B), more capable, and more expensive. Mistral AI NeMo has the Tekken tokenizer's compression advantages for code and non-Latin scripts. The choice depends on capability requirements and token efficiency needs.