Skip to content

MiMo V2 Flash

MiMo V2 Flash is Xiaomi's MiMo v2 Flash MoE reasoning model with 309B total parameters and 15B active per forward pass, using hybrid attention and multi-token prediction for inference efficiency. It supports a context window of 262.1K tokens at $0.1 per million input tokens and $0.3 per million output tokens.

ReasoningTool Use
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'xiaomi/mimo-v2-flash',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does MiMo V2 Flash score well with a small active parameter count?

    MoE routes tokens to expert blocks and only activates part of the weights each step. That keeps compute low while the full weight count still holds broad knowledge.

  • What is hybrid sliding window attention?

    It mixes sliding-window and global attention on a fixed schedule with a 128-token window. MiMo V2 Flash uses much smaller KV caches than full attention, which helps on a context of 262.1K tokens.

  • How does the multi-token prediction module work?

    It adds a small MTP block per layer so the stack can propose several future tokens and verify them in fewer full steps, which raises output tokens per second during inference.

  • How do I authenticate requests to MiMo V2 Flash through AI Gateway?

    Add your API key in AI Gateway project settings. Use xiaomi/mimo-v2-flash in API calls. AI Gateway routes, retries, and fails over across novita, chutes, xiaomi.

  • What does MiMo V2 Flash cost?

    Check the pricing panel on this page for today's numbers. AI Gateway tracks rates across every provider that serves MiMo V2 Flash.

  • How does MiMo V2 Flash compare to DeepSeek-V3?

    DeepSeek-V3 uses a larger active parameter count from a larger total than MiMo V2 Flash. Compare published tables on each vendor's page; both are MoE stacks with different size and training choices.