A diffusion-based reasoning LLM that generates text via parallel refinement (not token-by-token), delivering real-time latency with ~1k tokens/sec plus 128K context and built-in tool/JSON support.
import { streamText } from 'ai'
const result = streamText({ model: 'inception/mercury-2', prompt: 'Why is the sky blue?'})Try out Mercury 2 by Inception. Usage is billed to your team at API rates. Free users get $5 of credits every 30 days, and you are considered a free user if you haven't made a payment.
Chat with
Powered by AI Gateway
| Provider |
|---|
| Model |
|---|