Mercury 2

inception/mercury-2

A diffusion-based reasoning LLM that generates text via parallel refinement (not token-by-token), delivering real-time latency with ~1k tokens/sec plus 128K context and built-in tool/JSON support.

Tool UseReasoning
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'inception/mercury-2',
prompt: 'Why is the sky blue?'
})