Skip to content

Nova Micro

Nova Micro delivers text-only inference at high throughput with per-token pricing below multimodal Nova models in the same generation, purpose-built for latency-sensitive applications at scale.

index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'amazon/nova-micro',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • Why would I choose Nova Micro over Nova Lite for text tasks?

    Nova Micro is priced below Nova Lite's multimodal rate and is tuned for speed on pure text. If you never send images or video, Micro is usually the cheaper fit.

  • Can Nova Micro handle long documents?

    Keep the prompt within 128K tokens. If you exceed that, split the document, summarize in chunks, or switch to Nova 2 Lite for a 1M-token window.

  • Is Nova Micro good for structured output like JSON?

    Yes. It follows instructions well for classification, tagging, and structured extraction. Its speed makes it especially efficient for pipelines that process many short requests.

  • What is the maximum output length?

    Nova Micro generates up to 8.2K tokens per response.

  • How does Nova Micro perform on reasoning benchmarks?

    Nova Micro isn't designed for complex reasoning. It excels at speed and cost efficiency for routine language tasks. For reasoning-heavy workloads, consider Nova 2 Lite or Nova Pro.

  • Do I need separate AWS credentials?

    No. AI Gateway handles authentication with Amazon Bedrock. You only need a gateway API key or OIDC token.