Skip to content

Llama 3.2 1B Instruct

Llama 3.2 1B Instruct is Meta's smallest openly available model, with a context window of 128K tokens. It delivers text generation, summarization, and tool calling with minimal memory and compute requirements.

index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'meta/llama-3.2-1b',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • What is the context window for a 1B model?

    128K tokens, covering summarization, instruction following, rewriting, and tool use.

  • Does Llama 3.2 1B Instruct support tool calling?

    Yes. Tool calling is one of the trained capabilities included alongside summarization, instruction following, and rewriting, making it suitable for lightweight agentic applications that need to invoke external actions.

  • Is there a companion safety model for the 1B?

    Yes. Llama Guard 3 1B was built from this checkpoint for content moderation in memory-constrained environments.

  • How does Llama 3.2 1B Instruct compare to Gemma at similar scales?

    Llama 3.2 1B Instruct matches Gemma on summarization, instruction following, and tool use. The 3B beats Gemma 2 2.6B on those same tasks by a wider margin.