Skip to content

Trinity Large Preview

Trinity Large Preview is a 400B-parameter sparse mixture-of-experts model from Arcee AI that activates 13B parameters per forward pass, targeting math, coding, and multi-step agent workloads across a context window of 131K tokens.

Tool Use
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'arcee-ai/trinity-large-preview',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • What kinds of tasks is Trinity Large Preview explicitly designed for?

    Math, coding, and complex multi-step agent workflows. The release notes also emphasize efficient extended multi-turn use with high inference throughput.

  • Why is this model labeled a "preview" release?

    It ships before general availability. Expect changes while Arcee AI finalizes production behavior, pricing, and versioning.

  • How does Trinity Large Preview differ from Trinity Mini?

    Trinity Large Preview is a 400B-parameter MoE with 13B active per forward pass, targeting deep reasoning on math, coding, and multi-step agent tasks. Trinity Mini is a 26B-parameter MoE with 3B active parameters, tuned for lean inference and volume. Pick Mini when cost per token is the binding constraint; pick this model when you need the larger parameter space.

  • Do I need an Arcee AI account to use this model through AI Gateway?

    No. Use your AI Gateway API key or an OIDC token. You don't need a separate provider account.

  • Can I use this model with the AI SDK?

    Yes. Set model to arcee-ai/trinity-large-preview in the AI SDK's streamText or generateText call. AI Gateway also exposes OpenAI Chat Completions, OpenAI Responses, Anthropic Messages, and OpenResponses-compatible interfaces.

  • Does AI Gateway provide observability for requests to Trinity Large Preview?

    Yes. Token usage, latency, and cost show in your AI Gateway dashboard for each request without extra instrumentation.