Trinity Large Preview
Trinity Large Preview is a 400B-parameter sparse mixture-of-experts model from Arcee AI that activates 13B parameters per forward pass, targeting math, coding, and multi-step agent workloads across a context window of 131K tokens.
import { streamText } from 'ai'
const result = streamText({ model: 'arcee-ai/trinity-large-preview', prompt: 'Why is the sky blue?'})Frequently Asked Questions
What kinds of tasks is Trinity Large Preview explicitly designed for?
Math, coding, and complex multi-step agent workflows. The release notes also emphasize efficient extended multi-turn use with high inference throughput.
Why is this model labeled a "preview" release?
It ships before general availability. Expect changes while Arcee AI finalizes production behavior, pricing, and versioning.
How does Trinity Large Preview differ from Trinity Mini?
Trinity Large Preview is a 400B-parameter MoE with 13B active per forward pass, targeting deep reasoning on math, coding, and multi-step agent tasks. Trinity Mini is a 26B-parameter MoE with 3B active parameters, tuned for lean inference and volume. Pick Mini when cost per token is the binding constraint; pick this model when you need the larger parameter space.
Do I need an Arcee AI account to use this model through AI Gateway?
No. Use your AI Gateway API key or an OIDC token. You don't need a separate provider account.
Can I use this model with the AI SDK?
Yes. Set
modeltoarcee-ai/trinity-large-previewin the AI SDK'sstreamTextorgenerateTextcall. AI Gateway also exposes OpenAI Chat Completions, OpenAI Responses, Anthropic Messages, and OpenResponses-compatible interfaces.Does AI Gateway provide observability for requests to Trinity Large Preview?
Yes. Token usage, latency, and cost show in your AI Gateway dashboard for each request without extra instrumentation.