Announced December 2, 2025, Mistral Large 3 marks Mistral AI's return to the mixture-of-experts (MoE) architecture that defined their earlier Mixtral series, now at a larger scale. With 675B total parameters and 41B active per forward pass, Mistral Large 3 represents a substantial architectural evolution from the dense models that preceded it in the Large lineage.
The sparse MoE design lets Mistral Large 3 maintain inference efficiency comparable to a smaller dense model while drawing on a large total parameter pool for complex tasks. This architecture offers a tradeoff between capability and inference cost.
Through AI Gateway, you can access Mistral Large 3 without separate Mistral AI API credentials. Built-in observability gives you cost and latency visibility across every request.