Released April 17, 2024, Mixtral MoE 8x22B Instruct is Mistral AI's largest open-weight model. Mixtral MoE 8x22B Instruct uses a sparse Mixture-of-Experts (SMoE) architecture with eight expert networks totaling 141B parameters, of which 39B activate per forward pass. This architecture lets Mixtral MoE 8x22B Instruct outperform dense 70B models in inference speed while drawing on a larger total parameter space for complex reasoning.
The instruct variant adds native function calling and a constrained output mode on La Plateforme, both important for structured, agentic applications. Multilingual fluency covers English, French, Italian, German, and Spanish. Mixtral MoE 8x22B Instruct scores 90.8% on GSM8K (maj@8) and 44.6% on the MATH benchmark (maj@4).
Released under Apache 2.0, Mistral AI's most permissive license, Mixtral MoE 8x22B Instruct can be used commercially, modified, and redistributed without restriction.