Mistral 3 family of models released
Recorded: Dec. 3, 2025, 3:04 a.m.
| Original | Summarized |
Introducing Mistral 3 | Mistral AIProductsSolutionsResearchResourcesPricingCompanyTry AI Studio Talk to sales Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Mistral 3Introducing Mistral 3The next generation of open multimodal and multilingual AIToday, we announce Mistral 3, the next generation of Mistral models. Mistral 3 includes three state-of-the-art small, dense models (14B, 8B, and 3B) and Mistral Large 3 – our most capable model to date – a sparse mixture-of-experts trained with 41B active and 675B total parameters. All models are released under the Apache 2.0 license. Open-sourcing our models in a variety of compressed formats empowers the developer community and puts AI in people’s hands through distributed intelligence. Mistral Large 3 is one of the best permissive open weight models in the world, trained from scratch on 3000 of NVIDIA’s H200 GPUs. Mistral Large 3 is Mistral’s first mixture-of-experts model since the seminal Mixtral series, and represents a substantial step forward in pretraining at Mistral. After post-training, the model achieves parity with the best instruction-tuned open-weight models on the market on general prompts, while also demonstrating image understanding and best-in-class performance on multilingual conversations (i.e., non-English/Chinese). We release both the base and instruction fine-tuned versions of Mistral Large 3 under the Apache 2.0 license, providing a strong foundation for further customization across the enterprise and developer communities. A reasoning version is coming soon! For edge and local use cases, we release the Ministral 3 series, available in three model sizes: 3B, 8B, and 14B parameters. Furthermore, for each model size, we release base, instruct, and reasoning variants to the community, each with image understanding capabilities, all under the Apache 2.0 license. When married with the models’ native multimodal and multilingual capabilities, the Ministral 3 family offers a model for all enterprise or developer needs. Frontier performance, open access: Achieve closed-source-level results with the transparency and control of open-source models. Multimodal and multilingual: Build applications that understand text, images, and complex logic across 40+ native languages. Scalable efficiency: From 3B to 675B parameters, choose the model that fits your needs, from edge devices to enterprise workflows. Agentic and adaptable: Deploy for coding, creative collaboration, document analysis, or tool-use workflows with precision. Next Steps Explore the model documentation: Ministral 3 3B-25-12 Ministral 3 8B-25-12 Ministral 3 14B-25-12 Mistral Large 3 Technical documentation for customers is available on our AI Governance Hub Start building: Ministral 3 and Large 3 on Hugging Face, or deploy via Mistral AI’s platform for instant API access and API pricing Customize for your needs: Need a tailored solution? Contact our team to explore fine-tuning or enterprise-grade training. Share your projects, questions, or breakthroughs with us: Twitter/X, Discord, or GitHub. We believe that the future of AI should be built on transparency, accessibility, and collective progress. With this release, we invite the world to explore, build, and innovate with us, unlocking new possibilities in reasoning, efficiency, and real-world applications. Build on AI Studio Talk to an expert Mistral AI © 2025Why MistralAbout usOur customersCareersContact usExploreAI solutionsPartnersResearchDocumentationBuildAI StudioLe ChatMistral CodeMistral ComputeLegalTerms of servicePrivacy policyPrivacy choicesData processing agreementLegal noticeBrandenMistral AI © 2025 |
The newly released Mistral 3 represents a significant advancement in open-source AI models, spearheaded by Mistral AI. The core of the offering includes three distinct model sizes – 3B, 8B, and 14B parameters – alongside the flagship Mistral Large 3 model, boasting 41B active and 675B total parameters. These models are released under the Apache 2.0 license, emphasizing open access and community development. Mistral Large 3 distinguishes itself with a mixture-of-experts (MoE) architecture designed to achieve performance comparable to state-of-the-art instruction-tuned models while demonstrating strong capabilities in multilingual conversations and image understanding, evidenced by its placement as number two in the OSS non-reasoning models category on the LMArena leaderboard. The development of Mistral 3 is underpinned by a strategic partnership with NVIDIA, leveraging NVIDIA Hopper GPUs to train the models, capitalizing on the high-bandwidth HBM3e memory. This collaboration, alongside work with vLLM and Red Hat, dramatically improves accessibility through optimized checkpoints in NVFP4 format, enabling efficient execution on systems like Blackwell NVL72 and leveraging technologies such as TensorRT-LLM and SGLang. A key innovation is the support for prefill/decode disaggregated serving and speculative decoding, facilitating high-throughput, long-context workloads. Furthermore, the models’ deployment extends to edge devices, including DGX Spark, RTX PCs, and Jetson devices. The Mistral 3 family caters to diverse needs, offering base, instruct, and reasoning variants with image understanding across all model sizes. The 3B, 8B, and 14B models are specifically designed to offer the best cost-to-performance ratio, with instruction-tuned variants achieving comparable, and often lower, token generation counts. The reasoning variants, particularly the 14B model, are optimized for accuracy, achieving 85% on AIME ‘25. Beyond the models themselves, Mistral AI provides custom model training services, adapting models to specific enterprise needs regarding domain-specific tasks, optimization of performance on proprietary datasets, and deployment in unique environments. This customization service offers a powerful tool for organizations seeking tailored AI solutions that deliver maximum impact. The availability of Mistral 3 extends to major AI platforms including Mistral AI Studio, Amazon Bedrock, Azure Foundry, Hugging Face, Modal, IBM WatsonX, OpenRouter, Fireworks, Unsloth AI, and Together AI, and upcoming integrations with NVIDIA NIM and AWS SageMaker. By offering a comprehensive portfolio of deployment options, Mistral AI aims to democratize access to advanced AI capabilities. The overarching vision is to establish a future of AI built on transparency, accessibility, and collective progress, inviting the world to explore, build, and innovate. |