Mistral Medium 3.1
Mistral Medium 3.1 is the latest frontier-class multimodal foundation model released in August 2025, designed to deliver advanced reasoning, coding, and multimodal capabilities while dramatically reducing deployment complexity and costs. It builds on the highly efficient architecture of Mistral Medium 3, renowned for offering state-of-the-art performance at up to 8-times lower cost than leading large models, enhancing tone consistency, responsiveness, and accuracy across diverse tasks and modalities. The model supports deployment across hybrid environments, on-premises systems, and virtual private clouds, and it achieves competitive performance relative to high-end models such as Claude Sonnet 3.7, Llama 4 Maverick, and Cohere Command A. Ideal for professional and enterprise use cases, Mistral Medium 3.1 excels in coding, STEM reasoning, language understanding, and multimodal comprehension, while maintaining broad compatibility with custom workflows and infrastructure.
Learn more
Mistral AI
Mistral AI is a pioneering artificial intelligence startup specializing in open-source generative AI. The company offers a range of customizable, enterprise-grade AI solutions deployable across various platforms, including on-premises, cloud, edge, and devices. Flagship products include "Le Chat," a multilingual AI assistant designed to enhance productivity in both personal and professional contexts, and "La Plateforme," a developer platform that enables the creation and deployment of AI-powered applications. Committed to transparency and innovation, Mistral AI positions itself as a leading independent AI lab, contributing significantly to open-source AI and policy development.
Learn more
Mistral NeMo
Mistral NeMo, our new best small model. A state-of-the-art 12B model with 128k context length, and released under the Apache 2.0 license. Mistral NeMo is a 12B model built in collaboration with NVIDIA. Mistral NeMo offers a large context window of up to 128k tokens. Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. As it relies on standard architecture, Mistral NeMo is easy to use and a drop-in replacement in any system using Mistral 7B. We have released pre-trained base and instruction-tuned checkpoints under the Apache 2.0 license to promote adoption for researchers and enterprises. Mistral NeMo was trained with quantization awareness, enabling FP8 inference without any performance loss. The model is designed for global, multilingual applications. It is trained on function calling and has a large context window. Compared to Mistral 7B, it is much better at following precise instructions, reasoning, and handling multi-turn conversations.
Learn more
Mistral Large 3
Mistral Large 3 is a next-generation, open multimodal AI model built with a powerful sparse Mixture-of-Experts architecture featuring 41B active parameters out of 675B total. Designed from scratch on NVIDIA H200 GPUs, it delivers frontier-level reasoning, multilingual performance, and advanced image understanding while remaining fully open-weight under the Apache 2.0 license. The model achieves top-tier results on modern instruction benchmarks, positioning it among the strongest permissively licensed foundation models available today. With native support across vLLM, TensorRT-LLM, and major cloud providers, Mistral Large 3 offers exceptional accessibility and performance efficiency. Its design enables enterprise-grade customization, letting teams fine-tune or adapt the model for domain-specific workflows and proprietary applications. Mistral Large 3 represents a major advancement in open AI, offering frontier intelligence without sacrificing transparency or control.
Learn more