French AI startup Mistral made a significant leap in the competitive landscape of artificial intelligence with the release of its latest model family on Tuesday. Known for its role as the underdog in a field largely dominated by American and Chinese firms, Mistral’s new offerings are positioned to challenge existing open-source models, providing them free of charge.
The new lineup includes four models, ranging from compact personal assistants to a cutting-edge system boasting 675 billion parameters. All models are available under the permissive Apache 2.0 open-source license, allowing users to download, modify, and fine-tune them for various applications on compatible hardware.
At the forefront is the flagship model, Mistral Large 3, which utilizes a sparse Mixture-of-Experts architecture, activating only 41 billion parameters for each token processed. This engineering choice allows the model to achieve performance comparable to much larger systems while operating at a level typically associated with 40 billion parameter models.
Trained from scratch using 3,000 NVIDIA H200 GPUs, Mistral Large 3 debuted impressively, ranking second among open-source, non-reasoning models on the LMArena leaderboard. In terms of benchmark comparisons, Mistral’s leading model surpasses DeepSeek V3.1 across several metrics but trails slightly behind the newer V3.2 version.
When it comes to general knowledge and expert reasoning tasks, Mistral’s offerings hold their ground, although DeepSeek maintains an edge in coding speed and mathematical logic. Notably, this new release does not incorporate reasoning models, which limits its cognitive capabilities compared to competitors.
The smaller models in the lineup, referred to as “Ministral,” are particularly noteworthy for developers. Available in three sizes—3 billion, 8 billion, and 14 billion parameters—these models come with both base and instruct variants and support native vision input. The 3B model has garnered attention from AI researcher Simon Willison, who highlighted its capability to run entirely within a browser using WebGPU.
This capability offers unique opportunities for developers and hobbyists alike, making it suitable for applications in drones, robots, and even offline systems in vehicles. Early testing has revealed a distinctive character across the Mistral lineup; the Mistral 3 Large demonstrates conversational fluency, often mirroring the style of GPT-5 but with a more natural cadence.
However, it has also shown a tendency for repetition and overreliance on common phrases, especially in its 14B instruct variant, which users have flagged on platforms like Reddit. Despite these issues, its ability to generate long-form content remains a highlight for its size.
The smaller 3B and 8B models, while functional, sometimes produce formulaic outputs on creative tasks, although their compact size allows them to run on less powerful hardware, such as smartphones. The only other competitive option in this niche is Google’s smallest version of Gemma 3.
Enterprise interest in Mistral is already materializing, as demonstrated by HSBC‘s announcement of a multi-year partnership to implement generative AI within its operations. The bank plans to self-host the models on its infrastructure, aligning Mistral’s expertise with its internal technical capabilities—a choice particularly appealing for organizations managing sensitive customer data.
In collaboration with NVIDIA, Mistral has developed a compressed checkpoint called NVFP4 that enables Mistral Large 3 to operate on a single node powered by eight high-end NVIDIA cards. NVIDIA claims that the Ministral 3B model achieves approximately 385 tokens per second on an RTX 5090, and around 50 tokens per second on Jetson Thor for robotics applications, highlighting its efficiency and speed without compromising quality.
Future developments include a reasoning-optimized version of Large 3, although competitors like DeepSeek R1 and various Chinese models retain their advantages in explicit reasoning tasks for now. For enterprises prioritizing cutting-edge capabilities, open-source flexibility, multilingual support, and compliance with European regulations, Mistral’s emergence marks a pivotal expansion of options in the AI landscape.
See also
Unlock B2B Growth: Leverage Operational AI to Boost PR Efficiency and Strategy
Globant Converge 2025: Transform AI Ideas into Action with Global Leaders, Dec 10-11
Anthropic Warns: Humanity Faces AI Evolution Decision by 2027 Amidst Existential Risks
NVIDIA Expands AI Robotics Hub in London, Boosting Market Confidence and Stock Growth
BCC Research Reveals AI’s $1.5 Trillion Market Surge and Transformative Industry Disruptions


















































