Connect with us

Hi, what are you looking for?

Top Stories

Mistral Launches AI Family with 675B Parameters, Competes Head-to-Head with DeepSeek

French startup Mistral launches four AI models, including the flagship Large 3 with 675 billion parameters, challenging DeepSeek’s dominance in the open-source arena.

French AI startup Mistral made a significant leap in the competitive landscape of artificial intelligence with the release of its latest model family on Tuesday. Known for its role as the underdog in a field largely dominated by American and Chinese firms, Mistral’s new offerings are positioned to challenge existing open-source models, providing them free of charge.

The new lineup includes four models, ranging from compact personal assistants to a cutting-edge system boasting 675 billion parameters. All models are available under the permissive Apache 2.0 open-source license, allowing users to download, modify, and fine-tune them for various applications on compatible hardware.

At the forefront is the flagship model, Mistral Large 3, which utilizes a sparse Mixture-of-Experts architecture, activating only 41 billion parameters for each token processed. This engineering choice allows the model to achieve performance comparable to much larger systems while operating at a level typically associated with 40 billion parameter models.

Trained from scratch using 3,000 NVIDIA H200 GPUs, Mistral Large 3 debuted impressively, ranking second among open-source, non-reasoning models on the LMArena leaderboard. In terms of benchmark comparisons, Mistral’s leading model surpasses DeepSeek V3.1 across several metrics but trails slightly behind the newer V3.2 version.

When it comes to general knowledge and expert reasoning tasks, Mistral’s offerings hold their ground, although DeepSeek maintains an edge in coding speed and mathematical logic. Notably, this new release does not incorporate reasoning models, which limits its cognitive capabilities compared to competitors.

The smaller models in the lineup, referred to as “Ministral,” are particularly noteworthy for developers. Available in three sizes—3 billion, 8 billion, and 14 billion parameters—these models come with both base and instruct variants and support native vision input. The 3B model has garnered attention from AI researcher Simon Willison, who highlighted its capability to run entirely within a browser using WebGPU.

This capability offers unique opportunities for developers and hobbyists alike, making it suitable for applications in drones, robots, and even offline systems in vehicles. Early testing has revealed a distinctive character across the Mistral lineup; the Mistral 3 Large demonstrates conversational fluency, often mirroring the style of GPT-5 but with a more natural cadence.

However, it has also shown a tendency for repetition and overreliance on common phrases, especially in its 14B instruct variant, which users have flagged on platforms like Reddit. Despite these issues, its ability to generate long-form content remains a highlight for its size.

The smaller 3B and 8B models, while functional, sometimes produce formulaic outputs on creative tasks, although their compact size allows them to run on less powerful hardware, such as smartphones. The only other competitive option in this niche is Google’s smallest version of Gemma 3.

Enterprise interest in Mistral is already materializing, as demonstrated by HSBC‘s announcement of a multi-year partnership to implement generative AI within its operations. The bank plans to self-host the models on its infrastructure, aligning Mistral’s expertise with its internal technical capabilities—a choice particularly appealing for organizations managing sensitive customer data.

In collaboration with NVIDIA, Mistral has developed a compressed checkpoint called NVFP4 that enables Mistral Large 3 to operate on a single node powered by eight high-end NVIDIA cards. NVIDIA claims that the Ministral 3B model achieves approximately 385 tokens per second on an RTX 5090, and around 50 tokens per second on Jetson Thor for robotics applications, highlighting its efficiency and speed without compromising quality.

Future developments include a reasoning-optimized version of Large 3, although competitors like DeepSeek R1 and various Chinese models retain their advantages in explicit reasoning tasks for now. For enterprises prioritizing cutting-edge capabilities, open-source flexibility, multilingual support, and compliance with European regulations, Mistral’s emergence marks a pivotal expansion of options in the AI landscape.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

DeepSeek is set to launch its V4 AI model, enhanced by Huawei's latest chips, with major orders from Alibaba, ByteDance, and Tencent fueling anticipation.

AI Education

Alpha School's AI-driven education model faces scrutiny in Canada for prioritizing tech over student privacy and well-being, raising critical questions about its efficacy.

Top Stories

DeepSeek forecasts Nvidia's stock will surge 50% to $265 by 2026, driven by new technology and strong institutional confidence amid market challenges.

AI Generative

Google launches Gemma 4, an open-source AI suite with 26B and 31B models for local deployment, enhancing privacy and multimodal reasoning capabilities.

Top Stories

Microsoft unveils three new MAI models enhancing productivity, including MAI-Transcribe-1, which boasts 2.5x faster speech-to-text transcription than Azure Fast.

Top Stories

DeepSeek's seven-hour outage disrupts millions, revealing critical infrastructure gaps in AI reliability and raising stakes for developers dependent on its API.

Top Stories

DeepSeek shifts to Huawei chips, revealing a 50% spike in Chinese representation in US AI research, as Western firms struggle with $15M daily costs...

AI Technology

Rebellions secures $400M for scalable inference solutions, while Mistral AI raises $830M for a new Paris data center, underscoring a shift in AI infrastructure.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.