Connect with us

Hi, what are you looking for?

Top Stories

DeepSeek Launches V4 Models with 1.6 Trillion Parameters, Closing Gap with Leading AI

DeepSeek launches its V4 models, featuring 1.6 trillion parameters, significantly underpricing rivals while enhancing efficiency and performance in AI reasoning.

Chinese AI lab DeepSeek has launched two preview versions of its newest large language model, DeepSeek V4, a much-awaited update to last year’s V3.2 model and the accompanying R1 reasoning model that took the AI world by storm.

The company announced that both DeepSeek V4 Flash and V4 Pro are mixture-of-experts models featuring context windows of 1 million tokens each, allowing for the integration of extensive codebases or documents into prompts. This mixture-of-experts approach activates only a selected number of parameters per task, which helps to reduce inference costs.

The Pro model boasts a total of 1.6 trillion parameters, with 49 billion active, making it the largest open-weight model currently available. This surpasses rivals such as Moonshot AI’s Kimi K 2.6 with 1.1 trillion, MiniMax’s M1 at 456 billion, and more than doubling the 671 billion parameters of DeepSeek V3.2. The smaller, V4 Flash model contains 284 billion parameters, with 13 billion active.

DeepSeek asserts that both V4 models are more efficient and high-performing than V3.2 due to architectural enhancements, claiming they have nearly “closed the gap” with the leading models, whether open or closed, on reasoning benchmarks. The company maintains that its new V4-Pro-Max model surpasses its open-source competitors across reasoning benchmarks, even outpacing OpenAI’s GPT-5.2 and Gemini 3.0 Pro in certain tasks. In coding competition benchmarks, DeepSeek claims the performance of both V4 models is “comparable to GPT-5.4.”

Despite these advancements, the models appear to lag behind frontier models in knowledge tests, particularly OpenAI’s GPT-5.4 and Google’s Gemini 3.1 Pro. This discrepancy suggests a “developmental trajectory that trails state-of-the-art frontier models by approximately 3 to 6 months,” according to the lab.

Both V4 Flash and V4 Pro are limited to text-only capabilities, contrasting with many closed-source competitors that offer multi-modal functionalities, including audio, video, and image processing.

In terms of pricing, DeepSeek V4 is significantly more affordable than existing frontier models. The smaller V4 Flash is priced at $0.14 per million input tokens and $0.28 per million output tokens, undercutting GPT-5.4 Nano, Gemini 3.1 Flash, GPT-5.4 Mini, and Claude Haiku 4.5. The larger V4 Pro has a cost of $0.145 per million input tokens and $3.48 per million output tokens, also underpricing competitors like Gemini 3.1 Pro, GPT-5.5, Claude Opus 4.7, and GPT-5.4.

The announcement follows a day after the U.S. government accused China of large-scale theft of American AI intellectual property through numerous proxy accounts. DeepSeek itself faces allegations from Anthropic and OpenAI regarding the “distillation,” or copying, of their AI models.

As DeepSeek enters the competitive landscape with its V4 models, the implications for its positioning in the rapidly evolving AI sector remain significant, particularly as the race for dominance intensifies amid allegations of intellectual property disputes and advancing technological capabilities.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Research

DeepSeek unveils AI-2023, boosting processing speed by 50%, positioning the company to capture a $1 trillion AI market by 2025.

Top Stories

DeepSeek unveils its V4 AI model, outpacing open-source rivals and attracting funding discussions from Alibaba and Tencent, with a projected valuation over $20 billion.

Top Stories

DeepSeek's V4 API launches with a groundbreaking 2-million-token context window, challenging OpenAI and Anthropic while offering competitive pricing at $2.80 per million input tokens.

Top Stories

Tencent aims for a 20% stake in $40B AI startup DeepSeek as Alibaba joins funding talks, intensifying the competition in China's AI landscape

Top Stories

Xiaomi's MiMo-V2.5 series achieves a groundbreaking 50% token efficiency gain over competitors while introducing advanced models for intelligent agent applications.

Top Stories

Chinese AI models, led by DeepSeek's R1, capture 17.1% of global downloads, surpassing the U.S. as open-source innovation reshapes AI development.

AI Generative

Moonshot AI releases Kimi-K2.6, an open-source LLM surpassing GPT-5.4 with 1 trillion parameters and achieving a benchmark score of 54 on the challenging HLE-Full...

Top Stories

Stanford's 2026 AI Index reveals the China-US AI performance gap has narrowed to just 2.7%, as Nvidia captures 60% of the global AI computing...

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.