Connect with us

Hi, what are you looking for?

Top Stories

DeepSeek V4 Set to Launch February 17, Promises to Outperform Claude and ChatGPT in Coding

DeepSeek’s V4 model, launching February 17, aims to surpass Claude and GPT in coding performance, leveraging a $6 million development cost and innovative mHC training.

DeepSeek, a rising AI startup based in Hangzhou, is set to release its V4 model around mid-February, with sources suggesting a launch date of February 17—coinciding with the Lunar New Year. The new model is reportedly designed for coding tasks and, according to insiders, outperforms major players like Anthropic’s Claude and OpenAI’s GPT series in internal benchmarks, particularly in handling lengthy code prompts.

While no public benchmarks or details about V4 have been shared, the developer community appears undeterred. Discussions in Reddit’s r/DeepSeek and r/LocalLLaMA have intensified, with users amassing API credits and speculating that V4 could position DeepSeek as a formidable contender against Silicon Valley’s entrenched giants. This would not be DeepSeek’s first disruption; its January 2025 release of the R1 reasoning model sparked a $1 trillion sell-off in global markets after it matched OpenAI’s o1 model on math and reasoning benchmarks at a fraction of the development cost.

The R1 model was reportedly developed for just $6 million, approximately 68 times less than competitors’ expenditures. Following that, the V3 model achieved a score of 90.2% on the MATH-500 benchmark, surpassing Claude’s 78.3%. The subsequent “V3.2 Speciale” further enhanced this performance, indicating a consistent upward trajectory for the company.

Targeting the enterprise developer market, V4 represents a strategic pivot for DeepSeek. Unlike the R1, which focused on pure reasoning tasks, V4 is a hybrid model that merges reasoning with non-reasoning tasks, which is crucial for generating high-accuracy code that can directly translate into revenue.

To claim dominance in the coding space, V4 would need to surpass Claude Opus 4.5, which currently holds the SWE-bench Verified record at 80.9%. However, if past performance is any indicator, this may well be within reach for DeepSeek, despite the limitations faced by a Chinese AI lab.

Technical Details

Assuming the reports are accurate, the secret behind DeepSeek’s potential success may lie in a groundbreaking training methodology known as Manifold-Constrained Hyper-Connections, or mHC. Co-authored by founder Liang Wenfeng, this approach addresses a key challenge in scaling large language models: how to expand a model’s capacity without resulting in instability during training.

Traditional architectures typically funnel all information through a single narrow pathway, whereas mHC broadens this pathway into multiple streams, enabling information exchange without risking training collapse. Wei Sun, principal analyst for AI at Counterpoint Research, described mHC as a “striking breakthrough,” suggesting it allows DeepSeek to overcome compute limitations and achieve significant advancements in intelligence, even under U.S. export restrictions.

Lian Jye Su, chief analyst at Omdia, pointed out that DeepSeek’s choice to publish its methods reflects a “newfound confidence in the Chinese AI industry.” This open-source approach has earned the company a following among developers, contrasting sharply with OpenAI’s closed models and extensive fundraising strategies.

Despite this momentum, skepticism persists within the developer community. Some Reddit users argue that DeepSeek’s reasoning models inefficiently allocate compute resources to simple tasks, while others contend that the company’s benchmarks do not accurately capture real-world complexities. One viral Medium post criticized DeepSeek’s models for producing “boilerplate nonsense with bugs” and “hallucinated libraries.”

DeepSeek also faces challenges related to privacy, as some governments have banned its native app. The company’s ties to China and concerns over censorship add a layer of geopolitical friction to the ongoing technical discussions. Nevertheless, DeepSeek’s influence continues to grow, particularly in Asia, and if V4 meets its coding aspirations, enterprise adoption in Western markets may soon follow.

Looking ahead, the timing of V4’s launch is noteworthy. DeepSeek had initially planned to introduce its R2 model in May 2025 but delayed its release due to dissatisfaction with performance. With V4 on the horizon and R2 potentially following in August, the company appears to be moving with both urgency and confidence, suggesting that it is keen to solidify its foothold in a rapidly evolving AI landscape.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Education

Renowned AI researcher Haibin Ling joins Westlake University to spearhead the Intelligent Computing Lab, aiming for groundbreaking advancements in artificial intelligence.

Top Stories

DeepSeek's V4 model, launching by February 17, aims to outperform Claude and ChatGPT in coding, leveraging innovative training to boost accuracy beyond 80.9%.

AI Research

Thinking Machines Lab secures $2B funding at a $12B valuation and launches Tinker, a groundbreaking tool for efficient AI model customization.

AI Technology

NVIDIA and AMD unveil a future where AI becomes the core operating system of life, with AMD predicting a thousandfold increase in AI chip...

Top Stories

Nvidia, Broadcom, and Amazon are set to lead the AI market's explosive growth, with Nvidia's EPS projected to soar 45% and Broadcom's AI revenue...

AI Business

As enterprises double down on AI investments, OpenAI faces intensified competition from Google's Gemini and Microsoft's Copilot, threatening its market dominance.

Top Stories

Anthropic seeks $10 billion in funding to boost its valuation to $350 billion amid rising concerns of an AI bubble, as competition with OpenAI...

Top Stories

China's AI-driven labor market saw recruitment for high-exposure roles plummet by 30%, while Singapore pivoted to resilience with a 200% rise in demand for...

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.