Connect with us

Hi, what are you looking for?

Top Stories

DeepSeek Launches V3.2, Achieves Gold Medal Performance with 25x Lower Costs than GPT-5

DeepSeek launches V3.2, achieving gold medal scores at prestigious competitions with 25x lower costs than competitors like GPT-5, revolutionizing AI efficiency.

DeepSeek has launched DeepSeek-V3.2, a new model that merges high computational efficiency with enhanced reasoning and performance, less than a year after the original DeepSeek was introduced. The company unveiled two variants: DeepSeek-V3.2 Standard, accessible via web and API, and DeepSeek-V3.2 Speciale, available only through API and optimized for deep reasoning tasks.

Both models have been open-sourced under the MIT license, with weights shared on HuggingFace. DeepSeek also released a technical report titled, “DeepSeek-V3.2: Pushing the Frontier of Open Large Language Models,” detailing the innovations in AI model architecture and training processes that enhanced both efficiency and performance.

Notably, the DeepSeek-V3.2 Speciale has achieved unprecedented feats in multiple high-profile competitions. It scored 35 out of 42 at the International Mathematical Olympiad (IMO), landing a gold medal, and scored 492 out of 600 at the International Olympiad in Informatics (IOI), placing it 10th globally, had it been a human competitor. Additionally, it solved 10 out of 12 problems in the equivalent of the ICPC World Finals, ranking second globally.

Significantly, DeepSeek-V3.2 Speciale accomplished these benchmarks at a cost that is 25 to 30 times lower than that of competitors such as GPT-5 and Gemini 3 Pro. Its focus on reasoning has yielded impressive results across various math and coding benchmarks, outperforming GPT-5 High and Gemini 3.0 Pro on AIME-25, as well as on CodeForces and LiveCodeBench. The Speciale version attained a score of 30% on Humanity’s Last Exam, while the Standard version recorded 25%.

Despite these advancements, models like Claude 4.5 Opus and Gemini 3.0 Pro still surpass DeepSeek V3.2 Speciale in specific reasoning and coding tasks. Although the cost per token is low, the Speciale model is less efficient in token usage, sometimes generating significantly more tokens to resolve complex problems than rivals like Gemini 3 Pro. This inefficiency can undermine its competitive advantage, even at a lower cost per token.

In contrast, the DeepSeek V3.2 Thinking model is tailored for general-purpose use and boasts strong performance metrics, achieving 73.1% on SWE-Bench and 46.4% on TerminalBench 2.0.

DeepSeek V3.2 models exhibit robust agentic capabilities, introducing interleaved tool usage during thinking traces—a concept previously seen in Claude models. This allows for the integration of tool use while reasoning, although the models discard historical thinking traces between turns, which might affect certain coding agents.

DeepSeek’s architecture for V3.2 is based on its predecessor, DeepSeek-V3.2-Exp, which introduced DeepSeek Sparse Attention (DSA). This innovative attention mechanism reduces computational complexity without sacrificing performance, particularly in long-context scenarios. By dynamically selecting relevant tokens, DSA minimizes both pre-filling and decoding costs, achieving a shift from quadratic to O(Lk) in attention computational complexity.

DeepSeek has also strengthened its approach by enhancing the post-training reinforcement learning (RL) components of V3.2. Following a strategy seen in other AI developments like Grok 4, the company allocated over 10% of compute for the standard model and 20% for the Speciale version, focusing on domain-specific expert training in areas like mathematics and programming.

This domain-specific training produces high-quality data for final model checkpoints, with experimental results showing that models trained on this distilled data perform nearly on par with specialized models after subsequent RL training.

To further support its reasoning capabilities in tool-use scenarios, DeepSeek developed a synthesized task pipeline that generates scalable training data. This approach improves compliance and generalization in complex interactive environments, enhancing the models’ performance in agentic tasks.

With the introduction of Gemini 3.0 Pro, GPT-5.1, and Claude 4.5 Opus, leading AI labs are pushing the boundaries of artificial intelligence. Yet, DeepSeek V3.2 demonstrates that open-source models can attain high levels of reasoning and coding performance while being significantly more economical than their closed-source counterparts. This release underscores the potential for open-source AI to rival proprietary models, contributing to the ongoing commoditization of advanced artificial intelligence.

While formidable, DeepSeek V3.2 still trails behind competitive models like Gemini 3.0 Pro in terms of overall token efficiency and specific benchmarks. The company acknowledges a gap in expansive world knowledge compared to larger proprietary models, which they plan to address in future iterations like V4. However, DeepSeek’s innovations in architecture and training strategies have positioned it as a leading contender in the evolving AI landscape.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

DeepSeek introduces the groundbreaking mHC method to enhance the scalability and stability of language models, positioning itself as a major AI contender.

AI Tools

MIT study reveals that 83% of students using ChatGPT for essays struggle to recall their work, highlighting significant cognitive deficits and reduced engagement.

Top Stories

DeepSeek launches its mHC architecture, enhancing large-model training efficiency while reducing computational costs, with consistent performance across 3-27 billion parameter models.

Top Stories

Micron Technology's stock soars 250% as it anticipates a 132% revenue surge to $18.7B, positioning itself as a compelling long-term investment in AI.

Top Stories

57-year-old consultant enhances AI skills through a $3,000 Johns Hopkins program, transforming a critical gap into a strategic partnership with an oil and gas...

AI Regulation

OpenAI accelerates GPT-5 development amid rising concerns over low-quality AI content, as "AI slop" is named 2025's word of the year.

Top Stories

China's CPC faces innovation stagnation despite rapid AI advancements, as state control hinders the dynamic growth needed for global tech leadership.

Top Stories

Fraudulent Chrome extensions impersonating AITOPIA have misled 900,000 users, raising serious concerns over data security and user privacy.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.