8 min read

DeepSeek Releases Free Models Rivaling GPT‑5 & Gemini‑3

AI

ThinkTools Team

AI Research Lead

Introduction

In a move that could reverberate across the global AI ecosystem, Hangzhou‑based DeepSeek has unveiled two new language models that the company claims are on par with, or even surpass, the most advanced proprietary systems from the United States and Google. The releases, DeepSeek‑V3.2 and its high‑performance cousin DeepSeek‑V3.2‑Speciale, are not only notable for their technical prowess but also for their open‑source MIT license, making them freely available to developers, researchers, and enterprises worldwide. This development arrives at a time when the AI race is increasingly defined by cost efficiency, regulatory constraints, and the ability to deploy models locally without relying on cloud APIs. By offering 685‑billion‑parameter models that can process 128,000‑token context windows and achieve gold‑medal performance in international competitions, DeepSeek is challenging the long‑standing assumption that frontier AI performance requires massive, proprietary infrastructure.

The announcement has sparked a flurry of reactions across social media, with some commentators declaring the end of the dominance of ChatGPT‑style models. Beyond the headline‑grabbing claims, the real significance lies in the architectural innovations that enable such performance at a fraction of the inference cost, the rigorous benchmarking against established leaders, and the strategic decision to release the models openly. Together, these factors suggest a potential shift in how AI capabilities are distributed, how businesses adopt them, and how governments regulate cross‑border data flows.

In this post, we unpack the technical breakthroughs behind DeepSeek’s new models, examine the evidence that supports their parity with GPT‑5 and Gemini‑3, explore the implications of an open‑source approach in a highly competitive market, and discuss the regulatory hurdles that could shape the future trajectory of this technology.

Main Content

Sparse Attention: Cutting Costs

At the heart of DeepSeek’s performance leap is a novel attention mechanism dubbed DeepSeek Sparse Attention (DSA). Traditional transformer attention scales quadratically with input length, meaning that doubling the token count quadruples the computational load. For long documents—think 300‑page books or extensive codebases—this quickly becomes prohibitive. DSA sidesteps this bottleneck by employing a “lightning indexer” that identifies only the most relevant segments of context for each query, effectively pruning the attention matrix. The result is a dramatic reduction in inference cost: decoding 128,000 tokens now costs roughly $0.70 per million tokens, compared to $2.40 for the previous V3.1‑Terminus model—a 70% savings.

This efficiency does not come at the expense of accuracy. Independent evaluations on long‑context benchmarks demonstrate that V3.2 performs on par with, or better than, its predecessor, even when using the sparse attention mechanism. The ability to handle such extensive context windows while maintaining performance is a key differentiator, especially for applications that require deep analysis of large documents, codebases, or research papers.

Benchmarking Against GPT‑5 and Gemini‑3

DeepSeek’s claims of parity are substantiated by a battery of rigorous tests across mathematics, coding, and reasoning. On the American Invitational Mathematics Examination (AIME) 2025, the Speciale variant achieved a 96.0% pass rate, outperforming GPT‑5‑High (94.6%) and Gemini‑3.0‑Pro (95.0%). Similarly, on the Harvard‑MIT Mathematics Tournament, the Speciale model scored 99.2%, eclipsing Gemini’s 97.5%. While the standard V3.2 model scored slightly lower—93.1% on AIME and 92.5% on HMMT—it did so with significantly fewer resources.

Perhaps the most striking evidence comes from international competitions. The Speciale model earned gold medals in the 2025 International Mathematical Olympiad (score 35/42), the International Olympiad in Informatics (score 492/600, ranking 10th overall), and the ICPC World Finals (solving 10 of 12 problems, placing second). These results were achieved without internet access or external tools, underscoring the model’s pure reasoning capability.

In coding benchmarks, V3.2 resolved 73.1% of real‑world software bugs on SWE‑Verified, closely matching GPT‑5‑High’s 74.9%. On Terminal Bench 2.0, which measures complex coding workflows, DeepSeek scored 46.4%—well above GPT‑5‑High’s 35.2%. The company acknowledges that token efficiency remains a challenge, noting that longer generation trajectories are sometimes required to match the output quality of Gemini‑3.0‑Pro.

Tool‑Aware Reasoning

Beyond raw reasoning, DeepSeek has introduced a “thinking in tool‑use” capability that allows the model to maintain a coherent chain of thought while executing code, searching the web, or manipulating files. Previous models would lose their reasoning trace each time they invoked an external tool, forcing a restart. DSA’s architecture preserves the reasoning context across multiple tool calls, enabling fluid multi‑step problem solving.

To train this capability, DeepSeek built a synthetic data pipeline that generated over 1,800 distinct task environments and 85,000 complex instructions. These ranged from multi‑day trip planning with budget constraints to software bug fixes across eight programming languages and web‑based research requiring dozens of searches. By training on real‑world tools—such as web search APIs, coding environments, and Jupyter notebooks—while generating synthetic prompts for diversity, the model generalizes to unseen tools and environments, a critical requirement for real‑world deployment.

Open‑Source Strategy and Business Implications

Unlike the proprietary models of OpenAI and Anthropic, DeepSeek has released both V3.2 and V3.2‑Speciale under the MIT license. This permissive framework allows developers, researchers, and companies to download, modify, and deploy the 685‑billion‑parameter models without restriction. Full weights, training code, and documentation are available on Hugging Face, the leading platform for AI model sharing.

The strategic implications are profound. By offering frontier‑capable models for free, DeepSeek undermines competitors that charge premium API prices. The Hugging Face model card notes that DeepSeek has provided Python scripts and test cases that demonstrate how to encode messages in an OpenAI‑compatible format, making migration from competing services straightforward. For enterprises, the value proposition is compelling: high performance at a fraction of the cost, coupled with deployment flexibility. However, data residency concerns and regulatory uncertainty—particularly given DeepSeek’s Chinese origins—may limit adoption in sensitive applications.

Regulatory Challenges and Export Controls

DeepSeek’s global expansion faces mounting regulatory resistance. In June, Berlin’s data protection commissioner declared that DeepSeek’s transfer of German user data to China is unlawful under EU rules, urging Apple and Google to consider blocking the app. Italy ordered the company to block its app in February, and U.S. lawmakers have moved to ban the service from government devices, citing national security concerns.

Questions also persist about U.S. export controls designed to limit China’s AI capabilities. DeepSeek has hinted that China will soon have next‑generation home‑grown chips to support its models, indicating that export controls alone cannot halt Chinese AI progress. The company’s original V3 model was reportedly trained on roughly 2,000 older Nvidia H800 chips—hardware that has been restricted for China export. While the training hardware for V3.2 remains undisclosed, the continued advancement of the models suggests that the combination of domestic chip development and open‑source strategies can circumvent many of the barriers imposed by export controls.

Conclusion

DeepSeek’s release of two open‑source models that rival the performance of GPT‑5 and Gemini‑3 represents a watershed moment in the AI landscape. By combining a sparse attention mechanism that slashes inference costs, rigorous benchmarking that demonstrates parity with the industry’s best, and an open‑source distribution model that democratizes access, DeepSeek challenges the prevailing narrative that frontier AI requires proprietary infrastructure and exorbitant capital. The company’s approach underscores the growing importance of efficiency innovations and the potential for open‑source models to disrupt established business models.

At the same time, the regulatory environment remains a significant hurdle. Data protection authorities in Europe and security agencies in the United States are scrutinizing the cross‑border flow of data and the potential national‑security implications of Chinese‑origin AI models. Whether these challenges will stifle DeepSeek’s global reach or simply accelerate the development of domestic alternatives remains to be seen.

What is clear, however, is that the AI race is entering a new phase. The ability to deliver high‑performance, cost‑efficient models for free threatens to level the playing field and could force incumbent players to rethink their pricing strategies, open‑source contributions, and engagement with regulatory bodies. As the world watches, the next few months will be critical in determining whether open‑source AI can sustain the momentum it has gained or whether geopolitical constraints will reassert dominance for a few key players.

Call to Action

If you’re a developer, researcher, or business leader interested in cutting‑edge AI, now is the time to explore DeepSeek’s V3.2 and V3.2‑Speciale models. Download the weights from Hugging Face, experiment with the sparse attention architecture, and evaluate how the models perform on your own workloads. For enterprises, consider the cost savings of deploying these models locally versus paying for premium APIs, and assess the regulatory implications for your industry. Engage with the community on Hugging Face and X to share insights, contribute improvements, and help shape the future of open‑source AI. By embracing these models today, you can position yourself at the forefront of a paradigm shift that may redefine how AI is built, distributed, and governed worldwide.

We value your privacy

We use cookies, including Google Analytics, to improve your experience on our site. By accepting, you agree to our use of these cookies. Learn more