
DeepSeek’s 685B Models Rival GPT‑5
DeepSeek’s 685B Models Rival GPT‑5: China’s Sparse Attention
DeepSeek’s new 685B‑parameter models, powered by sparse attention and released open‑source, rival GPT‑5 on reasoning and coding, potentially democratizing frontier AI access globally.
DeepSeek released its new 685‑billion‑parameter models - DeepSeek‑V3.2 and DeepSeek‑V3.2‑Speciale - claiming performance comparable to top-tier models like GPT‑5, while introducing a major efficiency upgrade through a novel “sparse attention” mechanism. Rather than hammering every token in long contexts equally, the model selectively attends only to the most relevant parts, dramatically cutting compute and inference cost when handling large documents or extended codebases.
The key technical lift is called DeepSeek Sparse Attention (DSA). By upending the typical quadratic cost growth of attention (where every token attends to every other token), DSA reduces long‑context complexity, letting V3.2 handle up to 128,000–token contexts (think hundreds of pages or massive codebases) with far less computational overhead. According to published numbers, this reduces inference cost roughly by half compared to previous versions, making heavy-duty workflows more feasible and cheaper.
The fact that DeepSeek released the models under an open-source license (MIT) makes this even more interesting. Instead of locking frontier models behind paywalls or usage quotas, this move lowers the barrier for researchers, developers, and educational institutions worldwide, enabling anyone to experiment, build, or deploy without paying premium API fees.
Real‑world implications could be significant: from analyzing huge technical documents, to building AI-assisted tools for coding, research, legal review or long‑form content generation - now at lower cost and with global access. For developers, educators, or institutions in places with fewer resources, this could shift frontier‑scale AI from being a big‑budget luxury to a practical utility. China’s AI scene just dropped a serious wildcard. Maybe more than one.
The Takeaway
DeepSeek‑V3.2 shows that high‑parameter, high‑performance AI doesn’t always need blockbuster compute or sky‑high fees. With smart design (sparse attention, open licensing, and long‑context scaling) frontier capabilities might finally become more global, more accessible, and more diverse in user base. It’s not hype. It’s a potentially meaningful shift in how AI spreads across economies and geographies.
Tags
Join the Discussion
Enjoyed this? Ask questions, share your take (hot, lukewarm, or undecided), or follow the thread with people in real time. The community’s open, join us.
Latest in AI

Stanford's 2026 AI Report Card: A+ in Math, F in Telling Time
Apr 14, 2026

The Gap Between Mythos and a $0.11 Model Isn't as Big as You Think
Apr 13, 2026

Project Glasswing: Anthropic's Restricted AI Security Model
Apr 7, 2026

ClawBot: Tencent's OpenClaw Agent Is Coming to WeChat
Mar 23, 2026

OpenAI Is Merging ChatGPT, Codex, and Its Browser Into One App
Mar 20, 2026
Right Now in Tech

PS5 Price Hike: $650 for Standard, $900 for Pro Starting April 2
Mar 28, 2026

Apple Discontinues Mac Pro, Ends Intel Era
Mar 27, 2026

OpenAI Is Pulling the Plug on Sora
Mar 26, 2026

Meta and YouTube Ordered to Pay $3M in Landmark Social Media Ruling
Mar 25, 2026

Your Galaxy S26 Can Finally AirDrop to an iPhone
Mar 23, 2026