
DeepSeek’s 685B Models Rival GPT‑5
DeepSeek’s 685B Models Rival GPT‑5: China’s Sparse Attention
DeepSeek’s new 685B‑parameter models, powered by sparse attention and released open‑source, rival GPT‑5 on reasoning and coding, potentially democratizing frontier AI access globally.
DeepSeek released its new 685‑billion‑parameter models - DeepSeek‑V3.2 and DeepSeek‑V3.2‑Speciale - claiming performance comparable to top-tier models like GPT‑5, while introducing a major efficiency upgrade through a novel “sparse attention” mechanism. Rather than hammering every token in long contexts equally, the model selectively attends only to the most relevant parts, dramatically cutting compute and inference cost when handling large documents or extended codebases.
The key technical lift is called DeepSeek Sparse Attention (DSA). By upending the typical quadratic cost growth of attention (where every token attends to every other token), DSA reduces long‑context complexity, letting V3.2 handle up to 128,000–token contexts (think hundreds of pages or massive codebases) with far less computational overhead. According to published numbers, this reduces inference cost roughly by half compared to previous versions, making heavy-duty workflows more feasible and cheaper.
The fact that DeepSeek released the models under an open-source license (MIT) makes this even more interesting. Instead of locking frontier models behind paywalls or usage quotas, this move lowers the barrier for researchers, developers, and educational institutions worldwide, enabling anyone to experiment, build, or deploy without paying premium API fees.
Real‑world implications could be significant: from analyzing huge technical documents, to building AI-assisted tools for coding, research, legal review or long‑form content generation - now at lower cost and with global access. For developers, educators, or institutions in places with fewer resources, this could shift frontier‑scale AI from being a big‑budget luxury to a practical utility. China’s AI scene just dropped a serious wildcard. Maybe more than one.
The Takeaway
DeepSeek‑V3.2 shows that high‑parameter, high‑performance AI doesn’t always need blockbuster compute or sky‑high fees. With smart design (sparse attention, open licensing, and long‑context scaling) frontier capabilities might finally become more global, more accessible, and more diverse in user base. It’s not hype. It’s a potentially meaningful shift in how AI spreads across economies and geographies.
Gallery
No additional images available.
Tags
Related Links
No related links available.
Join the Discussion
Enjoyed this? Ask questions, share your take (hot, lukewarm, or undecided), or follow the thread with people in real time. The community’s open — join us.
Published December 5, 2025 • Updated December 6, 2025
published
Latest in AI

Signal Warns Agentic AI Is a Security and Surveillance Risk
Jan 14, 2026

CES 2026 Is Live and AI Is Everywhere
Jan 2, 2026

Alphabet Spends $4.75B to Secure the One Thing AI Can’t Run Without
Dec 23, 2025

Businesses Are All-In on AI. The Payoff Is Still a Question.
Dec 22, 2025

Apple Quietly Pushes AI Deeper Into iOS Without Calling It AI
Dec 15, 2025
Right Now in Tech

Google Found Its Rhythm Again in the AI Race
Jan 8, 2026

AI Is Starting to Show Up Inside Our Chats
Jan 5, 2026

ChatGPT Rolls Out a Personalized Year in Review
Dec 23, 2025

California Judge Says Tesla’s Autopilot Marketing Went Too Far
Dec 17, 2025

Windows 11 Will Ask Before AI Touches Your Files
Dec 17, 2025