China’s tech wonder kid DeepSeek has launched a new experimental model, V3.2-Exp, as part of its attempt to challenge American dominance in AI. The release came on Monday and was first made public through a post on Hugging Face, a popular AI forum. DeepSeek claims that this latest version builds on its current model, V3.1-Terminus, […]China’s tech wonder kid DeepSeek has launched a new experimental model, V3.2-Exp, as part of its attempt to challenge American dominance in AI. The release came on Monday and was first made public through a post on Hugging Face, a popular AI forum. DeepSeek claims that this latest version builds on its current model, V3.1-Terminus, […]

What to know about DeepSeek's new V3.2-Exp model

China’s tech wonder kid DeepSeek has launched a new experimental model, V3.2-Exp, as part of its attempt to challenge American dominance in AI. The release came on Monday and was first made public through a post on Hugging Face, a popular AI forum.

DeepSeek claims that this latest version builds on its current model, V3.1-Terminus, but with a stronger emphasis on speed, cost, and memory handling.

According to Hugging Face’s Chinese community lead Adina Yakefu, the model features something called DeepSeek Sparse Attention, or DSA, which she said “makes the AI better at handling long documents and conversations” while also cutting operating costs in half.

If you recall, around a year ago, DeepSeek dropped and shook things up by dropping its first model, R1, without warning. That model showed it was possible to train a large language model using fewer chips and much less computing power. No one expected a Chinese team to pull that off under those constraints. With V3.2-Exp, the goal hasn’t changed: less hardware, more performance.

Adds DeepSeek Sparse Attention and reduces AI running cost

DSA is the big feature in this model. It changes how the AI picks which information to look at. Instead of scanning everything, DeepSeek trains the model to focus only on what seems useful for the task. Adina explained that the benefit here is twofold: “efficiency” and “cost reduction.”

By skipping irrelevant data, the model moves faster and requires less energy. She said the model was designed with open-source collaboration in mind.

Nick Patience, who leads AI research at The Futurum Group, told CNBC the model has the potential to open up powerful AI tools to developers who can’t afford to use more expensive models. “It should make the model faster and more cost-effective to use without a noticeable drop in performance,” Nick said. But that doesn’t mean there aren’t risks.

The way DeepSeek uses sparse attention is like how airlines pick flight routes. There might be hundreds of ways to get from one place to another, but only a few make sense. The model filters through the noise and focuses on what matters — or at least what it thinks matters.

But this comes with concerns. Ekaterina Almasque, who cofounded BlankPage Capital, explained it simply: “So basically, you cut out things that you think are not important.” But the issue, she said, is that there’s no guarantee the model is cutting the right things.

Ekaterina, who has backed companies like Dataiku, Darktrace, and Graphcore, warned that cutting corners might create problems later. “They [sparse attention models] have lost a lot of nuances,” she said. “And then the real question is, did they have the right mechanism to exclude not important data, or is there a mechanism excluding really important data, and then the outcome will be much less relevant?”

Connects to Chinese chips and releases open code

Despite those concerns, DeepSeek insists that V3.2-Exp performs just as well as V3.1-Terminus. The model can also run directly on domestic Chinese chips like Ascend and Cambricon, with no extra configurations required. That’s key in China’s broader effort to build AI on homegrown hardware and reduce dependency on foreign tech. “Right out of the box,” Adina said, DeepSeek works with these chips.

The company also made the model’s full code and tools public. That means anyone can download, run, modify, or build on top of V3.2-Exp. This move aligns with DeepSeek’s open-source strategy, but it raises another issue: patents. Since the model is open and the core idea, sparse attention, has been around since 2015, DeepSeek can’t lock it down legally.

“The approach is not super new,” said Ekaterina. For her, the only defensible part of the tech is how DeepSeek chooses what to keep and what to ignore.

That’s where the real competition lies now. Not just in making smarter models, but making them faster, cheaper, and leaner — without screwing up results. Even DeepSeek called this version “an intermediate step toward our next-generation architecture,” which suggests they’re already working on something bigger.

Nick said the model shows that efficiency is now just as important as raw power. And Adina believes the company has a long-term play in mind. “DeepSeek is playing the long game to keep the community invested in their progress,” she said. “People will always go for what is cheap, reliable, and effective.”

Join a premium crypto trading community free for 30 days - normally $100/mo.

Market Opportunity
Particl Logo
Particl Price(PART)
$0.3056
$0.3056$0.3056
-2.92%
USD
Particl (PART) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference

Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference

The post Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference appeared on BitcoinEthereumNews.com. Key Takeaways Ethereum’s new roadmap was presented by Vitalik Buterin at the Japan Dev Conference. Short-term priorities include Layer 1 scaling and raising gas limits to enhance transaction throughput. Vitalik Buterin presented Ethereum’s development roadmap at the Japan Dev Conference today, outlining the blockchain platform’s priorities across multiple timeframes. The short-term goals focus on scaling solutions and increasing Layer 1 gas limits to improve transaction capacity. Mid-term objectives target enhanced cross-Layer 2 interoperability and faster network responsiveness to create a more seamless user experience across different scaling solutions. The long-term vision emphasizes building a secure, simple, quantum-resistant, and formally verified minimalist Ethereum network. This approach aims to future-proof the platform against emerging technological threats while maintaining its core functionality. The roadmap presentation comes as Ethereum continues to compete with other blockchain platforms for market share in the smart contract and decentralized application space. Source: https://cryptobriefing.com/ethereum-roadmap-scaling-interoperability-security-japan/
Share
BitcoinEthereumNews2025/09/18 00:25
MMDA, sleep health organization launch drowsy driving campaign ahead of holidays

MMDA, sleep health organization launch drowsy driving campaign ahead of holidays

The Metro Manila Development Authority (MMDA) and the Philippine Society of Sleep Medicine (PSSM) on Wednesday launch an awareness campaign to prevent drowsy driving
Share
Bworldonline2025/12/18 12:05
A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release

A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release

The post A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release appeared on BitcoinEthereumNews.com. KPop Demon Hunters Netflix Everyone has wondered what may be the next step for KPop Demon Hunters as an IP, given its record-breaking success on Netflix. Now, the answer may be something exactly no one predicted. According to a new filing with the MPA, something called Debut: A KPop Demon Hunters Story has been rated PG by the ratings body. It’s listed alongside some other films, and this is obviously something that has not been publicly announced. A short film could be well, very short, a few minutes, and likely no more than ten. Even that might be pushing it. Using say, Pixar shorts as a reference, most are between 4 and 8 minutes. The original movie is an hour and 36 minutes. The “Debut” in the title indicates some sort of flashback, perhaps to when HUNTR/X first arrived on the scene before they blew up. Previously, director Maggie Kang has commented about how there were more backstory components that were supposed to be in the film that were cut, but hinted those could be explored in a sequel. But perhaps some may be put into a short here. I very much doubt those scenes were fully produced and simply cut, but perhaps they were finished up for this short film here. When would Debut: KPop Demon Hunters theoretically arrive? I’m not sure the other films on the list are much help. Dead of Winter is out in less than two weeks. Mother Mary does not have a release date. Ne Zha 2 came out earlier this year. I’ve only seen news stories saying The Perfect Gamble was supposed to come out in Q1 2025, but I’ve seen no evidence that it actually has. KPop Demon Hunters Netflix It could be sooner rather than later as Netflix looks to capitalize…
Share
BitcoinEthereumNews2025/09/18 02:23