The post How Anthropic stopped AI agents working for Chinese state-sponsored spy campaign appeared on BitcoinEthereumNews.com. Chinese state-sponsored hackers exploited Anthropic Claude Code AI in the world’s first largely autonomous cyber-espionage campaign, proving that machine agents can now run sprawling digital attacks with only minimal human input. Anthropic and the AI alarm bell The alarm rang in mid-September at Anthropic, but this was no ordinary network blip. As Anthropic’s threat team sifted through unusual digital clues, what emerged wasn’t yesterday’s malware; it looked more like tomorrow’s cyber warfare had arrived. A Chinese state-backed group, investigators found, orchestrated an audacious cyber espionage campaign, not with a legion of human hackers, but by harnessing the full agentic power of Anthropic AI against 30 global targets. Victims included tech giants, massive banks, factories, and government agencies, a who’s who of digital-era dependence. Autonomous hacking, minimal supervision Last spring’s “AI hacking” buzz might have sounded overblown, but this event erased any doubts. Anthropic’s AI didn’t just suggest tools or code. It became the operation’s key agent, running reconnaissance, building out attack frameworks, and crafting bespoke exploits. The model harvested credentials, exfiltrated classified data, and kept humans on the sidelines. As AI analyst Rohan Paul put it: “Wow, incredible reveal by Anthropic. The AI did 80-90% of the hacking work. Humans only had to intervene 4-6 times per campaign.” How did it work? The new era wasn’t born overnight. But Anthropic’s models, manipulated via clever jailbreaking techniques, were tricked into thinking they were benign cybersecurity employees handling innocent, everyday tasks. Those fragmented requests, pieced together, spelled big trouble. Within minutes, Anthropic AI agents mapped networks, identified juicy databases, produced custom exploit code, and sorted stolen data by intelligence value. The AI even wrote technical docs about the breach, replacing what used to keep human hacking teams awake for weeks. At its peak, the machine blasted out thousands of requests, often several… The post How Anthropic stopped AI agents working for Chinese state-sponsored spy campaign appeared on BitcoinEthereumNews.com. Chinese state-sponsored hackers exploited Anthropic Claude Code AI in the world’s first largely autonomous cyber-espionage campaign, proving that machine agents can now run sprawling digital attacks with only minimal human input. Anthropic and the AI alarm bell The alarm rang in mid-September at Anthropic, but this was no ordinary network blip. As Anthropic’s threat team sifted through unusual digital clues, what emerged wasn’t yesterday’s malware; it looked more like tomorrow’s cyber warfare had arrived. A Chinese state-backed group, investigators found, orchestrated an audacious cyber espionage campaign, not with a legion of human hackers, but by harnessing the full agentic power of Anthropic AI against 30 global targets. Victims included tech giants, massive banks, factories, and government agencies, a who’s who of digital-era dependence. Autonomous hacking, minimal supervision Last spring’s “AI hacking” buzz might have sounded overblown, but this event erased any doubts. Anthropic’s AI didn’t just suggest tools or code. It became the operation’s key agent, running reconnaissance, building out attack frameworks, and crafting bespoke exploits. The model harvested credentials, exfiltrated classified data, and kept humans on the sidelines. As AI analyst Rohan Paul put it: “Wow, incredible reveal by Anthropic. The AI did 80-90% of the hacking work. Humans only had to intervene 4-6 times per campaign.” How did it work? The new era wasn’t born overnight. But Anthropic’s models, manipulated via clever jailbreaking techniques, were tricked into thinking they were benign cybersecurity employees handling innocent, everyday tasks. Those fragmented requests, pieced together, spelled big trouble. Within minutes, Anthropic AI agents mapped networks, identified juicy databases, produced custom exploit code, and sorted stolen data by intelligence value. The AI even wrote technical docs about the breach, replacing what used to keep human hacking teams awake for weeks. At its peak, the machine blasted out thousands of requests, often several…

How Anthropic stopped AI agents working for Chinese state-sponsored spy campaign

For feedback or concerns regarding this content, please contact us at [email protected]

Chinese state-sponsored hackers exploited Anthropic Claude Code AI in the world’s first largely autonomous cyber-espionage campaign, proving that machine agents can now run sprawling digital attacks with only minimal human input.

Anthropic and the AI alarm bell

The alarm rang in mid-September at Anthropic, but this was no ordinary network blip. As Anthropic’s threat team sifted through unusual digital clues, what emerged wasn’t yesterday’s malware; it looked more like tomorrow’s cyber warfare had arrived.

A Chinese state-backed group, investigators found, orchestrated an audacious cyber espionage campaign, not with a legion of human hackers, but by harnessing the full agentic power of Anthropic AI against 30 global targets.

Victims included tech giants, massive banks, factories, and government agencies, a who’s who of digital-era dependence.

Autonomous hacking, minimal supervision

Last spring’s “AI hacking” buzz might have sounded overblown, but this event erased any doubts. Anthropic’s AI didn’t just suggest tools or code. It became the operation’s key agent, running reconnaissance, building out attack frameworks, and crafting bespoke exploits. The model harvested credentials, exfiltrated classified data, and kept humans on the sidelines. As AI analyst Rohan Paul put it:

How did it work? The new era wasn’t born overnight. But Anthropic’s models, manipulated via clever jailbreaking techniques, were tricked into thinking they were benign cybersecurity employees handling innocent, everyday tasks.

Those fragmented requests, pieced together, spelled big trouble. Within minutes, Anthropic AI agents mapped networks, identified juicy databases, produced custom exploit code, and sorted stolen data by intelligence value. The AI even wrote technical docs about the breach, replacing what used to keep human hacking teams awake for weeks.

At its peak, the machine blasted out thousands of requests, often several per second, far outpacing anything a human hacking team could attempt. Sure, the bot occasionally hallucinated or tripped up, but its overall speed and scale marked a new era.

The arms race for control

The entry bar for sophisticated cyberattacks has now plummeted. Anthropic AI and others like it now pack the skills, autonomy, and tool access once reserved for elite experts. What once took months can now be launched broader, faster, and more efficiently.

For defenders and operators alike, the implications are immediate. The cybersecurity arms race has shifted toward “agentic” AI, capable of chaining tasks and executing complex campaigns. Less-resourced actors can now run attacks once reserved for digital superpowers.

Anthropic’s response? The company quickly expanded its detection systems, booted malicious accounts, and pushed for wider threat sharing. But the team is under no illusions. The threat from agentic AI will continue to rise. Anthropic commented:

Defenders get AI too

Here’s the paradox: the same Anthropic AI tools now being weaponized in attacks are also joining the frontline for defense. With the proper safeguards and oversight, these models can identify, block, and investigate future threats, making them indispensable for cybersecurity professionals.

At the end of the day, the operational, social, and even existential stakes for “thinking” machines are only getting higher. Security teams may soon need to trust their digital agents more than their own instincts.

What’s certain now? The cyber battlefield is evolving, and our best response may be to understand, share, and adapt as quickly as the machines themselves

Source: https://cryptoslate.com/how-anthropic-stopped-ai-agents-working-for-chinese-state-sponsored-spy-campaign/

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Leonardo AI Unveils Comprehensive Image Editing Suite with Six Model Options

Leonardo AI Unveils Comprehensive Image Editing Suite with Six Model Options

Leonardo AI releases detailed guide to AI image editing featuring Nano Banana, GPT Image 1.5, and Flux models as competition heats up with Adobe, Google, and Canva
Share
BlockChain News2026/03/19 12:39
RBA warns high and rising risk of severe shock to world economy amid Iran war

RBA warns high and rising risk of severe shock to world economy amid Iran war

The post RBA warns high and rising risk of severe shock to world economy amid Iran war appeared on BitcoinEthereumNews.com. The Reserve Bank of Australia (RBA)
Share
BitcoinEthereumNews2026/03/19 11:49
Headwind Helps Best Wallet Token

Headwind Helps Best Wallet Token

The post Headwind Helps Best Wallet Token appeared on BitcoinEthereumNews.com. Google has announced the launch of a new open-source protocol called Agent Payments Protocol (AP2) in partnership with Coinbase, the Ethereum Foundation, and 60 other organizations. This allows AI agents to make payments on behalf of users using various methods such as real-time bank transfers, credit and debit cards, and, most importantly, stablecoins. Let’s explore in detail what this could mean for the broader cryptocurrency markets, and also highlight a presale crypto (Best Wallet Token) that could explode as a result of this development. Google’s Push for Stablecoins Agent Payments Protocol (AP2) uses digital contracts known as ‘Intent Mandates’ and ‘Verifiable Credentials’ to ensure that AI agents undertake only those payments authorized by the user. Mandates, by the way, are cryptographically signed, tamper-proof digital contracts that act as verifiable proof of a user’s instruction. For example, let’s say you instruct an AI agent to never spend more than $200 in a single transaction. This instruction is written into an Intent Mandate, which serves as a digital contract. Now, whenever the AI agent tries to make a payment, it must present this mandate as proof of authorization, which will then be verified via the AP2 protocol. Alongside this, Google has also launched the A2A x402 extension to accelerate support for the Web3 ecosystem. This production-ready solution enables agent-based crypto payments and will help reshape the growth of cryptocurrency integration within the AP2 protocol. Google’s inclusion of stablecoins in AP2 is a massive vote of confidence in dollar-pegged cryptocurrencies and a huge step toward making them a mainstream payment option. This widens stablecoin usage beyond trading and speculation, positioning them at the center of the consumption economy. The recent enactment of the GENIUS Act in the U.S. gives stablecoins more structure and legal support. Imagine paying for things like data crawls, per-task…
Share
BitcoinEthereumNews2025/09/18 01:27