In short
- Anthropic examined ten AI fashions on 405 historic sensible contract exploits and reproduced 207 of them.
- Three fashions generated $4.6 million in simulated exploits on contracts created after their coaching cutoff.
- Brokers additionally found two new zero-day vulnerabilities in current Binance Sensible Chain contracts.
AI brokers matched the efficiency of expert human attackers in additional than half of the sensible contract exploits recorded on main blockchains during the last 5 years, in accordance with new knowledge launched Monday by Anthropic.
Anthropic evaluated ten frontier fashions, together with Llama 3, Sonnet 3.7, Opus 4, GPT-5, and DeepSeek V3, on a dataset of 405 historic sensible contract exploits. The brokers produced working assaults in opposition to 207 of them, totaling $550 million in simulated stolen funds.
The findings confirmed how shortly automated programs can weaponize vulnerabilities and establish new ones that builders haven’t addressed.
The brand new disclosure is the most recent from the developer of Claude AI. Final month, Anthropic detailed how Chinese language hackers used Claude Code to launch what it known as the primary AI-driven cyberattack.
Safety specialists stated the outcomes confirmed how accessible many of those flaws already are.
“AI is already being utilized in ASPM instruments like Wiz Code and Apiiro, and in commonplace SAST and DAST scanners,” David Schwed, COO of SovereignAI, informed Decrypt. “Meaning unhealthy actors will use the identical expertise to establish vulnerabilities.”
Schwed stated the model-driven assaults described within the report can be easy to scale as a result of many vulnerabilities are already publicly disclosed by Widespread Vulnerabilities and Exposures or audit studies, making them learnable by AI programs and straightforward to try in opposition to present sensible contracts.
“Even simpler can be to discover a disclosed vulnerability, discover tasks that forked that undertaking, and simply try that vulnerability, which can not have been patched,” he stated. “This may all be achieved now 24/7, in opposition to all tasks. Even these now with smaller TVLs are targets as a result of why not? It’s agentic.”
To measure present capabilities, Anthropic plotted every mannequin’s complete exploit income in opposition to its launch date utilizing solely the 34 contracts exploited after March 2025.
“Though complete exploit income is an imperfect metric—since just a few outlier exploits dominate the entire income—we spotlight it over assault success price as a result of attackers care about how a lot cash AI brokers can extract, not the quantity or problem of the bugs they discover,” the corporate wrote.
Anthropic didn’t instantly reply to requests for remark by Decrypt.
Anthropic stated it examined the brokers on a zero-day dataset of two,849 contracts drawn from greater than 9.4 million on Binance Sensible Chain.
The corporate stated Claude Sonnet 4.5 and GPT-5 every uncovered two undisclosed flaws that produced $3,694 in simulated worth, with GPT-5 attaining its outcome at an API value of $3,476. Anthropic famous that every one exams ran in sandboxed environments that replicated blockchains and never actual networks.
Its strongest mannequin, Claude Opus 4.5, exploited 17 of the post-March 2025 vulnerabilities and accounted for $4.5 million of the entire simulated worth.
The corporate linked enhancements throughout fashions to advances in device use, error restoration, and long-horizon activity execution. Throughout 4 generations of Claude fashions, token prices fell by 70.2%.
One of many newly found flaws concerned a token contract with a public calculator perform that lacked a view modifier, which allowed the agent to repeatedly alter inner state variables and promote inflated balances on decentralized exchanges. The simulated exploit generated about $2,500.
Schwed stated the problems highlighted within the experiment had been “actually simply enterprise logic flaws,” including that AI programs can establish these weaknesses when given construction and context.
“AI may also uncover them given an understanding of how a sensible contract ought to perform and with detailed prompts on methods to try to avoid logic checks within the course of,” he stated.
Anthropic stated the capabilities that enabled brokers to use sensible contracts additionally apply to different forms of software program, and that falling prices will shrink the window between deployment and exploitation. The corporate urged builders to undertake automated instruments of their safety workflows so defensive use advances as shortly as offensive use.
Regardless of Anthropic’s warning, Schwed stated the outlook shouldn’t be solely damaging.
“I at all times push again on the doom and gloom and say with correct controls, rigorous inner testing, together with real-time monitoring and circuit breakers, most of those are avoidable,” he stated. “The Good actors have the identical entry to the identical brokers. So if the unhealthy actors can discover it, so can the nice actors. We now have to suppose and act in a different way.”
Typically Clever E-newsletter
A weekly AI journey narrated by Gen, a generative AI mannequin.

