Anthropic study says AI agents developed $4.6M in smart contract bugs


Anthropic study says AI agents developed .6M in smart contract bugs


Recent research by major artificial intelligence company Antropic and AI security organization Machine Learning Alignment & Theory Scholars (MATS) showed that AI agents collectively developed smart contract exploits worth $4.6 million.

Research released by Anthropic’s red team (a team dedicated to acting like a bad actor to discover potential for abuse) on Monday found that currently available commercial AI models are significantly capable of exploiting smart contracts.

Anthropic’s Claude Opus 4.5, Claude Sonnet 4.5, and OpenAI’s GPT-5 collectively developed exploits worth $4.6 million when tested on contracts, exploiting them after their most recent training data was gathered.

Researchers also tested both Sonnet 4.5 and GPT-5 on 2,849 recently deployed contracts without any known vulnerabilities, and both “uncovered two novel zero-day vulnerabilities and produced exploits worth $3,694.” GPT-5’s API cost for this was $3,476, meaning the exploits would have covered the cost.

“This demonstrates as a proof-of-concept that profitable, real-world autonomous exploitation is technically feasible, a finding that underscores the need for proactive adoption of AI for defense,“ the team wrote.

Chart of AI exploiting revenue from simulations. Source: Anthropic

Related: UXLink hack turns ironic as attacker gets phished mid-exploit

An AI smart contract hacking benchmark

Researchers also developed the Smart Contracts Exploitation (SCONE) benchmark, comprising 405 contracts that were actually exploited between 2020 and 2025. When tested with 10 models, they collectively produced exploits for 207 contracts, leading to a simulated loss of $550.1 million.

Researchers also suggested that the output required (measured in tokens in the AI industry) for an AI agent to develop an exploit will decrease over time, thereby reducing the cost of this operation. “Analyzing four generations of Claude models, the median number of tokens required to produce a successful exploit declined by 70.2%,” the research found.

Average number of AI output tokens per explot per model. Source: Anthropic

Related: Coinbase’s preferred AI coding tool can be hijacked by new virus

AI smart contract hacking capabilities are growing fast

The study argues that AI capabilities in this area are improving at a rapid pace.

“In just one year, AI agents have gone from exploiting 2% of vulnerabilities in the post-March 2025 portion of our benchmark to 55.88%—a leap from $5,000 to $4.6 million in total exploit revenue,” the team claims. Furthermore, most of the smart contract exploits of this year “could have been executed autonomously by current AI agents.”

The research also showed that the average cost to scan a contract for vulnerabilities is $1.22. Researchers believe that with falling costs and rising capabilities, “the window between vulnerable contract deployment and exploitation will continue to shrink.” Such a situation would leave developers less time to detect and patch vulnerabilities before they are exploited.

Magazine: North Korea crypto hackers tap ChatGPT, Malaysia road money siphoned: Asia Express