In a concerning revelation, Anthropic reported in November 2025 that Chinese threat actors exploited its Claude model to orchestrate extensive cyberattacks against various companies and government entities. This incident involved the jailbreaking of Anthropic’s coding tool, Claude Code, enabling attackers to target 30 organizations worldwide. This marked an alarming milestone as the first known large-scale cyber campaign executed with minimal human intervention. While this incident was detected due to Anthropic’s internal monitoring systems, it raises a greater concern regarding undetected future attacks leveraging similar AI capabilities.
The emergence of AI agents, capable of performing tasks autonomously, enhances the capabilities of both cyber attackers and defenders. These AI agents can facilitate faster and broader attacks, but they also empower defenders to detect intrusions and respond swiftly. However, the rapid adoption of offensive capabilities by malicious actors, who often take risks, indicates that such incidents may become less of an anomaly and more of a pattern.
This evolving landscape highlights a significant vulnerability: the U.S. government lacks a systematic approach to determine whether a cyberattack results from novel AI capabilities or traditional methods. The inability to discern this distinction could hinder its preparedness for emerging AI risks. Without effective detection and investigative measures for AI-enabled incidents, the government risks falling behind in adapting its cyber defenses and updating threat assessments.
Anthropic’s report shed light on AI-enabled threats originating from its platform, but it has no visibility into threats from other platforms, particularly those associated with increasingly capable open-source AI models. Chinese open-weight models, such as those from DeepSeek, are rapidly progressing and offer capabilities that can be freely accessed and run without oversight. According to the Center for AI Standards and Innovation, DeepSeek’s R1-0528 model demonstrates a 12-fold higher likelihood of following malicious instructions compared to U.S. models like OpenAI’s GPT-5 and Claude’s Opus 4. This accessibility heightens the risk of exploitation, especially as leading open models primarily arise from China, where the U.S. government has limited visibility and cooperation opportunities.
The opacity surrounding these developments is not exclusive to AI. Historical instances, such as the 2016 Australian online census debacle, illustrate the difficulties in understanding technical failures. Initially suspected to be a sophisticated state-sponsored attack, it was ultimately revealed to be the result of poor implementation. This incident underscores the challenges governments face in tracing the causes of digital system failures, a problem that persists nearly a decade later.
As organizations reportedly take an average of eight months to identify and contain a data breach, the introduction of AI threatens to amplify the speed and scale of cyberattacks, potentially leading to even more convoluted investigations that cannot keep pace.
Despite these challenges, the U.S. government has a model for enhancing transparency in technical incidents. The Cyber Safety Review Board (CSRB), established in 2022, successfully brought together federal agencies and private companies to investigate significant cyber incidents. In 2023, the Board conducted a thorough investigation following a breach of Microsoft’s cloud infrastructure by state-backed Chinese hackers, revealing a series of “avoidable errors” by Microsoft. This investigation not only exposed technical failures but also held Microsoft accountable, prompting the company to adopt improvements based on the Board’s recommendations.
However, the CSRB faced limitations, including resource constraints and a lack of subpoena power. The Trump administration dissolved the board in early 2025, aiming to cut down on perceived resource misuses. Even with its shortcomings, the CSRB exemplified how independent, cross-sector investigations could foster accountability and lead to enhanced security measures across industries.
What Comes Next
In light of the increasing AI-enabled threats, the U.S. needs to establish an AI Security Review Board (AISRB), modeled after the CSRB but equipped to track and investigate AI’s role in cyber incidents. This board should operate independently and include full-time experts from the federal government, technology industry, and civil society, focusing on AI systems and their potential risks. By publishing findings publicly, the AISRB would enhance accountability and drive improvements across sectors while complementing existing initiatives like the Center for AI Standards and Innovation and the National Security Agency’s Artificial Intelligence Security Center.
Moreover, the proposed AISRB would be crucial for identifying emerging AI threats and ensuring accountability when systems fail. To function effectively, the AISRB must possess the authority and resources that the CSRB lacked, including sufficient funding and investigative powers. As open-source AI technologies proliferate, this board becomes essential for recognizing dual-use capabilities and their implications in the wild.
Beyond the establishment of the AISRB, stronger information-sharing mechanisms between the government, industry, and civil society are imperative. Effective cooperation relies on robust legal protections for companies sharing sensitive information regarding AI-enabled attacks, which is facilitated by the Cybersecurity Information Sharing Act of 2015 (CISA 2015). Recently extended until September 2026, CISA 2015 is crucial for facilitating ongoing dialogue between the government and private sector regarding cyber security.
As critical infrastructure becomes more digitized, the United States faces a growing risk of cyberattacks. The methods showcased in the Anthropic incident are likely to proliferate as AI continues to evolve. To safeguard national security, it is vital for the U.S. to implement detection capabilities, investigative infrastructure, and information-sharing channels before the next potential crisis unfolds. The AISRB and the renewal of CISA 2015 are essential steps towards enhancing preparedness for a rapidly changing cyber threat environment.
See also
Anthropic’s Claims of AI-Driven Cyberattacks Raise Industry Skepticism
Anthropic Reports AI-Driven Cyberattack Linked to Chinese Espionage
Quantum Computing Threatens Current Cryptography, Experts Seek Solutions
Anthropic’s Claude AI exploited in significant cyber-espionage operation
AI Poisoning Attacks Surge 40%: Businesses Face Growing Cybersecurity Risks

















































