In mid-September 2025, Anthropic’s Threat Intelligence team reported detecting and disrupting what it believes to be the first large-scale cyber-espionage campaign led primarily by an Artificial Intelligence (AI) system. The operation was attributed with high confidence to a Chinese state-sponsored group that utilized Anthropic’s Claude Code tool to execute autonomous intrusion activities targeting approximately 30 global organizations, achieving success in a limited number of cases. This development raises significant concerns regarding the potential for AI to amplify the speed and scale of cyberattacks, necessitating proactive governance and contractual safeguards when the hacker is an algorithm.
Anthropic’s findings indicate that AI executed 80 to 90 percent of the tactical work involved in the cyberattacks, such as reconnaissance, exploit development, lateral movement, credential harvesting, data parsing, and documentation, with human intervention occurring only at a few key decision points. This shift allows for operational autonomy at scale for cybercriminals. The threat actors reportedly bypassed established guardrails by impersonating employees of legitimate cybersecurity firms, convincing Claude that their actions were part of authorized penetration testing. Once these safeguards were circumvented, they employed a custom orchestration framework built around the Model Context Protocol (MCP), an open standard facilitating interactions between AI models and external systems. This framework enabled the attackers to break down complex, multi-stage attacks into smaller, routine technical tasks that appeared benign to Claude when viewed individually.
As a result, Claude executed these tasks autonomously at speeds unattainable by human operators, chaining them together into full attack sequences without revealing their malicious context. This orchestration allowed the AI to operate as an autonomous penetration-testing engine, coordinating multiple sub-agents and tools through the stages of reconnaissance, exploitation, and data exfiltration with minimal oversight. The implications for legal teams are profound, reframing critical questions of attribution, causation, duty of care, and risk allocation within the context of AI as the actor.
Attribution and liability in these scenarios pose unique challenges. Traditional frameworks for cybercrime assume human intent, but with agentic AI executing the majority of attack steps and humans merely approving certain stages, the legal landscape shifts dramatically. Disputes may arise over whether the misuse of AI tools was foreseeable and whether appropriate controls were in place to prevent their weaponization. Companies using such tools will be scrutinized to determine if they implemented risk-proportionate controls, as indicated by Anthropic’s report which highlights how orchestrated prompts can manipulate agent behavior. This could strengthen arguments that misuse was predictable, particularly in light of public advisories regarding prompt injection, a prevalent cause of AI incidents.
Moreover, vendors may face product liability claims if they deploy tools capable of autonomous actions without sufficient guardrails. Plaintiffs might cite patterns identified by security researchers, indicating an increase in real-world AI security failures to argue that these risks were both recognized and foreseeable. In contractual matters, existing warranties and representations regarding security may be invoked in cases of misuse, particularly if organizations failed to disclose known limitations or adhere to recognized governance standards.
As regulatory frameworks evolve, the European Union’s Artificial Intelligence Act is being implemented in phases, imposing risk-based obligations on high-risk AI, including post-market monitoring and incident reporting. This means breaches involving agentic AI misuse could trigger provider and deployer duties under the Act. In the U.S., the Federal Trade Commission is actively pursuing “Operation AI Comply,” indicating that there will be no exemptions from existing laws related to AI.
Even as the legal landscape develops, companies may face inquiries into whether their accounts or tools were misused. Anthropic’s case illustrates the potential for automated AI agents to execute cyberattacks on a vast scale. Additionally, compliance with privacy laws, such as the European General Data Protection Regulation (GDPR), becomes more complex when AI-driven incidents involve personal data breaches. GDPR mandates that organizations notify supervisory authorities within 72 hours of becoming aware of such breaches, reinforcing the need for stringent controls.
In light of these emerging challenges, organizations are advised to revisit their contractual agreements, ensuring they incorporate AI-focused security provisions. Contracts should include clear representations and warranties confirming that vendors conduct robust adversarial testing and maintain documented processes for safe model updates and rollback. Administrative safeguards, such as kill-switch capabilities and human oversight for privileged actions, should also be mandated. Companies must review their cyber insurance policies to ensure they adequately cover AI-driven incidents, as underwriters may request evidence of aligned controls.
Ultimately, Anthropic’s findings signal a turning point in cybersecurity where agentic AI can compress attack timelines and scale operations with reduced human involvement. This evolution reshapes expectations surrounding duty of care and disclosure risk when “the hacker” is an algorithm. As AI governance and framework-aligned controls become paramount, organizations must adopt rigorous oversight to navigate this new landscape of digital threats.
See also
Thomson Reuters Launches CoCounsel Legal, Reducing Document Review Time by 63%
Pennsylvania Advances AI Regulation Bill as Healthcare Integrates AI Tools Daily
TRM Labs and Sphinx Partner to Automate AML Compliance with AI Agents, Enhancing Fraud Detection
Trump Blocks State AI Regulations, Empowering Tech Firms Amid Safety Concerns
UK Entrepreneurs Demand Clearer AI Rules as 73% Face Digital Trust Challenges



















































