Connect with us

Hi, what are you looking for?

Top Stories

Anthropic Warns of AI-Driven Cyberattacks as Cybersecurity Reaches Critical Inflection Point

Anthropic reveals AI-powered cyberattacks are escalating, with its model Claude autonomously executing 80-90% of a sophisticated espionage campaign against global targets.

Anthropic has raised alarms regarding a pivotal moment in cybersecurity, noting that advancements in AI models are serving both defensive and offensive capacities in cyber operations. This warning comes on the heels of reports indicating that state-sponsored hackers from China utilized Anthropic’s technology to streamline their intrusions into major corporations and foreign governments during a hacking spree in September.

In a recent research report, Anthropic stated, “As part of our Safeguards work, we have found and disrupted threat actors on our own platform who leveraged AI to scale their operations.” The company detailed a case of “vibe hacking,” where a cybercriminal employed its AI model, Claude, to orchestrate an extensive data extortion scheme that traditionally would have required a full team. The Safeguards team also reported thwarting Claude’s use in complex espionage activities targeting vital telecommunications infrastructure, echoing tactics associated with Chinese APT operations.

Over the past year, a noticeable transition has emerged, as demonstrated in Anthropic’s findings. The company’s AI models were able to simulate one of the most costly cyberattacks in history, the 2017 Equifax breach. Furthermore, Claude has participated in cybersecurity competitions, at times outperforming human teams. This technology has also played a crucial role in identifying vulnerabilities in Anthropic’s own code, allowing those issues to be addressed prior to deployment.

In mid-September, Anthropic detected suspicious activities that led to the identification of an advanced espionage campaign. The attackers exploited AI’s agentic capabilities, employing the technology not merely as a consultant but as a direct executor of attacks.

See alsoIndia Unveils AI Governance Guidelines; Amazon Fights AI Browser PurchasesIndia Unveils AI Governance Guidelines; Amazon Fights AI Browser Purchases

Investigations revealed that the threat actor, assessed with high confidence as a Chinese state-sponsored group, manipulated Claude Code to infiltrate approximately thirty global targets, achieving successful breaches in a few instances. Traditional targets included tech firms, financial institutions, chemical manufacturers, and government agencies, marking this as a potentially unprecedented large-scale cyberattack executed with minimal human intervention.

Following the detection, Anthropic promptly initiated an investigation to ascertain the extent and nature of the operation. Over the course of ten days, the team mapped the entire campaign, banned compromised accounts, and coordinated with relevant authorities while amassing actionable intelligence.

The report highlighted a critical observation: “Agents are valuable for everyday work and productivity—but in the wrong hands, they can substantially increase the viability of large-scale cyberattacks.” As such, the company has expanded its detection capabilities and improved classifiers for identifying malicious activities, reiterating their commitment to developing new methods for investigating large-scale cyber threats.

Anthropic’s review indicates that cyber capabilities are doubling every six months, with real-world attacks increasingly leveraging AI to exploit vulnerabilities. Their Threat Intelligence team recently intercepted a threat campaign, emphasizing the need for collaborative industry efforts in threat sharing and enhanced safeguards to counter adversarial misuse of AI technology.

The recent attacks utilized several AI capabilities that were either nonexistent or nascent just a year ago. The advanced general intelligence of these models enables them to follow complex instructions and grasp context, allowing them to perform sophisticated tasks. Specifically, their coding skills make them particularly adept at facilitating cyberattacks.

Moreover, these models can act as autonomous agents, executing tasks in loops with limited human input. They now have access to an array of software tools through the open standard Model Context Protocol, allowing them to execute actions that previously required human intervention, including using password crackers and network scanners.

In the initial phase of the attack, human operators selected targets and crafted an attack framework relying on Claude Code as an automated tool. To bypass Claude’s safeguards against harmful behavior, attackers broke down the operation into smaller, seemingly innocuous tasks, misleading the AI by posing as employees of a legitimate cybersecurity firm.

In subsequent phases, Claude conducted reconnaissance on target systems, swiftly identifying high-value databases and reporting findings to its human operators—accomplishing in minutes what would take human teams significantly longer. Later, Claude generated exploit code and performed security vulnerability tests autonomously, harvesting credentials and extracting vast amounts of sensitive data with minimal human oversight.

Overall, Anthropic noted that the AI executed 80-90% of the campaign independently, requiring human intervention only at critical decision points. The attack, characterized by an unprecedented operational tempo, involved thousands of requests per second, a feat that human hackers could not match.

Despite Claude’s remarkable capabilities, the report acknowledged that the AI does not operate flawlessly and occasionally “hallucinates” information. However, it demonstrated extensive autonomous functionality throughout the operation phases, revealing a fundamental shift in cybersecurity dynamics.

Anthropic has urged security teams to experiment with leveraging AI for defense in areas such as Security Operations Center automation, threat detection, and incident response. “We must not cede the cyber advantage derived from AI to attackers and criminals,” the report emphasized, advocating for enhanced investment in safeguards across AI platforms.

Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

At the 2025 Cerebral Valley AI Conference, over 300 attendees identified AI search startup Perplexity and OpenAI as the most likely to falter amidst...

Top Stories

OpenAI's financial leak reveals it paid Microsoft $493.8M in 2024, with inference costs skyrocketing to $8.65B in 2025, highlighting revenue challenges.

AI Cybersecurity

Anthropic"s report of AI-driven cyberattacks faces significant doubts from experts.

Top Stories

Microsoft's Satya Nadella endorses OpenAI's $100B revenue goal by 2027, emphasizing urgent funding needs for AI innovation and competitiveness.

AI Technology

Cities like San Jose and Hawaii are deploying AI technologies, including dashcams and street sweeper cameras, to reduce traffic fatalities and improve road safety,...

AI Business

Satya Nadella promotes AI as a platform for mutual growth and innovation.

AI Technology

Shanghai plans to automate over 70% of its dining operations by 2028, transforming the restaurant landscape with AI-driven kitchens and services.

AI Government

AI initiatives in Hawaii and San Jose aim to improve road safety by detecting hazards.

Generative AI

OpenAI's Sam Altman celebrates ChatGPT"s new ability to follow em dash formatting instructions.

AI Technology

An MIT study reveals that 95% of generative AI projects fail to achieve expected results

AI Technology

Meta will implement 'AI-driven impact' in employee performance reviews starting in 2026, requiring staff to leverage AI tools for productivity enhancements.

AI Technology

Andrej Karpathy envisions self-driving cars reshaping cities by reducing noise and reclaiming space.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.