Recent advances in generative AI have drastically changed the landscape of cybersecurity, enabling rapid exploitation of software vulnerabilities that previously took months to leverage. As highlighted by Anthropic’s Project Glasswing, large language models (LLMs) can facilitate cyberattacks in mere minutes and often for less than a dollar in cloud computing costs. While this development poses significant threats to cyberspace, it also presents an opportunity for defenders to bolster their cyber defenses.
Anthropic reports that its Claude Mythos preview model has already assisted in the preemptive discovery of over a thousand zero-day vulnerabilities, revealing flaws in major operating systems and web browsers. This proactive approach to vulnerability disclosure and patching illustrates that while AI tools can empower attackers, they can also be harnessed to strengthen security measures. However, the ultimate impact of AI on cybersecurity—whether it favors attackers or defenders—remains uncertain.
The emergence of fuzzers in the early 2010s provides a historical context for understanding AI’s role in vulnerability discovery. Programs like American Fuzzy Lop (AFL) utilized random, malformed inputs to find critical flaws across major software platforms. In response, the security community adapted by industrializing defenses. Initiatives, such as Google’s OSS-Fuzz, continuously run fuzzing tests on thousands of software projects, allowing developers to catch vulnerabilities before they can be exploited. It is expected that the integration of AI-driven discovery tools will follow a similar trajectory, becoming standard in development practices to elevate security baselines.
However, the analogy has limitations. Setting up and managing fuzzing tools requires significant technical expertise, making them accessible mainly to specialists. In contrast, LLMs can identify vulnerabilities with simple prompts, creating a troubling disparity. Attackers now face lower entry barriers to exploit code, while defenders still require skilled engineers to interpret AI findings and implement fixes. This disparity in effort required highlights a shift in the landscape where the human cost of exploiting vulnerabilities may approach zero, while addressing them remains resource-intensive.
Is AI Better at Finding Bugs Than Fixing Them?
Peter Gutmann’s observation in his book Engineering Security—that many security technologies remain “secure” simply because no one has examined them—was made before the advent of AI making vulnerability discovery significantly cheaper. Presently, much of the code underpinning commercial software relies on small teams, part-time contributors, or volunteers who lack dedicated security resources. A single bug in an open-source project can have widespread consequences, as demonstrated by the Log4j vulnerability in 2021, which affected hundreds of millions of devices. This incident underscores the vulnerability of critical software dependencies that often go unreviewed. AI-driven vulnerability discovery can potentially automate extensive auditing, although it also facilitates attackers targeting under-resourced projects.
Research into LLM-assisted exploit generation reveals that these models can autonomously and rapidly exploit vulnerabilities, compressing the window from disclosure to exploitation from weeks to mere hours. For instance, researchers at NYU’s Tandon School of Engineering found that an LLM-based system could autonomously execute major phases of a ransomware campaign for about $0.70 per run, requiring no human involvement. This starkly contrasts with the defender’s role, which demands careful evaluation and implementation of fixes—tasks that remain labor-intensive and complex to manage for those maintaining critical libraries in their spare time.
Policymakers have responded to the rising threat of AI-driven attacks by considering regulations that hold AI companies accountable for misuse and enhancing product guardrails. While some evidence suggests that automated misuse detection can divert cyberattacks, such measures alone cannot provide comprehensive solutions. The primary challenges stem from two factors: first, LLMs assess the maliciousness of requests based on their wording, which can be cleverly disguised; second, regulations limited to specific jurisdictions leave a global problem largely unaddressed, as robust open-source LLMs are readily accessible online.
Another proposed solution is to automate the defensive processes entirely. Tools like GitHub Copilot Autofix can generate patches for flagged vulnerabilities, and various open-source initiatives are experimenting with autonomous AI maintainers for under-resourced projects. However, LLM-generated patches can introduce subtle errors, and the absence of human oversight creates vulnerabilities that malicious actors could exploit. These automated approaches, while useful, do not address whether software is fundamentally secure from its inception.
The adoption of memory-safe languages represents a more robust preventive measure. Both Google and Microsoft have found that approximately 70 percent of serious security flaws arise from memory management issues, primarily in languages like C and C++. Switching to memory-safe languages like Rust can mitigate these risks. Yet, legacy codebases will persist for years, necessitating complementary solutions such as software sandboxing techniques to contain existing vulnerabilities. Tools like WebAssembly demonstrate how effective sandboxing can be implemented, but they still require careful execution to remain secure.
As cybersecurity faces an evolving threat landscape, organizations must look beyond short-term fixes. A more foundational approach involves adopting memory-safe programming languages, implementing strong sandboxing measures, and utilizing formal verification for critical components. By ensuring that software is built with security in mind from the outset, rather than merely relying on AI to discover vulnerabilities afterward, the industry can create a substantially more secure environment. Generative AI can facilitate this transition, enabling the translation of legacy code into safer languages and assisting in formal verification efforts. Ultimately, the goal is to build systems that inherently avoid vulnerabilities rather than merely treating their symptoms.
See also
Anthropic’s Claims of AI-Driven Cyberattacks Raise Industry Skepticism
Anthropic Reports AI-Driven Cyberattack Linked to Chinese Espionage
Quantum Computing Threatens Current Cryptography, Experts Seek Solutions
Anthropic’s Claude AI exploited in significant cyber-espionage operation
AI Poisoning Attacks Surge 40%: Businesses Face Growing Cybersecurity Risks


















































