Connect with us

Hi, what are you looking for?

AI Cybersecurity

AI Cyberattacks Now Cost Under $50, Raising Security Risks for Every Company

AI-driven cyberattacks can now be executed for under $50, exposing organizations to unprecedented security risks, according to Wiz and Irregular’s groundbreaking study.

Artificial intelligence is reshaping the landscape of cybersecurity, making cyberattacks more accessible and cost-effective for malicious actors, according to a recent study by cloud security startup Wiz and AI security lab Irregular. The findings underscore a significant shift in the economic dynamics of cybercrime, where AI-driven attacks can now be executed for as little as $50, a stark contrast to the nearly $100,000 typically spent on human-driven efforts to identify security vulnerabilities.

In an interview, Gal Nagli, head of threat exposure at Wiz, and Omer Nevo, cofounder and CTO of Irregular, detailed the implications of their joint research, which challenges traditional perceptions of the costs associated with cyberattacks. AI agents successfully completed complex offensive security tasks in controlled environments, solving 90% of real-world attack scenarios modeled in their tests. Nevo noted that even seasoned cybersecurity professionals were taken aback by the rapid capabilities of AI models, particularly in managing multi-step challenges without losing focus. “We’re seeing more and more that models are able to solve challenges that are genuine expert level,” he said.

This emergence of affordable cyber threats is compounded by the increasing involvement of non-technical employees in developing applications using accessible coding tools like Anthropic’s Claude Code and OpenAI’s Codex. As Nagli pointed out, these individuals often lack essential security knowledge, creating a vast attack surface that cybercriminals are poised to exploit. “They don’t know anything about security,” he explained, “and they use sensitive data exposed to the public Internet, and then they are super easy to exploit.”

The research indicates that the traditional cat-and-mouse game of cybersecurity is no longer limited by cost barriers. Hackers are no longer required to selectively target their efforts; AI-driven tools enable them to probe vulnerabilities across numerous systems for just a few dollars. This new economic reality implies that every exposed system is a potential target, raising the stakes for organizations, regardless of size.

While the study noted a drop in performance and a doubling of costs in more realistic conditions, the overarching message is clear—cyberattacks are becoming cheaper and faster to deploy. Nevo cautioned that if AI can conduct sophisticated attacks at scale, the potential for harm could increase dramatically. “If we reach the point where AI is able to conduct sophisticated attacks, and it’s able to do that at scale, suddenly a lot more people will be exposed,” he warned, prompting a need for greater cybersecurity awareness across industries.

In light of this evolving threat landscape, the integration of AI into defensive strategies is increasingly crucial. Nevo raised an important question: “Are we helping defenders utilize AI fast enough to be able to keep up with what offensive actors are already doing?” The urgency of addressing this imbalance is paramount as organizations navigate a new era of technological vulnerability.

In other AI-related news, a U.S. lawmaker has raised concerns regarding Nvidia‘s alleged assistance to the Chinese AI startup DeepSeek, suggesting that the company aided in refining AI models subsequently used by the Chinese military. This revelation has sparked discussions about the need for stricter export controls to prevent American technology from being repurposed for military applications.

Additionally, Dow Chemical announced plans to cut 4,500 jobs as part of an AI-driven overhaul aimed at enhancing productivity amid rising costs and declining revenue. The company’s “Transform to Outperform” initiative is expected to generate significant operational earnings while incurring considerable one-time charges, including severance costs. This move follows a broader trend of corporate layoffs across industries as companies pivot toward AI and automation.

Finally, Anthropic has started a controversial project called “Project Panama,” involving the destructive scanning of millions of books to gather data for training AI models. Internal documents suggest that the company has invested heavily in acquiring physical books, raising ethical questions about copyright and the lengths to which AI companies will go to secure quality data.

As the landscape of cybersecurity continues to evolve with AI, the interplay of offensive and defensive capabilities will shape future strategies and policies. The implications of these developments extend beyond corporate boardrooms, signaling a need for comprehensive awareness and adaptation in a world increasingly susceptible to AI-driven threats.

See also
Rachel Torres
Written By

At AIPressa, my work focuses on exploring the paradox of AI in cybersecurity: it's both our best defense and our greatest threat. I've closely followed how AI systems detect vulnerabilities in milliseconds while attackers simultaneously use them to create increasingly sophisticated malware. My approach: explaining technical complexities in an accessible way without losing the urgency of the topic. When I'm not researching the latest AI-driven threats, I'm probably testing security tools or reading about the next attack vector keeping CISOs awake at night.

You May Also Like

AI Cybersecurity

AI-driven cloud security tools like Wiz and Prisma Cloud empower businesses to enhance protection across AWS, Azure, and Google Cloud, streamlining threat responses significantly.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.