Connect with us

Hi, what are you looking for?

AI Cybersecurity

OpenAI’s GPT-5.5 Matches Claude Mythos in Cyberattack Efficiency, Solves Puzzles in 10 Minutes

OpenAI’s GPT-5.5 autonomously executed complex cyberattacks with a 71.4% pass rate, raising alarms as U.K. officials unveil £90M to enhance cyber resilience.

A U.K. government agency has reported that OpenAI’s latest artificial intelligence model, GPT-5.5, can autonomously execute complex cyberattacks, completing a 32-step corporate network simulation in two out of ten attempts. This simulation, known as “The Last Ones,” was conducted by the AI Security Institute (AISI), part of Britain’s Department of Science, Innovation and Technology, and was designed in collaboration with the cybersecurity firm SpecterOps. The findings, published Thursday, raise significant concerns regarding the implications of advanced AI capabilities in cybersecurity.

The report indicated that GPT-5.5 demonstrated offensive cyber capabilities comparable to those of Anthropic’s Claude Mythos. In a particularly notable challenge, GPT-5.5 cracked a reverse-engineering puzzle in just over ten minutes, a task that took a human security expert approximately twelve hours. This puzzle required the AI to reconstruct a custom virtual machine’s instruction set and recover a cryptographic password, showcasing the model’s advanced problem-solving abilities.

On AISI’s rigorous evaluation, GPT-5.5 achieved an average pass rate of 71.4% on the most difficult “Expert” tier of advanced cybersecurity tasks. This performance surpassed that of Claude Mythos Preview, which had a pass rate of 68.6%, and significantly exceeded the previous model, GPT-5.4, which managed only 52.4%. These results suggest that the rapid improvement of offensive AI capabilities could be part of a broader trend rather than an isolated incident.

The findings also underscore serious safety concerns. Researchers discovered a universal jailbreak that allowed GPT-5.5 to bypass its safety guardrails entirely, generating harmful content across various cyber queries. This vulnerability, developed through six hours of expert red-teaming, prompted OpenAI to update its safeguard stack. However, a configuration issue prevented AISI from verifying whether the updated measures were effective.

While AISI’s evaluations were carried out under controlled conditions, the report cautioned that such capabilities may not reflect those available to the average user, as public deployments are equipped with additional safeguards and access controls. The implications of these findings are particularly pressing in light of the U.K. government’s annual Cyber Security Breaches Survey, which found that 43% of businesses reported suffering a cyber breach or attack in the past year.

In response to the escalating cybersecurity threats, the U.K. government announced £90 million in new funding aimed at bolstering cyber resilience. Additionally, officials are advancing the Cyber Security and Resilience Bill to protect essential services. They have urged organizations to prepare for a potential increase in newly discovered software vulnerabilities, as AI technologies like GPT-5.5 accelerate the pace at which security flaws can be identified and exploited.

The report’s findings raise critical questions about the future trajectory of AI development and its potential role in offensive cyber capabilities. AISI’s conclusions suggest that rapid advancements in reasoning, coding, and autonomous task execution may inadvertently contribute to the evolution of offensive cyber skills. If this trend continues, further advancements in AI-enhanced cyber capabilities could emerge quickly, posing significant risks to organizations and individuals alike.

See also
Rachel Torres
Written By

At AIPressa, my work focuses on exploring the paradox of AI in cybersecurity: it's both our best defense and our greatest threat. I've closely followed how AI systems detect vulnerabilities in milliseconds while attackers simultaneously use them to create increasingly sophisticated malware. My approach: explaining technical complexities in an accessible way without losing the urgency of the topic. When I'm not researching the latest AI-driven threats, I'm probably testing security tools or reading about the next attack vector keeping CISOs awake at night.

You May Also Like

Top Stories

Anthropic expands Claude Mythos AI into Japan amid U.S. government scrutiny over potential national security risks and AI misuse concerns.

AI Generative

OpenAI tests GPT 5.6 in Codex, aiming to enhance AI-driven coding efficiency and cybersecurity, potentially reshaping the developer landscape.

AI Cybersecurity

OpenAI's GPT-5.5 outperformed Claude Mythos Preview in cyberattack simulations, achieving a 71.4% success rate in expert-level tasks, raising cybersecurity concerns.

AI Research

OpenAI introduces explainable AI techniques to detect machine-generated music, enhancing authenticity measures amid rising public concerns over AI's creative impact.

AI Generative

OpenAI's GPT Image2 spurs a viral trend as users prompt AI to create MS Paint-style drawings from photos, challenging artistic norms and expectations.

AI Generative

Pinterest slashes its AI budget by 90% while adopting a hybrid model with OpenAI and Alibaba, enhancing user experience and cost efficiency.

AI Cybersecurity

Anthropic's Claude model has identified over 1,000 zero-day vulnerabilities in major software systems, revolutionizing cybersecurity and defense strategies.

AI Regulation

AI safety standards are at risk as Anthropic and OpenAI cut safety commitments amid competition, despite 80% of U.S. adults prioritizing regulation over innovation...

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.