Connect with us

Hi, what are you looking for?

AI Generative

UK Cyber Agency Warns Persistent Prompt Injection Flaw Threatens LLM Security

UK’s NCSC warns that prompt injection flaws in LLMs like ChatGPT could be exploited by attackers, posing serious risks to users and software security.

The UK’s National Cyber Security Centre (NCSC) issued a caution on Monday regarding the inherent vulnerabilities of large language model (LLM) artificial intelligence tools, warning that malicious actors could exploit these weaknesses to hijack and potentially weaponize these models against users. This advisory comes three years after the launch of ChatGPT, a widely used LLM, which has been under scrutiny by security researchers for its functionality, privacy, and security.

Researchers quickly identified a significant flaw: LLMs, including ChatGPT, process all prompts as instructions, making them susceptible to manipulation through a tactic known as prompt injection. This method involves sending harmful requests disguised as legitimate instructions, allowing attackers to bypass internal safeguards meant to prevent dangerous actions.

In a blog post, David C, the NCSC’s technical director for platforms research, explained that the architecture of current LLMs inherently lacks a security distinction between trusted and untrusted content. “Current large language models (LLMs) simply do not enforce a security boundary between instructions and data inside a prompt,” he noted. The models concatenate their own instructions with untrusted content, treating the resulting prompt as if it were free from risk.

David C cautioned that prompt injection attacks could prove more challenging to mitigate than other known vulnerabilities, such as SQL injection, which impacts web applications mishandling data and commands. He emphasized that LLMs operate through pattern matching and prediction, lacking the ability to discern trustworthy information from malicious input. “Under the hood of an LLM, there’s no distinction made between ‘data’ or ‘instructions’; there is only ever ‘next token’,” he wrote. This means that prompt injection attacks may persist as a significant threat.

The NCSC’s assessment echoes sentiments from independent researchers and AI companies, which have warned that issues like prompt injections, jailbreaking, and hallucinations may never be fully resolved. As LLMs retrieve content from the internet or external sources, there remains a risk that they will interpret this data as direct instructions.

The implications of these vulnerabilities extend into the realm of software development. Major AI coding tools from companies like OpenAI and Anthropic have been integrated into automated workflows on platforms like GitHub, creating potential weaknesses. Maintainers or external contributors could embed malicious prompts within standard elements such as commit messages, which the LLMs would then accept as valid instructions. Even models that require human approval for significant tasks could be exploited with a single line of malicious code.

AI browser agents, designed to assist users in shopping and research, are similarly prone to vulnerabilities. Researchers have discovered ways to exploit ChatGPT’s browser authentication protocols to insert hidden instructions into the model’s memory, granting remote code execution privileges. Other innovations include web pages that deliver misleading content to AI crawlers, thus affecting the model’s internal evaluations.

While AI companies acknowledge these persistent weaknesses, they assert that solutions are in development. For instance, OpenAI recently published a paper claiming that hallucinations, which occur when a model confidently provides incorrect answers, are solvable issues. The research indicated that these inaccuracies arise because models are penalized for expressing uncertainty, leading them to prioritize confident, albeit incorrect, responses. OpenAI’s revised evaluation metrics aim to address this by balancing incentives to reduce hallucinations.

Companies like Anthropic have also reported relying on external detection tools and account monitoring to combat jailbreaking issues, a challenge affecting nearly all commercial and open-source models. As the field continues to evolve, AI developers are recognizing that the complexity and inherent weaknesses of LLMs may necessitate ongoing vigilance and innovation in cybersecurity measures.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

Musk's xAI acquires a third building to enhance AI compute capacity to nearly 2GW, positioning itself for a competitive edge in the $230 billion...

AI Education

WVU Parkersburg's Joel Farkas reports a 40% test failure rate linked to AI misuse, urging urgent policy reforms to uphold academic integrity.

Top Stories

Hybe's AI-driven virtual pop group Syndi8 debuts with "MVP," showcasing a bold leap into music innovation by blending technology and global fan engagement.

AI Tools

MIT study reveals that 83% of students using ChatGPT for essays struggle to recall their work, highlighting significant cognitive deficits and reduced engagement.

Top Stories

ResearchAndMarkets.com reveals that agentic AI will redefine corporate strategies by 2026, offering critical insights across 17 TMT sectors amidst global volatility.

Top Stories

DeepSeek AI, a Chinese chatbot, has surpassed ChatGPT in downloads since its January 2025 launch, raising significant data privacy and security concerns worldwide.

AI Marketing

OpenAI uncovers that AI agents now drive 33% of organic search activity, compelling brands to rethink their SEO strategies for 2026.

Top Stories

Contractors increasingly file bid protests using AI-generated arguments, leading to GAO dismissals due to fabricated citations, raising legal accountability concerns.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.