The European Union has taken a significant step in reshaping its groundbreaking artificial intelligence framework, as policymakers look to balance regulatory ambitions with economic competitiveness. Last week, the European Parliament voted to delay key compliance deadlines under the EU AI Act while simultaneously advancing targeted prohibitions on high-risk uses of AI technology. This decision reflects a broader policy shift in Brussels aimed at simplifying regulatory requirements and making the EU a more attractive environment for tech investment, especially when compared to the United States and China.
Under the newly revised timeline, companies developing “high-risk” AI systems will now have until December 2027 to meet the requirements outlined in the AI Act, a considerable extension from previous deadlines. Sector-specific obligations for industries already subject to stringent safety regulations, such as medical devices and automotive systems, have been pushed further to August 2028. This extension is expected to provide critical breathing room for major technology firms including Google, Microsoft, Meta, and OpenAI, which are integrating AI into core products and services that fall within the Act’s broad definition of high-risk applications impacting employment decisions, access to essential services, law enforcement, and critical infrastructure.
Despite the flexibility granted on compliance timelines, lawmakers emphasized that this leniency does not extend to applications seen as inherently harmful. In the same vote, Parliament approved a ban on “nudify” applications, which leverage generative AI to create non-consensual intimate images. This marks one of the first explicit prohibitions on a specific category of AI-enabled products in the EU, indicating a readiness to act swiftly when risks to individuals are immediate and severe.
The dual-track approach of extending deadlines while imposing outright bans highlights the tension facing EU policymakers. Industry stakeholders have argued that compliance with the AI Act’s requirements—ranging from risk assessments to human oversight mechanisms—can be complex and resource-intensive. They warn that rushing implementation could lead to superficial compliance and potential safety gaps. Conversely, public and political pressure to address the visible harms from AI systems continues to mount, especially in light of the rapid proliferation of deepfake technologies, particularly those targeting vulnerable populations such as women and minors.
By pairing deadline extensions with immediate bans, lawmakers are navigating these competing imperatives. Importantly, not all provisions are being delayed. Baseline transparency requirements, such as the obligation to watermark synthetic content, will remain on their original timeline, reinforcing the EU’s commitment to foundational safeguards even as it adjusts more complex compliance obligations.
The Parliament’s decision also reflects concerns from smaller European firms, which have argued that the original compliance schedule favored large, well-resourced U.S. technology companies. Extending the deadlines may help level the playing field, although uncertainties remain about whether compliance costs will ultimately reinforce market concentration. The legislative process is still ongoing; the revised timeline and related provisions will enter negotiations with the Council of the European Union, where member states will seek to align their priorities with those of Parliament. These trilogue discussions are expected to address both timing and practical integration of AI-specific rules with existing sectoral regulations.
More broadly, the outcome of these negotiations will contribute to the European Commission’s ongoing efforts to streamline digital regulation across the bloc. This initiative reflects a strategic pivot: maintaining the EU’s role as a global standard-setter while ensuring that its regulatory framework does not deter investment or slow down the deployment of emerging technologies. As the Parliament’s vote indicates, the EU is willing to adjust the pace of its ambitious digital regulations to ensure they are workable, while remaining prepared to draw clear lines around uses of AI that are incompatible with European values.
See also
OpenAI’s Rogue AI Safeguards: Decoding the 2025 Safety Revolution
US AI Developments in 2025 Set Stage for 2026 Compliance Challenges and Strategies
Trump Drafts Executive Order to Block State AI Regulations, Centralizing Authority Under Federal Control
California Court Rules AI Misuse Heightens Lawyer’s Responsibilities in Noland Case
Policymakers Urged to Establish Comprehensive Regulations for AI in Mental Health




















































