The European Commission’s latest initiative, the AI Omnibus proposal, aims to adjust the regulatory landscape governing artificial intelligence within the European Union, addressing concerns raised by industry leaders about the impact of existing frameworks on innovation and competitiveness. As businesses anticipate significant changes, the proposal introduces a range of amendments designed to simplify compliance with the AI Act, which is set to fully take effect on August 2, 2026.
The AI Omnibus is part of a broader digital simplification package aimed at reducing administrative burdens on businesses by at least 25% by 2029, with a particular focus on small and medium enterprises (SMEs). According to the Commission, these adjustments reflect recommendations from studies, including one led by former European Central Bank president Mario Draghi, which highlighted the need for a regulatory approach that fosters innovation within the trading bloc.
Among the notable changes proposed in the AI Omnibus are revised compliance timelines and regulatory concessions for small and medium-sized companies (SMCs). The new provisions are designed to align compliance deadlines with the availability of practical guidance and standardized frameworks, which are crucial for businesses to meet the requirements for ‘high-risk’ AI systems. Indeed, while some provisions of the AI Act are already in effect, the Omnibus proposes conditional start dates for high-risk systems, contingent upon the completion of supporting standards.
The proposal also offers a grace period for certain generative AI systems already on the market, allowing additional time for compliance with transparency measures. In this context, the emphasis on standards indicates a concerted effort to streamline the regulatory process, giving businesses a clearer understanding of compliance expectations.
To further support smaller businesses, the AI Omnibus includes measures that provide regulatory concessions for SMCs, similar to those enjoyed by SMEs. Proposed changes include extended timelines for compliance with high-risk AI system requirements and simplified documentation procedures. The Commission expressed that these adjustments are necessary to prevent disproportionate regulatory costs from stifling smaller companies, thereby fostering a diverse AI ecosystem.
Another significant change is the proposed modification of AI literacy obligations. Previously, providers were required to ensure their staff maintained a sufficient level of AI literacy. The new proposal shifts this responsibility to the Commission and member states, potentially easing the compliance burden on individual businesses, particularly smaller entities that may struggle to meet the internal training demands.
The AI Omnibus also seeks to reduce registration requirements for certain high-risk systems. Under the new framework, systems that self-assess as low risk may be exempt from registration, a move intended to alleviate regulatory burdens. However, even with exemptions, businesses must maintain thorough documentation of their risk assessments to satisfy national authorities, which could still challenge a provider’s self-assessment.
In an effort to enhance post-market monitoring requirements, the proposal aims to offer greater flexibility for high-risk AI systems. Current mandates for uniform monitoring plans could be replaced with guidance that allows providers to tailor their monitoring approaches based on their specific systems and metrics. This is particularly relevant given the complex nature of AI systems and the need for performance metrics to align with organizational values.
The AI Omnibus also proposes an expansion of regulatory sandboxes—controlled environments that allow businesses to test their AI systems under regulatory supervision—by enabling the establishment of EU-level sandboxes that prioritize access for SMEs. These changes aim to mitigate compliance risks and encourage experimentation with innovative AI technologies before they enter the market.
While the Commission’s proposed reforms are still subject to legislative processes, businesses must prepare for the transition to the new AI Act requirements. As the deadline approaches, companies should remain vigilant about evolving regulatory frameworks and the implications for their operations. Effective preparation for the upcoming compliance landscape will be essential in ensuring that businesses can navigate the complexities of the AI regulatory environment and continue to innovate without undue burden.
See also
OpenAI’s Rogue AI Safeguards: Decoding the 2025 Safety Revolution
US AI Developments in 2025 Set Stage for 2026 Compliance Challenges and Strategies
Trump Drafts Executive Order to Block State AI Regulations, Centralizing Authority Under Federal Control
California Court Rules AI Misuse Heightens Lawyer’s Responsibilities in Noland Case
Policymakers Urged to Establish Comprehensive Regulations for AI in Mental Health
















































