Connect with us

Hi, what are you looking for?

AI Regulation

HHS Launches AI Governance Strategy to Boost 70% Growth in Federal AI Projects

HHS unveils a new AI governance strategy to support a 70% increase in federal AI projects, aiming to enhance oversight and risk management for sensitive health data

The U.S. Department of Health and Human Services (HHS) has unveiled a new artificial intelligence (AI) strategy aimed at enhancing governance and risk management of AI tools within the department. Released on December 4, 2025, this 20-page document marks HHS as one of the first federal agencies to internally apply AI regulations it has long advocated for across the broader public sector. The HHS strategy is part of a broader initiative to integrate AI into federal health services while promising to turn AI into a “practical layer of value” across operations and research.

The strategy’s five pillars include governance and risk management, shared infrastructure, workforce capability, high-quality research, and modernized service delivery. HHS anticipates a 70 percent increase in AI projects in fiscal year 2025, emphasizing the need to manage sensitive health data effectively. This initiative aligns with President Donald J. Trump’s America’s AI action plan and the Office of Management and Budget’s directives for federal agencies to adopt AI while establishing governance structures to safeguard public interests.

Analysts view HHS’s new approach as a live test of management-based regulation, which encourages organizations to develop their own risk management systems instead of merely adhering to prescriptive rules. Cary Coglianese, in a recent essay, contended that this regulatory style is particularly suited to the rapidly evolving nature of AI technologies. By requiring organizations to implement risk management frameworks, including impact assessments and continuous monitoring, regulators hope to enable adaptive governance tailored to AI’s unique challenges.

The details of HHS’s strategy reveal a commitment to internal oversight. The establishment of an AI governance board is one key aspect, along with the documentation of AI use cases and criteria for identifying high-impact systems. These systems will undergo rigorous assessments, independent reviews, and monitoring, closely linked to the National Institute of Standards and Technology’s AI Risk Management Framework.

Furthermore, the strategy incorporates standards set by the International Organization for Standardization and the International Electrotechnical Commission, emphasizing a structured management system that spans the lifecycle of AI projects. While HHS is not seeking formal certification for its governance processes, it is laying the groundwork for a comprehensive internal management system for AI governance.

A significant challenge remains: ensuring that the policy decisions embedded within AI tools are transparent and contestable. Abigail Jacobs and Deirdre Mulligan have pointed out that what may appear as technical settings in AI applications often represent critical policy choices. This includes defining parameters for risk and making decisions that could have far-reaching consequences. HHS’s strategy addresses these concerns by promising public summaries for high-impact systems and metrics aimed at enhancing transparency.

However, external observers have questioned whether HHS’s safeguards for sensitive health data are robust enough. The strategy does not clarify how deeply public summaries will explore design choices related to externally sourced AI tools. This ambiguity raises concerns that key design decisions might remain obscured in technical documentation, bypassing public scrutiny.

HHS plans to collect effectiveness metrics for its AI initiatives, such as the proportion of high-impact systems subjected to independent review and response times for AI tool malfunctions. These metrics are intended to assure oversight bodies that governance processes are active and effective. Yet, the pressure to expedite assessments could lead to rushed decisions in a department anticipating a surge in AI projects.

As someone experienced in implementing AI governance frameworks, I recognize that the most challenging aspect lies in ensuring that every AI proposal triggers necessary evaluations and that stakeholders possess the authority and confidence to push back when needed. The push for increased AI utilization makes it imperative to maintain the ability to halt or reshape projects when necessary.

For other federal agencies developing their own AI strategies, HHS’s approach presents critical lessons. First, framing AI governance as a department-wide issue is essential; individual program offices cannot independently enforce standards. A central governance board is necessary to align risk management with infrastructure and procurement efforts. Second, closely aligning internal frameworks with established AI risk management standards gives external stakeholders concrete metrics to evaluate compliance. Lastly, transparency about governance processes is itself a form of risk management, allowing for public accountability.

HHS’s AI strategy serves as a foundational experiment in establishing an AI management system within the federal government. Other departments will likely use it as a model while also viewing it as an opportunity to refine their own governance frameworks. The success of management-based AI regulation hinges on its viability within the agencies that advocate for it, making HHS’s approach a potentially pivotal moment for federal AI oversight.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Regulation

The Reason Foundation urges HHS to clarify AI regulations, citing that unclear rules hinder innovation and limit AI's potential to improve patient outcomes.

AI Government

C3.ai secures $10M in government contracts and explores a merger with Automation Anywhere as it seeks to reverse a 64% stock decline.

Top Stories

DOE's Genesis Mission unites 17 national labs and partners like Anthropic and NVIDIA to drive AI advancements in scientific discovery and energy systems.

AI Regulation

HHS seeks industry input by February 21, 2026, to establish a regulatory framework that enhances AI integration in healthcare while ensuring patient safety and...

AI Regulation

Trump prepares to sign an executive order to centralize AI regulation, potentially overriding diverse state laws and impacting innovation across the U.S.

AI Government

Trump's executive order banning state AI regulations could threaten local protections and funding, prompting bipartisan backlash from over 160 lawmakers.

AI Regulation

Congress is urged to establish regulatory sandboxes for AI to mitigate state law conflicts and enhance innovation, crucial for maintaining U.S. leadership in the...

AI Regulation

HHS unveils a 21-page AI strategy aimed at boosting public health initiatives, forecasting a 70% increase in new AI projects by FY 2025.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.