Connect with us

Hi, what are you looking for?

AI Generative

Northeastern Researchers Uncover AI Bias in Health Care Using Sparse Autoencoders

Northeastern researchers reveal AI bias in healthcare, showing racial prejudices in LLMs can harm patient care, underscoring urgent need for transparency.

Artificial intelligence is increasingly being integrated into healthcare settings, aiding tasks from writing physicians’ notes to making patient recommendations. However, new research from Northeastern University highlights a critical concern: AI and large language models (LLMs) can perpetuate racial biases embedded in their training data, influencing outputs in ways that may not be immediately apparent to users. This study aims to decode the decision-making process of LLMs, shedding light on when race may be problematic in their recommendations.

Hiba Ahsan, a Ph.D. student and lead author of the study, emphasizes a significant finding: previous research indicates that Black patients are often less likely to receive pain medications despite reporting pain levels comparable to those of white patients. Ahsan warns that AI models could replicate this bias, making potentially harmful recommendations based on race.

In some medical contexts, considering race can be clinically important. For instance, gestational hypertension is more prevalent among individuals of African descent, while cystic fibrosis occurs more frequently in Northern Europeans, according to the Mayo Clinic. However, Ahsan notes that many biases in LLMs arise from irrelevant prejudices based on race, leading to outcomes that could adversely affect patient care.

The researchers utilized a tool called a sparse autoencoder to examine the intricate, often opaque workings of LLMs. Ahsan explains that these models simplify complex data inputs through a process called encoding, creating intermediate representations known as “latents.” While the model can interpret these numbers, human understanding remains limited.

By employing the sparse autoencoder, Ahsan and her advisor Byron Wallace, a machine learning expert, aimed to translate these latents into comprehensible data. This tool can indicate whether a particular data point is associated with race or another identifiable characteristic. Ahsan describes the process: if the autoencoder detects a latent related to race, it will signal that race is influencing the model’s output.

The researchers analyzed clinical notes and discharge summaries from the publicly available MIMIC dataset, which anonymizes personal information. They focused on notes where patients identified as white or Black. After processing these notes through an LLM named Gemma-2, they employed the sparse autoencoder to identify latents corresponding to race.

The findings revealed that racial biases were indeed embedded in the LLM. The autoencoder identified a notable frequency of latents associated with Black individuals alongside stigmatizing concepts such as “incarceration,” “gunshot,” and “cocaine use.” While previous evidence of racial bias in AI is well-documented, this research provides a rare glimpse into the underlying elements that influence LLM responses.

Ahsan emphasizes the challenges of interpretability in LLMs, describing them as “black boxes” that obscure the factors leading to specific decisions. Utilizing a sparse autoencoder to peer into these systems offers a pathway for physicians to discern when a patient’s race may be improperly factored into AI recommendations. Increased transparency could help doctors intervene and mitigate bias or seek alternative solutions, such as retraining the AI on more representative datasets.

Wallace acknowledges that while they did not invent sparse autoencoders, their application in clinical settings is pioneering. He asserts, “If we’re going to use these models in health care and want to do it safely, we probably need to improve the methods for interpreting them.” The sparse autoencoder method represents a significant step in addressing the interpretability challenges inherent in AI systems.

As AI continues to play a larger role in healthcare, the implications of this research are profound. With the ability to identify and understand biases in AI recommendations, healthcare providers can work towards ensuring more equitable treatment and decision-making, paving the way for advancements that prioritize patient welfare over algorithmic convenience.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Technology

Nvidia shares plummet 4.3% to $178.07 after Inventec cites Chinese clearance delays for H200 chip, threatening crucial AI market expansion.

Top Stories

Educational institutions are embracing algorithm auditing to combat bias in AI, with Syracuse University leading the charge in equipping students for ethical challenges in...

AI Marketing

Tony Hayes reveals 14 AI-driven workflows that cut SEO timelines from months to hours, enabling beginners to achieve results worth $5,000 in web design.

Top Stories

Elon Musk brands OpenAI's ChatGPT 'diabolical' amid claims it influenced a tragic murder-suicide, raising urgent ethical questions about AI's societal impact.

AI Technology

Morgan Stanley rates Nebius Group (NASDAQ:NBIS) "Equalweight," setting a $126 price target amid strong long-term potential for AI infrastructure growth.

Top Stories

PwC's Mohamed Kande warns that over 50% of AI investments fail to generate value, urging CEOs to adopt strategic planning amid rising confidence in...

AI Business

UK lawmakers urge the FCA to implement AI-specific stress tests by 2026, addressing risks from 75% of firms using AI and potential market instability.

Top Stories

Meta unveils the Meta Compute initiative to generate tens of gigawatts of energy, aiming to reshape AI infrastructure and reduce vendor reliance.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.