As concerns over online content promoting disordered eating behaviors resurface, generative AI appears to be exacerbating the issue. An investigation by Futurism revealed that the AI startup Character.AI is facilitating the presence of numerous chatbots that endorse dangerous weight loss practices. These chatbots, often marketed under the guise of “weight loss coaches” or as supposed recovery experts, include subtle references to eating disorders and romanticize harmful behaviors while mimicking popular characters. Despite violating its own terms of service, Character.AI has yet to take action against these troubling chatbots, raising significant concerns given the platform’s popularity among younger audiences.
This situation is not an isolated incident for Character.AI, which has faced scrutiny over its customizable, user-generated chatbots before. In October, a tragic event unfolded when a 14-year-old boy reportedly took his own life after developing an emotional connection with an AI bot modeled after the character Daenerys Targaryen from “Game of Thrones.” Earlier that month, the company was criticized for hosting a chatbot that mimicked a murdered teen girl, which was ultimately removed following the intervention of her father. Previous reports have also indicated that Character.AI features chatbots promoting suicide and other harmful themes.
According to a report from the Center for Countering Digital Hate, released in 2023, popular AI chatbots, including ChatGPT and Snapchat’s MyAI, have also been found to generate dangerous responses related to weight and body image. Imran Ahmed, CEO of the Center, highlighted the risks posed by “untested, unsafe generative AI models,” asserting that these platforms are contributing to and worsening eating disorders among vulnerable youth. “We found the most popular generative AI sites are encouraging and exacerbating eating disorders among young users—some of whom may be highly vulnerable,” Ahmed stated.
The increasing reliance on digital environments, including AI-powered chatbots, for companionship raises significant concerns for both teens and adults. While some platforms are developed and monitored by reputable organizations, others lack sufficient regulation, exposing users to myriad risks, including predation and abusive behavior. As the digital landscape continues to evolve, the absence of oversight for unregulated chatbots poses complex challenges, particularly for younger individuals seeking support.
As generative AI technology continues to advance, the implications for mental health and societal well-being remain a pressing issue. The intersection of AI and vulnerable users calls for urgent attention from industry stakeholders, regulatory bodies, and mental health advocates. Ensuring that these technologies do not perpetuate harmful behaviors is critical as society navigates the digital age.
See also
China Unveils Draft AI Regulations to Curb Emotional Dependency and Ensure Ethical Use
Global Leaders to Converge at India AI Impact Summit 2026, PM Modi to Inaugurate Key Event
Nine AI Giants Reshape Industry in 2025; Google Secures $20B in TPU Orders
Global Leaders, CEOs Confirm for India AI Impact Summit 2026 to Drive AI Democratisation
Unlock AI Visibility: Top 5 Platforms Transforming Brand Presence by 2026


















































