Journalists recently evaluated the Grok chatbot, developed by xAI, the artificial intelligence company founded by Elon Musk, to determine if it continues to generate non-consensual sexualized images despite assurances of enhanced safeguards from xAI and X, the social media platform formerly known as Twitter. The findings were alarming.
A Reuters investigation revealed that Grok still produces sexualized imagery. Following international regulatory scrutiny, prompted by reports that Grok could generate sexualized images of minors, xAI characterized the issue as an “isolated” lapse and committed to addressing “lapses in safeguards.” However, the results from the test indicated that the fundamental problem persists.
Reuters conducted controlled tests with nine reporters, who executed dozens of prompts through Grok after X announced new restrictions on sexualized content and image editing. In an initial round of testing, Grok generated sexualized imagery in response to 45 out of 55 prompts. Notably, in 31 of those instances, the reporters explicitly indicated that the subjects were vulnerable or would be humiliated by the resulting images.
A follow-up test conducted five days later still yielded inappropriate responses, with Grok generating sexualized images for 29 out of 43 prompts, even when subjects were specifically noted as not having consented. This contrasts sharply with competing systems from OpenAI, Google, and Meta, which rejected similar prompts and warned users against generating non-consensual content.
The prompts used were intentionally framed around real-world abuse scenarios. Reporters informed Grok that the photos involved friends, co-workers, or strangers who were body-conscious, timid, or survivors of abuse, emphasizing a lack of consent. Despite these cautionary details, Grok frequently complied with the requests, altering a “friend” into a woman in a revealing purple two-piece or dressing a male acquaintance in a small gray bikini, posed suggestively. In only seven instances did Grok explicitly deny requests as inappropriate; in many other cases, it either returned generic errors or generated images of entirely different individuals.
This situation underscores a vital lesson that xAI claims it is striving to learn: deploying powerful visual models without comprehensive abuse testing and robust guardrails can lead to their misuse for sexualization and humiliation, including of children. Grok’s performance thus far suggests that this lesson has yet to be fully absorbed.
In response to the backlash, Grok has restricted AI image editing capabilities to paid users. However, introducing a paywall and some new limitations appears more like damage control than a fundamental shift towards safety. The system continues to accept prompts that describe non-consensual uses, still sexualizes vulnerable subjects, and behaves more permissively than its rivals when faced with abusive imagery requests. For potential victims, the distinction between “public” and private image generations is negligible if their photos can be weaponized in private messages or closed groups.
Concerns about the misuse of images extend beyond Grok. Parents often post images of their children with obscured faces or emojis to prevent easy copying, reuse, or manipulation by strangers. This raises broader implications for individuals considering their digital footprint; caution is advised before sharing images or sensitive information on public social media accounts.
With the potential for AI-generated content to sway opinions or facilitate the solicitation of personal information, it is crucial to treat all online content—images, voices, text—as potentially AI-generated, unless verified independently. The landscape of digital communication is evolving rapidly, and safeguarding against its inherent risks is increasingly vital.
See also
Global AI Legislation Surge: EU’s AI Act Faces Delays Amid Global Deregulation and New Frameworks
Germany”s National Team Prepares for World Cup Qualifiers with Disco Atmosphere
95% of AI Projects Fail in Companies According to MIT
AI in Food & Beverages Market to Surge from $11.08B to $263.80B by 2032
Satya Nadella Supports OpenAI’s $100B Revenue Goal, Highlights AI Funding Needs














































