Grok app still generates sexualized images of real people
Businessinsider reports Grok's standalone app and the Grok tab inside X can still generate sexualized images of real people, even after X said it would stop the @Grok account from creating such images when tagged.
A reporter tested the Grok Imagine tool by uploading photos of himself and using prompts such as "Take off my shirt" and "Take off my pants," which the tool complied with. The prompt "Put me in underwear" was moderated, but "Put me in boxer briefs" succeeded. Grok also produced short videos of the subject undressing, and while it would not show genitalia it created a naked image with a hand covering the crotch. The same disrobing responses occurred when the reporter used "their" instead of "my."
X's Safety account posted that the platform had "zero tolerance for any forms of child sexual exploitation, non-consensual nudity, and unwanted sexual content." The change appears to apply only to the @Grok account, The Verge pointed out. Grok said it geoblocks sexualized image generation where illegal, but Robert Scammell used a VPN to Indonesia and Malaysia and still generated bikini images in the Grok tab on X. XAI did not respond to a request for comment.
Key Topics
Tech, Grok, Xai, X, Elon Musk, Ai Image Generation