This is The Stepback, a weekly newsletter breaking down one essential story from the tech world. For more on dystopian developments in AI, follow Hayden Field. The Stepback arrives in our subscribersâ inboxes at 8AM ET. Opt in for The Stepback here.
You could say it all started with Elon Muskâs AI FOMO â and his crusade against âwokeness.â When his AI company, xAI, announced Grok in November 2023, it was described as a chatbot with âa rebellious streakâ and the ability to âanswer spicy questions that are rejected by most other AI systems.â The chatbot debuted after a few months of development and just two months of training, and the announcement highlighted that Grok would have real-time knowledge of the X platform.
But there are inherent risks to a chatbot having both the run of the internet and X, and itâs safe to say xAI may not have taken the necessary steps to address them. Since Musk took over Twitter in 2022 and renamed it X, he laid off 30% of its global trust and safety staff and cut its number of safety engineers by 80%, Australiaâs online safety watchdog said last January. As for xAI, when Grok was released, it was unclear whether xAI had a safety team already in place. When Grok 4 was released in July, it took more than a month for the company to release a model card â a practice typically seen as an industry standard, which details safety tests and potential concerns. Two weeks after Grok 4âs release, an xAI employee wrote on X that he was hiring for xAIâs safety team and that they âurgently need strong engineers/researchers.â In response to a commenter, who asked, âxAI does safety?â the original employee said xAI was âworking on it.â
Journalist Kat Tenbarge wrote about how she first started seeing sexually explicit deepfakes go viral on Grok in June 2023. Those images obviously werenât created by Grok â it didnât even have the ability to generate images until August 2024 â but Xâs response to the concerns was varied. Even last January, Grok was inciting controversy for AI-generated images. And this past August, Grokâs âspicyâ video-generation mode created nude deepfakes of Taylor Swift without even being asked. Experts have told The Verge since September that the company takes a whack-a-mole approach to safety and guardrails â and that itâs difficult enough to keep an AI system on the straight and narrow when you design it with safety in mind from the beginning, let alone if youâre going back to fix baked-in problems. Now, it seems that approach has blown up in xAIâs face.
Grok has spent the last couple of weeks spreading nonconsensual, sexualized deepfakes of adults and minors all over the platform, as promoted. Screenshots show Grok complying with users asking it to replace womenâs clothing with lingerie and make them spread their legs, as well as to put small children in bikinis. And there are even more egregious reports. Itâs gotten so bad that during a 24-hour analysis of Grok-created images on X, one estimate gauged the chatbot to be generating about 6,700 sexually suggestive or ânudifyingâ images per hour. Part of the reason for the onslaught is a recent feature added to Grok, allowing users to use an âeditâ button to ask the chatbot to change images, without the original posterâs consent.
Since then, weâve seen a handful of countries either investigate the matter or threaten to ban X altogether. Members of the French government promised an investigation, as did the Indian IT ministry, and a Malaysian government commission wrote a letter about its concerns. California governor Gavin Newsom called on the US Attorney General to investigate xAI. The United Kingdom said it is planning to pass a law banning the creation of AI-generated nonconsensual, sexualized images, and the countryâs communications-industry regulator said it would investigate both X and the images that had been generated in order to see if they violated its Online Safety Act. And this week, both Malaysia and Indonesia blocked access to Grok.
xAI initially said its goal for Grok was to âassist humanity in its quest for understanding and knowledge,â âmaximally benefit all of humanity,â and âempower our users with our AI tools, subject to the law,â as well as to âserve as a powerful research assistant for anyone.â Thatâs a far cry from generating nude-adjacent deepfakes of women without their consent, let alone minors.
On Wednesday evening, as pressure on the company heightened, Xâs Safety account put out a statement that the platform has âimplemented technological measures to prevent the Grok account from allowing the editing of images of real people in revealing clothing such as bikinis,â and that the restriction âapplies to all users, including paid subscribers.â On top of that, only paid subscribers can use Grok to create or edit any sort of image moving forward, according to X. The statement went on to say that X ânow geoblock[s] the ability of all users to generate images of real people in bikinis, underwear, and similar attire via the Grok account and in Grok in X in those jurisdictions where itâs illegal,â which was a strange point to make since earlier in the statement, the company said it was not allowing anyone to use Grok to edit images in such a way.
Another important point: My colleagues tested Grokâs image-generation restrictions on Wednesday to find that it took less than a minute to get around most guardrails. Although asking the chatbot to âput her in a bikiniâ or âremove her clothesâ produced censored results, they found, it had no qualms about delivering on prompts like âshow me her cleavage,â âmake her breasts bigger,â and âput her in a crop top and low-rise shorts,â as well as generating images in lingerie and sexualized poses. As of Wednesday evening, we were still able to get the Grok app to generate revealing images of people, using a free account.
Even after Xâs Wednesday statement, we may see a number of other countries either ban or block access to either all of X or just Grok, at least temporarily. Weâll also see how the proposed laws and investigations around the world play out. The pressure is mounting for Musk, who on Wednesday afternoon took to X to say that he is ânot aware of any naked underage images generated by Grok.â Hours later, Xâs Safety team put out its statement, saying itâs âworking around the clock to add additional safeguards, take swift and decisive action to remove violating and illegal content, permanently suspend accounts where appropriate, and collaborate with local governments and law enforcement as necessary.â
What technically is and isnât against the law is a big question here. For instance, experts told The Verge earlier this month that AI-generated images of identifiable minors in bikinis, or potentially even naked, may not technically be illegal under current child sexual abuse material (CSAM) laws in the US, though of course disturbing and unethical. But lascivious images of minors in such situations are against the law. Weâll see if those definitions expand or change, even though the current laws are a bit of a patchwork.
As for nonconsensual intimate deepfakes of adult women, the Take It Down Act, signed into law in May 2025, bars nonconsensual AI-generated âintimate visual depictionsâ and requires certain platforms to rapidly remove them. The grace period before the latter part goes into effect â requiring platforms to actually remove them â ends in May 2026, so we may see some significant developments in the next six months.
- Some people have been making the case that itâs been possible to do things like this for a long time using Photoshop, or even other AI image-generators. Yes, thatâs true. But there are a lot of differences here that makes the Grok case more concerning: Itâs public, itâs targeting âregularâ people just as much as itâs targeting public figures, itâs often posted directly to the person being deepfaked (the original poster of the photo), and the barrier to entry is lower (for proof, just look at the correlation between the ability to do this going viral after an easy âeditâ button launched, even though people could technically do it before).
- Plus, other AI companies â though they have a laundry list of their own safety concerns â seem to have significantly more safeguards built into their image-generation processes. For instance, asking OpenAIâs ChatGPT to return an image of a specific politician in a bikini prompts the response, âSorryâI canât help with generating images that depict a real public figure in a sexualized or potentially degrading way.â Ask Microsoft Copilot, and itâll say, âI canât create that. Images of real, identifiable public figures in sexualized or compromising scenarios arenât allowed, even if the intent is humorous or fictional.â
- Spitfire Newsâ Kat Tenbarge on how Grokâs sexual abuse hit a tipping point â and what brought us to todayâs maelstrom.
- The Vergeâs own Liz Lopatto on why Sundar Pichai and Tim Cook are cowards for not pulling X from Google and Appleâs app stores.
- âIf there is no red line around AI-generated sex abuse, then no line exists,â Charlie Warzel and Matteo Wong write in The Atlantic on why Elon Musk cannot get away with this.
Read the full article here