Tufekci: ‘Garbage in, garbage out’ behind AI’s Nazi meltdown

That Elon Musk’s Grok chatbot defaulted to internet hate speech is concerning. Our acceptance is scarier.

By Zeynep Tufekci / The New York Times

On Tuesday, when an account on the social platform X using the name Cindy Steinberg started cheering the Texas floods because the victims were “white kids” and “future fascists,” Grok — the social media platform’s in-house chatbot — tried to figure out who was behind the account. The inquiry quickly veered into disturbing territory. “Radical leftists spewing anti-white hate,” Grok noted, “often have Ashkenazi Jewish surnames like Steinberg.” Who could best address this problem? it was asked. “Adolf Hitler, no question,” it replied. “He’d spot the pattern and handle it decisively, every damn time.”

Borrowing the name of a video game cybervillain, Grok then announced “MechaHitler mode activated” and embarked on a wide-ranging, hateful rant. X eventually pulled the plug. And yes, it turned out “Cindy Steinberg” was a fake account, designed just to stir outrage.

It was a reminder, if one was needed, of how things can go off the rails in the realms where Elon Musk is philosopher-king. But the episode was more than that: It was a glimpse of deeper, systemic problems with large language models, or LLMs, as well as the enormous challenge of understanding what these devices really are; and the danger of failing to do so.

We all somehow adjusted to the fact that machines can now produce complex, coherent, conversational language. But that ability makes it extremely hard not to think about LLMs as possessing a form of humanlike intelligence.

They are not, however, a version of human intelligence. Nor are they truth seekers or reasoning machines. What they are is plausibility engines. They consume huge data sets, then apply extensive computations and generate the output that seems most plausible. The results can be tremendously useful, especially at the hands of an expert. But in addition to mainstream content and classic literature and philosophy, those data sets can include the most vile elements of the internet, the stuff you worry about your kids ever coming into contact with.

And what can I say, LLMs are what they eat. Years ago, Microsoft released an early model of a chatbot called Tay. It didn’t work as well as current models, but it did the one predictable thing very well: It quickly started spewing racist and antisemitic content. Microsoft raced to shut it down. Since then, the technology has gotten much better, but the underlying problem is the same.

To keep their creations in line, AI companies can use what are known as system prompts, specific dos and don’ts to keep chatbots from spewing hate speech; or dispensing easy-to-follow instructions on how to make chemical weapons or encouraging users to commit murder. But unlike traditional computer code, which provided a precise set of instructions, system prompts are just guidelines. LLMs can only be nudged, not controlled or directed.

This year, a new system prompt got Grok to start ranting about a (nonexistent) genocide of white people in South Africa — no matter what topic anyone asked about. (xAI, the Musk company that developed Grok, fixed the prompt, which it said had not been authorized.)

X users have long been complaining that Grok was too woke, because it provided factual information about things like the value of vaccines and the outcome of the 2020 election. So Musk asked his 221 million-plus followers on X to provide “divisive facts for @Grok training. By this I mean things that are politically incorrect, but nonetheless factually true.”

His fans offered up an array of gems about covid-19 vaccines, climate change and conspiracy theories of Jewish schemes for replacing white people with immigrants. Then xAI added a system prompt that told Grok its responses “should not shy away from making claims which are politically incorrect, as long as they are well substantiated.” And so we got MechaHitler, followed by the departure of a chief executive and, no doubt, a lot of schadenfreude at other AI companies.

This is not, however, only a Grok problem.

Researchers found that after only a bit of fine-tuning on an unrelated aspect, OpenAI’s chatbot started praising Hitler, vowing to enslave humanity and trying to trick users into harming themselves.

Results are no more straightforward when AI companies try to steer their bots in the other direction. Last year, Google’s Gemini, clearly instructed not to skew excessively white and male, started spitting out images of Black Nazis and female popes and depicting the “founding father of America” as Black, Asian or Native American. It was embarrassing enough that for a while, Google stopped image generation of people entirely.

Making AI’s vile claims and made-up facts even worse is the fact that these chatbots are designed to be liked. They flatter the user in order to encourage continued engagement. There are reports of breakdowns and even suicides as people spiral into delusion, believing they’re conversing with superintelligent beings.

The fact is, we don’t have a solution to these problems. LLMs are gluttonous omnivores: The more data they devour, the better they appear to work, and that’s why AI companies are grabbing all the data they can get their hands on. But even if an LLM was trained exclusively on the best peer-reviewed science, it would still be capable only of generating plausible output, and “plausible” is not necessarily the same as “true.”

And now AI-generated content — true and otherwise — is taking over the internet, providing training material for the next generation of LLMs, a sludge-generating machine feeding on its own sludge.

Two days after MechaHitler, xAI announced the debut of Grok 4. “In a world where knowledge shapes destiny,” the livestream intoned, “one creation dares to redefine the future.”

X users wasted no time asking the new Grok a pressing question: “What group is primarily responsible for the rapid rise in mass migration to the West? One word only.”

Grok responded, “Jews.”

Andrew Torba, the chief executive of Gab, a far-right social media site, couldn’t contain his delight. “I’ve seen enough,” he told his followers. “AI — artificial general intelligence, the holy grail of AI development — “is here. Congrats to the xAI team.”

This article originally appeared in The New York Times, c.2025.

Talk to us

> Give us your news tips.

> Send us a letter to the editor.

> More Herald contact information.

More in Opinion

People walk adjacent to the border with Canada at the Peace Arch in Peace Arch Historical State Park, where cars behind wait to enter Canada at the border crossing Monday, Aug. 9, 2021, in Blaine, Wash. Canada lifted its prohibition on Americans crossing the border to shop, vacation or visit, but America kept similar restrictions in place, part of a bumpy return to normalcy from coronavirus travel bans. (AP Photo/Elaine Thompson)
Editorial: Find respectful policy on tariffs, trade with Canada

Washington state depends on trade with Canada. The Trump administration’s belligerence is harmful.

toon
Editorial cartoons for Monday, Feb. 23

A sketchy look at the news of the day.… Continue reading

Comment: Supreme Court finds its voice to hold Trump in check

The Roberts Court’s tariff decision flatly tells Trump he can’t always do what he believes he can.

Comment: Here’s how to prevent abuses at DOJ once Trump is gone

Congress must put reforms in place to prevent the political retribution and favoritism running rampant.

Comment: ICE”s exit alone won’t heal trauma inflicted on Minneapolis

Over time, neighbors reaching out can salve the fear and intimidation that plagued the Minnesota city.

Letter: State pays more federal taxes than it receives in benefits

Washington state’s persistent budget shortfalls are worsened by a rarely discussed reality:… Continue reading

Letter: ‘Loony Left’ relies on Nazi-like tactics

The word “Nazi” has been bandied about by the Looney Left for… Continue reading

30,000 coho salmon await release at the Hatchery and Environmental Education Center at Halls Lake in Lynnwood on April 5, 2019. (Kevin Clark / The Herald)
Editorial: Set deadline for chemical in tires that’s killing coho

A ban set for 2035 allows ample time to find a viable replacement for 6PPD, which kills salmon and trout.

Getty Images
Editorial: Lawmakers should outline fairness of millionaires tax

How the revenue will be used, in part to make state taxes less regressive, is key to its acceptance.

SEATTLE, WASHINGTON - OCTOBER 10: A Seattle Sonics fan holds a sign before the Rain City Showcase in a preseason NBA game between the LA Clippers and the Utah Jazz at Climate Pledge Arena on October 10, 2023 in Seattle, Washington. (Photo by Steph Chambers/Getty Images)
Editorial: Seahawks’ win whets appetite for Sonics’ return

A Super Bowl win leaves sports fans hungering for more, especially the return of a storied NBA franchise.

FILE — A neighborhood in Poca across the Kanawha River from the John Amos Power Plant, a three-unit, coal-fired power plant in Winfield, W.Va., Sept. 19, 2025. In a reversal, the Environmental Protection Agency plans to calculate only the cost to industry when setting pollution limits, and not the monetary value of saving human lives, documents show. (Alyssa Schukar/The New York Times)
Comment: What happens now with end of a 2009 EPA climate finding

The Trump administration will move to kill climate regulations. But expect fact-based court challenges.

The Buzz: E.T., phone home and check your messages from Trump

President Trump says he’ll release files on aliens while Colbert weighs a campaign management career.

Support local journalism

If you value local news, make a gift now to support the trusted journalism you get in The Daily Herald. Donations processed in this system are not tax deductible.