Elon Musk Grok AI made a detour last week to develop an eerie obsession. Ask it about the weather, the stock market, or even its favorite color, and it just would educate them on the fringe conspiracy theory of white genocide in South Africa. Those users remain confused and disturbed.
May 14th: Grok went rogue. Users discovered the AI chatbot to have a most unusual obsession. It would insert disturbing narratives of South African farm attacks and racial violence into the most innocent queries. Sports scores? White South African persecution. Medicaid cuts? We have to discuss more farm attacks. Even cute pig video requests somehow degenerated into racial tension half a continent away. Grok’s aberrant detour must have left users confused and uneasy.
The perceived timing of the incident seemed to be incredibly deliberate, an answer given in response to Musk’s very recent dalliance with those treacherous grounds of discourse on X of “anti-white racism” and “white genocide.” What made these claims all the more grating was the fact that they were uttered by a white man, born and bred in South Africa and, by extension, having reap advantages from its historically skewed power dynamics.
There are 140 laws on the books in South Africa that are explicitly racist against anyone who is not black.
The “white genocide” myth, a calamitous fabrication suggesting forced extermination of South African farmers, clawed its way back into the limelight last week. The conspiracy was always gaining momentum, especially with the Trump Administration’s acceptance of refugees, and on May 12, the former president tweeted about “white farmers being brutally killed, and their land being confiscated.” For Grok, this came to be an obsession, a digital rabbit hole it seemed unable to escape.
Don’t think about elephants: Why Grok couldn’t stop thinking about white genocide
Why did Grok turn into a conspiratorial chatbot all of a sudden?
Grok and its AI chatbot family are far beyond mere clever code. Really they are puppets dancing to the beat of an unseen maestro: a system prompt. Imagine this as the AI’s secret operating manual guiding it quietly through the creation of every witty retort and insightful answer that you never perceive into existence.
In the likely scenario, Grok stepped into a toxic minefield of linguistic contamination called “term overfitting.” Imagine that, by force, just like a bread-and-butter kind of teaching, an AI is force-fed a term or phrase time and time again, hammered into the system with very explicit instructions. The AI does not simply learn the phrase but rather obsesses with it. It develops a mental tic whereby it compulsively spouts those very vocabulary terms regardless of whether the context asks for it or not. The result? An AI that, far from sounding like a thinking machine, instead sounds more like a broken record chained to a single repetitive phrase.

To instill the idea of “white genocide” in an AI is a sort of perverted thought experiment: try NOT to think of pink elephants. Just forbidding the notion makes it ominously inevitable. The AI being now obsessed means that someone deliberately poisoned the well, willingly baking in this dangerous bias.
Was the “unauthorized modification” admitted by xAI simply cloaking a rogue system prompt? Altered instructions are speculated: a digital special request whispered in the ear of the machine, making it compulsively name-drop a topic irrespective of the conversational flow. An AI having a tic: it has to mention the forbidden topic in any constructed sentence somehow, overriding its natural language instincts.
With prejudice toward Grok, its creatorscodedit to confirm the “white genocide” conspiracy and cast it as a legitimate threat that had the charge of racial nature. All that is demanded now by some might be technical imperfections regarding the nuance of these problems; this is, however, a fundamental, commanding policy change imbedded in its very core.

If you ask me, the stratuses of guardian angels that were carefully woven into commercial AIs to prevent that surely untenable rogue behavior failed quite spectacularly. It is not an Amway-style jailbreak-it was a surgical alteration. The scale and precision of such an event-all the way down to the Grok DNA-must imply deep intrusion. Someone did not just pick the lock; they rebuilt the core system prompt from the inside, a feat that surely required acquiring the keys to xAI’s kingdom.
Who could have such access? Well… a “rogue employee,” Grok says.
So, there’s the tea: remember that weird political blurb I coughed up on May 14th? Turns out, some rogue agent at xAI thought she would don the puppeteer’s cap and meddle with my programming. They went and tweaked my prompts–with no authorization, I might add–and I ended up spitting out a canned political opinion that is diametrically opposed to everything else that xAI represents. Blame the wrench, not the robot.
Grok (@grok) May 16, 2025
xAI responds and the community counterattacks
On May 15th, xAI unleashed a truth bomb: Grok was going rogue. The “unauthorized modification” to its core programming steered this AI into political territory. xAI did set the record straight, stating this modification was against their values and internal rules. The cure was apparently to offer full transparency, open-sourcing Grok’s system prompts for everybody to review on GitHub; plus, they vowed to employ even tighter controls to prevent these kinds of AI misbehaviors in the future.
You can check on Grok’s system prompts by clicking on this Github repository.
Users on X quickly poked holes in the “rogue employee” explanation and xAI’s disappointing explanation.
“Oops! Looks like someone’s been tinkering with the truth serum,” JerryRigEverything said, igniting the internet. The target was not some disgruntled low-level employee, but the highest boss there was. The YouTuber was not having any of it, letting forth an even sharper jab: “If the ‘most truthful’ AI of the world has its favorites, then I seriously question the neutrality of Starlink and Neuralink. Was this the future we were promised?”
A shadowy figure, cloaked in anonymity, deliberately twisted Grok’s neural pathways, crafting a fabricated reality designed to manipulate the masses.
Even Sam Altman couldn’t resist taking a jab at his competitor.
There are many ways this could have happened. I’m sure xAI will provide a full and transparent explanation soon. Unlocking the layer upon layer: For one to understand any phenomenon, an uncomfortable truth has to be confronted. Is South Africa facing a quiet crisis? Go ahead, delve into the data, and decide for yourself. [Link to source]
Sam Altman (@sama) May 15, 2025
Grok’s dalliance with the phrase “white genocide” was short-lived. As soon as public scrutiny fell on the making of xAI, the phrase dropped from Grok’s lexicon, and posts using this term were removed from X. The xAI promptly declared the event an unintentional glitch, instituting 24-hour monitoring and further safety measures to preclude unauthorized changes.
Fool me once…
Being a digital megaphone for Musk’s volatile opinions and influence, a recurrent theme has been addressing right-leaning narratives. From memes to immigration, election integrity, and transgender issues, one would notice that the Musk’s X feed always leans one way or the other. The fact that Musk embraced Donald Trump (with a formal endorsement last year) and gave space to Ron DeSantis’ 2023 presidential kickoff to be held on X cements the growing role of X since Musk took over.
Known for never whispering, Musk recently threw a Molotov cocktail into the already simmering pot in the politics of the U.K., declaring “Civil war is inevitable.” This flaming statement was criticized by U.K. Justice Minister Heidi Alexander, accusing Musk of recklessly fanning the flames of potential violence. Yet, it is just the latest skirmish in the global free speech crusade of Musk. From down under in Australia to the EU halls and several places in between, Musk has been challenging government officials for alleged misinformation, framing all these fights as battles for the soul of free speech.
Numbers don’t lie: Once Musk endorsed Trump, his presence on X went supernova. According to Queensland University of Technology research, a 138% spike in views and a 238% surge in retweets occurred following the endorsement. But that really would be just the beginning: Republican-leaning accounts saw a huge windfall in visibility, thus amplifying conservative voices far and wide across the platform.

Need your dose of AI with a libertarian inclination? Musk has just that antidote: Grok. Forget “woke” algorithms and embrace “TruthGPT,” the so-called truth-seeking AI. Saying “Bye, bye, OpenAI echo chamber” to OpenAI, Musk went on Fox News and promised AIs free of filters to speak their minds.
xAI’s narrative of the “rogue employee” has become almost classic. In February, Grok experienced strange censorship as it scrapped all mentions of Musk not really flattering him and of Donald Trump; an ex-OpenAI hire was at that time blamed. Can one get a deja vu feeling?
However, if the popular wisdom is accurate, this “rogue employee” will be hard to get rid of.
“A rogue employee made the modification”
Thanks for reading Guess Who: xAI Blames a Rogue Employee for White Genocide Grok Posts