xAI blamed an “unauthorized modification” for a bug in its AI-powered Grok chatbot that caused Grok to repeatedly refer to “white genocide in South Africa” when invoked in certain contexts on X. On Wednesday, Grok began replying to dozens of posts on X with information about white genocide in South Africa, even in response to […]

xAI blames Grok’s obsession with white genocide on an ‘unauthorized modification’


xAI blamed an “unauthorized modification” for a bug in its AI-powered Grok chatbot that caused Grok to repeatedly refer to “white genocide in South Africa” when invoked in certain contexts on X.

On Wednesday, Grok began replying to dozens of posts on X with information about white genocide in South Africa, even in response to unrelated subjects. The strange replies stemmed from the X account for Grok, which responds to users with AI-generated posts whenever a person tags “@grok.”

According to a post Thursday from xAI’s official X account, a change was made Wednesday morning to the Grok bot’s system prompt — the high-level instructions that guide the bot’s behavior — that directed Grok to provide a “specific response” on a “political topic.” xAI says that the tweak “violated [its] internal policies and core values,” and that the company has “conducted a thorough investigation.”

It’s the second time xAI has publicly acknowledged an unauthorized change to Grok’s code caused the AI to respond in controversial ways.

In February, Grok briefly censored unflattering mentions of Donald Trump and Elon Musk, the billionaire founder of xAI and owner of X. Igor Babuschkin, an xAI engineering lead, said that Grok had been instructed by a rogue employee to ignore sources that mentioned Musk or Trump spreading misinformation, and that xAI reverted the change as soon as users began pointing it out.

xAI said on Thursday that it’s going to make several changes to prevent similar incidents from occurring in the future.

Beginning today, xAI will publish Grok’s system prompts on GitHub as well as a changelog. The company says it’ll also “put in place additional checks and measures” to ensure that xAI employees can’t modify the system prompt without review and establish a “24/7 monitoring team to respond to incidents with Grok’s answers that are not caught by automated systems.”

Despite Musk’s frequent warnings of the dangers of AI gone unchecked, xAI has a poor AI safety track record. A recent report found that Grok would undress photos of women when asked. The chatbot can also be considerably more crass than AI like Google’s Gemini and ChatGPT, cursing without much restraint to speak of.

A study by SaferAI, a nonprofit aiming to improve the accountability of AI labs, found xAI ranks poorly on safety among its peers, owing to its “very weak” risk management practices. Earlier this month, xAI missed a self-imposed deadline to publish a finalized AI safety framework.