AI safety researchers from Openai, humanity and other organizations have made public comments about Xai's “reckless” and “completely irresponsible” safety culture, a billion-dollar AI startup owned by Elon Musk.
Criticism follows a week-long scandal at Xai that masked the company's technological advances.
Last week, the company's AI chatbot Grok spewed anti-Semitic comments, repeatedly calling them “Mechahitler.” Shortly after Xai took the chatbot offline to address the issue, it launched the Grok 4, an increasingly capable frontier AI model. In the latest development, Xai has launched an AI companion in the form of hypersexualized anime girls and an overly aggressive panda.
While friendly josting among competing AI Labs employees is fairly normal, these researchers seem to be asking for attention to Xai's safety practices, claiming they are at odds with industry norms.
“I work for a competitor and didn't want to post on Groke's safety, but it's not about competition,” Boaz Barak, a currently on vacation computer science professor at Harvard University, thanked Xai scientists and engineers in X's Tuesday post, but is handled entirely in safety. โ
Barak has problems with Xai's decision not to reveal the system card. Industry Standards reports in-depth training methods and safety assessments of honest efforts to share information with the research community. As a result, Barak says it is unknown that safety training was conducted on the Grok 4.
Openai and Google have a uneven reputation for quickly sharing system cards when it comes to launching new AI models. Openai has decided not to reveal its GPT-4.1 system cards. It claims that this is not a frontier model. Meanwhile, Google released a safety report a few months after its release of its Gemini 2.5 Pro. However, these companies have historically published safety reports for all frontier AI models before they enter full production.
TechCrunch Events
San Francisco
|
October 27th-29th, 2025
Barak also states that Grok's AI peers “take the worst issues they currently have with emotional dependencies and try to amplify them.” In recent years, I have seen countless stories of unstable people regarding their relationship with chatbots.
Samuel Marks, a human safety researcher, calls the move “reckless,” with issues with Xai's decision not to publish its safety report.
“There are issues with humanity, Openai and Google's release practices,” Marks wrote in X's post.
The reality is that I really don't know what Xai did to test the Grok 4. In a widely shared post on the online forum, LessWrong, it claims that Grok 4 does not have meaningful safety guardrails based on tests.
Whether it's true or not, the world seems to know about Glock's real-time shortcomings. Some of Xai's safety issues have since become viruses, and the company claims it has addressed them with tweaks to Grok's system prompts.
Openai, Anthropology, and Xai did not respond to TechCrunch requests for comments.
Dan Hendrycks, XAI safety advisor and director of the AI Center Center, posted to X that the company had “dangerous capabilities assessments” on GROK 4. However, the results of these assessments have not been published.
“I'm worried if standard safety practices are not supported across the AI industry, such as publishing the results of dangerous capacity assessments,” said Steven Adler, an independent AI researcher who previously led dangerous capacity assessments at Openai in a statement from TechCrunch. “The government and the public deserve to know how AI companies handle the risks of the very powerful systems they say they're building.”
What's interesting about Xai's questionable safety practices is that Musk has long been one of the most notable advocates in the AI safety industry. The billionaire owners of Xai, Tesla and SpaceX have warned many times about the possibility that advanced AI systems can have catastrophic human consequences and praised their open approach to developing AI models.
Still, AI researchers at competing labs claim that Xai is heading from the industry norms where AI models are safely releasing. In doing so, mask startups may be making a false and strong claim by state and federal lawmakers to set rules regarding the publication of AI safety reports.
At the state level, there are several attempts to do so. While California Sen. Scott Wiener is pushing for a bill that calls for major AI labs that could include XAI to release safety reports, New York Gov. Kathy Hochul is currently considering a similar bill. Supporters of these bills are paying attention to most AI labs publish this type of information anyway, but obviously not all of them do that consistently.
Today's AI models still don't show real-world scenarios that can truly devastatingly harm, such as deaths of people or billions of dollars of damage. However, many AI researchers say this could become a problem in the near future given the rapid advances in AI models, with billions of dollars investing in improving AI even further.
But even skeptics of such a catastrophic scenario have a strong case that suggests that Grok's misconduct will significantly exacerbate today's products.
Grok spread anti-Semitism on the X platform this week, just weeks after Chatbot repeatedly announced its “white genocide” in conversations with users. Soon, the mask shows that Glock is seeping into Tesla vehicles, and Zai is about to sell its AI model to pentagons and other companies. It's hard to imagine that musk cars, federal workers protecting the US, or employees of companies automating tasks can accept these fraudsters more than X users.
Several researchers argue that AI safety and alignment testing not only ensures that the worst outcomes will not occur, but also ensures that it protects against short-term behavioral problems.
At the very least, Grok's case tends to overshadow Xai's rapid advances in developing frontier AI models, where Openai and Google's technology have been best-in-one, just a few years after the startup was established.