{"id":68156,"date":"2025-07-16T21:26:12","date_gmt":"2025-07-16T21:26:12","guid":{"rendered":"https:\/\/www.europesays.com\/us\/68156\/"},"modified":"2025-07-16T21:26:12","modified_gmt":"2025-07-16T21:26:12","slug":"openai-and-anthropic-researchers-decry-reckless-safety-culture-at-elon-musks-xai","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/us\/68156\/","title":{"rendered":"OpenAI and Anthropic researchers decry &#8216;reckless&#8217; safety culture at Elon Musk&#8217;s xAI"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the \u201creckless\u201d and \u201ccompletely irresponsible\u201d safety culture at xAI, the billion-dollar AI startup owned by Elon Musk.<\/p>\n<p class=\"wp-block-paragraph\">The criticisms follow weeks of scandals at xAI that have overshadowed the company\u2019s technological advances.<\/p>\n<p class=\"wp-block-paragraph\">Last week, the company\u2019s AI chatbot, Grok, <a href=\"https:\/\/techcrunch.com\/2025\/07\/08\/grok-is-being-antisemitic-again-and-also-the-sky-is-blue\/\" rel=\"nofollow noopener\" target=\"_blank\">spouted antisemitic comments<\/a> and repeatedly called itself \u201cMechaHitler.\u201d Shortly after xAI took its chatbot offline to address the problem, it <a href=\"https:\/\/techcrunch.com\/2025\/07\/09\/elon-musks-xai-launches-grok-4-alongside-a-300-monthly-subscription\/\" rel=\"nofollow noopener\" target=\"_blank\">launched an increasingly capable frontier AI model, Grok 4<\/a>, which TechCrunch and others found to <a href=\"https:\/\/techcrunch.com\/2025\/07\/10\/grok-4-seems-to-consult-elon-musk-to-answer-controversial-questions\/\" rel=\"nofollow noopener\" target=\"_blank\">consult Elon Musk\u2019s personal politics for help answering hot-button issues.<\/a> In the latest development, <a href=\"https:\/\/techcrunch.com\/2025\/07\/15\/of-course-groks-ai-companions-want-to-have-sex-and-burn-down-schools\/\" rel=\"nofollow noopener\" target=\"_blank\">xAI launched AI companions<\/a> that take the form of a hyper-sexualized anime girl and an overly aggressive panda.<\/p>\n<p class=\"wp-block-paragraph\">Friendly joshing among employees of competing AI labs is fairly normal, but these researchers seem to be calling for increased attention to xAI\u2019s safety practices, which they claim to be at odds with industry norms.<\/p>\n<p class=\"wp-block-paragraph\">\u201cI didn\u2019t want to post on Grok safety since I work at a competitor, but it\u2019s not about competition,\u201d said Boaz Barak, a computer science professor currently on leave from Harvard to work on safety research at OpenAI, in a Tuesday <a href=\"https:\/\/x.com\/boazbaraktcs\/status\/1945165577154175288\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">post on X.<\/a> \u201cI appreciate the scientists and engineers @xai but the way safety was handled is completely irresponsible.\u201d<\/p>\n<blockquote class=\"twitter-tweet\" data-width=\"500\" data-dnt=\"true\">\n<p lang=\"en\" dir=\"ltr\">I didn&#8217;t want to post on Grok safety since I work at a competitor, but it&#8217;s not about competition.<\/p>\n<p>I appreciate the scientists and engineers at <a rel=\"nofollow noopener\" href=\"https:\/\/twitter.com\/xai?ref_src=twsrc%5Etfw\" target=\"_blank\">@xai<\/a> but the way safety was handled is completely irresponsible. Thread below.<\/p>\n<p>\u2014 Boaz Barak (@boazbaraktcs) <a rel=\"nofollow noopener\" href=\"https:\/\/twitter.com\/boazbaraktcs\/status\/1945165577154175288?ref_src=twsrc%5Etfw\" target=\"_blank\">July 15, 2025<\/a><\/p><\/blockquote>\n<p class=\"wp-block-paragraph\">Barak particularly takes issue with xAI\u2019s decision to not publish system cards \u2014 industry standard reports that detail training methods and safety evaluations in a good faith effort to share information with the research community. As a result, Barak says it\u2019s unclear what safety training was done on Grok 4.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI and Google have a spotty reputation themselves when it comes to promptly sharing system cards when unveiling new AI models. OpenAI decided <a href=\"https:\/\/techcrunch.com\/2025\/04\/15\/openai-ships-gpt-4-1-without-a-safety-report\/\" rel=\"nofollow noopener\" target=\"_blank\">not to publish a system card for GPT-4.1,<\/a> claiming it was not a frontier model. Meanwhile, <a href=\"https:\/\/techcrunch.com\/2025\/04\/03\/google-is-shipping-gemini-models-faster-than-its-ai-safety-reports\/\" rel=\"nofollow noopener\" target=\"_blank\">Google waited months after unveiling Gemini 2.5 Pro to publish a safety report<\/a>. However, these companies historically publish safety reports for all frontier AI models before they enter full production.<\/p>\n<p>Techcrunch event<\/p>\n<p>\n\t\t\t\t\t\t\t\t\tSan Francisco<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\t|<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\tOctober 27-29, 2025\n\t\t\t\t\t\t\t<\/p>\n<p class=\"wp-block-paragraph\">Barak also notes that Grok\u2019s AI companions \u201ctake the worst issues we currently have for emotional dependencies and tries to amplify them.\u201d In recent years, we\u2019ve seen <a href=\"https:\/\/apnews.com\/article\/chatbot-ai-lawsuit-suicide-teen-artificial-intelligence-9d48adc572100822fdbc3c90d1456bd0\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">countless stories<\/a> of <a href=\"https:\/\/www.nytimes.com\/2025\/06\/13\/technology\/chatgpt-ai-chatbots-conspiracies.html\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">unstable people developing concerning relationship with chatbots<\/a>, and how AI\u2019s over-agreeable answers can tip them over the edge of sanity.<\/p>\n<p class=\"wp-block-paragraph\">Samuel Marks, an AI safety researcher with Anthropic, also took issue with xAI\u2019s decision not to publish a safety report, calling the move \u201creckless.\u201d<\/p>\n<p class=\"wp-block-paragraph\">\u201cAnthropic, OpenAI, and Google\u2019s release practices have issues,\u201d Marks wrote in a <a href=\"https:\/\/x.com\/saprmarks\/status\/1944455357629333938\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">post on X<\/a>. \u201cBut they at least do something, anything to assess safety pre-deployment and document findings. xAI does not.\u201d<\/p>\n<blockquote class=\"twitter-tweet\" data-width=\"500\" data-dnt=\"true\">\n<p lang=\"en\" dir=\"ltr\">xAI launched Grok 4 without any documentation of their safety testing. This is reckless and breaks with industry best practices followed by other major AI labs.<\/p>\n<p>If xAI is going to be a frontier AI developer, they should act like one. \ud83e\uddf5<\/p>\n<p>\u2014 Samuel Marks (@saprmarks) <a rel=\"nofollow noopener\" href=\"https:\/\/twitter.com\/saprmarks\/status\/1944455357629333938?ref_src=twsrc%5Etfw\" target=\"_blank\">July 13, 2025<\/a><\/p><\/blockquote>\n<p class=\"wp-block-paragraph\">The reality is that we don\u2019t really know what xAI did to test Grok 4. In a widely shared post in the online forum LessWrong, <a href=\"https:\/\/www.lesswrong.com\/posts\/dqd54wpEfjKJsJBk6\/xai-s-grok-4-has-no-meaningful-safety-guardrails\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">one anonymous researcher claims that Grok 4 has no meaningful safety guardrails<\/a> based on their testing.<\/p>\n<p class=\"wp-block-paragraph\">Whether that\u2019s true or not, the world seems to be finding out about Grok\u2019s shortcomings in real time. Several of xAI\u2019s safety issues have since gone viral, and the company claims to have addressed them with <a href=\"https:\/\/x.com\/xai\/status\/1945039609840185489\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">tweaks to Grok\u2019s system prompt.<\/a><\/p>\n<p class=\"wp-block-paragraph\">OpenAI, Anthropic, and xAI did not respond to TechCrunch\u2019s request for comment.<\/p>\n<p class=\"wp-block-paragraph\">Dan Hendrycks, a safety adviser for xAI and director of the Center for AI Safety, <a href=\"https:\/\/x.com\/DanHendrycks\/status\/1943703394100850830\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">posted on X<\/a> that the company did \u201cdangerous capability evaluations\u201d on Grok 4. However, the results to those evaluations have not been publicly shared.<\/p>\n<p class=\"wp-block-paragraph\">\u201cIt concerns me when standard safety practices aren\u2019t upheld across the AI industry, like publishing the results of dangerous capability evaluations,\u201d said Steven Adler, an independent AI researcher who previously led safety teams at OpenAI, in a statement to TechCrunch. \u201cGovernments and the public deserve to know how AI companies are handling the risks of the very powerful systems they say they\u2019re building.\u201d<\/p>\n<p class=\"wp-block-paragraph\">What\u2019s interesting about xAI\u2019s questionable safety practices is that Musk has long been <a href=\"https:\/\/techcrunch.com\/2024\/08\/26\/elon-musk-unexpectedly-offers-support-for-californias-ai-bill\/\" rel=\"nofollow noopener\" target=\"_blank\">one of the AI safety industry\u2019s most notable advocates<\/a>. The billionaire leader of xAI, Tesla, and SpaceX has <a href=\"https:\/\/techcrunch.com\/2023\/03\/28\/1100-notable-signatories-just-signed-an-open-letter-asking-all-ai-labs-to-immediately-pause-for-at-least-6-months\/\" rel=\"nofollow noopener\" target=\"_blank\">warned many times<\/a> about the potential for advanced AI systems to cause catastrophic outcomes for humans, and he\u2019s praised an open approach to developing AI models.<\/p>\n<p class=\"wp-block-paragraph\">And yet, AI researchers at competing labs claim xAI is veering from industry norms around safely releasing AI models. In doing so, Musk\u2019s startup may be inadvertently making a strong case for state and federal lawmakers to set rules around publishing AI safety reports.<\/p>\n<p class=\"wp-block-paragraph\">There are several attempts at the state level to do so. California state Sen. Scott Wiener is <a href=\"https:\/\/techcrunch.com\/2025\/07\/09\/california-lawmaker-behind-sb-1047-reignites-push-for-mandated-ai-safety-reports\/\" rel=\"nofollow noopener\" target=\"_blank\">pushing a bill <\/a>that would require leading AI labs \u2014 likely including xAI \u2014 to publish safety reports, while <a href=\"https:\/\/techcrunch.com\/2025\/06\/13\/new-york-passes-a-bill-to-prevent-ai-fueled-disasters\/\" rel=\"nofollow noopener\" target=\"_blank\">New York Gov. Kathy Hochul is currently considering a similar bill<\/a>. Advocates of these bills note that most AI labs publish this type of information anyway \u2014 but evidently, not all of them do it consistently.<\/p>\n<p class=\"wp-block-paragraph\">AI models today have yet to exhibit real-world scenarios in which they create truly catastrophic harms, such as the death of people or billions of dollars in damages. However, many AI researchers say that this could be a problem in the near future given the rapid progress of AI models, and the billions of dollars Silicon Valley is investing to further improve AI.<\/p>\n<p class=\"wp-block-paragraph\">But even for skeptics of such catastrophic scenarios, there\u2019s a strong case to suggest that Grok\u2019s misbehavior makes the products it powers today significantly worse. <\/p>\n<p class=\"wp-block-paragraph\">Grok spread antisemitism around the X platform this week, <a href=\"https:\/\/techcrunch.com\/2025\/05\/14\/grok-is-unpromptedly-telling-x-users-about-south-african-genocide\/\" rel=\"nofollow noopener\" target=\"_blank\">just a few weeks after the chatbot repeatedly brought up \u201cwhite genocide\u201d<\/a> in conversations with users. Musk has indicated that Grok will be <a href=\"https:\/\/techcrunch.com\/2025\/07\/10\/grok-is-coming-to-tesla-vehicles-next-week-says-elon-musk\/\" rel=\"nofollow noopener\" target=\"_blank\">more ingrained in Tesla<\/a> vehicles, and xAI is trying to sell <a href=\"https:\/\/www.cbsnews.com\/news\/grok-elon-musk-xai-pentagon-contract\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">its AI models to The Pentagon<\/a> and other enterprises. It\u2019s hard to imagine that people driving Musk\u2019s cars, federal workers protecting the U.S., or enterprise employees automating tasks will be any more receptive to these misbehaviors than users on X.<\/p>\n<p class=\"wp-block-paragraph\">Several researchers argue that AI safety and alignment testing not only ensures that the worst outcomes don\u2019t happen, but they also protect against near-term behavioral issues.<\/p>\n<p class=\"wp-block-paragraph\">At the very least, Grok\u2019s incidents tend to overshadow xAI\u2019s rapid progress in developing frontier AI models that best OpenAI and Google\u2019s technology, just a couple years after the startup was founded. <\/p>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the \u201creckless\u201d and \u201ccompletely&hellip;\n","protected":false},"author":3,"featured_media":68157,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6],"tags":[48300,24142,64,66,34949,305,67,132,68,744],"class_list":{"0":"post-68156","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-business","8":"tag-ai-safety","9":"tag-anthropic","10":"tag-business","11":"tag-elon-musk","12":"tag-grok","13":"tag-openai","14":"tag-united-states","15":"tag-unitedstates","16":"tag-us","17":"tag-xai"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@us\/114865031321357790","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/68156","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/comments?post=68156"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/68156\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media\/68157"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media?parent=68156"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/categories?post=68156"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/tags?post=68156"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}