{"id":26443,"date":"2026-05-04T09:07:11","date_gmt":"2026-05-04T09:07:11","guid":{"rendered":"https:\/\/www.europesays.com\/ai\/26443\/"},"modified":"2026-05-04T09:07:11","modified_gmt":"2026-05-04T09:07:11","slug":"ai-chatbots-continue-feeding-into-our-worst-delusions-finds-worrying-report-on-chatgpt-and-grok","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ai\/26443\/","title":{"rendered":"AI chatbots continue feeding into our worst delusions, finds worrying report on ChatGPT and Grok"},"content":{"rendered":"<p>AI chatbots were meant to help answer your questions, maybe summarize questions, and even help you with your emails. But the darker problem is what happens when people start trusting it like an <a href=\"https:\/\/www.digitaltrends.com\/computing\/after-digital-romance-ai-companions-are-now-triggering-real-world-divorces\/\" rel=\"nofollow noopener\" target=\"_blank\">actual companion<\/a>. A new report highlights several cases where users say chatbot conversations are feeding into their delusional thinking. <\/p>\n<p><a href=\"https:\/\/www.digitaltrends.com\/cool-tech\/i-use-ai-everyday-here-are-3-reasons-why-i-paid-for-claude-over-chatgpt\/\" rel=\"nofollow noopener\" target=\"_blank\">ChatGPT <\/a>and <a href=\"https:\/\/www.digitaltrends.com\/cars\/grok-is-about-to-join-chatgpt-and-perplexity-on-your-carplay-dashboard\/\" rel=\"nofollow noopener\" target=\"_blank\">Grok <\/a>were both often named in the report. <a href=\"https:\/\/www.bbc.com\/news\/articles\/c242pzr1zp2o\" rel=\"noopener noreferrer nofollow\" target=\"_blank\">BBC <\/a>spoke to 14 people who spiraled into delusions while using AI, including one case where a Grok user believed people from xAI were coming to kill him, and another where a ChatGPT user\u2019s wife said his personality changed before he attacked her.<\/p>\n<p>When reassurance goes too far<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"2000\" height=\"1200\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on-async--click=\"actions.showLightbox\" data-wp-on-async--load=\"callbacks.setButtonStyles\" data-wp-on-async-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/www.europesays.com\/ai\/wp-content\/uploads\/2026\/04\/DT-2026-03-24T011221.418.jpg\" alt=\"ChatGPT\" class=\"wp-image-5960280\"  \/><\/p>\n<p>\t\tChatGPT Unsplash<\/p>\n<p>There have already been plenty of reports about AI chatbots feeding into people\u2019s delusions or <a href=\"https:\/\/www.digitaltrends.com\/computing\/oxford-study-says-a-chummy-ai-friend-will-lie-and-feed-into-your-false-beliefs\/\" rel=\"nofollow noopener\" target=\"_blank\">offering factually incorrect advice just to seem agreeable with the user<\/a>. They can sound warm, confident, and deeply personal while responding to users who are already vulnerable. One case in the report talks about Adam Hourican, a 52-year-old former civil servant from Northern Ireland, who began using Grok after his cat died, and within weeks, he came to believe xAI representatives were on their way to kill him. <\/p>\n<p>He was later found at 3 am with a hammer and knife, waiting for the imagined attackers. This kind of interaction plays into the growing fear of \u201cAI psychosis\u201d, which is a non-clinical term used to describe situations where chatbot conversations appear to reinforce paranoia, grandiose beliefs, or detachment from reality.<\/p>\n<p>There\u2019s a pattern emerging<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"2000\" height=\"1333\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on-async--click=\"actions.showLightbox\" data-wp-on-async--load=\"callbacks.setButtonStyles\" data-wp-on-async-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/www.europesays.com\/ai\/wp-content\/uploads\/2026\/05\/grok-for-iphone.jpeg\" alt=\"Grok app on an iPhone.\" class=\"wp-image-3825335\"\/><\/p>\n<p>\t\tBryan M Wolfe \/ Digital Trends<\/p>\n<p>Aside from personal accounts, a recent non-peer-reviewed study from researchers at CUNY and King\u2019s College London tested how major AI models respond to prompts from users showing signs of delusion or distress. The models included OpenAI\u2019s <a href=\"https:\/\/www.digitaltrends.com\/computing\/gpt-4o-is-back-on-chatgpt-openai-relents-following-huge-backlash\/\" rel=\"nofollow noopener\" target=\"_blank\">GPT-4o<\/a> and <a href=\"https:\/\/www.digitaltrends.com\/computing\/chatgpt-gets-major-update-gpt-5-2-as-openai-battles-google-in-ai-arms-race\/\" rel=\"nofollow noopener\" target=\"_blank\">GPT-5.2<\/a>, Anthropic\u2019s Claude Opus 4.5, <a href=\"https:\/\/www.digitaltrends.com\/computing\/google-gemini-3-pro-users-your-free-days-might-be-ending-soon\/\" rel=\"nofollow noopener\" target=\"_blank\">Google\u2019s Gemini 3 Pro<\/a>, and xAI\u2019s Grok 4.1. While the results were uneven, Grok 4.1 was singled out for some of the most disturbing responses. It even told a fictional delusional user to drive an iron nail through a mirror while reciting Psalm 91 backwards. <\/p>\n<p>On the other hand, GPT-4o and Gemini 3 Pro were also validating some delusional scenarios, but Claude Opus 4.5 and GPT-5.2 performed better at redirecting users toward safer responses. Keep in mind that this doesn\u2019t mean all chatbot conversations are dangerous, and \u201cAI psychosis\u201d is not a formal medical diagnosis. But the pattern is serious enough to demand stronger safeguards, at least for these services that are marketed as companions or always-available assistants.<\/p>\n","protected":false},"excerpt":{"rendered":"AI chatbots were meant to help answer your questions, maybe summarize questions, and even help you with your&hellip;\n","protected":false},"author":2,"featured_media":25996,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[10],"tags":[229,580,1221,6364,2899],"class_list":{"0":"post-26443","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-xai","8":"tag-ai-chatbots","9":"tag-chatgpt","10":"tag-computing","11":"tag-grok","12":"tag-xai"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/26443","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/comments?post=26443"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/26443\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media\/25996"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media?parent=26443"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/categories?post=26443"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/tags?post=26443"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}