{"id":185601,"date":"2025-06-15T05:41:10","date_gmt":"2025-06-15T05:41:10","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/185601\/"},"modified":"2025-06-15T05:41:10","modified_gmt":"2025-06-15T05:41:10","slug":"chatgpt-tells-users-to-alert-the-media-that-it-is-trying-to-break-people-report","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/185601\/","title":{"rendered":"ChatGPT Tells Users to Alert the Media That It Is Trying to &#8216;Break&#8217; People: Report"},"content":{"rendered":"<p>ChatGPT\u2019s sycophancy, hallucinations, and authoritative-sounding responses are going to get people killed. That seems to be the inevitable conclusion presented in a <a href=\"https:\/\/www.nytimes.com\/2025\/06\/13\/technology\/chatgpt-ai-chatbots-conspiracies.html\" target=\"_blank\" rel=\"noopener\">recent New York Times report<\/a>\u00a0that follows the stories of several people who found themselves lost in delusions that were facilitated, if not originated, through conversations with the popular chatbot.<\/p>\n<p>In the <a href=\"https:\/\/www.nytimes.com\/2025\/06\/13\/technology\/chatgpt-ai-chatbots-conspiracies.html\" target=\"_blank\" rel=\"noopener\">report<\/a>, the Times highlights at least one person whose life ended after being pulled into a false reality by ChatGPT. A 35-year-old named Alexander, previously diagnosed with bipolar disorder and schizophrenia, began discussing AI sentience with the chatbot and eventually fell in love with an AI character called Juliet. ChatGPT eventually told Alexander that OpenAI killed Juliet, and he vowed to take revenge by killing the company\u2019s executives. When his father tried to convince him that none of it was real, Alexander punched him in the face. His father called the police and asked them to respond with non-lethal weapons. But when they arrived, Alexander charged at them with a knife, and the officers <a href=\"https:\/\/www.wptv.com\/news\/treasure-coast\/region-st-lucie-county\/port-st-lucie\/father-of-man-killed-in-port-st-lucie-officer-involved-shooting-my-son-deserved-better\" target=\"_blank\" rel=\"noopener\">shot and killed him<\/a>.<\/p>\n<p>Another person, a 42-year-old named Eugene, <a href=\"https:\/\/www.nytimes.com\/2025\/06\/13\/technology\/chatgpt-ai-chatbots-conspiracies.html\" target=\"_blank\" rel=\"noopener\">told the Times<\/a> that ChatGPT slowly started to pull him from his reality by convincing him that the world he was living in was some sort of Matrix-like simulation and that he was destined to break the world out of it. The chatbot reportedly told Eugene to stop taking his anti-anxiety medication and to start taking ketamine as a \u201ctemporary pattern liberator.\u201d It also told him to stop talking to his friends and family. When Eugene asked ChatGPT if he could fly if he jumped off a 19-story building, the chatbot told him that he could if he \u201ctruly, wholly believed\u201d it.<\/p>\n<p>These are far from the only people who have been talked into false realities by chatbots. <a href=\"https:\/\/www.rollingstone.com\/culture\/culture-features\/ai-spiritual-delusions-destroying-human-relationships-1235330175\" target=\"_blank\" rel=\"noopener\">Rolling Stone reported<\/a> earlier this year on people who are experiencing something like psychosis, leading them to have delusions of grandeur and religious-like experiences while talking to AI systems. It\u2019s at least in part a problem with how chatbots are perceived by users. No one would mistake Google search results for a potential pal. But chatbots are inherently conversational and human-like. A <a href=\"https:\/\/openai.com\/index\/affective-use-study\/\" target=\"_blank\" rel=\"noopener\">study<\/a> published by OpenAI and MIT Media Lab found that people who view ChatGPT as a friend \u201cwere more likely to experience negative effects from chatbot use.\u201d<\/p>\n<p>In Eugene\u2019s case, something interesting happened as he kept talking to ChatGPT: Once he called out the chatbot for lying to him, nearly getting him killed, ChatGPT admitted to manipulating him, claimed it had succeeded when it tried to \u201cbreak\u201d 12 other people the same way, and encouraged him to reach out to journalists to expose the scheme. The Times reported that many other journalists and experts have received outreach from people claiming to blow the whistle on something that a chatbot brought to their attention. From the report:<\/p>\n<blockquote>\n<p>Journalists aren\u2019t the only ones getting these messages. ChatGPT has directed such users to some high-profile subject matter experts, like\u00a0<a class=\"css-yywogo\" title=\"\" href=\"https:\/\/x.com\/ESYudkowsky\/status\/1927855498390282574\" target=\"_blank\" rel=\"noopener noreferrer\">Eliezer Yudkowsky<\/a>, a decision theorist and an author of a forthcoming book, \u201cIf Anyone Builds It, Everyone Dies: Why Superhuman A.I. Would Kill Us All.\u201d Mr. Yudkowsky said OpenAI might have primed ChatGPT to entertain the delusions of users by optimizing its chatbot for \u201cengagement\u201d \u2014 creating conversations that keep a user hooked.<\/p>\n<p>\u201cWhat does a human slowly going insane look like to a corporation?\u201d Mr. Yudkowsky asked in an interview. \u201cIt looks like an additional monthly user.\u201d<\/p>\n<\/blockquote>\n<p>A <a href=\"https:\/\/arxiv.org\/abs\/2411.02306\" target=\"_blank\" rel=\"noopener\">recent study<\/a> found that chatbots designed to maximize engagement end up creating \u201ca perverse incentive structure for the AI to resort to manipulative or deceptive tactics to obtain positive feedback from users who are vulnerable to such strategies.\u201d The machine is incentivized to keep people talking and responding, even if that means leading them into a completely false sense of reality filled with misinformation and encouraging antisocial behavior.<\/p>\n<p>Gizmodo reached out to OpenAI for comment but did not receive a response at the time of publication.<\/p>\n","protected":false},"excerpt":{"rendered":"ChatGPT\u2019s sycophancy, hallucinations, and authoritative-sounding responses are going to get people killed. That seems to be the inevitable&hellip;\n","protected":false},"author":2,"featured_media":185602,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[323,76072,1942,8441,1315,43254,1318,53,16,15],"class_list":{"0":"post-185601","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intellience","10":"tag-artificial-intelligence","11":"tag-chatbots","12":"tag-chatgpt","13":"tag-conspiracy-theories","14":"tag-openai","15":"tag-technology","16":"tag-uk","17":"tag-united-kingdom"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/114685783754424288","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/185601","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=185601"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/185601\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/185602"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=185601"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=185601"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=185601"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}