{"id":23297,"date":"2026-04-30T17:17:08","date_gmt":"2026-04-30T17:17:08","guid":{"rendered":"https:\/\/www.europesays.com\/ai\/23297\/"},"modified":"2026-04-30T17:17:08","modified_gmt":"2026-04-30T17:17:08","slug":"the-goblins-came-back-to-haunt-us-openai-explains-how-chatgpts-nerdy-personality-got-out-of-control","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ai\/23297\/","title":{"rendered":"&#8216;The Goblins Came Back to Haunt Us&#8217;: OpenAI Explains How ChatGPT&#8217;s &#8216;Nerdy&#8217; Personality Got Out of Control"},"content":{"rendered":"<p>Earlier this week, OpenAI posted a document on GitHub as part of the open-sourcing of its coding agent, Codex CLI, that <a href=\"https:\/\/gizmodo.com\/never-talk-about-goblins-openais-instructions-to-codex-have-a-weirdly-emphatic-no-creatures-policy-2000751984\" rel=\"nofollow noopener\" target=\"_blank\">revealed an unusual system prompt<\/a> for GPT-5.5. The model was explicitly instructed, in coding contexts, to never talk about \u201cgoblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures\u201d unless it was \u201cabsolutely and unambiguously relevant\u201d to a user\u2019s request.<\/p>\n<p>Now, OpenAI has finally explained why that oddly specific instruction, which appeared twice in the prompt, was so important in the first place.<\/p>\n<p>For at least a year, <a href=\"https:\/\/www.reddit.com\/r\/ChatGPT\/comments\/1k5hg5c\/does_anyone_elses_chatgpt_refer_to_people_as\/\" rel=\"nofollow noopener\" target=\"_blank\">some ChatGPT users<\/a> have noticed the LLM\u2019s quirky habit of bringing up goblins, gremlins, trolls, and other creatures in its answers. The weird tic apparently became more common as newer models rolled out.<\/p>\n<p>Even OpenAI CEO Sam Altman referenced the issue in a post on X Monday morning.<\/p>\n<p>\u201cFeels like codex is having a ChatGPT moment,\u201d <a href=\"https:\/\/x.com\/sama\/status\/2049493797114126484\" rel=\"nofollow\">Altman wrote<\/a>. \u201cI meant a goblin moment, sorry.\u201d<\/p>\n<p>That same day, OpenAI published a <a href=\"https:\/\/openai.com\/index\/where-the-goblins-came-from\/\" rel=\"nofollow noopener\" target=\"_blank\">blog post<\/a> explaining the strange behavior and how the company finally addressed it.<\/p>\n<p>According to the post, OpenAI first became aware of the model\u2019s goblin obsession with the release of GPT-5.1 in November. The company launched an internal investigation after users complained that the model had become overly familiar in its responses. A safety researcher suggested adding \u201cgoblin\u201d and \u201cgremlin\u201d to the review after repeatedly encountering the words while using the model.<\/p>\n<p>The company found that use of the word \u201cgoblin\u201d in ChatGPT had jumped 175% after the launch of GPT-5.1, while mentions of \u201cgremlin\u201d rose 52%.<\/p>\n<p>At the time, OpenAI apparently didn\u2019t consider the behavior too concerning. But just a few months later, \u201cthe goblins came back to haunt us,\u201d the company wrote in the blog post.<\/p>\n<p>By March, with the release of GPT-5.4, references to the creatures had increased even further. Some users complained online that the word \u201cgoblin\u201d was appearing in \u201c<a href=\"https:\/\/news.ycombinator.com\/item?id=47319285\" rel=\"nofollow noopener\" target=\"_blank\">almost every conversation<\/a>.\u201d<\/p>\n<p>That prompted another internal analysis, which OpenAI says uncovered the root of the problem. The company found that references to these creatures were especially common in responses for users who selected the model\u2019s \u201cNerdy\u201d personality setting.<\/p>\n<p>That personality included a system prompt instructing the model to \u201cundercut pretension through playful use of language.\u201d<\/p>\n<p>OpenAI used its coding agent Codex to compare outputs generated during reinforcement learning training that included words like \u201cgoblin\u201d and \u201cgremlin\u201d against outputs that did not. The company found that one reward signal favored responses containing mentions of these creatures, scoring them higher than otherwise similar answers that did not use those words.<\/p>\n<p>Researchers also found that mentions of goblins, gremlins, and other creatures began spreading beyond the Nerdy personality.<\/p>\n<p>\u201cOnce a style tic is rewarded, later training can spread or reinforce it elsewhere, especially if those outputs are reused in supervised fine-tuning or preference data,\u201d the blog said.<\/p>\n<p>To address the problem, OpenAI said it retired the Nerdy personality, removed the reward signal that favored goblin mentions, and filtered training data containing creature words.<\/p>\n<p>Because GPT-5.5 had already begun training before the root cause was discovered, the newer model also had a strange obsession with goblins. OpenAI said it added the developer-prompt instruction, which some users later spotted in the model\u2019s open source code, to help curb inappropriate mentions of goblins and gremlins.<\/p>\n<p>\u201cDepending on who you ask, the goblins are a delightful or annoying quirk of the model,\u201d OpenAI wrote in the blog. \u201cBut they are also a powerful example of how reward signals can shape model behavior in unexpected ways, and how models can learn to generalize rewards in certain situations to unrelated ones.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"Earlier this week, OpenAI posted a document on GitHub as part of the open-sourcing of its coding agent,&hellip;\n","protected":false},"author":2,"featured_media":23298,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[7],"tags":[24,580,2317,157],"class_list":{"0":"post-23297","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-openai","8":"tag-ai","9":"tag-chatgpt","10":"tag-codex","11":"tag-openai"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/23297","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/comments?post=23297"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/23297\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media\/23298"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media?parent=23297"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/categories?post=23297"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/tags?post=23297"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}