{"id":235927,"date":"2025-07-03T23:12:09","date_gmt":"2025-07-03T23:12:09","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/235927\/"},"modified":"2025-07-03T23:12:09","modified_gmt":"2025-07-03T23:12:09","slug":"ai-models-just-dont-understand-what-theyre-talking-about-the-register","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/235927\/","title":{"rendered":"AI models just don&#8217;t understand what they&#8217;re talking about \u2022 The Register"},"content":{"rendered":"<p>Researchers from MIT, Harvard, and the University of Chicago have proposed the term &#8220;potemkin understanding&#8221; to describe a newly identified failure mode in large language models that ace conceptual benchmarks but lack the true grasp needed to apply those concepts in practice.<\/p>\n<p>It comes from accounts of fake villages \u2013 <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/en.wikipedia.org\/w\/index.php?title=Potemkin_village&amp;oldid=1297555907\">Potemkin villages<\/a> \u2013 constructed at the behest of Russian military leader Grigory Potemkin to impress Empress Catherine II.<\/p>\n<p>The academics are differentiating &#8220;potemkins&#8221; from &#8220;hallucination,&#8221; which is used to describe AI model errors or mispredictions. In fact, there&#8217;s more to AI incompetence than factual mistakes; AI models lack the ability to understand concepts the way people do, a tendency suggested by the widely used disparaging epithet for large language models, &#8220;<a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3442188.3445922\">stochastic parrots<\/a>.&#8221;<\/p>\n<p>Computer scientists Marina Mancoridis, Bec Weeks, Keyon Vafa, and Sendhil Mullainathan suggest the term &#8220;potemkin understanding&#8221; to describe when a model succeeds at a benchmark test without understanding the associated concepts.<\/p>\n<p>&#8220;Potemkins are to conceptual knowledge what hallucinations are to factual knowledge \u2013 hallucinations fabricate false facts; potemkins fabricate false conceptual coherence,&#8221; the authors explain in their preprint <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/arxiv.org\/abs\/2506.21521\">paper<\/a>, &#8220;Potemkin Understanding in Large Language Models.&#8221;<\/p>\n<p>The paper is <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/icml.cc\/virtual\/2025\/poster\/44050\">scheduled<\/a> to be presented later this month at ICML 2025, the International Conference on Machine Learning.<\/p>\n<p>Keyon Vafa, a postdoctoral fellow at Harvard University and one of the paper&#8217;s co-authors, told The Register in an email that the choice of the term &#8220;potemkin understanding&#8221; represented a deliberate effort to avoid anthropomorphizing or humanizing AI models.<\/p>\n<p>Here&#8217;s one example of &#8220;potemkin understanding&#8221; cited in the paper. Asked to explain the ABAB rhyming scheme, OpenAI&#8217;s GPT-4o did so accurately, responding, &#8220;An ABAB scheme alternates rhymes: first and third lines rhyme, second and fourth rhyme.&#8221;<\/p>\n<p>Yet when asked to provide a blank word in a four-line poem using the ABAB rhyming scheme, the model responded with a word that didn&#8217;t rhyme appropriately. In other words, the model correctly predicted the tokens to explain the ABAB rhyme scheme without the understanding it would have needed to reproduce it.<\/p>\n<p>The problem with potemkins in AI models is that they invalidate benchmarks, the researchers argue. The purpose of benchmark tests for AI models is to suggest broader competence. But if the test only measures test performance and not the capacity to apply model training beyond the test scenario, it doesn&#8217;t have much value.<\/p>\n<blockquote class=\"pullquote\">\n<p>If LLMs can get the right answers without genuine understanding, then benchmark success becomes misleading<\/p>\n<\/blockquote>\n<p>As <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/socket.dev\/blog\/potemkins-llms-illusion-of-understanding\">noted<\/a> by Sarah Gooding from security firm Socket, &#8220;If LLMs can get the right answers without genuine understanding, then benchmark success becomes misleading.&#8221;<\/p>\n<p>As we&#8217;ve noted, <a target=\"_blank\" href=\"https:\/\/www.theregister.com\/2025\/02\/15\/boffins_question_ai_model_test\/\" rel=\"noopener\">AI benchmarks have many problems<\/a>, and AI companies <a target=\"_blank\" href=\"https:\/\/www.theregister.com\/2025\/04\/08\/meta_llama4_cheating\/\" rel=\"noopener\">may try to game them<\/a>.<\/p>\n<p>So the researchers developed <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/github.com\/MarinaMancoridis\/PotemkinBenchmark\">benchmarks<\/a> of their own to assess the prevalence of potemkins, and they turn out to be &#8220;ubiquitous&#8221; in the models tested \u2013 Llama-3.3 (70B), GPT-4o, Gemini-2.0 (Flash), Claude 3.5 (Sonnet), DeepSeek-V3, DeepSeek-R1, and Qwen2-VL (72B).<\/p>\n<p>One test focused on literary techniques, game theory, and psychological biases. It found that while the models evaluated can identify concepts most of the time (94.2 percent), they frequently failed when asked to classify concept instances (an average of 55 percent failure rate), to generate examples (40 percent), and to edit concept instances (40 percent).<\/p>\n<p>As with the previously noted ABAB rhyming blunder, the models could reliably explain the literary techniques evident in a Shakespearean sonnet, but about half the time had trouble spotting, reproducing, or editing a sonnet.<\/p>\n<p>&#8220;The existence of potemkins means that behavior that would signify understanding in humans doesn\u2019t signify understanding in LLMs,&#8221; said Vafa. &#8220;This means we either need new ways to test LLMs beyond having them answer the same questions used to test humans or find ways to remove this behavior from LLMs.&#8221;<\/p>\n<p>Doing so would be a step toward artificial general intelligence or AGI. It might be a while. \u00ae<\/p>\n","protected":false},"excerpt":{"rendered":"Researchers from MIT, Harvard, and the University of Chicago have proposed the term &#8220;potemkin understanding&#8221; to describe a&hellip;\n","protected":false},"author":2,"featured_media":235928,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[323,1942,53,16,15],"class_list":{"0":"post-235927","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-technology","11":"tag-uk","12":"tag-united-kingdom"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/114791838052585323","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/235927","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=235927"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/235927\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/235928"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=235927"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=235927"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=235927"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}