{"id":349849,"date":"2025-08-16T19:06:51","date_gmt":"2025-08-16T19:06:51","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/349849\/"},"modified":"2025-08-16T19:06:51","modified_gmt":"2025-08-16T19:06:51","slug":"contra-dwarkesh-on-continual-learning","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/349849\/","title":{"rendered":"Contra Dwarkesh on Continual Learning"},"content":{"rendered":"<p>\u2019s now well-read <a href=\"https:\/\/www.dwarkesh.com\/p\/timelines-june-2025\" rel=\"noopener\" target=\"_blank\">post<\/a> on why he is extending his AI timelines focuses on the idea of continual learning. If you ask me, <a href=\"https:\/\/www.interconnects.ai\/p\/agi-is-what-you-want-it-to-be\" rel=\"noopener\" target=\"_blank\">what we have already is AGI<\/a>, so the core question is: Is continual learning a bottleneck on AI progress?<\/p>\n<p>In this post, I argue that continual learning as he describes it actually doesn\u2019t matter for the trajectory of AI progress that we are on. Continual learning will eventually be solved, but in the sort of way that a new type of AI will emerge from it, rather than continuing to refine what it means to host ever more powerful LLM-based systems. <\/p>\n<p>Continual learning is the ultimate algorithmic nerd snipe for AI researchers, when in reality all we need to do is keep scaling systems and we\u2019ll get something indistinguishable from how humans do it, for free.<\/p>\n<p data-attrs=\"{&quot;url&quot;:&quot;https:\/\/www.interconnects.ai\/p\/contra-dwarkesh-on-continual-learning?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}\" data-component-name=\"ButtonCreateButton\" class=\"button-wrapper\"><a href=\"https:\/\/www.interconnects.ai\/p\/contra-dwarkesh-on-continual-learning?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share\" rel=\"noopener\" class=\"button primary\" target=\"_blank\">Share<\/a><\/p>\n<p>To start, here\u2019s the core of the Dwarkesh piece as a refresher for what he means by continual learning.<\/p>\n<blockquote>\n<p>Sometimes people say that even if all AI progress totally stopped, the systems of today would still be far more economically transformative than the internet. I disagree. I think the LLMs of today are magical. But the reason that the Fortune 500 aren\u2019t using them to transform their workflows isn\u2019t because the management is too stodgy. Rather, I think it\u2019s genuinely hard to get normal humanlike labor out of LLMs. And this has to do with some fundamental capabilities these models lack.<\/p>\n<p>I like to think I\u2019m \u201cAI forward\u201d here at the Dwarkesh Podcast. I\u2019ve probably spent over a hundred hours trying to build little LLM tools for my post production setup. And the experience of trying to get them to be useful has extended my timelines. I\u2019ll try to get the LLMs to rewrite autogenerated transcripts for readability the way a human would. Or I\u2019ll try to get them to identify clips from the transcript to tweet out. Sometimes I\u2019ll try to get them to co-write an essay with me, passage by passage. These are simple, self contained, short horizon, language in-language out tasks &#8211; the kinds of assignments that should be dead center in the LLMs\u2019 repertoire. And they&#8217;re 5\/10 at them. Don\u2019t get me wrong, that\u2019s impressive.<\/p>\n<p>But the fundamental problem is that LLMs don\u2019t get better over time the way a human would. The lack of continual learning is a huge huge problem. The LLM baseline at many tasks might be higher than an average human&#8217;s. But there\u2019s no way to give a model high level feedback. You\u2019re stuck with the abilities you get out of the box. You can keep messing around with the system prompt. In practice this just doesn\u2019t produce anything even close to the kind of learning and improvement that human employees experience.<\/p>\n<\/blockquote>\n<p>The core issue I have with this argument is the dream of making the LLMs we\u2019re building today look more like humans. In many ways I\u2019m surprised that Dwarkesh and other very AGI-focused AI researchers or commentators believe this \u2014 it\u2019s the same root argument that AI critics use when they say AI models don\u2019t reason. The goal to make AI more human is constraining the technological progress to a potentially impossible degree. <\/p>\n<p>Human intelligence has long been the inspiration for AI, but we have long surpassed it being the mirror we look to for inspiration. Now the industry is all in on the expensive path to make the best language models it possibly can. We\u2019re no longer trying to build the bird, we\u2019re trying to transition the Wright Brothers\u2019 invention into the 737 in the shortest time frame possible.<\/p>\n<p>To put it succinctly. My argument very much rhymes with some of my past writing. <\/p>\n<p>Do language models reason like humans? No. <br \/>Do language models reason? Yes. <\/p>\n<p>Will language model systems continually learn like humans? No.<br \/>Will language model systems continually learn? Of course.<\/p>\n<p>Dwarkesh writes \u201cRather, I think it\u2019s genuinely hard to get normal humanlike labor out of LLMs.\u201d This is because we\u2019re still early on the buildout of the technology. Human labor takes an immense amount of context and quick thinking, both of which we\u2019re starting to unlock with our language models. On top of this, human labor may not be what we want to create \u2014 we want to augment it. <\/p>\n<p>Using LLMs as drop in replacements for humans is not a requirement for AGI nor is what Dwarkesh describes a fundamental limitation on AI progress. Francois Chollet cleverly poked at this weakness in his recent <a href=\"https:\/\/youtu.be\/1if6XbzD5Yg\" rel=\"noopener\" target=\"_blank\">conversation<\/a> with Dwarkesh at an ARC-AGI event:<\/p>\n<blockquote>\n<p>Well, how do you define the difference between the ability to adapt to a new task and learning on the fly? It&#8217;s, it sounds like the same thing to me.<\/p>\n<\/blockquote>\n<p>Language models can already pick up subtle context extremely fast. ChatGPT\u2019s memory feature has gotten far better for me. When we\u2019re using the far more powerful models we can expect in the next 18 months this\u2019ll already start to appear magical. Language models are extremely apt at inferring context even without us giving it to them. Soon we\u2019ll be unlocking that subtle connection engine by providing immense, explicit context. <\/p>\n<p>I don\u2019t know of anyone who has actually thoroughly digitized all the relevant context of their job and formatted it in a way that is easily readable by an LLM. GPT-5 Pro <a href=\"https:\/\/chatgpt.com\/share\/689e683a-e28c-8005-b85d-5ae07fe0830b\" rel=\"noopener\" target=\"_blank\">estimates<\/a> that all of the writing on Interconnects would be only 500K tokens. That would fit into an existing LLM with no extra system, but I\u2019ve never tried it.<\/p>\n<p>The problem that Dwarkesh is facing is that we\u2019re still using LLMs primarily in a single generation manner, which got far better with the introduction of reasoning models, but the economically useful way to use current tools in more complex intellectual domains will require a deep-research style approach over all of your recent work interactions. No one is giving language models that kind of context. None of the tools we use are set up properly to accumulate this type of context.<\/p>\n<p>I expect this to change rapidly. ChatGPT, Claude, and the likes are all <a href=\"https:\/\/www.theverge.com\/news\/757743\/anthropic-claude-ai-search-past-chats\" rel=\"noopener\" target=\"_blank\">adding memory features<\/a> across chats and countless connectors to other pieces of information in your professional life. These memory features will be omnimodal and essential to extracting the type of value Dwarkesh wants. Without them, I agree language models in their current form are hopeless at solving continual learning.<\/p>\n<p>This is what I would expect the rumored $2000\/month ChatGPT level subscriptions to work with. Each of these bespoke tasks needs to absorb a ton of context and reasoning tokens in order to make a directionally right output. If someone built the Claude Code equivalent for my Substack, with every post tagged by topic and performance metrics, I bet the AI could easily make useful suggestions on how to format my content.<\/p>\n<p>Continual learning in how Dwarkesh presents it is a systems problem rather than a learning problem. I expect better context management over my information ecosystem to exist in 2026, but more work to be needed for the AI companies to know how best to reference it and unlock in-context learning that feels like rapid adaptation. Call that 2027.<\/p>\n<p>The models that have been released in 2025 will make this far more tractable in the near future. Reasoning models have made in-context learning far more powerful, resulting in rapid progress on held-out and complex domains such as <a href=\"https:\/\/x.com\/fchollet\/status\/1953511631054680085\" rel=\"\">ARC-AGI<\/a>. These models also have come with massive improvements in context length. Claude and Gemini have 1M+ token context lengths and GPT-5\u2019s is at 400K \u2014 they\u2019re all growing steadily. What is important with the context length numbers is that evaluations are showing that these are meaningful improvements that the models can leverage intelligently.<\/p>\n<p>With these reasoning models and smart retrieval of context, the systems we are building will look indistinguishable from continual learning. This will definitely be multiple LLMs working together and will operate very differently than the first versions of ChatGPT we were given (and often still use today).<\/p>\n<p>The path to continual learning is more context and more horsepower. This is directly in line with the direction AI investment is going. This doesn\u2019t feel like a bottleneck, rather another product problem that we are going to solve. This sort of continual learning may not enable the type of raw intelligence and autonomy that many vocal leaders in AI describe as \u201csuperintelligence.\u201d <\/p>\n<p>Training models to be smarter on even more complex tasks \u2014 e.g. novel biological research \u2014 requires mastering agentic behaviors that need to be learned from scratch, as discussed in my post on \u201c<a href=\"https:\/\/www.interconnects.ai\/p\/what-comes-next-with-reinforcement\" rel=\"noopener\" target=\"_blank\">What comes next with RL<\/a>\u201d. There\u2019s no internet scale pretraining data for such agentic tasks. My point is that not all jobs that require continual learning will require the frontiers of intelligence. I\u2019m excited to write blog posts with the bliss of my ChatGPT 6 co-editor.<\/p>\n<p>This technology coming soon will not be without its challenges. My first reaction to the continual learning post was more in line with \u201csociety isn\u2019t ready for this\u201d rather than commentary on its feasibility. I\u2019ll repeat <a href=\"https:\/\/www.interconnects.ai\/i\/164894857\/is-continual-learning-something-we-should-want\" rel=\"noopener\" target=\"_blank\">my warning<\/a>:<\/p>\n<blockquote>\n<p>For a long time I\u2019ve written that AI models have a higher risk potential in terms of social outcomes because the modalities they interact with us in are far more personal\u2026 As AI is going to be so powerful as a standalone entity, breaking some of the symbiotic links will be good for adding friction that makes the technology easier to steer towards good outcomes. In short, be wary of wishing for end-to-end (reinforcement) learning when you\u2019re part of the environment.<a href=\"https:\/\/www.interconnects.ai\/p\/what-comes-next-with-reinforcement#footnote-2-164894857\" rel=\"noopener\" target=\"_blank\">2<\/a> It\u2019s a destiny to dystopia.<\/p>\n<\/blockquote>\n<p>What we have today is a form of AGI and it\u2019ll soon get much better with better context and memory. The industrialization of language models is giving us incredible improvements across a wide swath of use-cases. These will blow past many basic primitives of intelligence in humans that have motivated AI for decades. First was models reasoning, then will come systems with continual learning. This is exactly what most AI companies are actually building \u2014 regardless of what their superintelligence messaging is.<\/p>\n<p>Comments are open on this post, please continue the debate!<\/p>\n","protected":false},"excerpt":{"rendered":"\u2019s now well-read post on why he is extending his AI timelines focuses on the idea of continual&hellip;\n","protected":false},"author":2,"featured_media":349850,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[323,1942,53,16,15],"class_list":{"0":"post-349849","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-technology","11":"tag-uk","12":"tag-united-kingdom"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/115040016315667048","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/349849","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=349849"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/349849\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/349850"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=349849"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=349849"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=349849"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}