{"id":315290,"date":"2025-08-03T19:15:20","date_gmt":"2025-08-03T19:15:20","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/315290\/"},"modified":"2025-08-03T19:15:20","modified_gmt":"2025-08-03T19:15:20","slug":"inside-openais-quest-to-make-ai-do-anything-for-you","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/315290\/","title":{"rendered":"Inside OpenAI\u2019s quest to make AI do anything for you"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">Shortly after Hunter Lightman joined OpenAI as a researcher in 2022, he watched his colleagues launch ChatGPT, one of the fastest-growing products ever. Meanwhile, Lightman quietly worked on a team teaching OpenAI\u2019s models to solve high school math competitions.\u00a0<\/p>\n<p class=\"wp-block-paragraph\">Today that team, known as MathGen, is considered instrumental to OpenAI\u2019s industry-leading effort to create AI reasoning models: the core technology behind AI agents that can do tasks on a computer like a human would.<\/p>\n<p class=\"wp-block-paragraph\">\u201cWe were trying to make the models better at mathematical reasoning, which at the time they weren\u2019t very good at,\u201d Lightman told TechCrunch, describing MathGen\u2019s early work.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI\u2019s models are far from perfect today \u2014 the company\u2019s latest AI systems <a href=\"https:\/\/techcrunch.com\/2025\/04\/18\/openais-new-reasoning-ai-models-hallucinate-more\/\" target=\"_blank\" rel=\"noopener\">still hallucinate<\/a> and its agents <a rel=\"nofollow noopener\" href=\"https:\/\/www.wired.com\/story\/browser-haunted-by-ai-agents\/\" target=\"_blank\">struggle with complex tasks.<\/a><\/p>\n<p class=\"wp-block-paragraph\">But its state-of-the-art models have improved significantly on mathematical reasoning. One of OpenAI\u2019s models recently won a <a href=\"https:\/\/techcrunch.com\/2025\/07\/21\/openai-and-google-outdo-the-mathletes-but-not-each-other\/\" target=\"_blank\" rel=\"noopener\">gold medal <\/a>at the International Math Olympiad, a math competition for the world\u2019s brightest high school students. OpenAI believes these reasoning capabilities will translate to other subjects, and ultimately power general-purpose agents that the company has always dreamed of building.<\/p>\n<p class=\"wp-block-paragraph\">ChatGPT was a happy accident \u2014 a lowkey research preview turned viral consumer business \u2014 but OpenAI\u2019s agents are the product of a years-long, deliberate effort within the company.\u00a0<\/p>\n<p class=\"wp-block-paragraph\">\u201cEventually, you\u2019ll just ask the computer for what you need and it\u2019ll do all of these tasks for you,\u201d said OpenAI CEO Sam Altman at the company\u2019s <a href=\"https:\/\/techcrunch.com\/2023\/11\/06\/everything-announced-at-openais-first-developer-event\/\" target=\"_blank\" rel=\"noopener\">first developer conference<\/a> in 2023. \u201cThese capabilities are often talked about in the AI field as agents. The upsides of this are going to be tremendous.\u201d<\/p>\n<p>Techcrunch event<\/p>\n<p>\n\t\t\t\t\t\t\t\t\tSan Francisco<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\t|<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\tOctober 27-29, 2025\n\t\t\t\t\t\t\t<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"382\" width=\"680\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/08\/GettyImages-1778704897.jpg\" alt=\"OpenAI CEO Sam Altman speaks during the OpenAI DevDay event on November 06, 2023 in San Francisco, California.\" class=\"wp-image-2625318\"  \/>OpenAI CEO Sam Altman speaks during the OpenAI DevDay event on November 06, 2023 in San Francisco, California.(Photo by Justin Sullivan\/Getty Images)<strong>Image Credits:<\/strong>Justin Sullivan \/ Getty Images<\/p>\n<p class=\"wp-block-paragraph\">Whether agents will meet Altman\u2019s vision remains to be seen, but OpenAI shocked the world with the release of <a href=\"https:\/\/techcrunch.com\/2024\/09\/12\/openai-unveils-a-model-that-can-fact-check-itself\/\" target=\"_blank\" rel=\"noopener\">its first AI reasoning model, o1,<\/a> in the fall of 2024. Less than a year later, the 21 foundational researchers behind that breakthrough are the most highly sought-after talent in Silicon Valley.<\/p>\n<p class=\"wp-block-paragraph\">Mark Zuckerberg <a href=\"https:\/\/techcrunch.com\/2025\/06\/26\/meta-hires-key-openai-researcher-to-work-on-ai-reasoning-models\/\" target=\"_blank\" rel=\"noopener\">recruited<\/a> five of the o1 researchers to work on Meta\u2019s new superintelligence-focused unit, offering some compensation packages north of $100 million. One of them, Shengjia Zhao, was recently <a href=\"https:\/\/techcrunch.com\/2025\/07\/25\/meta-names-shengjia-zhao-as-chief-scientist-of-ai-superintelligence-unit\/\" target=\"_blank\" rel=\"noopener\">named<\/a> chief scientist of Meta Superintelligence Labs.<\/p>\n<p>The reinforcement learning renaissance<\/p>\n<p class=\"wp-block-paragraph\">The rise of OpenAI\u2019s reasoning models and agents are tied to a machine learning training technique known as reinforcement learning (RL). RL provides feedback to an AI model on whether its choices were correct or not in simulated environments.<\/p>\n<p class=\"wp-block-paragraph\">RL has been used for decades. For instance, in 2016, about a year after OpenAI was founded in 2015, an AI system created by Google DeepMind using RL, <a href=\"https:\/\/techcrunch.com\/2016\/03\/15\/google-ai-beats-go-world-champion-again-to-complete-historic-4-1-series-victory\/\" target=\"_blank\" rel=\"noopener\">AlphaGo<\/a>, gained global attention after beating a world champion in the board game, Go.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"436\" width=\"680\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/08\/GettyImages-515358462.jpg\" alt=\"\" class=\"wp-image-3033346\"  \/>South Korean professional Go player Lee Se-Dol (R) prepares for his fourth match against Google\u2019s artificial intelligence program, AlphaGo, during the Google DeepMind Challenge Match on March 13, 2016 in Seoul, South Korea. Lee Se-dol played a five-game match against a computer program developed by a Google, AlphaGo.  (Photo by Google via Getty Images)<\/p>\n<p class=\"wp-block-paragraph\">Around that time, one of OpenAI\u2019s first employees, Andrej Karpathy, began pondering how to leverage RL to create an AI agent that could use a computer. But it would take years for OpenAI to develop the necessary models and training techniques.<\/p>\n<p class=\"wp-block-paragraph\">By 2018, OpenAI pioneered its first large language model in the GPT series, pretrained on massive amounts of internet data and a large clusters of GPUs. GPT models excelled at text processing, eventually leading to ChatGPT, but struggled with basic math.\u00a0<\/p>\n<p class=\"wp-block-paragraph\">It took until 2023 for OpenAI to achieve a breakthrough, initially dubbed \u201cQ*\u201d and then \u201cStrawberry,\u201d by combining LLMs, RL, and a technique called test-time computation. The latter gave the models extra time and computing power to plan and work through problems, verifying its steps, before providing an answer. <\/p>\n<p class=\"wp-block-paragraph\">This allowed OpenAI to introduce a new approach called \u201cchain-of-thought\u201d (CoT), which improved AI\u2019s performance on math questions the models hadn\u2019t seen before.<\/p>\n<p class=\"wp-block-paragraph\">\u201cI could see the model starting to reason,\u201d said El Kishky. \u201cIt would notice mistakes and backtrack, it would get frustrated. It really felt like reading the thoughts of a person.\u201d\u00a0<\/p>\n<p class=\"wp-block-paragraph\">Though individually these techniques weren\u2019t novel, OpenAI uniquely combined them to create Strawberry, which directly led to the development of o1. OpenAI quickly identified that the planning and fact checking abilities of AI reasoning models could be useful to power AI agents.<\/p>\n<p class=\"wp-block-paragraph\">\u201cWe had solved a problem that I had been banging my head against for a couple of years,\u201d said Lightman. \u201cIt was one of the most exciting moments of my research career.\u201d<\/p>\n<p>Scaling reasoning<\/p>\n<p class=\"wp-block-paragraph\">With AI reasoning models, OpenAI determined it had two new axes that would allow it to improve AI models: using more computational power during the post-training of AI models, and giving AI models more time and processing power while answering a question.<\/p>\n<p class=\"wp-block-paragraph\">\u201cOpenAI, as a company, thinks a lot about not just the way things are, but the way things are going to scale,\u201d said Lightman.<\/p>\n<p class=\"wp-block-paragraph\">Shortly after the 2023 Strawberry breakthrough, OpenAI spun up an \u201cAgents\u201d team led by OpenAI researcher Daniel Selsam to make further progress on this new paradigm, two sources told TechCrunch. Although the team was called \u201cAgents,\u201d\u00a0 OpenAI didn\u2019t initially differentiate between reasoning models and agents as we think of them today. The company just wanted to make AI systems capable of completing complex tasks.<\/p>\n<p class=\"wp-block-paragraph\">Eventually, the work of Selsam\u2019s Agents team became part of a larger project to develop the o1 reasoning model, with leaders including OpenAI co-founder Ilya Sutskever, chief research officer Mark Chen, and chief scientist Jakub Pachocki.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"453\" width=\"680\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/08\/019A7E29-3482-4F68-9A29-624BA22F0334.jpeg\" alt=\"Ilya Sutskever, Russian Israeli-Canadian computer scientist and co-founder and Chief Scientist of OpenAI.\" class=\"wp-image-2797714\"  \/>Ilya Sutskever, Russian Israeli-Canadian computer scientist and co-founder and Chief Scientist of OpenAI, speaks at Tel Aviv University in Tel Aviv on June 5, 2023. (Photo by JACK GUEZ \/ AFP)<strong>Image Credits:<\/strong>Getty Images<\/p>\n<p class=\"wp-block-paragraph\">OpenAI would have to divert precious resources \u2014 mainly talent and GPUs \u2014 to create o1. Throughout OpenAI\u2019s history, researchers have had to negotiate with company leaders to obtain resources; demonstrating breakthroughs was a surefire way to secure them.<\/p>\n<p class=\"wp-block-paragraph\">\u201cOne of the core components of OpenAI is that everything in research is bottom up,\u201d said Lightman. \u201cWhen we showed the evidence [for o1], the company was like, \u2018This makes sense, let\u2019s push on it.\u2019\u201d<\/p>\n<p class=\"wp-block-paragraph\">Some former employees say that the startup\u2019s mission to develop AGI was the key factor in achieving breakthroughs around AI reasoning models. By focusing on developing the smartest-possible AI models, rather than products, OpenAI was able to prioritize o1 above other efforts.\u00a0That type of large investment in ideas wasn\u2019t always possible at competing AI labs.<\/p>\n<p class=\"wp-block-paragraph\">The decision to try new training methods proved prescient. By late 2024, several leading AI labs started seeing<a href=\"https:\/\/techcrunch.com\/2024\/11\/20\/ai-scaling-laws-are-showing-diminishing-returns-forcing-ai-labs-to-change-course\/\" target=\"_blank\" rel=\"noopener\"> diminishing returns<\/a> on models created through traditional pretraining scaling. Today, much of the AI field\u2019s momentum comes from advances in reasoning models.<\/p>\n<p><strong>What does it mean for an AI to \u201creason?\u201d<\/strong><\/p>\n<p class=\"wp-block-paragraph\">In many ways, the goal of AI research is to recreate human intelligence with computers. Since the launch of o1, ChatGPT\u2019s UX has been filled with more human-sounding features such as \u201cthinking\u201d and \u201creasoning.\u201d<\/p>\n<p class=\"wp-block-paragraph\">When asked whether OpenAI\u2019s models were truly reasoning, El Kishky hedged, saying he thinks about the concept in terms of computer science. <\/p>\n<p class=\"wp-block-paragraph\">\u201cWe\u2019re teaching the model how to efficiently expend compute to get an answer. So if you define it that way, yes, it is reasoning,\u201d said El Kishky.<\/p>\n<p class=\"wp-block-paragraph\">Lightman takes the approach of focusing on the model\u2019s results and not as much on the means or their relation to human brains.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"454\" width=\"680\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/08\/img_5619.jpg\" alt=\"The OpenAI logo on screen at their developer day stage.\" class=\"wp-image-2625161\"  \/>The OpenAI logo on screen at their developer day stage. (Credit: Devin Coldeway)<strong>Image Credits:<\/strong>Devin Coldewey<\/p>\n<p class=\"wp-block-paragraph\">\u201cIf the model is doing hard things, then it is doing whatever necessary approximation of reasoning it needs in order to do that,\u201d said Lightman. \u201cWe can call it reasoning, because it looks like these reasoning traces, but it\u2019s all just a proxy for trying to make AI tools that are really powerful and useful to a lot of people.\u201d<\/p>\n<p class=\"wp-block-paragraph\">OpenAI\u2019s researchers note people may disagree with their nomenclature or definitions of reasoning \u2014 and surely, <a rel=\"nofollow noopener\" href=\"https:\/\/garymarcus.substack.com\/p\/a-knockout-blow-for-llms?r=8tdk6&amp;utm_campaign=post&amp;utm_medium=web&amp;triedRedirect=true\" target=\"_blank\">critics have emerged<\/a> \u2014 but they argue it\u2019s less important than the capabilities of their models. Other AI researchers tend to agree.<\/p>\n<p class=\"wp-block-paragraph\">Nathan Lambert, an AI researcher with the non-profit AI2, compares AI reasoning modes to airplanes in a <a rel=\"nofollow noopener\" href=\"https:\/\/www.interconnects.ai\/p\/the-rise-of-reasoning-machines\" target=\"_blank\">blog post<\/a>. Both, he says, are manmade systems inspired by nature \u2014 human reasoning and bird flight, respectively \u2014 but they operate through entirely different mechanisms. That doesn\u2019t make them any less useful, or any less capable of achieving similar outcomes.<\/p>\n<p class=\"wp-block-paragraph\">A group of AI researchers from OpenAI, Anthropic, and Google DeepMind agreed in a recent <a href=\"https:\/\/techcrunch.com\/2025\/07\/15\/research-leaders-urge-tech-industry-to-monitor-ais-thoughts\/\" target=\"_blank\" rel=\"noopener\">position paper<\/a> that AI reasoning models are not well understood today, and more research is needed. It may be too early to confidently claim what exactly is going on inside them.<\/p>\n<p><strong>The next frontier: AI agents for subjective tasks<\/strong><\/p>\n<p class=\"wp-block-paragraph\">The AI agents on the market today work best for well-defined, verifiable domains such as coding. OpenAI\u2019s <a href=\"https:\/\/techcrunch.com\/2025\/05\/16\/openai-launches-codex-an-ai-coding-agent-in-chatgpt\/\" target=\"_blank\" rel=\"noopener\">Codex agent<\/a> aims to help software engineers offload simple coding tasks. Meanwhile, Anthropic\u2019s models have become particularly <a href=\"https:\/\/techcrunch.com\/2025\/07\/31\/enterprises-prefer-anthropics-ai-models-over-anyone-elses-including-openais\/\" target=\"_blank\" rel=\"noopener\">popular<\/a> in AI coding tools like Cursor and Claude Code \u2014 these are some of the first AI agents that people are willing to <a href=\"https:\/\/techcrunch.com\/2025\/06\/05\/cursors-anysphere-nabs-9-9b-valuation-soars-past-500m-arr\/\" target=\"_blank\" rel=\"noopener\">pay up for<\/a>.<\/p>\n<p class=\"wp-block-paragraph\">However, general purpose AI agents like OpenAI\u2019s <a href=\"https:\/\/techcrunch.com\/2025\/07\/17\/openai-launches-a-general-purpose-agent-in-chatgpt\/\" target=\"_blank\" rel=\"noopener\">ChatGPT Agent<\/a> and Perplexity\u2019s <a href=\"https:\/\/techcrunch.com\/2025\/07\/09\/perplexity-launches-comet-an-ai-powered-web-browser\/\" target=\"_blank\" rel=\"noopener\">Comet<\/a> struggle with many of the complex, subjective tasks people want to automate. When trying to use these tools for online shopping or finding a long-term parking spot, I\u2019ve found the agents take longer than I\u2019d like and make <a href=\"https:\/\/techcrunch.com\/2025\/02\/04\/openais-operator-agent-helped-me-move-but-i-had-to-help-it-too\/?_thumbnail_id=2593103\" target=\"_blank\" rel=\"noopener\">silly mistakes<\/a>.<\/p>\n<p class=\"wp-block-paragraph\">Agents are, of course, early systems that will undoubtedly improve. But researchers must first figure out how to better train the underlying models to complete tasks that are more subjective.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"453\" width=\"680\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/08\/GettyImages-2197821080.jpg\" alt=\"\" class=\"wp-image-3033358\"  \/>AI applications (Photo by Jonathan Raa\/NurPhoto via Getty Images)<\/p>\n<p class=\"wp-block-paragraph\">\u201cLike many problems in machine learning, it\u2019s a data problem,\u201d said Lightman, when asked about the limitations of agents on subjective tasks. \u201cSome of the research I\u2019m really excited about right now is figuring out how to train on less verifiable tasks. We have some leads on how to do these things.\u201d\u00a0<\/p>\n<p class=\"wp-block-paragraph\">Noam Brown, an OpenAI researcher who helped create the IMO model and o1, told TechCrunch that OpenAI has new general-purpose RL techniques which allow them to teach AI models skills that aren\u2019t easily verified. This was how the company built the model which achieved a gold medal at IMO, he said.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI\u2019s IMO model was a newer AI system that spawns multiple agents, which then simultaneously explore several ideas, and then choose the best possible answer. These types of AI models are becoming more popular; <a href=\"https:\/\/techcrunch.com\/2025\/08\/01\/google-rolls-out-gemini-deep-think-ai-a-reasoning-model-that-tests-multiple-ideas-in-parallel\/\" target=\"_blank\" rel=\"noopener\">Google<\/a> and <a href=\"https:\/\/techcrunch.com\/2025\/07\/09\/elon-musks-xai-launches-grok-4-alongside-a-300-monthly-subscription\/\" target=\"_blank\" rel=\"noopener\">xAI<\/a> have recently released state-of-the-art models using this technique.<\/p>\n<p class=\"wp-block-paragraph\">\u201cI think these models will become more capable at math, and I think they\u2019ll get more capable in other reasoning areas as well,\u201d said Brown. \u201cThe progress has been incredibly fast. I don\u2019t see any reason to think it will slow down.\u201d<\/p>\n<p class=\"wp-block-paragraph\">These techniques may help OpenAI\u2019s models become more performant, gains that could show up in the company\u2019s upcoming GPT-5 model. OpenAI hopes to assert its dominance over competitors with the launch of GPT-5, ideally offering the <a rel=\"nofollow noopener\" href=\"https:\/\/www.theinformation.com\/articles\/inside-openais-rocky-path-gpt-5?rc=dp0mql\" target=\"_blank\">best AI model<\/a> to power agents for developers and consumers. <\/p>\n<p class=\"wp-block-paragraph\">But the company also wants to make its products simpler to use. El Kishky says OpenAI wants to develop AI agents that intuitively understand what users want, without requiring them to select specific settings. He says OpenAI aims to build AI systems that understand when to call up certain tools, and how long to reason for.<\/p>\n<p class=\"wp-block-paragraph\">These ideas paint a picture of an ultimate version of ChatGPT: an agent that can do anything on the internet for you, and understand how you want it to be done. That\u2019s a much different product than what ChatGPT is today, but the company\u2019s research is squarely headed in this direction.<\/p>\n<p class=\"wp-block-paragraph\">While OpenAI undoubtedly led the AI industry a few years ago, the company now faces a tranche of worthy opponents. The question is no longer just whether OpenAI can deliver its agentic future, but can the company do so before Google, Anthropic, xAI, or Meta beat them to it?<\/p>\n","protected":false},"excerpt":{"rendered":"Shortly after Hunter Lightman joined OpenAI as a researcher in 2022, he watched his colleagues launch ChatGPT, one&hellip;\n","protected":false},"author":2,"featured_media":315291,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[15624,323,1942,1315,2915,1318,53,16,15],"class_list":{"0":"post-315290","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-agents","9":"tag-ai","10":"tag-artificial-intelligence","11":"tag-chatgpt","12":"tag-exclusive","13":"tag-openai","14":"tag-technology","15":"tag-uk","16":"tag-united-kingdom"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/114966437889570946","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/315290","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=315290"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/315290\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/315291"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=315290"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=315290"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=315290"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}