{"id":66665,"date":"2025-05-01T21:13:08","date_gmt":"2025-05-01T21:13:08","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/66665\/"},"modified":"2025-05-01T21:13:08","modified_gmt":"2025-05-01T21:13:08","slug":"ai2s-new-small-ai-model-outperforms-similarly-sized-models-from-google-meta","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/66665\/","title":{"rendered":"Ai2&#8217;s new small AI model outperforms similarly-sized models from Google, Meta"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">\u2018Tis the week for small AI models, it seems. <\/p>\n<p class=\"wp-block-paragraph\">On Thursday, Ai2, the nonprofit AI research institute, <a rel=\"nofollow noopener\" href=\"https:\/\/huggingface.co\/allenai\/OLMo-2-0425-1B\" target=\"_blank\">released<\/a> Olmo 2 1B, a 1-billion-parameter model that Ai2 claims beats similarly-sized models from Google, Meta, and Alibaba on several benchmarks. Parameters, sometimes referred to as weights, are the internal components of a model that guide its behavior.<\/p>\n<p class=\"wp-block-paragraph\">Olmo 2 1B is available under a permissive Apache 2.0 license on the AI dev platform Hugging Face. Unlike most models, Olmo 2 1B can be replicated from scratch; Ai2 has provided the code and data sets (<a rel=\"nofollow noopener\" href=\"https:\/\/huggingface.co\/datasets\/allenai\/olmo-mix-1124\" target=\"_blank\">Olmo-mix-1124<\/a>, <a rel=\"nofollow noopener\" href=\"https:\/\/huggingface.co\/datasets\/allenai\/dolmino-mix-1124\" target=\"_blank\">Dolmino-mix-1124<\/a>) used to develop it. <\/p>\n<p class=\"wp-block-paragraph\">Small models might not be as capable as their behemoth counterparts, but importantly, they don\u2019t require beefy hardware to run. That makes them much more accessible for developers and hobbyists contending with the limitations of lower-end and consumer machines. <\/p>\n<p class=\"wp-block-paragraph\">There\u2019s been a raft of small model launches over the past few days, from Microsoft\u2019s <a href=\"https:\/\/techcrunch.com\/2025\/04\/30\/microsofts-most-capable-new-phi-4-ai-model-rivals-the-performance-of-far-larger-systems\/\" target=\"_blank\" rel=\"noopener\">Phi 4 reasoning family<\/a> to <a rel=\"nofollow noopener\" href=\"https:\/\/venturebeat.com\/ai\/qwen-swings-for-a-double-with-2-5-omni-3b-model-that-runs-on-consumer-pcs-laptops\/\" target=\"_blank\">Qwen\u2019s 2.5 Omni 3B<\/a>. Most of these \u2014 and Olmo 2 1B \u2014 can easily run on a modern laptop or even a mobile device.<\/p>\n<p class=\"wp-block-paragraph\">Ai2 says that Olmo 2 1B was trained on a data set of 4 trillion tokens from publicly available, AI-generated, and manually created sources. Tokens are the raw bits of data models ingest and generate \u2014 1 million tokens is equivalent to about 750,000 words. <\/p>\n<p class=\"wp-block-paragraph\">On a benchmark measuring arithmetic reasoning, GSM8K, Olmo 2 1B scores better than Google\u2019s Gemma 3 1B, Meta\u2019s Llama 3.2 1B, and Alibaba\u2019s Qwen 2.5 1.5B. Olmo 2 1B also eclipses the performance of those three models on TruthfulQA, a test for evaluating factual accuracy.  <\/p>\n<p>Techcrunch event<\/p>\n<p>\n\t\t\t\t\t\t\t\t\tBerkeley, CA<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\t|<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\tJune 5\n\t\t\t\t\t\t\t<\/p>\n<p>\t\t\t\t\t\t\t<a href=\"https:\/\/techcrunch.com\/events\/tc-sessions-ai\/exhibit\/?promo=tc_inline_exhibit&amp;utm_campaign=tcsessionsai2025&amp;utm_content=exhibit&amp;utm_medium=ad&amp;utm_source=tc\" class=\"inline-cta__register-button\" target=\"_blank\" rel=\"noopener\"><br \/>\n\t\t\t\t\tBOOK NOW<br \/>\n\t\t\t\t<\/a><\/p>\n<blockquote class=\"twitter-tweet\" data-width=\"500\" data-dnt=\"true\">\n<p lang=\"en\" dir=\"ltr\">This model was pretrained on 4T tokens of high-quality data, following the same standard pretraining into high-quality annealing of our 7, 13, &amp; 32B models. We upload intermediate checkpoints from every 1000 steps in training.<\/p>\n<p>Access the base model: <a rel=\"nofollow\" href=\"https:\/\/t.co\/xofyWJmo85\">https:\/\/t.co\/xofyWJmo85<\/a> <a rel=\"nofollow\" href=\"https:\/\/t.co\/7uSJ6sYMdL\">pic.twitter.com\/7uSJ6sYMdL<\/a><\/p>\n<p>\u2014 Ai2 (@allen_ai) <a rel=\"nofollow noopener\" href=\"https:\/\/twitter.com\/allen_ai\/status\/1917927467056058665?ref_src=twsrc%5Etfw\" target=\"_blank\">May 1, 2025<\/a><\/p><\/blockquote>\n<p class=\"wp-block-paragraph\">Ai2 warns that that Olmo 2 1B carries risks, however. Like all AI models, it can produce \u201cproblematic outputs\u201d including harmful and \u201csensitive\u201d content, the organization says, as well as factually inaccurate statements. For these reasons, Ai2 recommends against deploying Olmo 2 1B in commercial settings. <\/p>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"\u2018Tis the week for small AI models, it seems. On Thursday, Ai2, the nonprofit AI research institute, released&hellip;\n","protected":false},"author":2,"featured_media":66666,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[323,33867,1942,10733,53,16,15],"class_list":{"0":"post-66665","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-ai2","10":"tag-artificial-intelligence","11":"tag-open-source","12":"tag-technology","13":"tag-uk","14":"tag-united-kingdom"},"share_on_mastodon":{"url":"","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/66665","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=66665"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/66665\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/66666"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=66665"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=66665"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=66665"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}