{"id":213168,"date":"2025-12-03T12:46:07","date_gmt":"2025-12-03T12:46:07","guid":{"rendered":"https:\/\/www.europesays.com\/ie\/213168\/"},"modified":"2025-12-03T12:46:07","modified_gmt":"2025-12-03T12:46:07","slug":"ai-companies-safety-practices-fail-to-meet-global-standards-study-shows","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ie\/213168\/","title":{"rendered":"AI companies&#8217; safety practices fail to meet global standards, study shows"},"content":{"rendered":"<p>Item 1 of 2 Open AI and Anthropic logos are seen in this illustration taken on September 12, 2025. REUTERS\/Dado Ruvic\/Illustration<\/p>\n<p><b data-testid=\"Body\" class=\"text-module__text__0GDob text-module__dark-grey__UFC18 text-module__medium__2Rl30 text-module__inherit-size__EyiQW body-module__base__o--Cl body-module__ultra_small_body__WelUC primary-gallery-module__pagination__3IXiB\">[1\/2]<\/b>Open AI and Anthropic logos are seen in this illustration taken on September 12, 2025. REUTERS\/Dado Ruvic\/Illustration <a data-testid=\"Link\" target=\"_blank\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reutersconnect.com\/item\/illustration-shows-open-ai-and-anthropic-logos\/dGFnOnJldXRlcnMuY29tLDIwMjU6bmV3c21sX1JDMkVRR0FTSEVQRQ%3D%3D\/?utm_medium=rcom-article-media&amp;utm_campaign=rcom-rcp-lead\" rel=\"noopener nofollow\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC link-module__with-icon__qlg76 collapsible-caption-module__license__CPQvV\">Purchase Licensing Rights, opens new tab<\/a><\/p>\n<p>Dec 3 (Reuters) &#8211; The safety practices of major artificial intelligence companies, such as Anthropic, OpenAI, xAI and Meta, are &#8220;far short of emerging global standards,&#8221; according to a new edition of Future of Life Institute&#8217;s AI safety index released on Wednesday.<\/p>\n<p>The institute said the safety evaluation, conducted by an independent panel of experts, found that while the companies were busy racing to develop superintelligence, none had a robust strategy for controlling such advanced systems.<\/p>\n<p data-testid=\"promo-box\" class=\"text-module__text__0GDob text-module__dark-grey__UFC18 text-module__regular__qJJtA text-module__small__sph8i body-module__base__o--Cl body-module__small_body__gOmDf article-body-module__promo-box__hVl8h\"> Sign up  <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/business\/ai-companies-safety-practices-fail-meet-global-standards-study-shows-2025-12-03\/undefined?location=article-paragraph&amp;redirectUrl=%2Fbusiness%2Fai-companies-safety-practices-fail-meet-global-standards-study-shows-2025-12-03%2F\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">here.<\/a><\/p>\n<p>The study comes amid heightened public concern about the societal impact of smarter-than-human systems capable of reasoning and logical thinking, after several cases of <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/sustainability\/boards-policy-regulation\/google-ai-firm-must-face-lawsuit-filed-by-mother-over-suicide-son-us-court-says-2025-05-21\/\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">suicide<\/a> and <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/sustainability\/boards-policy-regulation\/openai-altman-sued-over-chatgpts-role-california-teens-suicide-2025-08-26\/\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">self-harm<\/a> were tied to <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/investigates\/special-report\/meta-ai-chatbot-death\/\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">AI chatbots<\/a>.<\/p>\n<p>&#8220;Despite recent uproar over AI-powered hacking and AI driving people to psychosis and self-harm, US AI companies remain less regulated than restaurants and continue lobbying against binding safety standards,&#8221; said Max Tegmark, MIT Professor and Future of Life President.<\/p>\n<p>The AI race also shows no signs of slowing, with major tech companies committing <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/business\/autos-transportation\/companies-pouring-billions-advance-ai-infrastructure-2025-10-06\/\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">hundreds of billions of dollars<\/a> to upgrading and expanding their machine learning efforts.The Future of Life Institute is a non-profit organization that has raised concerns about the risks intelligent machines pose to humanity. Founded in 2014, it was supported early on by Tesla <a data-testid=\"Link\" target=\"_blank\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/markets\/companies\/TSLA.O\" rel=\"noopener nofollow\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC link-module__with-icon__qlg76\">(TSLA.O), opens new tab<\/a> CEO Elon Musk.In October, a group including scientists Geoffrey Hinton and Yoshua Bengio <a data-testid=\"Link\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.reuters.com\/business\/media-telecom\/us-right-wing-media-figures-tech-pioneers-call-superintelligent-ai-ban-2025-10-22\/\" class=\"text-module__text__0GDob text-module__inherit-color__PhuPF text-module__inherit-font__1P1hv text-module__inherit-size__EyiQW link-module__link__INqxZ link-module__underline_default__-okuC\" rel=\"nofollow noopener\" target=\"_blank\">called for a ban<\/a> on developing superintelligent artificial intelligence until the public demands it and science paves a safe way forward.<\/p>\n<p>XAI said &#8220;Legacy media lies&#8221;, in what seemed to be an automated response, while Anthropic, OpenAI, Google DeepMind, Meta, Z.ai, DeepSeek and Alibaba Cloud did not immediately respond to request for comments on the study.<\/p>\n<p data-testid=\"SignOff\" class=\"text-module__text__0GDob text-module__dark-grey__UFC18 text-module__regular__qJJtA text-module__extra_small__8Buss body-module__full_width__kCIGb body-module__extra_small_body__Bfz20 sign-off-module__text__LQAMP\">Reporting by Zaheer Kachwala in Bengaluru, additional reporting by Arnav Mishra in Bengaluru; Editing by Shinjini Ganguli<\/p>\n<p data-testid=\"Body\" dir=\"ltr\" class=\"text-module__text__0GDob text-module__dark-grey__UFC18 text-module__regular__qJJtA text-module__small__sph8i body-module__base__o--Cl body-module__small_body__gOmDf article-body-module__element__5eCce article-body-module__trust-badge__5mS3f\">Our Standards: <a data-testid=\"Link\" target=\"_blank\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/www.thomsonreuters.com\/en\/about-us\/trust-principles.html\" rel=\"noopener nofollow\" class=\"text-module__text__0GDob text-module__dark-grey__UFC18 text-module__medium__2Rl30 text-module__small__sph8i link-module__link__INqxZ link-module__underline_default__-okuC link-module__with-icon__qlg76\">The Thomson Reuters Trust Principles., opens new tab<\/a><\/p>\n<p><a href=\"https:\/\/www.reutersagency.com\/en\/licensereuterscontent\/?utm_medium=rcom-article-media&amp;utm_campaign=rcom-rcp-lead\" target=\"_blank\" dir=\"ltr\" class=\"button-module__link__A3sD0 button-module__secondary__70gBu button-module__round__QDFgq button-module__w_auto__Sem-F\" data-testid=\"LicenceContentButton\" rel=\"nofollow noopener\">Purchase Licensing Rights<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"Item 1 of 2 Open AI and Anthropic logos are seen in this illustration taken on September 12,&hellip;\n","protected":false},"author":2,"featured_media":213169,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[261],"tags":[291,82519,82793,289,290,2786,82703,84105,82500,53515,82543,82538,82546,82549,82541,82550,82776,82530,83399,82531,82529,82779,82555,82542,82777,82547,82539,82557,82540,82548,82551,85719,82702,93732,18,82578,45748,59750,19,17,82797,84435,93734,82524,82799,82780,82510,82512,87107,84106,82800,82801,753,82,30150,82534,82533,107],"class_list":{"0":"post-213168","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-amers","10":"tag-arti","11":"tag-artificial-intelligence","12":"tag-artificialintelligence","13":"tag-asia","14":"tag-asxpac","15":"tag-blr","16":"tag-cmpny","17":"tag-cn","18":"tag-destabn","19":"tag-destabx","20":"tag-destbnx","21":"tag-destdnp","22":"tag-deste","23":"tag-destfun","24":"tag-destg","25":"tag-destocabsm","26":"tag-destocatpm","27":"tag-destousbsm","28":"tag-destoustcm","29":"tag-destoustpm","30":"tag-destpco","31":"tag-destpcu","32":"tag-destpge","33":"tag-destpsc","34":"tag-destrast","35":"tag-destrbn","36":"tag-destrnp","37":"tag-destu","38":"tag-destucdptest","39":"tag-destz","40":"tag-easia","41":"tag-ecomma","42":"tag-eire","43":"tag-emrg","44":"tag-gen","45":"tag-hk","46":"tag-ie","47":"tag-ireland","48":"tag-itec","49":"tag-itse","50":"tag-mlaisv","51":"tag-namer","52":"tag-netsv","53":"tag-packageus-top-news","54":"tag-publ","55":"tag-sci","56":"tag-socmed","57":"tag-sofw","58":"tag-srchen","59":"tag-swit","60":"tag-tech","61":"tag-technology","62":"tag-tmt","63":"tag-topcmb","64":"tag-topnws","65":"tag-us"},"share_on_mastodon":{"url":"","error":"Validation failed: Text character limit of 500 exceeded"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts\/213168","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/comments?post=213168"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts\/213168\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/media\/213169"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/media?parent=213168"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/categories?post=213168"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/tags?post=213168"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}