{"id":23885,"date":"2025-04-16T05:01:07","date_gmt":"2025-04-16T05:01:07","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/23885\/"},"modified":"2025-04-16T05:01:07","modified_gmt":"2025-04-16T05:01:07","slug":"openai-may-adjust-its-safeguards-if-rivals-release-high-risk-ai","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/23885\/","title":{"rendered":"OpenAI may &#8216;adjust&#8217; its safeguards if rivals release &#8216;high-risk&#8217; AI"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">OpenAI has <a href=\"https:\/\/openai.com\/index\/updating-our-preparedness-framework\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">updated<\/a> its Preparedness Framework \u2014 the internal system it uses to assess the safety of AI models and determine necessary safeguards during development and deployment. In the update, OpenAI stated that it may \u201cadjust\u201d its safety requirements if a competing AI lab releases a \u201chigh-risk\u201d system without similar protections in place.<\/p>\n<p class=\"wp-block-paragraph\">The change reflects the increasing competitive pressures on commercial AI developers to deploy models quickly. OpenAI has been <a href=\"https:\/\/www.ft.com\/content\/8253b66e-ade7-4d1f-993b-2d0779c7e7d8\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">accused of lowering safety standards<\/a> in favor of faster releases, and of failing to deliver <a href=\"https:\/\/techcrunch.com\/2025\/04\/15\/openai-ships-gpt-4-1-without-a-safety-report\/\" target=\"_blank\" rel=\"noopener\">timely reports detailing its safety testing<\/a>.\u00a0Last week, 12 former OpenAI employees <a href=\"https:\/\/techcrunch.com\/2025\/04\/11\/ex-openai-staff-file-amicus-brief-opposing-the-companys-for-profit-transition\/\" target=\"_blank\" rel=\"noopener\">filed a brief<\/a> in Elon Musk\u2019s case against OpenAI, arguing the company would be encouraged to cut even more corners on safety should it complete its planned corporate restructuring.<\/p>\n<p class=\"wp-block-paragraph\">Perhaps anticipating criticism, OpenAI claims that it wouldn\u2019t make these policy adjustments lightly, and that it would keep its safeguards at \u201ca level more protective.\u201d<\/p>\n<p class=\"wp-block-paragraph\">\u201cIf another frontier AI developer releases a high-risk system without comparable safeguards, we may adjust our requirements,\u201d wrote OpenAI in a <a href=\"https:\/\/openai.com\/index\/updating-our-preparedness-framework\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">blog post<\/a> published Tuesday afternoon. \u201cHowever, we would first rigorously confirm that the risk landscape has actually changed, publicly acknowledge that we are making an adjustment, assess that the adjustment does not meaningfully increase the overall risk of severe harm, and still keep safeguards at a level more protective.\u201d<\/p>\n<p class=\"wp-block-paragraph\">The refreshed Preparedness Framework also makes clear that OpenAI is relying more heavily on automated evaluations to speed up product development. The company says that while it hasn\u2019t abandoned human-led testing altogether, it has built \u201ca growing suite of automated evaluations\u201d that can supposedly \u201ckeep up with [a] faster [release] cadence.\u201d<\/p>\n<p class=\"wp-block-paragraph\">Some reports contradict this. <a href=\"https:\/\/www.ft.com\/content\/8253b66e-ade7-4d1f-993b-2d0779c7e7d8\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">According to the Financial Times<\/a>, OpenAI gave testers less than a week for safety checks for an upcoming major model \u2014 a compressed timeline compared to previous releases. The publication\u2019s sources also alleged that many of OpenAI\u2019s safety tests are now conducted on earlier versions of models rather than the versions released to the public.<\/p>\n<p class=\"wp-block-paragraph\">In statements, OpenAI has disputed the notion that it\u2019s compromising on safety.<\/p>\n<blockquote class=\"wp-block-quote twitter-tweet is-layout-flow wp-block-quote-is-layout-flow\">\n<p class=\"wp-block-paragraph\">OpenAI is quietly reducing its safety commitments.<\/p>\n<p class=\"wp-block-paragraph\">Omitted from OpenAI\u2019s list of Preparedness Framework changes:<\/p>\n<p class=\"wp-block-paragraph\">No longer requiring safety tests of finetuned models <a href=\"https:\/\/t.co\/oTmEiAtSjS\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">https:\/\/t.co\/oTmEiAtSjS<\/a><\/p>\n<p class=\"wp-block-paragraph\">\u2014 Steven Adler (@sjgadler) <a href=\"https:\/\/twitter.com\/sjgadler\/status\/1912242577723781258?ref_src=twsrc%5Etfw\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">April 15, 2025<\/a><\/p>\n<\/blockquote>\n<p class=\"wp-block-paragraph\">Other changes to OpenAI\u2019s framework pertain to how the company categorizes models according to risk, including models that can conceal their capabilities, evade safeguards, prevent their shutdown, and even self-replicate. OpenAI says that it\u2019ll now focus on whether models meet one of two thresholds: \u201chigh\u201d capability or \u201ccritical\u201d capability. <\/p>\n<p class=\"wp-block-paragraph\">OpenAI\u2019s definition of the former is a model that could \u201camplify existing pathways to severe harm.\u201d The latter are models that \u201cintroduce unprecedented new pathways to severe harm,\u201d per the company.<\/p>\n<p class=\"wp-block-paragraph\">\u201cCovered systems that reach high capability must have safeguards that sufficiently minimize the associated risk of severe harm before they are deployed,\u201d wrote OpenAI in its blog post. \u201cSystems that reach critical capability also require safeguards that sufficiently minimize associated risks during development.\u201d<\/p>\n<p class=\"wp-block-paragraph\">The updates are the first OpenAI has made to the Preparedness Framework since 2023.<\/p>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"OpenAI has updated its Preparedness Framework \u2014 the internal system it uses to assess the safety of AI&hellip;\n","protected":false},"author":2,"featured_media":23886,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3163],"tags":[323,1942,1318,5549,53,16,15],"class_list":{"0":"post-23885","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-openai","11":"tag-safety","12":"tag-technology","13":"tag-uk","14":"tag-united-kingdom"},"share_on_mastodon":{"url":"","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/23885","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=23885"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/23885\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/23886"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=23885"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=23885"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=23885"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}