{"id":29065,"date":"2026-05-06T06:41:16","date_gmt":"2026-05-06T06:41:16","guid":{"rendered":"https:\/\/www.europesays.com\/ai\/29065\/"},"modified":"2026-05-06T06:41:16","modified_gmt":"2026-05-06T06:41:16","slug":"what-we-know-about-us-stress-tests-of-google-xai-and-microsoft-ai-models-firstpost","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ai\/29065\/","title":{"rendered":"What we know about US stress tests of Google, xAI and Microsoft AI models \u2013 Firstpost"},"content":{"rendered":"<p>The Trump administration on Tuesday announced it had expanded a program to give US government scientists access to unreleased artificial intelligence models to conduct risk assessments to include Google\u2019s DeepMind, xAI and Microsoft.<\/p>\n<p>The Trump administration on Tuesday announced it had<br \/>\n<a href=\"https:\/\/www.firstpost.com\/tech\/us-pushes-stricter-ai-checks-as-trump-team-weighs-pentagon-role-14007626.html\" id=\"https:\/\/www.firstpost.com\/tech\/us-pushes-stricter-ai-checks-as-trump-team-weighs-pentagon-role-14007626.html\" class=\"body_anchor\" rel=\"nofollow noopener\" target=\"_blank\">expanded a program<\/a> to give US government scientists access to unreleased artificial intelligence models to conduct risk assessments to include Google\u2019s DeepMind, xAI and Microsoft.<\/p>\n<p>ChatGPT maker<br \/>\n<a href=\"https:\/\/www.firstpost.com\/tech\/anthropic-out-openai-in-sam-altman-announces-deal-with-pentagon-to-deploy-ai-models-13984577.html\" id=\"https:\/\/www.firstpost.com\/tech\/anthropic-out-openai-in-sam-altman-announces-deal-with-pentagon-to-deploy-ai-models-13984577.html\" class=\"body_anchor\" rel=\"nofollow noopener\" target=\"_blank\">OpenAI<\/a> and Claude owner Anthropic had already been voluntarily working with the US Center for AI Standards and Innovation, the team of US government scientists, to test unreleased models for vulnerabilities, according to the companies.<\/p>\n<p>Here is what we know about the reviews:<\/p>\n<p>STORY CONTINUES BELOW THIS ADWhat risks are the US focused on?<\/p>\n<p>US government scientists are focused on \u201cdemonstrable risks,\u201d such as the risk that advanced models can be used to launch cyberattacks on American infrastructure, according to the CAISI website. They want to limit opportunities for US adversaries to use AI to develop chemical or biological weapons, or corrupt the data used to train American AI models.<\/p>\n<p>What will companies hand over?<\/p>\n<p>OpenAI is working with the group to test<br \/>\n<a href=\"https:\/\/www.firstpost.com\/tech\/openai-launches-gpt-5-5-moving-closer-to-its-vision-of-a-unified-ai-super-app-14003778.html\" id=\"https:\/\/www.firstpost.com\/tech\/openai-launches-gpt-5-5-moving-closer-to-its-vision-of-a-unified-ai-super-app-14003778.html\" class=\"body_anchor\" rel=\"nofollow noopener\" target=\"_blank\">GPT-5.5-Cyber<\/a>, said Chris Lehane, head of global affairs at OpenAI, in a LinkedIn post on Tuesday. GPT-5.5-Cyber is a variant of its latest model designed for defensive cybersecurity work.<\/p>\n<p>Microsoft will work with the scientists to build shared datasets and workflows to assess advanced AI models, the company said in a statement. Microsoft did not specify which models.<\/p>\n<p>Anthropic gave CAISI access to both publicly available and unreleased models, allowing researchers to probe for vulnerabilities in a process known as \u201cred-teaming,\u201d or simulating the behavior of malicious actors, the company said in September. The company also gave CAISI detailed documentation on known vulnerabilities and safety mechanisms.<\/p>\n<p>Google DeepMind, Alphabet\u2019s AI research arm, will provide access to its \u201cproprietary models\u201d and data, a spokesperson said.<\/p>\n<p>xAI did not immediately respond to a request for comment from Reuters.<\/p>\n<p>What has the US found so far?<\/p>\n<p>Anthropic\u2019s work with CAISI revealed that tricks such as claiming that human review had occurred, or substituting characters, could get around safety mechanisms, the company said, adding that it had patched the vulnerabilities.<\/p>\n<p>OpenAI said in September that it worked with CAISI to probe vulnerabilities in its ChatGPT Agent that could have allowed sophisticated actors to bypass OpenAI\u2019s cybersecurity measures. The exploit would have allowed the attacker to \u201cremotely control the computer systems the agent could access for that session and successfully impersonate the user for other websites they\u2019d logged into,\u201d the company said.<\/p>\n<p>The companies, along with Meta, Amazon and Inflection AI, agreed in 2023 to allow independent experts to check their models for biosecurity and cybersecurity risks.<\/p>\n<p>The US government scientists, organized under a different name during former US President Joe Biden\u2019s tenure, also released voluntary guidelines to protect against the risk of AI models leaking private health information or producing incorrect answers.<\/p>\n<p>STORY CONTINUES BELOW THIS AD<\/p>\n<p>The scientists are now working on guidelines for critical infrastructure providers, such as the communications and emergency services sectors, to test their AI systems, according to their website.<\/p>\n<p class=\"first-published\">First Published:<br \/>\nMay 06, 2026, 08:32 IST<\/p>\n<p><a href=\"https:\/\/www.firstpost.com\/\" rel=\"nofollow noopener\" target=\"_blank\">Home<\/a><a href=\"https:\/\/www.firstpost.com\/tech\/\" title=\"Tech\" rel=\"nofollow noopener\" target=\"_blank\">Tech<\/a>What we know about US stress tests of Google, xAI and Microsoft AI modelsEnd of Article<\/p>\n","protected":false},"excerpt":{"rendered":"The Trump administration on Tuesday announced it had expanded a program to give US government scientists access to&hellip;\n","protected":false},"author":2,"featured_media":29066,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[9],"tags":[11611,19061,19059,5044,132,7543,19062,19060,19058,19056,19057],"class_list":{"0":"post-29065","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-google","8":"tag-ai-cybersecurity-risks","9":"tag-anthropic-ai-vulnerabilities","10":"tag-caisi-ai-standards","11":"tag-deepmind","12":"tag-google","13":"tag-google-deepmind","14":"tag-google-deepmind-proprietary-models","15":"tag-openai-gpt-5-5-cyber","16":"tag-unreleased-ai-models","17":"tag-us-ai-risk-assessments","18":"tag-us-government-scientists"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/29065","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/comments?post=29065"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/29065\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media\/29066"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media?parent=29065"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/categories?post=29065"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/tags?post=29065"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}