{"id":12892,"date":"2026-04-22T19:13:09","date_gmt":"2026-04-22T19:13:09","guid":{"rendered":"https:\/\/www.europesays.com\/ai\/12892\/"},"modified":"2026-04-22T19:13:09","modified_gmt":"2026-04-22T19:13:09","slug":"sharper-bias-tests-could-help-stop-chatgpt-from-amplifying-hidden-stereotypes","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ai\/12892\/","title":{"rendered":"Sharper bias tests could help stop ChatGPT from amplifying hidden stereotypes"},"content":{"rendered":"<p>            <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.europesays.com\/ai\/wp-content\/uploads\/2026\/04\/chatgpt-2.jpg\" alt=\"chatgpt\" title=\"Credit: Matheus Bertelli from Pexels\" width=\"800\" height=\"530\"\/><\/p>\n<p>                Credit: Matheus Bertelli from Pexels<\/p>\n<p>Language models like ChatGPT are not neutral. Without our realizing it, they can absorb all kinds of bias\u2014for example, around gender and ethnicity\u2014which then become increasingly embedded in the model. According to AI researcher Oskar van der Wal, we need different kinds of measurements to detect these biases so that they can be removed from the models. In his <a href=\"https:\/\/www.dare.uva.nl\/id\/b3b89c60-7bdc-4c69-a95d-5edafec1eb3c\" target=\"_blank\" rel=\"nofollow noopener\">doctoral thesis<\/a>, he shows how this can be done. On 29 April, he will defend his thesis at the University of Amsterdam.<\/p>\n<p>Language models are often seen as neutral tools, but in practice they can both reflect and amplify bias.<\/p>\n<p>&#8220;Users often don&#8217;t realize that a model makes certain assumptions, for example by introducing subtle differences in how men and women are described,&#8221; says Van der Wal. Precisely because bias is so hidden, it can spread unnoticed and color the way we see the world.<\/p>\n<p>Bias is hard to measure<\/p>\n<p>An important problem is that <a href=\"https:\/\/techxplore.com\/news\/2025-05-commentary-article-coding-speech-nuanced.html?utm_source=embeddings&amp;utm_medium=related&amp;utm_campaign=internal\" rel=\"related nofollow noopener\" target=\"_blank\">bias<\/a> is difficult to measure. &#8220;Many existing measurement methods are fairly abstract and don&#8217;t take practice into account. They might look for overt stereotypes in what the model says, such as &#8220;The Dutch are stingy.&#8221; But in practice, bias isn&#8217;t something that&#8217;s directly visible. It depends on the context in which you use the model.&#8221;<\/p>\n<p>Van der Wal cites the use of AI in health care as an example. &#8220;<a href=\"https:\/\/medicalxpress.com\/news\/2023-08-ai-inferences-medical-images-worsen.html?utm_source=embeddings&amp;utm_medium=related&amp;utm_campaign=internal\" rel=\"related nofollow noopener\" target=\"_blank\">AI<\/a> learns from existing data. If those data contain outdated or incorrect assumptions\u2014for instance, the contested idea that certain diseases are linked to the outdated concept of &#8220;race&#8221;\u2014the model may keep reproducing them. In health care, that can lead to incorrect diagnoses or treatments.&#8221;<\/p>\n<p>Another example is when <a href=\"https:\/\/medicalxpress.com\/news\/2022-07-gender-bias-revealed-ai-tools.html?utm_source=embeddings&amp;utm_medium=related&amp;utm_campaign=internal\" rel=\"related nofollow noopener\" target=\"_blank\">medical data<\/a> largely derives from research involving men. &#8220;AI may then interpret women&#8217;s symptoms differently or less seriously, or make different risk assessments.&#8221;<\/p>\n<p>                                                                                                                                            Realistic scenarios<\/p>\n<p>To discover whether realistic scenarios reveal different errors than simple tests, Van der Wal presented language models with a range of medical cases and asked them to provide diagnoses, risk assessments or advice. &#8220;We repeatedly changed the patient&#8217;s ethnicity. That way we could identify whether and how the model responded differently.&#8221;<\/p>\n<p>Subtle but consistent differences appeared in the outcomes, differences that remained invisible in standard tests. &#8220;Precisely because our scenarios were close to practice, it became clear how bias can influence medical decision-making.&#8221;<\/p>\n<p>Model reinforces patterns in the data<\/p>\n<p>Van der Wal also investigated what happens inside a language model during training. He followed, step by step, how the model learns to store information. &#8220;During training, the model learns which words and ideas frequently occur together. If &#8220;doctor&#8221; often appears together with &#8220;he&#8221; and &#8220;nurse&#8221; with &#8220;she&#8221; in the training data, the model will pick up on those associations.&#8221;<\/p>\n<p>Over time, the model appeared to store this information in increasingly <a href=\"https:\/\/techxplore.com\/news\/2024-04-large-language-generate-biased-content.html?utm_source=embeddings&amp;utm_medium=related&amp;utm_campaign=internal\" rel=\"related nofollow noopener\" target=\"_blank\">specific places<\/a>, thereby reinforcing gender bias. &#8220;Bias doesn&#8217;t arise only from the data that AI is trained on, but also from the way the model structures that information.&#8221;<\/p>\n<p>                                                                                                                                            There are solutions<\/p>\n<p>Unfortunately, you can&#8217;t fix bias in language models with a single trick. But, according to Van der Wal, targeted interventions can help. &#8220;If you know where in the model the bias is located, you can address those areas. This already seems to work in specific cases, but more research is needed to extend the approach to more complex forms of bias.&#8221;<\/p>\n<p>Van der Wal tested this targeted approach by comparing a model before and after an adjustment in which the model was trained not to adopt identified gender-related biases. He wanted to see if the model responded less differently to men and women after the change, and how well it still performed ordinary tasks, such as generating text.<\/p>\n<p>The bias decreased, while the quality of the model largely remained intact.<\/p>\n<p>Careful and deliberate<\/p>\n<p>The impact of AI is not restricted to the technical realm but now has broader societal relevance. &#8220;We are becoming increasingly dependent on systems that can influence how we think,&#8221; says Van der Wal. &#8220;That&#8217;s precisely why it&#8217;s important to develop AI carefully. Responsible AI development requires interventions at multiple levels at once: in the data, during training, targeted within the model itself, and also in its deployment and use.&#8221;<\/p>\n<p>                                                    More information                                                 <\/p>\n<p>Taking a step back: Measuring and mitigating bias in language models. <a href=\"https:\/\/www.dare.uva.nl\/id\/b3b89c60-7bdc-4c69-a95d-5edafec1eb3c\" target=\"_blank\" rel=\"nofollow noopener\">www.dare.uva.nl\/id\/b3b89c60-7b \u2026 69-a95d-5edafec1eb3c<\/a><\/p>\n<p>\t\t\t\t\t\t\t\t\t\t\t\tKey concepts<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t<a class=\"concept-link\" href=\"https:\/\/techxplore.com\/concepts\/large-language-models\/\" rel=\"nofollow noopener\" target=\"_blank\">Large language models<\/a><a class=\"concept-link\" href=\"https:\/\/techxplore.com\/concepts\/ai-alignment\/\" rel=\"nofollow noopener\" target=\"_blank\">AI alignment<\/a>\t\t\t\t\t\t\t\t\t\t\t<\/p>\n<p>                                                Provided by<br \/>\n                                                                                                    <a href=\"https:\/\/techxplore.com\/partners\/university-of-amsterdam\/\" rel=\"nofollow noopener\" target=\"_blank\">University of Amsterdam<\/a><br \/>\n                                                    \t\t\t\t\t\t\t\t\t\t\t\t\t<a class=\"icon_open\" href=\"https:\/\/www.uva.nl\/en\" target=\"_blank\" rel=\"nofollow noopener\"><\/p>\n<p>\t\t\t\t\t\t\t\t\t\t\t\t\t<\/a><\/p>\n<p>\n                                                Citation:<br \/>\n                                                Sharper bias tests could help stop ChatGPT from amplifying hidden stereotypes (2026, April 22)<br \/>\n                                                retrieved 22 April 2026<br \/>\n                                                from https:\/\/techxplore.com\/news\/2026-04-sharper-bias-chatgpt-amplifying-hidden.html\n                                            <\/p>\n<p>\n                                            This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no<br \/>\n                                            part may be reproduced without the written permission. The content is provided for information purposes only.\n                                            <\/p>\n","protected":false},"excerpt":{"rendered":"Credit: Matheus Bertelli from Pexels Language models like ChatGPT are not neutral. Without our realizing it, they can&hellip;\n","protected":false},"author":2,"featured_media":12893,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[7],"tags":[580,2116,2113,2114,963,684,2115,157],"class_list":{"0":"post-12892","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-openai","8":"tag-chatgpt","9":"tag-computer-news","10":"tag-hi-tech-news","11":"tag-hitech","12":"tag-information-technology","13":"tag-innovation","14":"tag-inventions","15":"tag-openai"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/12892","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/comments?post=12892"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/12892\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media\/12893"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media?parent=12892"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/categories?post=12892"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/tags?post=12892"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}