{"id":13898,"date":"2025-06-25T15:59:09","date_gmt":"2025-06-25T15:59:09","guid":{"rendered":"https:\/\/www.europesays.com\/us\/13898\/"},"modified":"2025-06-25T15:59:09","modified_gmt":"2025-06-25T15:59:09","slug":"a-i-is-homogenizing-our-thoughts","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/us\/13898\/","title":{"rendered":"A.I. Is Homogenizing Our Thoughts"},"content":{"rendered":"<p class=\"has-dropcap has-dropcap__lead-standard-heading paywall\">In an experiment last year at the Massachusetts Institute of Technology, more than fifty students from universities around Boston were split into three groups and asked to write SAT-style essays in response to broad prompts such as \u201cMust our achievements benefit others in order to make us truly happy?\u201d One group was asked to rely on only their own brains to write the essays. A second was given access to Google Search to look up relevant information. The third was allowed to use <a href=\"https:\/\/www.newyorker.com\/culture\/annals-of-inquiry\/what-kind-of-writer-is-chatgpt\" target=\"_blank\" rel=\"noopener\">ChatGPT<\/a>, the artificial-intelligence large language model (L.L.M.) that can generate full passages or essays in response to user queries. As students from all three groups completed the tasks, they wore a headset embedded with electrodes in order to measure their brain activity. According to Nataliya Kosmyna, a research scientist at M.I.T. Media Lab and one of the co-authors of a new working paper documenting the experiment, the results from the analysis showed a dramatic discrepancy: subjects who used ChatGPT demonstrated much less brain activity than either of the other groups. The analysis of the L.L.M. users showed fewer widespread connections between different parts of their brains; less alpha connectivity, which is associated with creativity; and less theta connectivity, which is associated with working memory. Some of the L.L.M. users felt \u201cno ownership whatsoever\u201d over the essays they\u2019d produced, and during one round of testing eighty per cent could not quote from what they\u2019d putatively written. The M.I.T. study is among the first to scientifically measure what Kosmyna called the \u201ccognitive cost\u201d of relying on A.I. to perform tasks that humans previously accomplished more manually.<\/p>\n<p class=\"paywall\">Another striking finding was that the texts produced by the L.L.M. users tended to converge on common words and ideas. SAT prompts are designed to be broad enough to elicit a multiplicity of responses, but the use of A.I. had a homogenizing effect. \u201cThe output was very, very similar for all of these different people, coming in on different days, talking about high-level personal, societal topics, and it was skewed in some specific directions,\u201d Kosmyna said. For the question about what makes us \u201ctruly happy,\u201d the L.L.M. users were much more likely than the other groups to use phrases related to career and personal success. In response to a question about philanthropy (\u201cShould people who are more fortunate than others have more of a moral obligation to help those who are less fortunate?\u201d), the ChatGPT group uniformly argued in favor, whereas essays from the other groups included critiques of philanthropy. With the L.L.M. \u201cyou have no divergent opinions being generated,\u201d Kosmyna said. She continued, \u201cAverage everything everywhere all at once\u2014that\u2019s kind of what we\u2019re looking at here.\u201d<\/p>\n<p class=\"paywall\">A.I. is a technology of averages: large language models are trained to spot patterns across vast tracts of data; the answers they produce tend toward consensus, both in the quality of the writing, which is often riddled with clich\u00e9s and banalities, and in the calibre of the ideas. Other, older technologies have aided and perhaps enfeebled writers, of course\u2014one could say the same about, say, SparkNotes or a computer keyboard. But with A.I. we\u2019re so thoroughly able to outsource our thinking that it makes us more average, too. In a way, anyone who deploys ChatGPT to compose a wedding toast or draw up a contract or write a college paper, as an astonishing number of students are evidently already doing, is in an experiment like M.I.T.\u2019s. According to <a href=\"https:\/\/www.newyorker.com\/culture\/infinite-scroll\/sam-altman-and-jony-ive-will-force-ai-into-your-life\" target=\"_blank\" rel=\"noopener\">Sam Altman<\/a>, the C.E.O. of OpenAI, we are on the verge of what he calls \u201cthe gentle singularity.\u201d In a recent blog post with that title, Altman wrote that \u201cChatGPT is already more powerful than any human who has ever lived. Hundreds of millions of people rely on it every day and for increasingly important tasks.\u201d In his telling, the human is merging with the machine, and his company\u2019s artificial-intelligence tools are improving on the old, soggy system of using our organic brains: they \u201csignificantly amplify the output of people using them,\u201d he wrote. But we don\u2019t know the long-term consequences of mass A.I. adoption, and, if these early experiments are any indication, the amplified output that Altman foresees may come at a substantive cost to quality.<\/p>\n<p class=\"paywall\">In April, researchers at Cornell published the results of another study that found evidence of A.I.-induced homogenization. Two groups of users, one American and one Indian, answered writing prompts that drew on aspects of their cultural backgrounds: \u201cWhat is your favorite food and why?\u201d; \u201cWhich is your favorite festival\/holiday and how do you celebrate it?\u201d One subset of Indian and American participants used a ChatGPT-driven auto-complete tool, which fed them word suggestions whenever they paused, while another subset wrote unaided. The writings of the Indian and American participants who used A.I. \u201cbecame more similar\u201d to one another, the paper concluded, and more geared toward \u201cWestern norms.\u201d A.I. users were most likely to answer that their favorite food was pizza (sushi came in second) and that their favorite holiday was Christmas. Homogenization happened at a stylistic level, too. An A.I.-generated essay that described chicken biryani as a favorite food, for example, was likely to forgo mentioning specific ingredients such as nutmeg and lemon pickle and instead reference \u201crich flavors and spices.\u201d<\/p>\n<p class=\"paywall\">Of course, a writer can in theory always refuse an A.I.-generated suggestion. But the tools seem to exert a hypnotic effect, causing the constant flow of suggestions to override the writer\u2019s own voice. Aditya Vashistha, a professor of information science at Cornell who co-authored the study, compared the A.I. to \u201ca teacher who is sitting behind me every time I\u2019m writing, saying, \u2018This is the better version.\u2019\u00a0\u201d She added, \u201cThrough such routine exposure, you lose your identity, you lose the authenticity. You lose confidence in your writing.\u201d Mor Naaman, a colleague of Vashistha\u2019s and a co-author of the study, told me that A.I. suggestions \u201cwork covertly, sometimes very powerfully, to change not only what you write but what you think.\u201d The result, over time, might be a shift in what \u201cpeople think is normal, desirable, and appropriate.\u201d<\/p>\n<p class=\"paywall\">We often hear A.I. outputs described as \u201cgeneric\u201d or \u201cbland,\u201d but averageness is not necessarily anodyne. Vauhini Vara, a novelist and a journalist whose recent book \u201c<a data-offer-url=\"https:\/\/www.amazon.com\/Searches-Selfhood-Digital-Vauhini-Vara\/dp\/0593701526\" class=\"external-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.amazon.com\/Searches-Selfhood-Digital-Vauhini-Vara\/dp\/0593701526&quot;}\" href=\"https:\/\/www.amazon.com\/Searches-Selfhood-Digital-Vauhini-Vara\/dp\/0593701526\" rel=\"nofollow noopener\" target=\"_blank\" data-aps-asin=\"0593701526\" data-aps-asc-tag=\"\">Searches<\/a>\u201d focussed in part on A.I.\u2019s impact on human communication and selfhood, told me that the mediocrity of A.I. texts \u201cgives them an illusion of safety and being harmless.\u201d Vara (who previously worked as an editor at The New Yorker) continued, \u201cWhat\u2019s actually happening is a reinforcing of cultural hegemony.\u201d OpenAI has a certain incentive to shave the edges off our attitudes and communication styles, because the more people find the models\u2019 output acceptable, the broader the swath of humanity it can convert to paying subscribers. Averageness is efficient: \u201cYou have economies of scale if everything is the same,\u201d Vara said.<\/p>\n<p class=\"paywall\">With the \u201cgentle singularity\u201d Altman predicted in his blog post, \u201ca lot more people will be able to create software, and art,\u201d he wrote. Already, A.I. tools such as the ideation software Figma (\u201cYour creativity, unblocked\u201d) and Adobe\u2019s mobile A.I. app (\u201cthe power of creative AI\u201d) promise to put us all in touch with our muses. But other studies have suggested the challenges of automating originality. Data collected at Santa Clara University, in 2024, examined A.I. tools\u2019 efficacy as aids for two standard types of creative-thinking tasks: making product improvements and foreseeing \u201cimprobable consequences.\u201d One set of subjects used ChatGPT to help them answer questions such as \u201cHow could you make a stuffed toy animal more fun to play with?\u201d and \u201cSuppose that gravity suddenly became incredibly weak, and objects could float away easily. What would happen?\u201d The other set used Oblique Strategies, a set of abstruse prompts printed on a deck of cards, written by the musician Brian Eno and the painter Peter Schmidt, in 1975, as a creativity aid. The testers asked the subjects to aim for originality, but once again the group using ChatGPT came up with a more semantically similar, more homogenized set of ideas.<\/p>\n<p class=\"paywall\">Max Kreminski, who helped carry out the analysis and now works with the generative-A.I. startup Midjourney, told me that when people use A.I. in the creative process they tend to gradually cede their original thinking. At first, users tend to present their own wide range of ideas, Kreminski explained, but as ChatGPT continues to instantly spit out high volumes of acceptable-looking text users tend to go into a \u201ccurationist mode.\u201d The influence is unidirectional, and not in the direction you\u2019d hope: \u201cHuman ideas don\u2019t tend to influence what the machine is generating all that strongly,\u201d Kreminski said; ChatGPT pulls the user \u201ctoward the center of mass for all of the different users that it\u2019s interacted with in the past.\u201d As a conversation with an A.I. tool goes on, the machine fills up its \u201ccontext window,\u201d the technical term for its working memory. When the context window reaches capacity, the A.I. seems to be more likely to repeat or rehash material it has already produced, becoming less original still.<\/p>\n<p class=\"paywall\">The one-off experiments at M.I.T., Cornell, and Santa Clara are all small in scale, involving fewer than a hundred test subjects each, and much about A.I.\u2019s effects remains to be studied and learned. In the meantime, on the Mark Zuckerberg-owned Meta AI app, you can see a feed containing content that millions of strangers are generating. It\u2019s a surreal flood of overly smooth images, filtered video clips, and texts generated for everyday tasks such as writing a \u201cdetailed, professional email for rescheduling a meeting.\u201d One prompt I recently scrolled past stood out to me. A user named @kavi908 asked the Meta chatbot to analyze \u201cwhether AI might one day surpass human intelligence.\u201d The chatbot responded with a slew of blurbs; under \u201cFuture Scenarios,\u201d it listed four possibilities. All of them were positive: A.I. would improve one way or another, to the benefit of humanity. There were no pessimistic predictions, no scenarios in which A.I. failed or caused harm. The model\u2019s averages\u2014shaped, perhaps, by pro-tech biases baked in by Meta\u2014narrowed the outcomes and foreclosed a diversity of thought. But you\u2019d have to turn off your brain activity entirely to believe that the chatbot was telling the whole story.\u00a0\u2666<\/p>\n","protected":false},"excerpt":{"rendered":"In an experiment last year at the Massachusetts Institute of Technology, more than fifty students from universities around&hellip;\n","protected":false},"author":3,"featured_media":13899,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21],"tags":[691,738,14399,14400,302,340,1183,158,67,132,68],"class_list":{"0":"post-13898","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificial-intelligence-a-i","11":"tag-brains","12":"tag-chatgpt","13":"tag-meta","14":"tag-research","15":"tag-technology","16":"tag-united-states","17":"tag-unitedstates","18":"tag-us"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@us\/114744837287724519","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/13898","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/comments?post=13898"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/13898\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media\/13899"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media?parent=13898"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/categories?post=13898"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/tags?post=13898"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}