{"id":480138,"date":"2025-12-30T13:42:11","date_gmt":"2025-12-30T13:42:11","guid":{"rendered":"https:\/\/www.europesays.com\/us\/480138\/"},"modified":"2025-12-30T13:42:11","modified_gmt":"2025-12-30T13:42:11","slug":"patients-are-consulting-ai-doctors-should-too-ai-needs-to-be-part-of-doctor-training","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/us\/480138\/","title":{"rendered":"Patients are consulting AI. Doctors should, too AI needs to be part of doctor training"},"content":{"rendered":"<p>Ask most physicians today, and they\u2019ll describe some version of this scene: In the middle of an appointment, a patient says, \u201cI asked ChatGPT about the treatment you recommended.\u201d\u00a0<\/p>\n<p>A few years ago, doctors might have bristled. Today, this is the new reality. And yet, it\u2019s exactly what tens of thousands of medical students and residents applying to programs this fall have been forbidden to do.<\/p>\n<p>As an academic physician and a medical school professor, I watch schools and health systems around the country wrestle with an uncomfortable truth: Health care is training doctors for a world that no longer exists. There are some forward-thinking institutions. At Dartmouth\u2019s Geisel School of Medicine, we\u2019re building artificial intelligence literacy into clinical training. Harvard Medical School offers a Ph.D. track in AI Medicine. But all of us must move faster.<\/p>\n<p>The numbers illustrate the problem. Every day, hundreds of medical studies appear in <a href=\"https:\/\/urldefense.com\/v3\/__https:\/\/oncodaily.com\/oncolibrary\/cancer-research-trends-2005-2025__;!!BspMT6SJLSDJ!NUjP2PMJn7Jyq0VZOKJU45T_Zl8rIRvnFCemXB9HJushS8tolofcBTZqqxtsIrCsIUm6NqW_BMq5XJTdAq4bGB8BkHe_%24\" target=\"_blank\" rel=\"noopener\">oncology<\/a> alone. The volume across all specialties has become impossible for any individual to absorb. Within a decade, clinicians who treat patients without consulting validated, clinically appropriate AI tools will find their decisions increasingly difficult to defend in malpractice proceedings. The gap between what one person can know and what medicine collectively knows has grown too wide to bridge alone.<\/p>\n<p>Our patients aren\u2019t waiting. They have already consulted ChatGPT or other AI chatbots before they arrive at appointments. They ask questions that assume their physician has considered options that the doctor has never encountered. A colleague in Boston recently told me about a patient who had turned to a chatbot that raised three treatment options the doctor hadn\u2019t initially considered. They spent 20 minutes working through the alternatives together. When the doctor explained his recommendation to his patient, he noticed her hands trembling. The AI had given her information, but he gave her reassurance. The AI outlined probabilities. He held space for her fear.<\/p>\n<p>\t\t\t<img decoding=\"async\" width=\"768\" height=\"432\" src=\"https:\/\/www.europesays.com\/us\/wp-content\/uploads\/2025\/12\/TheWorkup_Illustration_MollyFerguson_093025-768x432.jpg\" class=\"attachment-article-main-medium-large size-article-main-medium-large\" alt=\"\" loading=\"lazy\"  \/>\t\t<\/p>\n<p>\t\t\t\t<img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.statnews.com\/wp-content\/themes\/stat\/images\/home\/statplus.svg\" width=\"19\" height=\"16\" alt=\"\"\/><br \/>\n\t\t\t\t<a href=\"https:\/\/www.statnews.com\/2025\/11\/19\/doctors-ai-human-in-loop-problem\/\" rel=\"noopener\" target=\"_blank\">STAT Plus: Every doctor needs a basic level of AI literacy<\/a><\/p>\n<p>Several months later, she was in remission. The AI had helped her advocate for herself. But only the doctor could answer what she really wanted to know: \u201cAm I going to be OK?\u201d<\/p>\n<p>This is the future of medicine: AI as consultant, not replacement. But some of our medical schools and health systems seem determined to prepare students for the past. Some schools have restricted AI use for coursework and <a href=\"https:\/\/urldefense.com\/v3\/__https:\/\/med.stanford.edu\/md\/mdhandbook\/section-3-md-requirements-procedures\/3-32--generative-artificial-intelligence--ai--policy.html__;!!BspMT6SJLSDJ!NUjP2PMJn7Jyq0VZOKJU45T_Zl8rIRvnFCemXB9HJushS8tolofcBTZqqxtsIrCsIUm6NqW_BMq5XJTdAq4bGGizNNXI%24\" target=\"_blank\" rel=\"noopener\">clinical write-ups<\/a>.\u00a0 The Association of American Medical Colleges <a href=\"https:\/\/students-residents.aamc.org\/applying-medical-school-amcas\/publication-chapters\/personal-comments-essay?utm_source=chatgpt.com\" target=\"_blank\" rel=\"noopener\">limits<\/a> the use of AI in residency applications; students and trainees often complain that they need to be competent in AI even as they\u2019re told they can\u2019t use it. The instinct to restrain new technology is understandable. It\u2019s also obsolete. What students need instead are more useful institutional requirements.<\/p>\n<p>First, <strong>AI verification protocols<\/strong>. Medical schools already run morbidity and mortality conferences where physicians review cases that went wrong. We need AI rounds where students present: Which model did you consult? What did it recommend? Where did you override it and why? This should be a standard part of clinical training, documented in the medical record, reviewed by attending physicians.<\/p>\n<p>Second, <strong>transparency standards<\/strong>. The Accreditation Council for Graduate Medical Education (ACGME) should require residents to document AI consultation the way we document any specialist consultation. What question was asked? What answer was returned? What clinical judgment led to the final decision? This creates an auditable trail and teaches habits that will define their careers.<\/p>\n<p>Third, <strong>competency assessments<\/strong>. Medical licensing boards should test AI literacy the way they test pharmacology. Which models have been validated for which clinical questions? What are the known error rates? When should an algorithm be trusted, and when should it be questioned? These aren\u2019t theoretical questions. They\u2019re the foundation of every treatment decision that learners will make.<\/p>\n<p>Finally, <a href=\"https:\/\/urldefense.com\/v3\/__https:\/\/jamanetwork.com\/journals\/jama\/article-abstract\/2836687__;!!BspMT6SJLSDJ!NUjP2PMJn7Jyq0VZOKJU45T_Zl8rIRvnFCemXB9HJushS8tolofcBTZqqxtsIrCsIUm6NqW_BMq5XJTdAq4bGN_vXs7a%24\" target=\"_blank\" rel=\"noopener\"><strong>patient consent frameworks<\/strong><\/a>. When AI informs clinical decisions, patients deserve to know. Not because the technology is inherently experimental, but because transparency is part of partnership and many deployments are still being evaluated for safety, privacy, and effectiveness. Students need practice with these kinds of conversations: \u201cI consulted a clinical decision support tool that analyzes thousands of similar cases. Here\u2019s what it suggested, and here\u2019s why I agree or disagree.\u201d<\/p>\n<p>This matters most where American medicine is failing. At Dartmouth Health, we serve rural New Hampshire, Vermont, and Maine, where specialist shortages are severe in geriatrics, palliative care, and mental health. This fall, Geisel <a href=\"https:\/\/urldefense.com\/v3\/__https:\/\/dhgeiselgiving.org\/geisel-launches-ai-focused-medical-curriculum-train-digital-health-leaders__;!!BspMT6SJLSDJ!NUjP2PMJn7Jyq0VZOKJU45T_Zl8rIRvnFCemXB9HJushS8tolofcBTZqqxtsIrCsIUm6NqW_BMq5XJTdAq4bGEUy-sYm%24\" target=\"_blank\" rel=\"noopener\">launched<\/a> an AI curriculum that begins the moment students arrive, because we recognized a critical truth: If medical schools don\u2019t guide how students think about and use these tools, technology companies will drive both the curriculum and clinical practice. We\u2019re uniquely positioned to show how AI can bridge impossible gaps in underserved areas by training the first generation of clinicians who master the technology, rather than fear it.<\/p>\n<p>Through my work and research in end-of-life care, I\u2019ve held dying patients\u2019 hands, embraced families in their final moments, and sat in silence when silence was the only honest response. No algorithm can do this work. But AI can make us smarter and more effective. The stethoscope isn\u2019t obsolete. Neither is the held hand. They\u2019re irreplaceable.<\/p>\n<p>Right now, thousands of students and residents are interviewing at medical schools and training programs nationwide. To those learners: Ask about AI training. Ask how your education will prepare you for patients who arrive with AI-generated questions. Ask about clinical decision support tools. Ask how you\u2019ll learn to be the doctor that AI cannot replace: the one who holds the hand, interprets the fear, and answers the question beneath the question.<\/p>\n<p>To my colleagues in academic medicine: The ACGME should mandate AI competency standards by 2026. Medical licensing boards should add AI literacy to board exams within two years. Schools should replace their bans or restraints on AI with AI protocols. Our students deserve training for the medicine they\u2019ll practice, not the medicine we remember.<\/p>\n<p>To patients (in other words, all of us): The next time you pull out your phone and mention ChatGPT, your doctor should have a better answer than silence. Ask them: Have you consulted AI tools for my diagnosis? Were you trained to use them? If the answer is no, ask why not.<\/p>\n<p>The choice isn\u2019t between human doctors and artificial intelligence. It\u2019s between doctors who use all available tools \u2014 technological and human \u2014 to serve their patients, and doctors who face an impossible task alone. We know the future our patients need. It\u2019s time our medical schools and health systems caught up.<\/p>\n<p>Angelo Volandes is a professor at Dartmouth\u2019s Geisel School of Medicine, a clinician-investigator, and vice chair of research for the department of medicine at Dartmouth Health.<\/p>\n","protected":false},"excerpt":{"rendered":"Ask most physicians today, and they\u2019ll describe some version of this scene: In the middle of an appointment,&hellip;\n","protected":false},"author":3,"featured_media":480139,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[35],"tags":[738,210,1141,1142,153,881,1183,67,132,68],"class_list":{"0":"post-480138","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-health-care","8":"tag-artificial-intelligence","9":"tag-health","10":"tag-health-care","11":"tag-healthcare","12":"tag-policy","13":"tag-public-health","14":"tag-research","15":"tag-united-states","16":"tag-unitedstates","17":"tag-us"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@us\/115808812783697436","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/480138","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/comments?post=480138"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/posts\/480138\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media\/480139"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/media?parent=480138"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/categories?post=480138"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/us\/wp-json\/wp\/v2\/tags?post=480138"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}