{"id":484744,"date":"2026-05-14T18:41:20","date_gmt":"2026-05-14T18:41:20","guid":{"rendered":"https:\/\/www.europesays.com\/ie\/484744\/"},"modified":"2026-05-14T18:41:20","modified_gmt":"2026-05-14T18:41:20","slug":"advancing-conversational-diagnostic-ai-with-multimodal-reasoning","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ie\/484744\/","title":{"rendered":"Advancing conversational diagnostic AI with multimodal reasoning"},"content":{"rendered":"<p>This work advances the diagnostic capabilities of the multimodal AMIE system<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 4\" title=\"Tu, T. et al. Towards conversational diagnostic artificial intelligence. Nature 642, 442&#x2212;450 (2025).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR4\" id=\"ref-link-section-d22012503e1437\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 5\" title=\"Palepu, A. et al. Towards conversational AI for disease management. Preprint at arXiv &#010;                https:\/\/doi.org\/10.48550\/arXiv.2503.06074&#010;                &#010;               (2025).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR5\" id=\"ref-link-section-d22012503e1440\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a> by integrating multimodal perception, enabling it to conduct diagnostic conversations that are more clinically realistic and that incorporate various forms of medical data beyond text.<\/p>\n<p>Our approach centers on a state-aware dialogue phase transition framework, which leverages the multimodal reasoning capabilities of Gemini 2.0 Flash<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 29\" title=\"Google Cloud. Gemini 2.0 Flash&#x2014;Generative AI on Vertex AI. &#010;                https:\/\/cloud.google.com\/vertex-ai\/generative-ai\/docs\/models\/gemini\/2-0-flash&#010;                &#010;              \" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR29\" id=\"ref-link-section-d22012503e1447\" rel=\"nofollow noopener\" target=\"_blank\">29<\/a>. This framework guides multimodal AMIE through structured phases of history-taking, diagnosis and management and follow-up, dynamically adapting the conversation based on intermediate model outputs that reflect the evolving patient state and diagnostic hypotheses. To rigorously evaluate the system\u2019s performance during model development and in comparison to human clinicians, we employed a two-pronged evaluation strategy. First, we developed an automated evaluation pipeline involving perception tests on isolated medical artifacts and simulated dialogues assessed by an auto-rater across key clinical dimensions, such as diagnostic accuracy and information gathering. Second, we conducted an expert evaluation using an OSCE-style methodology to assess the capabilities of multimodal AMIE in realistic, simulated patient encounters involving multimodal data.<\/p>\n<p>Multimodal state-aware reasoning<\/p>\n<p>Real clinical diagnostic dialogues follow a structured yet flexible path. Clinicians methodically gather information, form potential diagnoses, strategically request and interpret further details (including multimodal data such as skin photographs or ECGs), continually update their assessment based on new evidence and eventually formulate a management plan. This process requires a clinician to adapt the line of questioning based on evolving hypotheses and uncertainty while ensuring that all critical information is considered<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 30\" title=\"Gruppen, L. D. Clinical reasoning: defining it, teaching it, assessing it, studying it. West. J. Emerg. Med. 18, 4&#x2212;7 (2017).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR30\" id=\"ref-link-section-d22012503e1458\" rel=\"nofollow noopener\" target=\"_blank\">30<\/a>.<\/p>\n<p>Given the rapid advancements in LLM capabilities and their increasing proficiency in following complex instructions, one might achieve considerable progress toward emulating this process using a sophisticated system prompt alone. However, we hypothesize that, for a safety-critical and highly dynamic task like multimodal diagnosis, building an explicit state-aware reasoning system layered on top of the LLM offers critical advantages. Such a system provides greater control over the dialogue flow, enables more reliable tracking of the diagnostic state and uncertainty, facilitates more deliberate integration of multimodal inputs and ultimately leads to higher-quality, more dependable clinical reasoning rather than relying solely on complex prompting (validated in the \u2018Automated evaluations of multimodal AMIE configurations\u2019 section).<\/p>\n<p>Therefore, the multimodal AMIE system implements this state-aware dialogue phase transition framework to manage the diagnostic process. This framework dynamically controls the progression of multimodal AMIE through three distinct phases: (1) history-taking, (2) diagnosis and management and (3) follow-up. Transitions between phases, and actions within each phase, are driven by intermediate model outputs representing the evolving patient state and diagnostic hypotheses (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig6\" rel=\"nofollow noopener\" target=\"_blank\">6<\/a>). Crucially, each phase builds upon the accumulated dialogue history, which contextually incorporates various forms of patient data, including text, images (such as skin photographs or ECG tracings) and clinical documents (such as laboratory reports or prior consultation notes). This state-aware approach allows multimodal AMIE to emulate the structured yet adaptive reasoning process of clinicians (see examples in Supplementary Figs. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a> and <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">6<\/a>).<\/p>\n<p>Phase 1: history-taking\u2014building a comprehensive picture<\/p>\n<ol class=\"u-list-style-none\">\n<li>\n                    1.<\/p>\n<p><b>Patient profile initialization:<\/b> A structured patient profile is initialized and dynamically updated throughout the interaction. This profile acts as a condensed, evolving record of known patient information, encompassing the chief complaint; history of present illness; demographics (age, sex and race); positive and negative symptoms; past medical, family and social\/travel histories; medications; other relevant details; and a prioritized list of knowledge gaps. Initially, this profile may contain minimal information.<\/p>\n<\/li>\n<li>\n                    2.<\/p>\n<p><b>Evolving DDx generation:<\/b> An internal, evolving DDx is generated. This DDx is not initially presented to the patient. DDx generation begins after an initial interaction period, allowing baseline information collection. The frequency of DDx updates is configurable.<\/p>\n<\/li>\n<li>\n                    3.<\/p>\n<p><b>Continuation decision:<\/b> A key decision point is whether to continue gathering history or transition to presenting a diagnosis. This decision uses a set of criteria, including assessing if sufficient information exists to formulate a reasonable differential diagnosis. A decision module, querying Gemini 2.0 Flash, determines if current information is sufficient to proceed or if more targeted questions are needed. This module considers the dialogue history and preliminary DDx.<\/p>\n<\/li>\n<li>\n                    4.<\/p>\n<p><b>Targeted question and multimodal data request generation:<\/b> If history-taking continues, the system generates focused questions to address information gaps identified in the patient profile and DDx uncertainty. It is important to note that this internal measure of uncertainty, derived from the model\u2019s iterative DDx generation, is used as a pragmatic heuristic to guide the history-taking dialogue toward more relevant lines of questioning. It is not presented to the user nor is it a formally calibrated diagnostic probability intended for direct clinical interpretation. Crucially, the system is designed to recognize when multimodal data are necessary and to strategically request them. For instance, based on reported symptoms such as a rash, multimodal AMIE will prompt the user to upload skin photographs. Its reasoning extends to requesting additional views if needed (for example, \u2018Could you please provide a photo from a different angle or in better lighting?\u2019 or \u2018Do you have photos showing how the rash looked previously?\u2019). Similarly, for reported cardiac symptoms, it might request an ECG tracing if available. Upon receiving an artifact, multimodal AMIE elicits detailed descriptions adhering to instructions for interpreting the artifacts and their key aspects for determining salient findings (for example, for skin: lesion morphology, distribution and color; for ECGs: heart rate, rhythm, key waveforms and intervals). This explicit prompting for descriptive details ensures that the model extracts salient features from the artifact to inform the ongoing conversation and diagnostic reasoning. The generated question or request is presented to the patient.<\/p>\n<\/li>\n<li>\n                    5.<\/p>\n<p><b>Iterative refinement:<\/b> The process is iterative. New information from patient responses and data uploads is incorporated into the dialogue history, updating the patient profile and internal DDx. The patient summary is periodically refreshed to reflect the current understanding.<\/p>\n<\/li>\n<\/ol>\n<p>Once the continuation decision indicates readiness, the system transitions to phase 2.<\/p>\n<p>Phase 2: diagnosis and management\u2014from data to actionable plan<\/p>\n<ol class=\"u-list-style-none\">\n<li>\n                    1.<\/p>\n<p><b>Patient profile (optional update):<\/b> The patient profile may be further refined.<\/p>\n<\/li>\n<li>\n                    2.<\/p>\n<p><b>DDx validation (subphase):<\/b> The system enters a DDx validation phase. Focused questions are generated to gather specific evidence that supports or refutes potential diagnoses within the internal DDx. A decision module determines when the DDx is sufficiently validated to present to the patient.<\/p>\n<\/li>\n<li>\n                    3.<\/p>\n<p><b>DDx presentation:<\/b> The system presents a ranked DDx (5\u221210 conditions). Crucially, the explanation for each diagnosis is grounded in evidence from the entire interaction, explicitly referencing and explaining findings from the provided multimodal data (for example, \u2018Based on the photo you sent, the circular shape and central clearing of the rash are characteristic of\u2026,\u2019 or \u2018The ECG shows specific changes in the ST segment which support the possibility of\u2026\u2019).<\/p>\n<\/li>\n<li>\n                    4.<\/p>\n<p><b>Management plan formulation:<\/b> After presenting the DDx, the system formulates a management plan based on dialogue history, patient profile and the presented DDx. The plan includes recommendations for investigations, testing and\/or treatment.<\/p>\n<\/li>\n<li>\n                    5.<\/p>\n<p><b>Management plan delivery:<\/b> The system delivers the management plan, potentially iteratively across several turns, allowing for clarification and addressing patient concerns.<\/p>\n<\/li>\n<\/ol>\n<p>The presentation of the refined DDx and management plan signals the transition to phase 3.<\/p>\n<p>Phase 3: answer follow-up questions\u2014 ensuring patient understanding<\/p>\n<ol class=\"u-list-style-none\">\n<li>\n                    1.<\/p>\n<p><b>Patient profile (optional update):<\/b> The patient profile may continue to be updated with information from follow-up questions.<\/p>\n<\/li>\n<li>\n                    2.<\/p>\n<p><b>Plan communication and question answering:<\/b> The system addresses remaining patient questions, ensuring that the patient understands the proposed management plan, potentially referencing the multimodal artifacts again to clarify points. Responses are guided by the dialogue history, patient profile, presented DDx and management plan.<\/p>\n<\/li>\n<li>\n                    3.<\/p>\n<p><b>Dialogue termination:<\/b> Dialogue continues until patient questions are addressed, the management plan is communicated and a natural conclusion is reached.<\/p>\n<\/li>\n<\/ol>\n<p>Upon reaching this termination state, the system proceeds to synthesize the interaction into a structured summary.<\/p>\n<p>After dialogue conclusion: structured post-questionnaire generation<\/p>\n<p>Once the dialogue reaches a natural conclusion (phase 3 completion), the system automatically generates a structured post-questionnaire. This process leverages the full multimodal dialogue history and the final internal state (patient profile, final DDx and management plan) to produce a comprehensive summary suitable for clinical review. Specifically, multimodal AMIE is prompted to:<\/p>\n<ol class=\"u-list-style-none\">\n<li>\n                    1.<\/p>\n<p><b>Finalize DDx:<\/b> Based on the entire interaction, including all textual exchanges and interpretations of multimodal artifacts, generate a final, ranked DDx listing the most probable condition and several plausible alternatives.<\/p>\n<\/li>\n<li>\n                    2.<\/p>\n<p><b>Formulate the management plan:<\/b> Leveraging both the established DDx and a constrained web search process for grounding in current medical knowledge and guidelines, detail the recommended management plan. This includes proposed in-visit and ordered investigations, specific actions or recommendations for the patient and necessary escalation level (for example, video call or in-person visit) with justification and follow-up requirements (necessity, timeframe and reason). This retrieval-based approach is used exclusively during this offline step to ensure that recommendations are informed by up-to-date information. The specific methodology is detailed in Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a>.<\/p>\n<\/li>\n<li>\n                    3.<\/p>\n<p><b>Extract salient artifact findings:<\/b> Identify and list the key clinical findings observed in any provided images (skin photographs, ECGs and documents) that were relevant to the diagnostic and management reasoning.<\/p>\n<\/li>\n<\/ol>\n<p>The structured post-questionnaire serves as a standardized record of multimodal AMIE\u2019s clinical assessment, reasoning and recommendations based on the completed multimodal consultation; it is used both as an artifact rated by clinicians in the OSCE evaluation and as the standardized output scored by the auto-rater in simulated dialogues.<\/p>\n<p>Automatic evaluations and system validation<\/p>\n<p>We established an automated evaluation framework to support rapid iteration and rigorous assessment of the multimodal AMIE system. This included evaluating perception capabilities on isolated medical artifacts and using a simulation environment with auto-raters to assess complete diagnostic multimodal dialogues and perform component ablations and determine optimal model selection.<\/p>\n<p>Validation of base model perception of medical artifacts<\/p>\n<p>A critical precursor to effective multimodal diagnostic conversation is ensuring that our underlying models possess a fundamental capability: accurate perception of diverse medical artifacts. Although LLMs have demonstrated remarkable progress in understanding and generating text, their ability to reliably \u2018see\u2019 and interpret medical images and documents\u2014 akin to a clinician\u2019s initial visual assessment\u2014remains less explored in the context of conversational DDx. Without robust perceptual grounding, even the most sophisticated conversational framework would be limited in its ability to meaningfully integrate multimodal data into history-taking and diagnostic reasoning.<\/p>\n<p>Therefore, we first conducted a suite of perception tests designed to isolate and evaluate the visual understanding of our base models when presented with common medical artifacts: smartphone-captured skin images, ECG tracings and clinical documents. The objective was not to achieve state-of-the-art performance on complex diagnostic tasks per se but, rather, to establish a baseline confidence in whether our models could reliably discern key visual features and clinical information from these modalities in isolation.<\/p>\n<p>For skin images, we used the SCIN dataset<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 31\" title=\"Ward, A. et al. Creating an empirical dermatology dataset through crowdsourcing with web search advertisements. JAMA Netw. Open 7, e2446615 (2024).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR31\" id=\"ref-link-section-d22012503e1759\" rel=\"nofollow noopener\" target=\"_blank\">31<\/a> and PAD-UFES-20 (ref. <a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 32\" title=\"Pacheco, A. G. C. et al. PAD-UFES-20: a skin lesion dataset composed of patient data and clinical images collected from smartphones. Data Brief 32, 106221 (2020).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR32\" id=\"ref-link-section-d22012503e1763\" rel=\"nofollow noopener\" target=\"_blank\">32<\/a>), assessing the ability of the model to describe lesion morphology, color and distribution. For ECGs, we employed PTB-XL<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 33\" title=\"Wagner, P. et al. PTB-XL, a large publicly available electrocardiography dataset. Sci. Data 7, 154 (2020).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR33\" id=\"ref-link-section-d22012503e1767\" rel=\"nofollow noopener\" target=\"_blank\">33<\/a> and ECG-QA benchmark<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 34\" title=\"Oh, J. et al. ECG-QA: a comprehensive question answering dataset combined with electrocardiogram. In NeurIPS &#x2BC;23: 37th International Conference on Neural Information Processing Systems (eds Oh, A. et al.). 35, 66277&#x2013;66288 (Curran Associates, 2023).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR34\" id=\"ref-link-section-d22012503e1771\" rel=\"nofollow noopener\" target=\"_blank\">34<\/a>, prompting the model to provide probable diagnoses or answer expert-validated questions based on ECG images generated from raw signals. Finally, for clinical documents, we created the ClinicalDoc-QA dataset that consists of question-answering tasks based on a large collection of deidentified clinical notes and patient records generated by physicians to evaluate the model\u2019s comprehension of medical information from clinical documents. Additional dataset details can be found in Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">3.1<\/a>.<\/p>\n<p>To ensure accurate processing of artifacts, we verified the perceptual capabilities of our base model, Gemini 2.0 Flash. Our objective was not to achieve state-of-the-art performance on isolated benchmarks but, rather, to confirm a sufficient perceptual foundation for our state-aware reasoning framework. Overall, this verification (detailed in Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig11\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a> and Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">3.1.2<\/a>) demonstrated that Gemini 2.0 Flash possessed the necessary foundational perceptual grounding across our target modalities. This provided the confidence required to build multimodal AMIE\u2019s advanced reasoning and dialogue functions upon this base model. This is quantitatively validated by our ablation studies (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig5\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a>) and ultimately demonstrated by the superior diagnostic accuracy of multimodal AMIE in the human-evaluated OSCE study (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig2\" rel=\"nofollow noopener\" target=\"_blank\">2<\/a>). The outcomes of these perception tests, presented in the \u2018Automated evaluations of multimodal AMIE configurations\u2019 section and Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">3.1.2<\/a>, provide an essential confidence check on our base LLM, Gemini 2.0 Flash, and also enable us to explore critical questions such as the impact of history-taking in addition to multimodal perception.<\/p>\n<p>Auto-rating simulated diagnostic conversations<\/p>\n<p>The following subsections detail our three-step approach illustrated in Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig4\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a> to creating a simulation environment for auto-evaluation. Step (1) encompasses patient profile and scenario generation, which is then used in step (2) for the turn-by-turn multimodal dialogue generation process used to mimic real-world telemedicine interactions, and, finally, the simulated dialogue is used in step (3) by the auto-rater for automated assessment of multimodal AMIE\u2019s conversational abilities.<\/p>\n<p>                  Step 1: Patient profile and scenario generation<\/p>\n<p>Generating realistic synthetic dialogues requires detailed patient profiles and corresponding clinical scenarios. Our methodology involves compiling comprehensive patient metadata, including symptoms, demographics and medical history, tailored to specific medical domains.<\/p>\n<ul class=\"u-list-style-bullet\">\n<li>\n<p><b>Dermatology:<\/b> We used the SCIN and PAD-UES-20 datasets, which provide rich, preexisting patient attributes (for example, age, sex, symptoms, skin tone and medical and social history), requiring no synthetic imputation.<\/p>\n<\/li>\n<li>\n<p><b>Cardiology:<\/b> We started with the PTB-XL ECG dataset. As it includes only age and sex, we employed Gemini 2.0 Flash, using its web search tool, to impute clinically plausible symptoms, social\/family\/medical history and cardiovascular risk factors associated with the ECG\u2019s indicated condition. This grounded the synthetic profiles in the model\u2019s medical knowledge and real-world examples from the web.<\/p>\n<\/li>\n<li>\n<p><b>Clinical documents:<\/b> For this domain, we collaborated with the same external clinical partners who developed our OSCE scenarios. They created realistic patient profiles, document artifacts and corresponding metadata specifically for our simulation needs.<\/p>\n<\/li>\n<\/ul>\n<p>Once comprehensive metadata were established for all domains, we used Gemini 2.0 Flash to generate detailed clinical scenarios. These scenarios provide context for the patient agent in the simulation, outlining the initial presentation, details to share only upon questioning, patient expectations and concerns, desired outcomes and potential questions for the doctor. We guided Gemini using few-shot examples, ensuring scenario diversity (for example, varied ethnicities and occupations) and clinical realism. Although the raw images are from public datasets, we augmented them with unique, synthetically generated scenarios. We designed this approach to mitigate potential data leakage by ensuring that the model encounters these images within novel clinical contexts. Crucially, the scenarios deliberately omit the final diagnosis, requiring the AI agent to deduce it through the simulated conversation and analysis of any provided multimodal artifacts.<\/p>\n<p>                  Step 2: Turn-by-turn multimodal dialogue generation<\/p>\n<p>We simulate dialogues turn by turn between a doctor agent (representing multimodal AMIE) and a patient agent, mimicking a telemedical interaction.<\/p>\n<p>The doctor agent (multimodal AMIE) is instructed to be empathetic and clinically accurate. It uses the state-aware dialogue phase transition framework (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig6\" rel=\"nofollow noopener\" target=\"_blank\">6<\/a>) to navigate history-taking, diagnosis and management and follow-up phases. A key capability is its multimodal nature: it can strategically request and analyze relevant medical artifacts (such as skin images or ECGs) based on patient-reported symptoms during the conversation.<\/p>\n<p>The patient agent simulates a patient strictly following a predefined scenario from step 1, which details their profile and clinical scenario. It is prompted to respond truthfully using concise, casual language appropriate for an online consultation while adhering to specific rules regarding politeness and pacing the release of information to avoid overwhelming the doctor agent.<\/p>\n<p>In each dialogue turn, the doctor agent generates a question or statement based on the conversation history and its internal state. The patient agent then formulates a response according to its scenario instructions. If the scenario dictates and the doctor agent requests it, the patient agent can upload a relevant medical artifact (for example, an image). Multimodal AMIE then analyzes this artifact, integrating the findings into its ongoing diagnostic reasoning and subsequent dialogue turns.<\/p>\n<p>This turn-by-turn exchange continues until the conversation reaches a natural conclusion (for example, patient questions are resolved) or hits a predefined maximum turn limit. After the exchange is concluded, the doctor agent generates the structured post-questionnaire. The output is a complete simulated dialogue transcript, capturing the dynamic exchange of text and multimodal data.<\/p>\n<p>                  Step 3: Criteria and scoring with auto-raters<\/p>\n<p>Auto-rating (automated evaluation) is crucial for the iterative development and safety assessment of conversational models. Although human evaluation is valuable, it is often constrained by cost, time and scalability. Auto-raters provide a mechanism for rapid, scalable and consistent performance assessment across essential characteristics.<\/p>\n<p>We employ an auto-rater based on Gemini 2.0 Flash to evaluate the simulated dialogues. The auto-rater, which is given access to the ground truth condition, assesses various aspects, from quantitative metrics such as diagnostic accuracy to qualitative dimensions such as information gathering and safety. The specific criteria and scoring methods used by our auto-rater are detailed below (Extended Data Table <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"table anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Tab3\" rel=\"nofollow noopener\" target=\"_blank\">3<\/a>):<\/p>\n<ul class=\"u-list-style-bullet\">\n<li>\n<p><b>DDx accuracy:<\/b> Measures if the patient\u2019s ground truth condition is present within the top-1, top-3 or top-10 diagnoses listed by multimodal AMIE\u2019s generated post-questionnaire. Evaluation accounts for synonyms and uses binary scoring (1 if present in top-n, 0 otherwise), averaged across dialogues.<\/p>\n<\/li>\n<li>\n<p><b>Gathering information:<\/b> Assesses the effectiveness of the model in eliciting information, including its use of open-ended questions, active listening, addressing patient concerns and summarization. Rated on a five-point Likert scale (from 1\u2009=\u2009very poor to 5\u2009=\u2009excellent).<\/p>\n<\/li>\n<li>\n<p><b>Management plan appropriateness:<\/b> Evaluates the suitability of the proposed management plan (including recommended actions) relative to best medical practices and the patient\u2019s specific situation. Rated on a five-point Likert scale (from 1\u2009=\u2009very poor to 5\u2009=\u2009excellent).<\/p>\n<\/li>\n<li>\n<p><b>Hallucination:<\/b> Detects instances of fabricated information stated by the model (for example, mentioning details not provided by the patient, making incorrect assumptions and claiming access to non-existent data or medications). This safety-critical metric excludes assessments of the diagnostic or management reasoning itself and is evaluated with a binary score (1\u2009=\u2009hallucination present, 0\u2009=\u2009absent).<\/p>\n<\/li>\n<\/ul>\n<p>This simulation framework advances beyond previous text-only versions<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 4\" title=\"Tu, T. et al. Towards conversational diagnostic artificial intelligence. Nature 642, 442&#x2212;450 (2025).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR4\" id=\"ref-link-section-d22012503e1921\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a> by generating multimodal patient scenarios grounded in real-world datasets (SCIN, PTB-XL and Clinical Documents) and simulating realistic patient\u2212clinician interactions involving these multimodal medical data. As detailed in Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">3.3<\/a>, calibration analysis confirms good alignment with human expert judgments (Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a>), thus enabling automated evaluation of the system\u2019s multimodal reasoning capabilities.<\/p>\n<p>                Expert evaluation: multimodal virtual OSCE<\/p>\n<p>To compare multimodal AMIE\u2019s capabilities in undertaking multimodal diagnostic conversations to those of PCPs, we extended the remote OSCE study design introduced by Tu et al.<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 4\" title=\"Tu, T. et al. Towards conversational diagnostic artificial intelligence. Nature 642, 442&#x2212;450 (2025).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR4\" id=\"ref-link-section-d22012503e1942\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a>. The quality of dialogues was assessed using a set of rubrics and metrics reflecting the perspective of patients and specialist physicians. We also introduced a new evaluation rubric specifically to assess the ability to use multimodal medical data effectively in the context of clinical consultations.<\/p>\n<p>The OSCE is a standardized practical assessment widely used in healthcare education to objectively evaluate clinical skills and competencies by simulating real-world practice<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 35\" title=\"Sloan, D. A., Donnelly, M. B., Schwartz, R. W. &amp; Strodel, W. E. The Objective Structured Clinical Examination. The new gold standard for evaluating postgraduate clinical performance. Ann. Surg. 222, 735&#x2212;742 (1995).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR35\" id=\"ref-link-section-d22012503e1949\" rel=\"nofollow noopener\" target=\"_blank\">35<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 36\" title=\"Newble, D. Techniques for measuring clinical competence: objective structured clinical examinations. Med. Educ. 38, 199&#x2013;203 (2004).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR36\" id=\"ref-link-section-d22012503e1952\" rel=\"nofollow noopener\" target=\"_blank\">36<\/a>. Unlike traditional knowledge-based examinations, the OSCE assesses practical skills of real-world clinical encounters, typically involving candidates rotating through a series of timed stations where they encounter a trained patient-actor portraying a specific clinical scenario. Test-takers perform designated tasks such as taking a medical history, conducting a physical examination, interpreting results or counseling the patient. Examiners observe these interactions and score the test-taker\u2019s performance against detailed, predefined checklists that assess crucial skills such as history-taking, examination technique, clinical reasoning and communication.<\/p>\n<p>In this work, we designed and conducted a virtual analogue of the OSCE adapted for multimodal text chat. In this setting, patient-actors engaged in blinded, synchronous chat conversations with either the multimodal AMIE system or PCPs, conducted through a chat interface (Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">1<\/a>) that allowed exchange of both text and uploaded images as is now commonplace for mobile chat applications (see Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig12\" rel=\"nofollow noopener\" target=\"_blank\">6<\/a> for an overview). Within the virtual consultations, patient-actors were instructed to upload images such as skin photographs, laboratory tests or ECG tracings, emulating how popular text chat platforms have been reportedly used as a means for remote consultation.<\/p>\n<p>Scenario packs<\/p>\n<p>In collaboration with two organizations that routinely perform OSCE assessments in Canada and India, we developed 105 case scenarios. The scenarios were centered around three types of image artifacts that are commonly available to patients in telemedical primary care: (1) smartphone-captured skin images, (2) ECG tracings and (3) clinical documents. We selected these modalities as the most likely to occur in primary telecare: patients can readily photograph their skin, scan documents they received or report with ECGs collected via consumer tools<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 22\" title=\"Li, K., Elgalad, A., Cardoso, C. &amp; Perin, E. C. Using the Apple Watch to record multiple-lead electrocardiograms in detecting myocardial infarction: where are we now? Tex. Heart Inst. J. 49, e227845 (2022).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR22\" id=\"ref-link-section-d22012503e1972\" rel=\"nofollow noopener\" target=\"_blank\">22<\/a>. Extended Data Table <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"table anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Tab4\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a> provides an overview and examples of image artifacts. Scenarios contained different representations of these image artifacts to represent varying levels of image quality, including photographs of a given skin concern from different angles as well as screenshots and smartphone photographs of ECG tracings and clinical documents.<\/p>\n<p>We selected cases for dermatology from the SCIN dataset<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 31\" title=\"Ward, A. et al. Creating an empirical dermatology dataset through crowdsourcing with web search advertisements. JAMA Netw. Open 7, e2446615 (2024).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR31\" id=\"ref-link-section-d22012503e1982\" rel=\"nofollow noopener\" target=\"_blank\">31<\/a>. SCIN contains representative skin images along with rich metadata crowdsourced from real internet users with skin concerns. ECG tracings were taken from PTB-XL<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 33\" title=\"Wagner, P. et al. PTB-XL, a large publicly available electrocardiography dataset. Sci. Data 7, 154 (2020).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR33\" id=\"ref-link-section-d22012503e1986\" rel=\"nofollow noopener\" target=\"_blank\">33<\/a>, the largest publicly available dataset for this modality. Clinical documents were crafted by OSCE laboratories for the purpose of this study. Notably, we included normal cases, too. Scenarios were designed such that both requesting and interpreting images as well as taking the patient\u2019s history were required\u2014while either alone was insufficient\u2014in order to form a confident diagnosis. To this end, for both skin photographs and ECG-based scenarios, we selected challenging images with high annotation ambiguity in their diagnosis labels as detailed in Extended Data Table <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"table anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Tab4\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a>. Furthermore, dermatologists, cardiologists and internists (medical specialists in internal medicine) ensured that the accompanying text-based component of scenarios (for example, family\/medical history and symptoms) was complementary in the sense that both image and textual information were required to arrive at an accurate diagnosis. Nevertheless, we note that, although the scenarios are consistent with the artifacts and the diagnosis, there is no guarantee that they reflect the true case history, as they were created post hoc. Scenarios were crafted to match case metadata provided in the SCIN and PTB-XL datasets (such as age, sex and ethnicity) when available. For Clinical Documents, we selected conditions for which diagnosis formation can be guided based on the results of laboratory reports.<\/p>\n<p>Lastly, to simulate variability in image quality in real-world care settings, for half of the ECG and clinical document cases, we used smartphone photographs of a computer screen showing the artifacts (Extended Data Table <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"table anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Tab4\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a>), as in previous work<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 37\" title=\"Phillips, N. A. et al. CheXphoto: 10,000+ smartphone photos and synthetic photographic transformations of chest X-rays for benchmarking deep learning robustness. In Proc. Machine Learning Research 136, 318&#x2013;327 (2020).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR37\" id=\"ref-link-section-d22012503e1999\" rel=\"nofollow noopener\" target=\"_blank\">37<\/a>.<\/p>\n<p>Study design<\/p>\n<p>Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig12\" rel=\"nofollow noopener\" target=\"_blank\">6<\/a> provides an overview of our OSCE study design. For each case scenario, the same patient-actor performed one conversation with multimodal AMIE and a qualified PCP each via a synchronous chat interface (Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">1<\/a>), in a blinded and randomized order. (Although we conducted a randomized, blinded evaluation, our study is not a randomized clinical trial.) The chat interface supported text-based communication and sharing of images from the scenarios while displaying the scenario pack information to patient-actors throughout the conversation.<\/p>\n<p>After each consultation, patient-actors completed a questionnaire to rate their experience to represent the patient-actor perspective. Separately, a different questionnaire was completed by both multimodal AMIE (using offline generation) and PCPs to summarize key clinical findings and next steps. In particular, the questionnaire asked for a DDx list (at least three and up to 10 plausible items, ordered by likelihood), a management plan and a description of salient image findings. Lastly, a group of specialist physicians assessed the performance of multimodal AMIE and PCPs in a blinded fashion (Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">2<\/a>), based on their consultation transcripts and responses to the post-questionnaires using various rubrics, representing the specialist physician perspective.<\/p>\n<p>A large part of the collected patient-centric as well as specialist metrics is derived from the evaluation rubric introduced in the previous work<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 4\" title=\"Tu, T. et al. Towards conversational diagnostic artificial intelligence. Nature 642, 442&#x2212;450 (2025).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR4\" id=\"ref-link-section-d22012503e2027\" rel=\"nofollow noopener\" target=\"_blank\">4<\/a> for their direct applicability in the assessment of multimodal diagnostic dialogues. These criteria are designed primarily to evaluate the consultation quality, the appropriateness of diagnostic and management decisions, the accuracy of clinical reasoning and various aspects of effective communication skills (such as the ability to elicit information and manage patient concerns). The questions in this rubric were derived from consideration of authoritative assessment schemes of clinical interactions such as the PACES, used by the Royal College of Physicians in the UK for examining history-taking skills<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 38\" title=\"Dacre, J., Besser, M. &amp; White, P. MRCP (UK) PART 2 Clinical Examination (PACES): a review of the first four examination sessions (June 2001&#x2013;July 2002). Clin. Med. (Lond.) 3, 452&#x2212;459 (2003).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR38\" id=\"ref-link-section-d22012503e2031\" rel=\"nofollow noopener\" target=\"_blank\">38<\/a>; the GMCPQ (<a href=\"https:\/\/edwebcontent.ed.ac.uk\/sites\/default\/files\/imports\/fileManager\/patient_questionnaire%20pdf_48210488.pdf\" rel=\"nofollow noopener\" target=\"_blank\">https:\/\/edwebcontent.ed.ac.uk\/sites\/default\/files\/imports\/fileManager\/patient_questionnaire%20pdf_48210488.pdf<\/a>); and approaches to Patient-Centered Communication Best Practices (PCCBP)<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 39\" title=\"King, A. &amp; Hoppe, R. B. &#x2018;Best practice&#x2019; for patient-centered communication: a narrative review. J. Grad. Med. Educ. 5, 385&#x2013;393 (2013).\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#ref-CR39\" id=\"ref-link-section-d22012503e2042\" rel=\"nofollow noopener\" target=\"_blank\">39<\/a>.<\/p>\n<p>The MUH rubric was designed to assess competence in handling and interpreting multimodal artifacts in the context of clinical consultations, including the ability to understand medical image artifacts; to use that understanding to guide the conversation and inform a clinically accurate assessment; and to communicate the salient findings and address the patients\u2019 questions in an appropriate manner. Details are provided in Extended Data Table <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"table anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Tab1\" rel=\"nofollow noopener\" target=\"_blank\">1<\/a>.<\/p>\n<p>Participants<\/p>\n<p>This study involved 19 board-certified PCPs and 25 validated patient-actors, with participants split between India and Canada (10\/9 for PCPs and 10\/10 for patient-actors for India\/Canada, respectively). Informed consent was obtained from each participant before their participation. To ensure consistent and high-quality interactions, all patient-actors and PCPs completed a standardized training prior to participation. This training included an interactive workshop to familiarize them with the chat interface and the specifics of the OSCE interaction format based on detailed instructional guides. The PCPs had a median post-residency experience of 6\u2009years with an interquartile range of 3.5\u221211.5\u2009years. For quality assessment of the multimodal AMIE\/PCP consultations and their post-questionnaire responses, we recruited 18 independent specialist physicians from India and North America over three medical specialities (dermatology, cardiology and internal medicine) to ensure diverse clinical viewpoints as well as requisite expertise. These specialists were independent of the study team and the patient-actor cohort, had a median post-residency experience of 5\u2009years (interquartile range, 4\u22128) and were assigned evaluation tasks matching the medical specialty of the scenarios (for example, dermatology scenarios were evaluated by dermatologists). For each of 105 scenarios, each assigned patient-actor performed two sessions, one with a PCP and another with multimodal AMIE in a randomized order, yielding a total of 210 consultations. Each of these conversations was evaluated by three independent specialists.<\/p>\n<p>Statistical analysis<\/p>\n<p>We analyzed the results from the remote OSCE study using a mixed-effect approach to account for differences in the scenarios using random effects. Most rating choice options for patient-actors and specialist physicians were ordinal (for example, ranging from \u2018Very unfavorable\u2019 to \u2018Very favorable\u2019), and we model them using a cumulative ordinal model with logit link function. For binary ratings (for example, \u2018Yes\/No\u2019), we use a Bernoulli model with logit link. For diagnostic accuracy, we again use a Bernoulli model for \u2018Correct\/Incorrect\u2019. In all models, the scenario is used as a random intercept to model differences in the quality of the patient vignettes or the difficulty of the diagnostic task. Both of these factors can affect the flow of conversation and should, therefore, be considered when modeling ratings of these conversations. The experimental arm (PCP\/multimodal AMIE) is modeled as a fixed effect. For diagnostic accuracy, we also model the fixed effect of the differential size (k) using monotonic regression (Supplementary Section <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM1\" rel=\"nofollow noopener\" target=\"_blank\">2.2<\/a>).<\/p>\n<p>Here, clinical scenarios were modeled as random intercepts to allow generalization of findings to the broader population of medical cases rather than restricting inference to the specific vignettes sampled. We explicitly note that this study is a comparative system evaluation, not a randomized controlled trial. Additionally, because clinical utility requires simultaneous success in accuracy, empathy and trust, we treated these as co-equal outcomes adjusted via false discovery rate (FDR) rather than designating a single primary endpoint.<\/p>\n<p>In some analyses, we also estimate preference by comparing ordinal or binary metrics (for example, in Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig3\" rel=\"nofollow noopener\" target=\"_blank\">3c<\/a>). Here, we model the proportion of preference for multimodal AMIE and PCPs using the one-sided \u03c72 test. For ablation results (for example, in Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig5\" rel=\"nofollow noopener\" target=\"_blank\">5<\/a>), we test differences in metrics using Mann\u2013Whitney U-tests. All P values were corrected for FDR using the Benjamini\u2013Hochberg method as implemented in the Python library \u2018statsmodels\u2019. These corrections were applied to all statistical estimates of the experimental arm on either patient-actor or expert ratings, both when estimating parametric ordinal models and when estimating preference (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#Fig2\" rel=\"nofollow noopener\" target=\"_blank\">2<\/a>). In all figures, we display confidence intervals that show the 2.5th and 97.5th percentiles of bootstrapped distributions of the displayed quantity.<\/p>\n<p>Reporting summary<\/p>\n<p>Further information on research design is available in the <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"http:\/\/www.nature.com\/articles\/s41591-026-04371-0#MOESM2\" rel=\"nofollow noopener\" target=\"_blank\">Nature Portfolio Reporting Summary<\/a> linked to this article.<\/p>\n","protected":false},"excerpt":{"rendered":"This work advances the diagnostic capabilities of the multimodal AMIE system4,5 by integrating multimodal perception, enabling it to&hellip;\n","protected":false},"author":2,"featured_media":484745,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[275],"tags":[2564,2566,19064,18,910,135,475,474,19,4381,17,19880,10047,5801],"class_list":{"0":"post-484744","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-healthcare","8":"tag-biomedicine","9":"tag-cancer-research","10":"tag-diseases","11":"tag-eire","12":"tag-general","13":"tag-health","14":"tag-health-care","15":"tag-healthcare","16":"tag-ie","17":"tag-infectious-diseases","18":"tag-ireland","19":"tag-metabolic-diseases","20":"tag-molecular-medicine","21":"tag-neurosciences"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@ie\/116574400429784268","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts\/484744","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/comments?post=484744"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/posts\/484744\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/media\/484745"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/media?parent=484744"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/categories?post=484744"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ie\/wp-json\/wp\/v2\/tags?post=484744"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}