<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v27i1e76557</article-id><article-id pub-id-type="doi">10.2196/76557</article-id><article-categories><subj-group subj-group-type="heading"><subject>Viewpoint</subject></subj-group></article-categories><title-group><article-title>Multimodal Integration in Health Care: Development With Applications in Disease Management</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Hao</surname><given-names>Yan</given-names></name><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib2">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Cheng</surname><given-names>Chao</given-names></name><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib2">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Li</surname><given-names>Juanjuan</given-names></name><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib2">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Li</surname><given-names>Hongwen</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Di</surname><given-names>Xingsi</given-names></name><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zeng</surname><given-names>Xiaoxia</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Jin</surname><given-names>Shoumei</given-names></name><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Han</surname><given-names>Xiaodong</given-names></name><xref ref-type="aff" rid="aff5">5</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Chongsong</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wang</surname><given-names>Qianqian</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Luo</surname><given-names>Bingying</given-names></name><xref ref-type="aff" rid="aff6">6</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zeng</surname><given-names>Xianhai</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Li</surname><given-names>Ke</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Otolaryngology, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute</institution><addr-line>186 Huangge Road, Longcheng Subdistrict, Longgang District</addr-line><addr-line>Shenzhen, Guangdong</addr-line><country>China</country></aff><aff id="aff2"><institution>Department of Dentistry, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute</institution><addr-line>Shenzhen, Guangdong</addr-line><country>China</country></aff><aff id="aff3"><institution>School of Law, Guangzhou University</institution><addr-line>Guangzhou, Guangdong</addr-line><country>China</country></aff><aff id="aff4"><institution>Department of Ophthalmology, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute</institution><addr-line>Shenzhen, Guangdong</addr-line><country>China</country></aff><aff id="aff5"><institution>Department of Medical Imaging, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute</institution><addr-line>Shenzhen, Guangdong</addr-line><country>China</country></aff><aff id="aff6"><institution>Department of Immunology, Tianjin Medical University</institution><addr-line>Tianjin</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Cahill</surname><given-names>Naomi</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Madu</surname><given-names>Chidinma</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Oluwagbade</surname><given-names>Emmanuel</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Ajibade</surname><given-names>Victoria</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Ke Li, Department of Otolaryngology, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute, 186 Huangge Road, Longcheng Subdistrict, Longgang District, Shenzhen, Guangdong, 518172, China, 86 (755)28989999; <email>jylike@163.com</email></corresp><fn fn-type="equal" id="equal-contrib2"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>21</day><month>8</month><year>2025</year></pub-date><volume>27</volume><elocation-id>e76557</elocation-id><history><date date-type="received"><day>26</day><month>04</month><year>2025</year></date><date date-type="rev-recd"><day>10</day><month>06</month><year>2025</year></date><date date-type="accepted"><day>27</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9; Yan Hao, Chao Cheng, Juanjuan Li, Hongwen Li, Xingsi Di, Xiaoxia Zeng, Shoumei Jin, Xiaodong Han, Chongsong Liu, Qianqian Wang, Bingying Luo, Xianhai Zeng, Ke Li. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 21.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2025/1/e76557"/><abstract><p>Multimodal data integration has emerged as a transformative approach in the health care sector, systematically combining complementary biological and clinical data sources such as genomics, medical imaging, electronic health records, and wearable device outputs. This approach provides a multidimensional perspective of patient health that enhances the diagnosis, treatment, and management of various medical conditions. This viewpoint presents an overview of the current state of multimodal integration in health care, spanning clinical applications, current challenges, and future directions. We focus primarily on its applications across different disease domains, particularly in oncology and ophthalmology. Other diseases are briefly discussed due to the few available literature. In oncology, the integration of multimodal data enables more precise tumor characterization and personalized treatment plans. Multimodal fusion demonstrates accurate prediction of anti&#x2013;human epidermal growth factor receptor 2 therapy response (area under the curve=0.91). In ophthalmology, multimodal integration through the combination of genetic and imaging data facilitates the early diagnosis of retinal diseases. However, substantial challenges remain regarding data standardization, model deployment, and model interpretability. We also highlight the future directions of multimodal integration, including its expanded disease applications, such as neurological and otolaryngological diseases, and the trend toward large-scale multimodal models, which enhance accuracy. Overall, the innovative potential of multimodal integration is expected to further revolutionize the health care industry, providing more comprehensive and personalized solutions for disease management.</p></abstract><kwd-group><kwd>multimodal integration</kwd><kwd>healthcare</kwd><kwd>personalized medicine</kwd><kwd>artificial intelligence</kwd><kwd>digital health</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>In the realm of computer science, the concept of multimodal data refers to the integration and analysis of information from multiple sources or modalities. These modalities can include text, images, audio, video, and sensor data, among others [<xref ref-type="bibr" rid="ref1">1</xref>]. The primary objective of multimodal data integration is to leverage the complementary strengths of different data types to gain a more comprehensive understanding of a given problem or phenomenon. By combining diverse data sources, multimodal approaches can enhance the accuracy, robustness, and depth of analysis [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>].</p><p>In the context of health care, the application of multimodal data integration becomes even more critical due to the diversity of medical information. The health care sector generates vast amounts of data from a wide array of sources, including medical imaging (such as magnetic resonance imaging [MRI], computed tomography [CT] scans, and x-rays), laboratory test results, electronic health records (EHRs), wearable devices, and environmental sensors [<xref ref-type="bibr" rid="ref4">4</xref>]. Medical imaging modalities provide detailed anatomical and functional views of the body. EHRs contain a wealth of clinical information, including patient history, diagnoses, treatments, and outcomes, which are essential for longitudinal health monitoring. Wearable devices continuously monitor physiological parameters, such as heart rate, blood pressure, and physical activity, providing real-time data on a patient&#x2019;s health status. Each of these data types provides unique and valuable insights into patient health, but when considered in isolation, they may offer an incomplete or fragmented view. The integration of these diverse data sources enables a more nuanced and comprehensive understanding of patient health.</p><p>However, the integration and analysis of multimodal data in health care present significant difficulties. The sheer volume and heterogeneity of the data require sophisticated methodologies capable of handling large, complex datasets. This is where artificial intelligence (AI) and machine learning come into play. The development of multimodal AI is a rapidly evolving field. This approach has already shown promise in various areas of health care [<xref ref-type="bibr" rid="ref5">5</xref>-<xref ref-type="bibr" rid="ref7">7</xref>]. Through AI-driven integration of multimodal data, health care providers can achieve a more comprehensive understanding of patient conditions, leading to more accurate diagnoses, personalized treatments, and improved patient outcomes [<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>The future of multimodal integration in health care is promising, with ongoing research and technological advancements poised to further enhance its capabilities and applications. Emerging technologies, such as advanced imaging modalities, next-generation sequencing, and novel wearable devices, are expected to provide even richer datasets for integration [<xref ref-type="bibr" rid="ref9">9</xref>]. In addition, the development of more sophisticated AI algorithms and data fusion techniques will enhance the ability to analyze and interpret complex multimodal data.</p><p>Despite the vast potential of multimodal integration in health care, several challenges remain to be addressed. First, data standardization and privacy protection require robust solutions while ensuring regulatory compliance. Second, model training and deployment face computational bottlenecks when processing large-scale and biased multimodal datasets. Third, model interpretability must be enhanced to provide clinically meaningful explanations that gain physician trust. Overcoming these barriers is critical for realizing the full clinical potential of multimodal health care systems.</p><p>The purpose of this viewpoint is to provide an overview of the current state of multimodal integration in health care, summarize its applications across key disease domains, and discuss the challenges and future directions in this rapidly evolving field. By examining the development and applications of multimodal integration across different disease domains, this viewpoint aims to offer insights into how this approach can further revolutionize the health care industry by providing more comprehensive and personalized solutions for disease management. The content of this study was informed by a systematic search of relevant studies (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p></sec><sec id="s2"><title>Applications</title><sec id="s2-1"><title>Overview</title><p>This section focuses on 2 clinical domains that have seen particularly robust development of multimodal AI applications&#x2014;oncology and ophthalmology. These specialties were selected due to their substantial body of published research and complex diagnostic requirements benefiting from multimodal data. As summarized in <xref ref-type="table" rid="table1">Table 1</xref>, we provide a summary of current multimodal developments in these fields.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Multimodal artificial intelligence applications across specialties.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Disease and application directions</td><td align="left" valign="bottom">Specific examples</td></tr></thead><tbody><tr><td align="left" valign="top">Oncology</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enhanced tumor characterization</td><td align="left" valign="top">Tumor subtype and tumor microenvironment</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Personalized treatment planning</td><td align="left" valign="top">Personalized radiotherapy and immunotherapy</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Early detection and diagnosis</td><td align="left" valign="top">Early cancer detection</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Predicting disease prognosis</td><td align="left" valign="top">Overall survival and progression-free survival</td></tr><tr><td align="left" valign="top">Ophthalmology</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Early diagnosis and risk stratification</td><td align="left" valign="top">Glaucoma and age-related macular degeneration</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Ophthalmology imaging as a noninvasive predictive tool for circulatory system disease</td><td align="left" valign="top">Cardiovascular disease</td></tr></tbody></table></table-wrap></sec><sec id="s2-2"><title>Application of Multimodal Data in Oncology</title><sec id="s2-2-1"><title>Overview</title><p>The integration of multimodal data in cancer care represents one of the most promising advancements in modern oncology. For example, advancements in quantitative multimodal imaging technologies involve the combination of multiple quantitative functional measurements, thereby providing a more comprehensive characterization of tumor phenotypes [<xref ref-type="bibr" rid="ref10">10</xref>]. In addition, integrated genomic analysis methods can reveal dysregulation in biological functions and molecular pathways, offering new opportunities for personalized treatment and monitoring [<xref ref-type="bibr" rid="ref11">11</xref>]. By combining diverse data sources, health care providers can achieve a more comprehensive understanding of cancer biology, leading to more accurate predictions of patient outcomes. This section explores the various applications of multimodal data in cancer care, highlighting specific case studies and the transformative impact of this approach.</p></sec><sec id="s2-2-2"><title>Enhanced Tumor Characterization</title><p>One of the primary objectives of integrating multimodal data in cancer care is to achieve enhanced tumor characterization. Tumor characterization involves understanding the genetic, molecular, and phenotypic features of a tumor [<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref14">14</xref>], which is essential for elucidating the nature and properties of the malignancy.</p><p>A key aspect of this process is the differentiation of tumor subtypes. Tumor subtypes refer to the classification of tumors into distinct categories. Differentiating tumor subtypes is essential because it allows for more precise diagnosis, prognosis, and the development of tailored treatment strategies, specific to the characteristics of each subtype [<xref ref-type="bibr" rid="ref15">15</xref>]. Previous cancer subtypes were often classified based on gene expression profiles, such as the PAM50 method [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. However, patients within the same group may still experience different outcomes [<xref ref-type="bibr" rid="ref18">18</xref>], indicating the need for more accurate subtype classification methods. Pathological images and omics data are commonly used for accurate tumor classification through multimodal integration. The features derived from the fusion of image modality data with genomic and other omics data can predict breast cancer subtypes [<xref ref-type="bibr" rid="ref19">19</xref>]. Typically, dedicated feature extractors are used for each modality. A trained convolutional neural network model captures deep features from pathological images, while a trained deep neural network model extracts features from genomic and other omics data. These multimodal features are then integrated through a fusion model to achieve an accurate prediction of breast cancer molecular subtypes. This integrative approach can also be extended to other tumor types and even pan-cancer studies to support the prediction of cancer subtypes and severity [<xref ref-type="bibr" rid="ref20">20</xref>-<xref ref-type="bibr" rid="ref22">22</xref>]. A large-scale study integrated transcriptome, exome, and pathology data from over 200,000 tumors to develop a multilineage cancer subtype classifier [<xref ref-type="bibr" rid="ref18">18</xref>].</p><p>The tumor microenvironment (TME) plays a crucial role in tumor initiation, progression, metastasis, and resistance to therapy [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref24">24</xref>]. In recent years, advancements in new technologies such as single-cell and spatial technologies [<xref ref-type="bibr" rid="ref25">25</xref>] have provided fine-grained resolution of TME, significantly enhancing our understanding of cellular interactions at both single-cell and spatial dimensions [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. Besides, the use of multimodal nanosensors can achieve real-time monitoring within the TME [<xref ref-type="bibr" rid="ref28">28</xref>]. Using multimodal features extracted from single-cell and spatial transcriptomics reveals immunotherapy-relevant non&#x2013;squamous cell carcinoma (non&#x2013;small cell lung cancer [NSCLC]) TME heterogeneity [<xref ref-type="bibr" rid="ref29">29</xref>]. The combination of the 2 modalities and multiplexed ion beam imaging identifies distinct tumor subgroups and a cancer-specific tumor-specific keratinocyte [<xref ref-type="bibr" rid="ref30">30</xref>]. Spatial multiomics delineate core and margin compartments in oral squamous cell carcinoma, with metabolically active margins demonstrating elevated adenosine triphosphate production to fuel invasion [<xref ref-type="bibr" rid="ref31">31</xref>]. In cross-modal applications, gene expression can be predicted from histopathological images of breast cancer tissue with a resolution of 100 &#x00B5;m [<xref ref-type="bibr" rid="ref32">32</xref>]. Conversely, spatial transcriptomic features can better characterize breast cancer tissue sections, revealing hidden histological features [<xref ref-type="bibr" rid="ref33">33</xref>]. By extracting interpretable features from pathological slides, it is also possible to predict different molecular phenotypes [<xref ref-type="bibr" rid="ref34">34</xref>]. These methods provide a comprehensive, quantitative, and interpretable window into the composition and spatial structure of the TME.</p></sec><sec id="s2-2-3"><title>Personalized Treatment Planning</title><p>Another critical objective of multimodal data integration in cancer care is personalized treatment planning. Personalized treatment involves tailoring medical interventions to the individual characteristics of each patient, taking into account their tumor biology and overall health status. By integrating data from multiple sources, health care providers can develop more precise and personalized treatment plans that improve patient outcomes.</p><p>In terms of radiation therapy, using multimodal scanning techniques and mathematical models, it is possible to design personalized radiotherapy plans for glioblastoma patients. By integrating high-resolution MRI scans and metabolic profiles, this approach enables more accurate inference of tumor cell density, thereby optimizing radiotherapy regimens and reducing damage to healthy tissue [<xref ref-type="bibr" rid="ref35">35</xref>]. The integration of biological information-driven multimodal imaging techniques allows physicians to better understand the spatial and temporal heterogeneity of tumors to develop personalized radiotherapy regimens [<xref ref-type="bibr" rid="ref36">36</xref>].</p><p>In the trend of precision medicine, another therapeutic approach is immunotherapy. Immune checkpoint blockade can unleash immune cells to reinvigorate antitumor immunity [<xref ref-type="bibr" rid="ref37">37</xref>]. Multiple phase III clinical trials have demonstrated that the anti&#x2013;programmed cell death protein 1 antibody nivolumab significantly improves overall survival with a favorable safety profile in patients with NSCLC [<xref ref-type="bibr" rid="ref38">38</xref>]. Although single-modality biomarkers can predict responses to immune checkpoint blockade, their predictive power is not always satisfactory. Activating an antitumor immune response through immunotherapy involves a series of complex events that require the interaction of multiple cell types [<xref ref-type="bibr" rid="ref39">39</xref>]. Therefore, achieving precision immunotherapy necessitates integrating multiple data modalities and adopting a holistic approach to analyze the human TME. Translating these multimodal factors into clinically usable predictive markers facilitates the selection of optimal immunotherapy. Combining the informational content present in routine diagnostic data, including annotated CT scans, digitized immunohistochemistry slides, and common genomic alterations in NSCLC, can improve the prediction of responses to programmed cell death protein 1 or programmed cell death-ligand 1 blockade [<xref ref-type="bibr" rid="ref40">40</xref>]. Multi-modal model by Chen et al [<xref ref-type="bibr" rid="ref41">41</xref>] can predict the response to anti&#x2013;human epidermal growth factor receptor 2 combined immunotherapy using multimodal radiology, pathology, and clinical information, achieving an area under the curve (AUC) of 0.91. Furthermore, the application of multimodal approaches in targeted cancer therapy has demonstrated significant potential. Integrating radiomic phenotypes with liquid biopsy data can enhance the predictive accuracy for the efficacy of epidermal growth factor receptor inhibitors [<xref ref-type="bibr" rid="ref42">42</xref>].</p></sec><sec id="s2-2-4"><title>Early Detection and Diagnosis</title><p>Early detection and diagnosis of cancer are crucial for improving patient outcomes, as early-stage cancers are often more treatable and have better prognoses [<xref ref-type="bibr" rid="ref43">43</xref>]. Multimodal data integration plays a vital role in enhancing the accuracy and timeliness of cancer detection and diagnosis.</p><p>Liquid biopsy is a noninvasive technique that involves the collection of nonsolid samples, providing possibilities for early cancer detection and longitudinal tracking [<xref ref-type="bibr" rid="ref44">44</xref>]. This technology includes circulating tumor cells shed from primary and metastatic tumors, as well as circulating tumor DNA (ctDNA) [<xref ref-type="bibr" rid="ref45">45</xref>]. ctDNA can detect trace amounts of tumor DNA even before the tumor manifests obvious symptoms or becomes visible through imaging. Numerous studies and articles have used ctDNA in combination with various other modalities for early cancer prediction, including lung cancer [<xref ref-type="bibr" rid="ref46">46</xref>], breast cancer [<xref ref-type="bibr" rid="ref47">47</xref>], and colorectal cancer [<xref ref-type="bibr" rid="ref48">48</xref>]. Cell-free DNA is a substance that is consistently present in plasma and has been receiving increasing attention. Combining cell-free DNA with other modalities can be used for highly specific early detection across multiple cancer types [<xref ref-type="bibr" rid="ref49">49</xref>-<xref ref-type="bibr" rid="ref51">51</xref>]. AutoCancer uses a transformer model to integrate multiple modalities, including liquid biopsy, mutation, and clinical data, achieving accurate early cancer detection in both lung cancer and pan-cancer analyses [<xref ref-type="bibr" rid="ref52">52</xref>]. Multimodal models that integrate genomic features and clinical data have also demonstrated excellent performance in the early detection of colorectal cancer, with an AUC of 0.98 in the validation set and a sensitivity and specificity of more than 90% [<xref ref-type="bibr" rid="ref49">49</xref>].</p></sec><sec id="s2-2-5"><title>Predicting Disease Prognosis</title><p>Prognosis involves assessing the risk of future outcomes based on an individual&#x2019;s clinical and nonclinical characteristics. These outcomes are typically specific events, such as death or complications, but they can also be quantitative measures, such as disease progression, changes in pain levels, or quality of life [<xref ref-type="bibr" rid="ref53">53</xref>]. Predicting disease prognosis is a critical aspect of cancer care, as it allows for timely interventions and improved long-term outcomes. Multimodal data integration enhances the ability to predict disease prognosis.</p><p>Prognosis in tumor research can be divided into 2 key areas: recurrence and survival. In the context of recurrence, a retrospective analysis and multicenter validation study involving over 2000 patients demonstrated that a multimodal recurrence score, which integrated clinical, genomic, and histopathological data, accurately predicted postoperative local recurrence of renal cell carcinoma [<xref ref-type="bibr" rid="ref54">54</xref>]. Combining the emerging tool of habitat imaging with traditional gene expression and clinical data enables noninvasive stratification of patients with NSCLC, enhancing the prediction of recurrence risk [<xref ref-type="bibr" rid="ref55">55</xref>]. In another study, algorithms were developed based on structured clinical and administrative data to detect recurrence in lung and colorectal cancer patients. By using EHRs and tumor registry data, these algorithms successfully improved the accuracy of recurrence detection [<xref ref-type="bibr" rid="ref56">56</xref>].</p><p>Regarding survival, an increasing number of studies have adopted multimodal approaches to predict patient survival [<xref ref-type="bibr" rid="ref57">57</xref>-<xref ref-type="bibr" rid="ref61">61</xref>]. By integrating data from various sources, these studies have achieved accurate survival predictions across multiple tumor types, including overall survival, 5-year survival rates, and progression-free survival.</p></sec></sec><sec id="s2-3"><title>Application of Multimodal Data in Ophthalmology</title><sec id="s2-3-1"><title>Overview</title><p>Ophthalmology, the medical specialty focused on the diagnosis and treatment of eye disorders, has experienced significant advancements through the integration of multimodal data. Advanced imaging techniques are central to ophthalmology, providing detailed visualizations of the retina, optic nerve, and other ocular structures [<xref ref-type="bibr" rid="ref62">62</xref>]. Optical coherence tomography (OCT) is a widely used imaging modality that offers high-resolution cross-sectional images of the retina, enabling the detection of structural abnormalities and disease progression. Fundus photography and fluorescein angiography provide additional insights into the retinal vasculature and blood flow, which are critical for diagnosing and managing conditions like diabetic retinopathy and retinal vein occlusion. These imaging techniques, when integrated, offer a comprehensive view of both the structural and genetic factors contributing to ocular diseases. The fusion of these data types enables early diagnosis, personalized treatment plans, and continuous monitoring of disease progression and response to therapy, particularly in conditions like age-related macular degeneration (AMD), diabetic retinopathy, and glaucoma [<xref ref-type="bibr" rid="ref63">63</xref>].</p></sec><sec id="s2-3-2"><title>Early Diagnosis and Risk Stratification</title><p>The integration of these diverse data types in ophthalmology achieves several important objectives. Early diagnosis and risk stratification are critical for managing ocular diseases, and the combination of genetic, imaging, and clinical data enables the identification of early signs of eye conditions and stratification of patients based on their risk profiles.</p><p>Color fundus photography and OCT are 2 of the most cost-effective tools for glaucoma screening. Mehta et al [<xref ref-type="bibr" rid="ref64">64</xref>] developed a high-performance multimodal glaucoma detection system by integrating OCT volumes, fundus photographs, and clinical data. Their approach combined features extracted from individual modalities, followed by gradient boosting decision trees for final multimodal construction. The model was rigorously developed and validated on a cohort of 96,020 UK Biobank participants, demonstrating both excellent discriminative performance (AUC=0.97). Importantly, the architecture maintained clinical interpretability through comprehensive feature importance analysis [<xref ref-type="bibr" rid="ref64">64</xref>]. Other multimodal models for glaucoma and its grading detection, based on modalities, such as OCT and fundus images, have also achieved AUC exceeding 0.90 [<xref ref-type="bibr" rid="ref65">65</xref>-<xref ref-type="bibr" rid="ref67">67</xref>]. By using a dual-stream convolutional neural network model to extract features from OCT and color fundus photographs, AMD can be classified into 3 categories&#x2014;normal fundus, dry AMD, and wet AMD [<xref ref-type="bibr" rid="ref68">68</xref>]. Another study enrolled 75 participants from optometry clinics in Auckland and Milford Eye Clinic, New Zealand. By stratifying subjects into young healthy controls, older adult healthy controls, and moderate dry AMD groups, the multimodal diagnostic system achieved 96% classification accuracy [<xref ref-type="bibr" rid="ref69">69</xref>]. In addition, the use of multimodal data can also identify polypoidal choroidal vasculopathy [<xref ref-type="bibr" rid="ref70">70</xref>], dry eye disease [<xref ref-type="bibr" rid="ref71">71</xref>], and diabetic retinopathy [<xref ref-type="bibr" rid="ref72">72</xref>-<xref ref-type="bibr" rid="ref75">75</xref>]. There is also comprehensive work demonstrating that multimodal deep learning (DL) models, which use combined color fundus photography and OCT image sequences as input, can be used to simultaneously detect multiple common retinal diseases [<xref ref-type="bibr" rid="ref76">76</xref>,<xref ref-type="bibr" rid="ref77">77</xref>].</p></sec><sec id="s2-3-3"><title>Ophthalmology Imaging as a Noninvasive Predictive Tool for Circulatory System Disease</title><p>Currently, the diagnosis and treatment of circulatory system disease primarily rely on imaging examinations such as MRI, coronary CT angiography, and coronary angiography. These examinations are not only expensive and time-consuming but also partially invasive and require a high level of professional expertise from the operators. Consequently, early screening and long-term follow-up examinations are challenging to implement in regions with limited medical resources. To better achieve early warning and assessment of circulatory system disease, there is a continuous need to develop new diagnostic tools that are noninvasive, convenient, and efficient.</p><p>The microcirculation of the retina is part of the body&#x2019;s microcirculation system and shares similar embryological origins and pathophysiological characteristics with the cardiovascular system [<xref ref-type="bibr" rid="ref78">78</xref>]. Numerous studies have identified retinal imaging biomarkers associated with early cardiovascular diseases (CVDs) lesions and prognosis, demonstrating the significant value of retinal imaging in CVD screening and prognostic evaluation [<xref ref-type="bibr" rid="ref79">79</xref>,<xref ref-type="bibr" rid="ref80">80</xref>].</p><p>Al-Absi et al [<xref ref-type="bibr" rid="ref81">81</xref>] used a multimodal approach integrating retinal images and dual-energy x-ray absorptiometry data to diagnose CVD in a Qatari cohort. The multimodal model achieved 78.3% accuracy, outperforming unimodal models [<xref ref-type="bibr" rid="ref81">81</xref>]. Notably, their model is interpretable, using Gradient-weighted Class Activation Mapping (Grad-CAM) to highlight the areas of interest in retinal images that most influenced the decisions of the proposed DL model. A study using clinical information and fundus photographs from the UK Biobank demonstrated a significant association between the incidence of CVD in high-risk patients and multimodal predicted risk (hazard ratio 6.28, 95% CI 4.72&#x2010;8.34), and visualized feature importance [<xref ref-type="bibr" rid="ref82">82</xref>].</p></sec></sec></sec><sec id="s3"><title>Challenges in Multimodal Health Care</title><p>While the integration of multimodal data in health care holds great promise, it also presents several significant challenges that need to be addressed.</p><sec id="s3-1"><title>Data Standardization and Privacy</title><p>One of the primary challenges in multimodal health care is integrating diverse medical data sources with varying formats, resolutions, and quality levels [<xref ref-type="bibr" rid="ref83">83</xref>]. Inconsistent data collection practices, missing entries, and recording errors can compromise model reliability, necessitating robust standardization protocols [<xref ref-type="bibr" rid="ref84">84</xref>]. Effective multimodal integration requires comprehensive data cleaning, validation, and preprocessing to create cohesive, high-quality datasets that support accurate predictive analytics. The growing availability of novel health care data sources presents both opportunities for personalized medicine and challenges for systematic integration.</p><p>The use of multimodal data in health care raises significant concerns about data privacy and security. Medical data is highly sensitive, and ensuring its protection is paramount. Regulatory frameworks, such as the Health Insurance Portability and Accountability Act in the United States and the General Data Protection Regulation in the European Union, are essential for protecting patient privacy and ensuring data security. But the concept of health information privacy continues to evolve over time. As new technologies and data sources emerge, it is essential to update and adapt these legal frameworks to reflect new realities [<xref ref-type="bibr" rid="ref85">85</xref>]. The use of multimodal data raises significant privacy concerns [<xref ref-type="bibr" rid="ref86">86</xref>]. Implementing robust data encryption, secure data storage, and strict access controls are essential measures to protect patient information [<xref ref-type="bibr" rid="ref87">87</xref>]. Comprehensive data governance frameworks must establish clear guidelines for responsible and transparent multimodal data usage, while carefully balancing potential risks and benefits for participants, researchers, and society at large [<xref ref-type="bibr" rid="ref88">88</xref>]. Effective implementation requires developing robust data sharing agreements, establishing independent oversight committees, and maintaining ongoing engagement with research participants and other stakeholders [<xref ref-type="bibr" rid="ref89">89</xref>]. In addition, developing secure data sharing protocols and anonymization techniques can help mitigate risks while enabling the effective use of multimodal data for research and clinical applications [<xref ref-type="bibr" rid="ref90">90</xref>]. Ensuring data privacy and security is fundamental to maintaining patient trust and the ethical use of medical data.</p><p>The initial phase of multimodal health care requires systematic collection of standardized data following heterogeneity resolution, coupled with privacy protection through secure protocols. Integrating rigorous data processing with ethically compliant governance frameworks enables usage of diverse datasets for precision medicine while safeguarding sensitive information. This equilibrium is critical for advancing research ethically and maintaining public trust in medical AI applications.</p></sec><sec id="s3-2"><title>Model Training and Deployment</title><p>Multimodal models demand substantial computational resources for both training and inference. The complexity of these models often results in extended training times and significant costs, which can be prohibitive for many health care institutions [<xref ref-type="bibr" rid="ref91">91</xref>,<xref ref-type="bibr" rid="ref92">92</xref>]. Training these models requires high-performance computing environments equipped with powerful Graphics Processing Units or Tensor Processing Units, which are not always accessible to all institutions. Furthermore, the inference phase, where the trained model is applied to new data, can also be resource-intensive, particularly when dealing with large-scale datasets or real-time applications [<xref ref-type="bibr" rid="ref93">93</xref>]. This computational burden can limit the scalability and practical deployment of multimodal models in clinical settings.</p><p>Beyond computational constraints, biases in training data pose a significant challenge to multimodal fusion. Biases may arise from uneven data distribution, inconsistent annotation quality, or systemic disparities in data collection. AI-driven decisions are fundamentally shaped by their initial training data. If the underlying datasets contain biases or inequities, the resulting algorithms risk perpetuating prejudice, incomplete representations, or discriminatory outcomes&#x2014;potentially amplifying systemic inequalities [<xref ref-type="bibr" rid="ref94">94</xref>]. To counteract these biases, strategies such as bias-aware sampling and fairness constraints during model optimization can be implemented. While some AI developers claim their algorithmic systems can mitigate biases, critics maintain that algorithms alone cannot eradicate discrimination, as they may inadvertently perpetuate existing bias in training data [<xref ref-type="bibr" rid="ref95">95</xref>]. This tension highlights the need for complementary strategies (ie, rigorous dataset curation to ensure diversity and continuous monitoring for disparate impacts) [<xref ref-type="bibr" rid="ref96">96</xref>].</p><p>Training and running multimodal models demand expensive hardware, limiting clinical adoption. Meanwhile, biased training data can perpetuate health care disparities. While optimization techniques and bias mitigation strategies help, robust data curation and ongoing monitoring of potentially biased data remain essential for practical, equitable deployment.</p></sec><sec id="s3-3"><title>Model Interpretability</title><p>While multimodal models can achieve high accuracy, their complexity often makes them difficult to interpret. This lack of interpretability poses a significant barrier to their adoption in clinical practice, as clinicians and patients need to understand the rationale behind model predictions to trust and effectively use these tools [<xref ref-type="bibr" rid="ref97">97</xref>]. Enhancing the interpretability and transparency of multimodal models is therefore crucial [<xref ref-type="bibr" rid="ref98">98</xref>]. Techniques, such as explainable artificial intelligence (XAI), can play a pivotal role in this regard [<xref ref-type="bibr" rid="ref99">99</xref>]. XAI methods aim to make the decision-making processes of AI models more understandable to humans by providing explanations that are both accurate and comprehensible. Classical XAI approaches include attention mechanisms and Grad-CAM. Attention scores highlight relevant regions through forward propagation, while Grad-CAM reveals feature significance by capturing gradient changes during backpropagation [<xref ref-type="bibr" rid="ref100">100</xref>].</p><p>Attention mechanisms were originally developed to help neural networks focus on the most relevant parts of input data when making predictions. The core principle involves calculating attention weights&#x2014;numerical scores that determine how much each input element (eg, words in text or regions in an image) should influence the model&#x2019;s output [<xref ref-type="bibr" rid="ref101">101</xref>]. MedFuseNet [<xref ref-type="bibr" rid="ref102">102</xref>] uses an image attention mechanism to dynamically focus on the most clinically relevant regions of medical images corresponding to the input textual queries. Visualization of the attention matrices reveals that the model consistently attends to anatomically discriminative regions of target organs, demonstrating its capability to identify pathologically significant features. StereoMM [<xref ref-type="bibr" rid="ref103">103</xref>] enables quantitative analysis of cross-attention matrices to determine the relative contribution weights of different modalities during fusion, thereby offering interpretable insights into the prioritization of modalities by the model in its decision-making process. Nevertheless, attention weights primarily reflect statistical correlations rather than causal relationships. The fact that a feature receives high attention does not necessarily imply it was determinative for the model&#x2019;s prediction. Compounding this issue, empirical studies have demonstrated that substantially different attention weight distributions can yield identical model outputs [<xref ref-type="bibr" rid="ref104">104</xref>]. These limitations raise questions about the validity of using attention mechanisms as reliable tools for explaining neural network behavior, making an ongoing subject of debate in the machine learning community [<xref ref-type="bibr" rid="ref105">105</xref>].</p><p>Grad-CAM generates explanations by computing gradients from the final convolutional layer, highlighting prediction-relevant regions [<xref ref-type="bibr" rid="ref106">106</xref>]. This interpretability method helps detect invalid decision patterns. For instance, if highest activations appear on imaging artifacts rather than anatomical structures, it exposes critical model flaws. In a clinical study using brain MRI for classification of multiple sclerosis subtypes, Grad-CAM&#x2013;generated heatmaps consistently and distinctly highlighted brain regions critical for differentiating between subtypes, thereby demonstrating the validity and explanatory power. Furthermore, Grad-CAM analysis identified previously unrecognized neuroanatomical loci, offering novel insights into disease progression mechanisms and potentially revealing new imaging biomarkers or therapeutic targets [<xref ref-type="bibr" rid="ref107">107</xref>]. It should be noted that Grad-CAM offers qualitative visualization of model decisions, not quantitative validation. Its clinical relevance must be determined through physician assessment of the identified features [<xref ref-type="bibr" rid="ref108">108</xref>].</p><p>Multimodal AI models face a key challenge&#x2014;balancing high accuracy with clinical interpretability. Current XAI methods offer partial solutions, but with important limitations. Both methods produce explanations that require clinical validation, and physician expertise remains essential to assess biological plausibility. These limitations highlight the need for XAI approaches that provide both technical transparency and clinically meaningful explanations to enable trustworthy AI adoption in health care.</p></sec></sec><sec id="s4"><title>The Development Direction of Multimodal Technology: Expanding Disease Applications</title><p>The development of multimodal technology encompasses broader applications across various diseases and the advancement of large-scale models. With technological progress, multimodal approaches are no longer limited to the diagnosis and prognosis of cancer and ophthalmic diseases but are expanding into CVD, neurological disorders, metabolic diseases, otolaryngology, and more.</p><p>In the field of CVD, multimodal technology can combine data from cardiac MRI, coronary CT, echocardiography, and biomarkers to provide a more comprehensive assessment of heart health [<xref ref-type="bibr" rid="ref87">87</xref>]. For example, integrating these data can more accurately predict the risk of ischemic heart disease [<xref ref-type="bibr" rid="ref109">109</xref>,<xref ref-type="bibr" rid="ref110">110</xref>], coronary artery disease [<xref ref-type="bibr" rid="ref111">111</xref>], assess cardiac function [<xref ref-type="bibr" rid="ref112">112</xref>], and detect disease subgroups plans [<xref ref-type="bibr" rid="ref113">113</xref>]. In addition, multimodal technology can be used to monitor the treatment effects and disease progression in heart disease patients, allowing timely adjustments to treatment strategies and improving patient survival rates and quality of life [<xref ref-type="bibr" rid="ref114">114</xref>].</p><p>In the realm of neurological disorders, multimodal technology also holds significant promise. A proposed model demonstrates robust multimodal integration capabilities, effectively combining both imaging and nonimaging clinical data to achieve accurate differential diagnosis of Alzheimer disease, with discriminative performance exceeding AUC values of 0.9 across multiple diagnostic tasks [<xref ref-type="bibr" rid="ref115">115</xref>]. By combining brain MRI, functional MRI, electroencephalography, and genomic data, researchers can gain a more comprehensive understanding of the pathophysiology of diseases such as Alzheimer [<xref ref-type="bibr" rid="ref116">116</xref>], Parkinson [<xref ref-type="bibr" rid="ref117">117</xref>], and multiple sclerosis [<xref ref-type="bibr" rid="ref118">118</xref>]. Integrating these data can aid in the early diagnosis of these diseases and assess disease severity.</p><p>In the field of metabolic diseases, multimodal technology also has important applications. Integrating clinical documentation with structured laboratory data significantly improves the predictive performance of unimodal machine learning models for early-stage type 2 diabetes mellitus detection. The model achieved an AUC greater than 0.70 for new-onset type 2 diabetes mellitus prediction [<xref ref-type="bibr" rid="ref119">119</xref>]. By integrating metabolomics, genomics, imaging, and clinical data, researchers can gain a more comprehensive understanding of the pathophysiology of diseases, such as obesity [<xref ref-type="bibr" rid="ref120">120</xref>] and fatty liver disease [<xref ref-type="bibr" rid="ref121">121</xref>]. Integrating these data can aid in the early diagnosis of these diseases and assess disease status.</p><p>In the field of otolaryngology, the automatic classification of parotid gland tumors based on multimodal MRI sequences shows promise for improving diagnostic decision-making in clinical settings [<xref ref-type="bibr" rid="ref122">122</xref>]. The integration of CT and MRI enables precise tumor segmentation of oropharyngeal squamous cell carcinoma, resulting in higher dice similarity coefficients and lower Hausdorff distances [<xref ref-type="bibr" rid="ref123">123</xref>]. Combining otoscopic images and wideband tympanometry enables the automatic detection of otitis media [<xref ref-type="bibr" rid="ref124">124</xref>]. Institutions have recognized the importance of collecting multimodal data for interdisciplinary audiology research and have developed a multimodal database that can be used for algorithm development [<xref ref-type="bibr" rid="ref125">125</xref>].</p></sec><sec id="s5"><title>The Trend Toward Large Language Models</title><p>Large language models (LLMs) are foundational pretrained AI systems capable of processing and generating human-like text [<xref ref-type="bibr" rid="ref126">126</xref>]. Their key advantage lies in capturing complex semantic relationships within language data. Building upon LLMs, large multimodal models extend these capabilities to integrate and analyze diverse data types (text, images, genomic data, etc), achieving significant advancements and breakthroughs, gradually forming the rudiments of artificial general intelligence [<xref ref-type="bibr" rid="ref127">127</xref>]. The trend toward LLM in multimodal technology enhances the accuracy and robustness of disease prediction and diagnosis by capturing complex relationships between different data types [<xref ref-type="bibr" rid="ref128">128</xref>,<xref ref-type="bibr" rid="ref129">129</xref>].</p><p>For example, transformer models, which have achieved remarkable success in natural language processing and computer vision, are now being applied to the integration and analysis of multimodal data [<xref ref-type="bibr" rid="ref130">130</xref>]. The transformer-based unified multimodal diagnostic transformer model is capable of directly generating diagnostic results for lung diseases from multimodal input data [<xref ref-type="bibr" rid="ref131">131</xref>].</p><p>Furthermore, LLMs have stronger generalization capabilities, allowing them to be applied across various diseases and populations. This general-purpose approach not only enhances diagnostic accuracy but also reduces the cost and complexity of training and deploying multiple specialized models. For instance, a single large multimodal model could be used for the diagnosis and prognosis of cancer, aging and age-related diseases [<xref ref-type="bibr" rid="ref132">132</xref>], CVDs, neurological disorders, and metabolic diseases, streamlining the process and improving efficiency.</p><p>Another important aspect of LLMs is their interpretability, primarily achieved through the use of attention weights. Although DL models are often considered &#x201C;black boxes,&#x201D; recent advancements have focused on improving model transparency. Attention mechanisms enhance interpretability by identifying and emphasizing the most critical features in the input data, allowing attention to be visualized as regions of information that contribute to decision-making [<xref ref-type="bibr" rid="ref133">133</xref>,<xref ref-type="bibr" rid="ref134">134</xref>]. By visualizing the distribution of attention weights, one can extract the content with high attention weights, which often have a greater impact on the final outcome prediction [<xref ref-type="bibr" rid="ref135">135</xref>].</p><p>In summary, the trend toward LLMs in multimodal development is poised to bring significant innovations and breakthroughs to the medical field. By leveraging the power of large-scale, multimodal datasets and advanced neural network architectures, researchers can achieve more accurate and comprehensive disease predictions and diagnoses.</p></sec></body><back><ack><p>This work was supported by Shenzhen Science and Technology Plan Projects (JCYJ20220530154200002 and JCYJ20230807091701004), Shenzhen Key Medical Discipline Construction Fund (SZXK039), and Longgang District Medical and Health Technology Attack Project (LGKCYLWS2023027).</p></ack><fn-group><fn fn-type="con"><p>YH, CC, JL, XH, BL, and SJ finished the writing-original draft. HL and Xianhai Z were involved in investigation and validation. CL, XD, and QW did conceptualization and editing. CC, Xianhai Z, and KL performed supervision and funding acquisition.</p><p>Xianhai Z is the co-corresponding author of this paper and can be reached at: Department of Otolaryngology, Shenzhen Longgang Otolaryngology Hospital &#x0026; Shenzhen Otolaryngology Research Institute; zxhklwx@163.com</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">AMD</term><def><p>age-related macular degeneration</p></def></def-item><def-item><term id="abb3">AUC</term><def><p>area under the curve</p></def></def-item><def-item><term id="abb4">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb5">ctDNA</term><def><p>circulating tumor DNA</p></def></def-item><def-item><term id="abb6">CVD</term><def><p>cardiovascular disease</p></def></def-item><def-item><term id="abb7">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb8">EHR</term><def><p>electronic health record</p></def></def-item><def-item><term id="abb9">Grad-CAM</term><def><p>Gradient-weighted Class Activation Mapping</p></def></def-item><def-item><term id="abb10">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb11">MRI</term><def><p>magnetic resonance imaging</p></def></def-item><def-item><term id="abb12">NSCLC</term><def><p>non&#x2013;small cell lung cancer</p></def></def-item><def-item><term id="abb13">OCT</term><def><p>optical coherence tomography</p></def></def-item><def-item><term id="abb14">TME</term><def><p>tumor microenvironment</p></def></def-item><def-item><term id="abb15">XAI</term><def><p>explainable artificial intelligence</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Baltrusaitis</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ahuja</surname><given-names>C</given-names> </name><name name-style="western"><surname>Morency</surname><given-names>LP</given-names> </name></person-group><article-title>Multimodal machine learning: a survey and taxonomy</article-title><source>IEEE Trans Pattern Anal Mach Intell</source><year>2019</year><month>02</month><volume>41</volume><issue>2</issue><fpage>423</fpage><lpage>443</lpage><pub-id pub-id-type="doi">10.1109/TPAMI.2018.2798607</pub-id><pub-id pub-id-type="medline">29994351</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>A comprehensive review on synergy of multi-modal data and AI technologies in medical diagnosis</article-title><source>Bioengineering (Basel)</source><year>2024</year><month>02</month><day>25</day><volume>11</volume><issue>3</issue><fpage>219</fpage><pub-id pub-id-type="doi">10.3390/bioengineering11030219</pub-id><pub-id pub-id-type="medline">38534493</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Atrey</surname><given-names>PK</given-names> </name><name name-style="western"><surname>Hossain</surname><given-names>MA</given-names> </name><name name-style="western"><surname>El Saddik</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kankanhalli</surname><given-names>MS</given-names> </name></person-group><article-title>Multimodal fusion for multimedia analysis: a survey</article-title><source>Multimedia Systems</source><year>2010</year><month>11</month><volume>16</volume><issue>6</issue><fpage>345</fpage><lpage>379</lpage><pub-id pub-id-type="doi">10.1007/s00530-010-0182-0</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dash</surname><given-names>S</given-names> </name><name name-style="western"><surname>Shakyawar</surname><given-names>SK</given-names> </name><name name-style="western"><surname>Sharma</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kaushik</surname><given-names>S</given-names> </name></person-group><article-title>Big data in healthcare: management, analysis and future prospects</article-title><source>J Big Data</source><year>2019</year><month>12</month><volume>6</volume><issue>1</issue><fpage>54</fpage><pub-id pub-id-type="doi">10.1186/s40537-019-0217-0</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>AP</given-names> </name><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>AI for science: predicting infectious diseases</article-title><source>Journal of Safety Science and Resilience</source><year>2024</year><month>06</month><volume>5</volume><issue>2</issue><fpage>130</fpage><lpage>146</lpage><pub-id pub-id-type="doi">10.1016/j.jnlssr.2024.02.002</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pinto-Coelho</surname><given-names>L</given-names> </name></person-group><article-title>How artificial intelligence is shaping medical imaging technology: a survey of innovations and applications</article-title><source>Bioengineering (Basel)</source><year>2023</year><month>12</month><day>18</day><volume>10</volume><issue>12</issue><fpage>1435</fpage><pub-id pub-id-type="doi">10.3390/bioengineering10121435</pub-id><pub-id pub-id-type="medline">38136026</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Acosta</surname><given-names>JN</given-names> </name><name name-style="western"><surname>Falcone</surname><given-names>GJ</given-names> </name><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>Multimodal biomedical AI</article-title><source>Nat Med</source><year>2022</year><month>09</month><volume>28</volume><issue>9</issue><fpage>1773</fpage><lpage>1784</lpage><pub-id pub-id-type="doi">10.1038/s41591-022-01981-2</pub-id><pub-id pub-id-type="medline">36109635</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moghadam</surname><given-names>MP</given-names> </name><name name-style="western"><surname>Moghadam</surname><given-names>ZA</given-names> </name><name name-style="western"><surname>Qazani</surname><given-names>MRC</given-names> </name><name name-style="western"><surname>P&#x0142;awiak</surname><given-names>P</given-names> </name><name name-style="western"><surname>Alizadehsani</surname><given-names>R</given-names> </name></person-group><article-title>Impact of artificial intelligence in nursing for geriatric clinical care for chronic diseases: a systematic literature review</article-title><source>IEEE Access</source><year>2024</year><volume>12</volume><fpage>122557</fpage><lpage>122587</lpage><pub-id pub-id-type="doi">10.1109/ACCESS.2024.3450970</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shaik</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tao</surname><given-names>X</given-names> </name><name name-style="western"><surname>Li</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>H</given-names> </name><name name-style="western"><surname>Vel&#x00E1;squez</surname><given-names>JD</given-names> </name></person-group><article-title>A survey of multimodal information fusion for smart healthcare: mapping the journey from data to wisdom</article-title><source>Information Fusion</source><year>2024</year><month>02</month><volume>102</volume><fpage>102040</fpage><pub-id pub-id-type="doi">10.1016/j.inffus.2023.102040</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yankeelov</surname><given-names>TE</given-names> </name><name name-style="western"><surname>Abramson</surname><given-names>RG</given-names> </name><name name-style="western"><surname>Quarles</surname><given-names>CC</given-names> </name></person-group><article-title>Quantitative multimodality imaging in cancer research and therapy</article-title><source>Nat Rev Clin Oncol</source><year>2014</year><month>11</month><volume>11</volume><issue>11</issue><fpage>670</fpage><lpage>680</lpage><pub-id pub-id-type="doi">10.1038/nrclinonc.2014.134</pub-id><pub-id pub-id-type="medline">25113842</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kristensen</surname><given-names>VN</given-names> </name><name name-style="western"><surname>Lingj&#x00E6;rde</surname><given-names>OC</given-names> </name><name name-style="western"><surname>Russnes</surname><given-names>HG</given-names> </name><name name-style="western"><surname>Vollan</surname><given-names>HKM</given-names> </name><name name-style="western"><surname>Frigessi</surname><given-names>A</given-names> </name><name name-style="western"><surname>B&#x00F8;rresen-Dale</surname><given-names>AL</given-names> </name></person-group><article-title>Principles and methods of integrative genomic analyses in cancer</article-title><source>Nat Rev Cancer</source><year>2014</year><month>05</month><volume>14</volume><issue>5</issue><fpage>299</fpage><lpage>313</lpage><pub-id pub-id-type="doi">10.1038/nrc3721</pub-id><pub-id pub-id-type="medline">24759209</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>S</given-names> </name></person-group><article-title>Tumor characterization and stratification by integrated molecular profiles reveals essential pan-cancer features</article-title><source>BMC Genomics</source><year>2015</year><month>07</month><day>7</day><volume>16</volume><issue>1</issue><fpage>503</fpage><pub-id pub-id-type="doi">10.1186/s12864-015-1687-x</pub-id><pub-id pub-id-type="medline">26148869</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jena</surname><given-names>B</given-names> </name><name name-style="western"><surname>Saxena</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nayak</surname><given-names>GK</given-names> </name><etal/></person-group><article-title>Brain tumor characterization using radiogenomics in artificial intelligence framework</article-title><source>Cancers (Basel)</source><year>2022</year><month>08</month><day>22</day><volume>14</volume><issue>16</issue><fpage>4052</fpage><pub-id pub-id-type="doi">10.3390/cancers14164052</pub-id><pub-id pub-id-type="medline">36011048</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hoffmann</surname><given-names>E</given-names> </name><name name-style="western"><surname>Masthoff</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kunz</surname><given-names>WG</given-names> </name><etal/></person-group><article-title>Multiparametric MRI for characterization of the tumour microenvironment</article-title><source>Nat Rev Clin Oncol</source><year>2024</year><month>06</month><volume>21</volume><issue>6</issue><fpage>428</fpage><lpage>448</lpage><pub-id pub-id-type="doi">10.1038/s41571-024-00891-1</pub-id><pub-id pub-id-type="medline">38641651</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yeo</surname><given-names>SK</given-names> </name><name name-style="western"><surname>Guan</surname><given-names>JL</given-names> </name></person-group><article-title>Breast cancer: multiple subtypes within a tumor?</article-title><source>Trends Cancer</source><year>2017</year><month>11</month><volume>3</volume><issue>11</issue><fpage>753</fpage><lpage>760</lpage><pub-id pub-id-type="doi">10.1016/j.trecan.2017.09.001</pub-id><pub-id pub-id-type="medline">29120751</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Messer</surname><given-names>K</given-names> </name><name name-style="western"><surname>Davies</surname><given-names>SR</given-names> </name><etal/></person-group><article-title>Research-based PAM50 signature and long-term breast cancer survival</article-title><source>Breast Cancer Res Treat</source><year>2020</year><month>01</month><volume>179</volume><issue>1</issue><fpage>197</fpage><lpage>206</lpage><pub-id pub-id-type="doi">10.1007/s10549-019-05446-y</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Parker</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Mullins</surname><given-names>M</given-names> </name><name name-style="western"><surname>Cheang</surname><given-names>MCU</given-names> </name><etal/></person-group><article-title>Supervised risk predictor of breast cancer based on intrinsic subtypes</article-title><source>J Clin Oncol</source><year>2009</year><month>03</month><day>10</day><volume>27</volume><issue>8</issue><fpage>1160</fpage><lpage>1167</lpage><pub-id pub-id-type="doi">10.1200/JCO.2008.18.1370</pub-id><pub-id pub-id-type="medline">19204204</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shergalis</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bankhead</surname><given-names>A</given-names>  <suffix>III</suffix></name><name name-style="western"><surname>Luesakul</surname><given-names>U</given-names> </name><name name-style="western"><surname>Muangsin</surname><given-names>N</given-names> </name><name name-style="western"><surname>Neamati</surname><given-names>N</given-names> </name></person-group><article-title>Current challenges and opportunities in treating glioblastoma</article-title><source>Pharmacol Rev</source><year>2018</year><month>07</month><volume>70</volume><issue>3</issue><fpage>412</fpage><lpage>445</lpage><pub-id pub-id-type="doi">10.1124/pr.117.014944</pub-id><pub-id pub-id-type="medline">29669750</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liao</surname><given-names>T</given-names> </name><name name-style="western"><surname>Pu</surname><given-names>R</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>Y</given-names> </name></person-group><article-title>A hybrid deep learning model for predicting molecular subtypes of human breast cancer using multimodal data</article-title><source>IRBM</source><year>2022</year><month>02</month><volume>43</volume><issue>1</issue><fpage>62</fpage><lpage>74</lpage><pub-id pub-id-type="doi">10.1016/j.irbm.2020.12.002</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Duroux</surname><given-names>D</given-names> </name><name name-style="western"><surname>Wohlfart</surname><given-names>C</given-names> </name><name name-style="western"><surname>Van Steen</surname><given-names>K</given-names> </name><name name-style="western"><surname>Vladimirova</surname><given-names>A</given-names> </name><name name-style="western"><surname>King</surname><given-names>M</given-names> </name></person-group><article-title>Graph-based multi-modality integration for prediction of cancer subtype and severity</article-title><source>Sci Rep</source><year>2023</year><month>11</month><day>10</day><volume>13</volume><issue>1</issue><fpage>19653</fpage><pub-id pub-id-type="doi">10.1038/s41598-023-46392-6</pub-id><pub-id pub-id-type="medline">37949935</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ding</surname><given-names>S</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ying</surname><given-names>S</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>J</given-names> </name></person-group><article-title>Multimodal co-attention fusion network with online data augmentation for cancer subtype classification</article-title><source>IEEE Trans Med Imaging</source><year>2024</year><month>11</month><volume>43</volume><issue>11</issue><fpage>3977</fpage><lpage>3989</lpage><pub-id pub-id-type="doi">10.1109/TMI.2024.3405535</pub-id><pub-id pub-id-type="medline">38801690</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>B</given-names> </name><name name-style="western"><surname>Nabavi</surname><given-names>S</given-names> </name></person-group><article-title>A multimodal graph neural network framework for cancer molecular subtype classification</article-title><source>BMC Bioinformatics</source><year>2024</year><month>01</month><day>15</day><volume>25</volume><issue>1</issue><fpage>27</fpage><pub-id pub-id-type="doi">10.1186/s12859-023-05622-4</pub-id><pub-id pub-id-type="medline">38225583</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Anderson</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Simon</surname><given-names>MC</given-names> </name></person-group><article-title>The tumor microenvironment</article-title><source>Curr Biol</source><year>2020</year><month>08</month><day>17</day><volume>30</volume><issue>16</issue><fpage>R921</fpage><lpage>R925</lpage><pub-id pub-id-type="doi">10.1016/j.cub.2020.06.081</pub-id><pub-id pub-id-type="medline">32810447</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Baghban</surname><given-names>R</given-names> </name><name name-style="western"><surname>Roshangar</surname><given-names>L</given-names> </name><name name-style="western"><surname>Jahanban-Esfahlan</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Tumor microenvironment complexity and therapeutic implications at a glance</article-title><source>Cell Commun Signal</source><year>2020</year><month>04</month><day>7</day><volume>18</volume><issue>1</issue><fpage>59</fpage><pub-id pub-id-type="doi">10.1186/s12964-020-0530-4</pub-id><pub-id pub-id-type="medline">32264958</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Walsh</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Quail</surname><given-names>DF</given-names> </name></person-group><article-title>Decoding the tumor microenvironment with spatial technologies</article-title><source>Nat Immunol</source><year>2023</year><month>12</month><volume>24</volume><issue>12</issue><fpage>1982</fpage><lpage>1993</lpage><pub-id pub-id-type="doi">10.1038/s41590-023-01678-9</pub-id><pub-id pub-id-type="medline">38012408</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sch&#x00FC;rch</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Bhate</surname><given-names>SS</given-names> </name><name name-style="western"><surname>Barlow</surname><given-names>GL</given-names> </name><etal/></person-group><article-title>Coordinated cellular neighborhoods orchestrate antitumoral immunity at the colorectal cancer invasive front</article-title><source>Cell</source><year>2020</year><month>09</month><day>3</day><volume>182</volume><issue>5</issue><fpage>1341</fpage><lpage>1359</lpage><pub-id pub-id-type="doi">10.1016/j.cell.2020.07.005</pub-id><pub-id pub-id-type="medline">32763154</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>C</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Spatially resolved multi-omics highlights cell-specific metabolic remodeling and interactions in gastric cancer</article-title><source>Nat Commun</source><year>2023</year><month>05</month><day>10</day><volume>14</volume><issue>1</issue><fpage>37164975</fpage><pub-id pub-id-type="doi">10.1038/s41467-023-38360-5</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hao</surname><given-names>L</given-names> </name><name name-style="western"><surname>Rohani</surname><given-names>N</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>RT</given-names> </name><etal/></person-group><article-title>Microenvironment-triggered multimodal precision diagnostics</article-title><source>Nat Mater</source><year>2021</year><month>10</month><volume>20</volume><issue>10</issue><fpage>1440</fpage><lpage>1448</lpage><pub-id pub-id-type="doi">10.1038/s41563-021-01042-y</pub-id><pub-id pub-id-type="medline">34267368</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lapuente-Santana</surname><given-names>&#x00D3;</given-names> </name><name name-style="western"><surname>Sturm</surname><given-names>G</given-names> </name><name name-style="western"><surname>Kant</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Multimodal analysis unveils tumor microenvironment heterogeneity linked to immune activity and evasion</article-title><source>iScience</source><year>2024</year><month>08</month><day>16</day><volume>27</volume><issue>8</issue><fpage>110529</fpage><pub-id pub-id-type="doi">10.1016/j.isci.2024.110529</pub-id><pub-id pub-id-type="medline">39161957</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ji</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Rubin</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Thrane</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Multimodal analysis of composition and spatial architecture in human squamous cell carcinoma</article-title><source>Cell</source><year>2020</year><month>07</month><day>23</day><volume>182</volume><issue>2</issue><fpage>497</fpage><lpage>514</lpage><pub-id pub-id-type="doi">10.1016/j.cell.2020.05.039</pub-id><pub-id pub-id-type="medline">32579974</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arora</surname><given-names>R</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Spatial transcriptomics reveals distinct and conserved tumor core and edge architectures that predict survival and targeted therapy response</article-title><source>Nat Commun</source><year>2023</year><month>08</month><day>18</day><volume>14</volume><issue>1</issue><fpage>37596273</fpage><pub-id pub-id-type="doi">10.1038/s41467-023-40271-4</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>B</given-names> </name><name name-style="western"><surname>Bergenstr&#x00E5;hle</surname><given-names>L</given-names> </name><name name-style="western"><surname>Stenbeck</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Integrating spatial gene expression and breast tumour morphology via deep learning</article-title><source>Nat Biomed Eng</source><year>2020</year><month>08</month><volume>4</volume><issue>8</issue><fpage>827</fpage><lpage>834</lpage><pub-id pub-id-type="doi">10.1038/s41551-020-0578-x</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Monjo</surname><given-names>T</given-names> </name><name name-style="western"><surname>Koido</surname><given-names>M</given-names> </name><name name-style="western"><surname>Nagasawa</surname><given-names>S</given-names> </name><name name-style="western"><surname>Suzuki</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Kamatani</surname><given-names>Y</given-names> </name></person-group><article-title>Efficient prediction of a spatial transcriptomics profile better characterizes breast cancer tissue sections without costly experimentation</article-title><source>Sci Rep</source><year>2022</year><month>03</month><day>8</day><volume>12</volume><issue>1</issue><fpage>35260632</fpage><pub-id pub-id-type="doi">10.1038/s41598-022-07685-4</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Diao</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>JK</given-names> </name><name name-style="western"><surname>Chui</surname><given-names>WF</given-names> </name><etal/></person-group><article-title>Human-interpretable image features derived from densely mapped cancer pathology slides predict diverse molecular phenotypes</article-title><source>Nat Commun</source><year>2021</year><month>03</month><day>12</day><volume>12</volume><issue>1</issue><fpage>1613</fpage><pub-id pub-id-type="doi">10.1038/s41467-021-21896-9</pub-id><pub-id pub-id-type="medline">33712588</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lipkova</surname><given-names>J</given-names> </name><name name-style="western"><surname>Angelikopoulos</surname><given-names>P</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Personalized radiotherapy design for glioblastoma: integrating mathematical tumor models, multimodal scans, and Bayesian inference</article-title><source>IEEE Trans Med Imaging</source><year>2019</year><month>08</month><volume>38</volume><issue>8</issue><fpage>1875</fpage><lpage>1884</lpage><pub-id pub-id-type="doi">10.1109/TMI.2019.2902044</pub-id><pub-id pub-id-type="medline">30835219</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Breen</surname><given-names>WG</given-names> </name><name name-style="western"><surname>Aryal</surname><given-names>MP</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>MM</given-names> </name></person-group><article-title>Integrating multi-modal imaging in radiation treatments for glioblastoma</article-title><source>Neuro-oncology</source><year>2024</year><month>03</month><day>4</day><volume>26</volume><issue>Supplement_1</issue><fpage>S17</fpage><lpage>S25</lpage><pub-id pub-id-type="doi">10.1093/neuonc/noad187</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>X</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>C</given-names> </name></person-group><article-title>Immune checkpoint signaling and cancer immunotherapy</article-title><source>Cell Res</source><year>2020</year><month>08</month><volume>30</volume><issue>8</issue><fpage>660</fpage><lpage>669</lpage><pub-id pub-id-type="doi">10.1038/s41422-020-0343-4</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vokes</surname><given-names>EE</given-names> </name><name name-style="western"><surname>Ready</surname><given-names>N</given-names> </name><name name-style="western"><surname>Felip</surname><given-names>E</given-names> </name><etal/></person-group><article-title>Nivolumab versus docetaxel in previously treated advanced non-small-cell lung cancer (CheckMate 017 and CheckMate 057): 3-year update and outcomes in patients with liver metastases</article-title><source>Ann Oncol</source><year>2018</year><month>04</month><day>1</day><volume>29</volume><issue>4</issue><fpage>959</fpage><lpage>965</lpage><pub-id pub-id-type="doi">10.1093/annonc/mdy041</pub-id><pub-id pub-id-type="medline">29408986</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roelofsen</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Kaptein</surname><given-names>P</given-names> </name><name name-style="western"><surname>Thommen</surname><given-names>DS</given-names> </name></person-group><article-title>Multimodal predictors for precision immunotherapy</article-title><source>Immuno-Oncology and Technology</source><year>2022</year><month>06</month><volume>14</volume><issue>100071</issue><fpage>100071</fpage><pub-id pub-id-type="doi">10.1016/j.iotech.2022.100071</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vanguri</surname><given-names>RS</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>J</given-names> </name><name name-style="western"><surname>Aukerman</surname><given-names>AT</given-names> </name><etal/></person-group><article-title>Multimodal integration of radiology, pathology and genomics for prediction of response to PD-(L)1 blockade in patients with non-small cell lung cancer</article-title><source>Nat Cancer</source><year>2022</year><month>10</month><volume>3</volume><issue>10</issue><fpage>1151</fpage><lpage>1164</lpage><pub-id pub-id-type="doi">10.1038/s43018-022-00416-8</pub-id><pub-id pub-id-type="medline">36038778</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Predicting gastric cancer response to anti-HER2 therapy or anti-HER2 combined immunotherapy based on multi-modal data</article-title><source>Signal Transduct Target Ther</source><year>2024</year><month>08</month><day>26</day><volume>9</volume><issue>1</issue><fpage>222</fpage><pub-id pub-id-type="doi">10.1038/s41392-024-01932-y</pub-id><pub-id pub-id-type="medline">39183247</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yousefi</surname><given-names>B</given-names> </name><name name-style="western"><surname>LaRiviere</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Cohen</surname><given-names>EA</given-names> </name><etal/></person-group><article-title>Combining radiomic phenotypes of non-small cell lung cancer with liquid biopsy data may improve prediction of response to EGFR inhibitors</article-title><source>Sci Rep</source><year>2021</year><month>05</month><day>11</day><volume>11</volume><issue>1</issue><fpage>9984</fpage><pub-id pub-id-type="doi">10.1038/s41598-021-88239-y</pub-id><pub-id pub-id-type="medline">33976268</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Crosby</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bhatia</surname><given-names>S</given-names> </name><name name-style="western"><surname>Brindle</surname><given-names>KM</given-names> </name><etal/></person-group><article-title>Early detection of cancer</article-title><source>Science</source><year>2022</year><month>03</month><day>18</day><volume>375</volume><issue>6586</issue><fpage>eaay9040</fpage><pub-id pub-id-type="doi">10.1126/science.aay9040</pub-id><pub-id pub-id-type="medline">35298272</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Crowley</surname><given-names>E</given-names> </name><name name-style="western"><surname>Di Nicolantonio</surname><given-names>F</given-names> </name><name name-style="western"><surname>Loupakis</surname><given-names>F</given-names> </name><name name-style="western"><surname>Bardelli</surname><given-names>A</given-names> </name></person-group><article-title>Liquid biopsy: monitoring cancer-genetics in the blood</article-title><source>Nat Rev Clin Oncol</source><year>2013</year><month>08</month><volume>10</volume><issue>8</issue><fpage>472</fpage><lpage>484</lpage><pub-id pub-id-type="doi">10.1038/nrclinonc.2013.110</pub-id><pub-id pub-id-type="medline">23836314</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lone</surname><given-names>SN</given-names> </name><name name-style="western"><surname>Nisar</surname><given-names>S</given-names> </name><name name-style="western"><surname>Masoodi</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Liquid biopsy: a step closer to transform diagnosis, prognosis and future of cancer treatments</article-title><source>Mol Cancer</source><year>2022</year><month>03</month><day>18</day><volume>21</volume><issue>1</issue><fpage>79</fpage><pub-id pub-id-type="doi">10.1186/s12943-022-01543-7</pub-id><pub-id pub-id-type="medline">35303879</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chabon</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Hamilton</surname><given-names>EG</given-names> </name><name name-style="western"><surname>Kurtz</surname><given-names>DM</given-names> </name><etal/></person-group><article-title>Integrating genomic features for non-invasive early lung cancer detection</article-title><source>Nature New Biol</source><year>2020</year><month>04</month><volume>580</volume><issue>7802</issue><fpage>245</fpage><lpage>251</lpage><pub-id pub-id-type="doi">10.1038/s41586-020-2140-0</pub-id><pub-id pub-id-type="medline">32269342</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pham</surname><given-names>TMQ</given-names> </name><name name-style="western"><surname>Phan</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Jasmine</surname><given-names>TX</given-names> </name><etal/></person-group><article-title>Multimodal analysis of genome-wide methylation, copy number aberrations, and end motif signatures enhances detection of early-stage breast cancer</article-title><source>Front Oncol</source><year>2023</year><volume>13</volume><issue>1127086</issue><fpage>1127086</fpage><pub-id pub-id-type="doi">10.3389/fonc.2023.1127086</pub-id><pub-id pub-id-type="medline">37223690</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bessa</surname><given-names>X</given-names> </name><name name-style="western"><surname>Vidal</surname><given-names>J</given-names> </name><name name-style="western"><surname>Balboa</surname><given-names>JC</given-names> </name><etal/></person-group><article-title>High accuracy of a blood ctDNA-based multimodal test to detect colorectal cancer</article-title><source>Ann Oncol</source><year>2023</year><month>12</month><volume>34</volume><issue>12</issue><fpage>1187</fpage><lpage>1193</lpage><pub-id pub-id-type="doi">10.1016/j.annonc.2023.09.3113</pub-id><pub-id pub-id-type="medline">37805131</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>D</given-names> </name><name name-style="western"><surname>Li</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Integration of multiomics features for blood-based early detection of colorectal cancer</article-title><source>Mol Cancer</source><year>2024</year><month>08</month><day>22</day><volume>23</volume><issue>1</issue><fpage>173</fpage><pub-id pub-id-type="doi">10.1186/s12943-024-01959-3</pub-id><pub-id pub-id-type="medline">39175001</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nguyen</surname><given-names>VTC</given-names> </name><name name-style="western"><surname>Nguyen</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Doan</surname><given-names>NNT</given-names> </name><etal/></person-group><article-title>Multimodal analysis of methylomics and fragmentomics in plasma cell-free DNA for multi-cancer early detection and localization</article-title><source>Elife</source><year>2023</year><month>10</month><day>11</day><volume>12</volume><fpage>RP89083</fpage><pub-id pub-id-type="doi">10.7554/eLife.89083</pub-id><pub-id pub-id-type="medline">37819044</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Dai</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Q</given-names> </name><etal/></person-group><article-title>Multimodal analysis of cfDNA methylomes for early detecting esophageal squamous cell carcinoma and precancerous lesions</article-title><source>Nat Commun</source><year>2024</year><month>05</month><day>2</day><volume>15</volume><issue>1</issue><fpage>38697989</fpage><pub-id pub-id-type="doi">10.1038/s41467-024-47886-1</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xiong</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>AutoCancer as an automated multimodal framework for early cancer detection</article-title><source>iScience</source><year>2024</year><month>07</month><volume>27</volume><issue>7</issue><fpage>110183</fpage><pub-id pub-id-type="doi">10.1016/j.isci.2024.110183</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moons</surname><given-names>KGM</given-names> </name><name name-style="western"><surname>Royston</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vergouwe</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Grobbee</surname><given-names>DE</given-names> </name><name name-style="western"><surname>Altman</surname><given-names>DG</given-names> </name></person-group><article-title>Prognosis and prognostic research: what, why, and how?</article-title><source>BMJ</source><year>2009</year><month>02</month><day>23</day><volume>338</volume><issue>feb23 1</issue><fpage>b375</fpage><pub-id pub-id-type="doi">10.1136/bmj.b375</pub-id><pub-id pub-id-type="medline">19237405</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gui</surname><given-names>CP</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>HW</given-names> </name><etal/></person-group><article-title>Multimodal recurrence scoring system for prediction of clear cell renal cell carcinoma outcome: a discovery and validation study</article-title><source>Lancet Digit Health</source><year>2023</year><month>08</month><volume>5</volume><issue>8</issue><fpage>e515</fpage><lpage>e524</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(23)00095-X</pub-id><pub-id pub-id-type="medline">37393162</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sujit</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>Aminu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Karpinets</surname><given-names>TV</given-names> </name><etal/></person-group><article-title>Enhancing NSCLC recurrence prediction with PET/CT habitat imaging, ctDNA, and integrative radiogenomics-blood insights</article-title><source>Nat Commun</source><year>2024</year><month>11</month><volume>15</volume><issue>1</issue><fpage>38605064</fpage><pub-id pub-id-type="doi">10.1038/s41467-024-47512-0</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hassett</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Uno</surname><given-names>H</given-names> </name><name name-style="western"><surname>Cronin</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Carroll</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Hornbrook</surname><given-names>MC</given-names> </name><name name-style="western"><surname>Ritzwoller</surname><given-names>D</given-names> </name></person-group><article-title>Detecting lung and colorectal cancer recurrence using structured clinical/administrative data to enable outcomes research and population health management</article-title><source>Med Care</source><year>2017</year><month>12</month><volume>55</volume><issue>12</issue><fpage>e88</fpage><lpage>e98</lpage><pub-id pub-id-type="doi">10.1097/MLR.0000000000000404</pub-id><pub-id pub-id-type="medline">29135771</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Steyaert</surname><given-names>S</given-names> </name><name name-style="western"><surname>Qiu</surname><given-names>YL</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Mukherjee</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vogel</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gevaert</surname><given-names>O</given-names> </name></person-group><article-title>Multimodal deep learning to predict prognosis in adult and pediatric brain tumors</article-title><source>Commun Med (Lond)</source><year>2023</year><month>03</month><day>29</day><volume>3</volume><issue>1</issue><fpage>44</fpage><pub-id pub-id-type="doi">10.1038/s43856-023-00276-y</pub-id><pub-id pub-id-type="medline">36991216</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Guo</surname><given-names>W</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Deng</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Zou</surname><given-names>X</given-names> </name></person-group><article-title>A multimodal affinity fusion network for predicting the survival of breast cancer patients</article-title><source>Front Genet</source><year>2021</year><volume>12</volume><issue>709027</issue><fpage>709027</fpage><pub-id pub-id-type="doi">10.3389/fgene.2021.709027</pub-id><pub-id pub-id-type="medline">34490038</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schulz</surname><given-names>S</given-names> </name><name name-style="western"><surname>Woerl</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Jungmann</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Multimodal deep learning for prognosis prediction in renal cancer</article-title><source>Front Oncol</source><year>2021</year><volume>11</volume><issue>788740</issue><fpage>788740</fpage><pub-id pub-id-type="doi">10.3389/fonc.2021.788740</pub-id><pub-id pub-id-type="medline">34900744</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cheerla</surname><given-names>A</given-names> </name><name name-style="western"><surname>Gevaert</surname><given-names>O</given-names> </name></person-group><article-title>Deep learning with multimodal representation for pancancer prognosis prediction</article-title><source>Bioinformatics</source><year>2019</year><month>07</month><day>15</day><volume>35</volume><issue>14</issue><fpage>i446</fpage><lpage>i454</lpage><pub-id pub-id-type="doi">10.1093/bioinformatics/btz342</pub-id><pub-id pub-id-type="medline">31510656</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Dong</surname><given-names>S</given-names> </name></person-group><article-title>A multi-modal fusion framework based on multi-task correlation learning for cancer prognosis prediction</article-title><source>Artif Intell Med</source><year>2022</year><month>04</month><volume>126</volume><issue>102260</issue><fpage>102260</fpage><pub-id pub-id-type="doi">10.1016/j.artmed.2022.102260</pub-id><pub-id pub-id-type="medline">35346442</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Saleh</surname><given-names>GA</given-names> </name><name name-style="western"><surname>Batouty</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Haggag</surname><given-names>S</given-names> </name><etal/></person-group><article-title>The role of medical image modalities and AI in the early detection, diagnosis and grading of retinal diseases: a survey</article-title><source>Bioengineering (Basel)</source><year>2022</year><month>08</month><day>4</day><volume>9</volume><issue>8</issue><fpage>366</fpage><pub-id pub-id-type="doi">10.3390/bioengineering9080366</pub-id><pub-id pub-id-type="medline">36004891</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><name name-style="western"><surname>He</surname><given-names>X</given-names> </name><name name-style="western"><surname>Jian</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Advances and prospects of multi-modal ophthalmic artificial intelligence based on deep learning: a review</article-title><source>Eye Vis (Lond)</source><year>2024</year><month>10</month><day>1</day><volume>11</volume><issue>1</issue><fpage>38</fpage><pub-id pub-id-type="doi">10.1186/s40662-024-00405-1</pub-id><pub-id pub-id-type="medline">39350240</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mehta</surname><given-names>P</given-names> </name><name name-style="western"><surname>Petersen</surname><given-names>CA</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>JC</given-names> </name><etal/></person-group><article-title>Automated detection of glaucoma with interpretable machine learning using clinical data and multimodal retinal images</article-title><source>Am J Ophthalmol</source><year>2021</year><month>11</month><volume>231</volume><fpage>154</fpage><lpage>169</lpage><pub-id pub-id-type="doi">10.1016/j.ajo.2021.04.021</pub-id><pub-id pub-id-type="medline">33945818</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xiong</surname><given-names>J</given-names> </name><name name-style="western"><surname>Li</surname><given-names>F</given-names> </name><name name-style="western"><surname>Song</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Multimodal machine learning using visual fields and peripapillary circular OCT scans in detection of glaucomatous optic neuropathy</article-title><source>Ophthalmology</source><year>2022</year><month>02</month><volume>129</volume><issue>2</issue><fpage>171</fpage><lpage>180</lpage><pub-id pub-id-type="doi">10.1016/j.ophtha.2021.07.032</pub-id><pub-id pub-id-type="medline">34339778</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Li</surname><given-names>F</given-names> </name><etal/></person-group><article-title>GAMMA challenge: Glaucoma grAding from Multi-Modality imAges</article-title><source>Med Image Anal</source><year>2023</year><month>12</month><volume>90</volume><issue>102938</issue><fpage>102938</fpage><pub-id pub-id-type="doi">10.1016/j.media.2023.102938</pub-id><pub-id pub-id-type="medline">37806020</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>G</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Ding</surname><given-names>D</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Zhao</surname><given-names>J</given-names> </name></person-group><article-title>Representation, alignment, fusion: a generic transformer-based framework for multi-modal glaucoma recognition</article-title><conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name><conf-date>Oct 1, 2023</conf-date><conf-loc>Vancouver Convention Centre, Canada</conf-loc><publisher-name>Springer</publisher-name><fpage>704</fpage><lpage>713</lpage><pub-id pub-id-type="doi">10.1007/978-3-031-43990-2_66</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>W</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>He</surname><given-names>F</given-names> </name></person-group><article-title>Two-stream CNN with loose pair training for multi-modal AMD categorization</article-title><conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name><conf-date>Oct 10, 2019</conf-date><conf-loc>Shenzhen, China</conf-loc><pub-id pub-id-type="doi">10.1007/978-3-030-32239-7_18</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vaghefi</surname><given-names>E</given-names> </name><name name-style="western"><surname>Hill</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kersten</surname><given-names>HM</given-names> </name><name name-style="western"><surname>Squirrell</surname><given-names>D</given-names> </name></person-group><article-title>Multimodal retinal image analysis via deep learning for the diagnosis of intermediate dry age-related macular degeneration: a feasibility study</article-title><source>J Ophthalmol</source><year>2020</year><volume>2020</volume><issue>7493419</issue><fpage>7493419</fpage><pub-id pub-id-type="doi">10.1155/2020/7493419</pub-id><pub-id pub-id-type="medline">32411434</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Automated diagnoses of age-related macular degeneration and polypoidal choroidal vasculopathy using bi-modal deep convolutional neural networks</article-title><source>Br J Ophthalmol</source><year>2021</year><month>04</month><volume>105</volume><issue>4</issue><fpage>561</fpage><lpage>566</lpage><pub-id pub-id-type="doi">10.1136/bjophthalmol-2020-315817</pub-id><pub-id pub-id-type="medline">32499330</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>MH</given-names> </name><name name-style="western"><surname>Xing</surname><given-names>L</given-names> </name><name name-style="western"><surname>Pan</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>AI-based advanced approaches and dry eye disease detection based on multi-source evidence: cases, applications, issues, and future directions</article-title><source>Big Data Min Anal</source><year>2024</year><volume>7</volume><issue>2</issue><fpage>445</fpage><lpage>484</lpage><pub-id pub-id-type="doi">10.26599/BDMA.2023.9020024</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>X</given-names> </name><name name-style="western"><surname>Deng</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Fang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>Q</given-names> </name></person-group><article-title>Multi-modal retinal image classification with modality-specific attention network</article-title><source>IEEE Trans Med Imaging</source><year>2021</year><month>06</month><volume>40</volume><issue>6</issue><fpage>1591</fpage><lpage>1602</lpage><pub-id pub-id-type="doi">10.1109/TMI.2021.3059956</pub-id><pub-id pub-id-type="medline">33625978</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hervella</surname><given-names>&#x00C1;S</given-names> </name><name name-style="western"><surname>Rouco</surname><given-names>J</given-names> </name><name name-style="western"><surname>Novo</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ortega</surname><given-names>M</given-names> </name></person-group><article-title>Multimodal image encoding pre-training for diabetic retinopathy grading</article-title><source>Comput Biol Med</source><year>2022</year><month>04</month><volume>143</volume><fpage>105302</fpage><pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.105302</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Atse</surname><given-names>YC</given-names> </name><name name-style="western"><surname>Le Boit&#x00E9;</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bonnin</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cosette</surname><given-names>D</given-names> </name><name name-style="western"><surname>Deman</surname><given-names>P</given-names> </name><name name-style="western"><surname>Borderie</surname><given-names>L</given-names> </name></person-group><article-title>Improved automatic diabetic retinopathy severity classification using deep multimodal fusion of UWF-CFP and OCTA images</article-title><conf-name>Ophthalmic Medical Image Analysis: 10th International Workshop, OMIA 2023, Held in Conjunction with MICCAI 2023</conf-name><conf-date>Oct 12, 2023</conf-date><conf-loc>Vancouver, BC, Canada</conf-loc></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Shang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Identification of diabetic retinopathy classification using machine learning algorithms on clinical data and optical coherence tomography angiography</article-title><source>Eye (Lond)</source><year>2024</year><month>10</month><volume>38</volume><issue>14</issue><fpage>2813</fpage><lpage>2821</lpage><pub-id pub-id-type="doi">10.1038/s41433-024-03173-3</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Mao</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Li</surname><given-names>B</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Bi-modal deep learning for recognizing multiple retinal diseases based on color fundus photos and OCT images</article-title><source>Invest Ophthalmol Vis Sci</source><year>2021</year><access-date>2025-08-14</access-date><volume>62</volume><issue>8</issue><comment><ext-link ext-link-type="uri" xlink:href="https://iovs.arvojournals.org/article.aspx?articleid=2773464">https://iovs.arvojournals.org/article.aspx?articleid=2773464</ext-link></comment></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Peng</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>R</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Development and evaluation of multimodal AI for diagnosis and triage of ophthalmic diseases using ChatGPT and anterior segment images: protocol for a two-stage cross-sectional study</article-title><source>Front Artif Intell</source><year>2023</year><volume>6</volume><issue>1323924</issue><fpage>1323924</fpage><pub-id pub-id-type="doi">10.3389/frai.2023.1323924</pub-id><pub-id pub-id-type="medline">38145231</pub-id></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flammer</surname><given-names>J</given-names> </name><name name-style="western"><surname>Konieczka</surname><given-names>K</given-names> </name><name name-style="western"><surname>Bruno</surname><given-names>RM</given-names> </name><name name-style="western"><surname>Virdis</surname><given-names>A</given-names> </name><name name-style="western"><surname>Flammer</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Taddei</surname><given-names>S</given-names> </name></person-group><article-title>The eye and the heart</article-title><source>Eur Heart J</source><year>2013</year><month>05</month><volume>34</volume><issue>17</issue><fpage>1270</fpage><lpage>1278</lpage><pub-id pub-id-type="doi">10.1093/eurheartj/eht023</pub-id><pub-id pub-id-type="medline">23401492</pub-id></nlm-citation></ref><ref id="ref79"><label>79</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Allon</surname><given-names>R</given-names> </name><name name-style="western"><surname>Aronov</surname><given-names>M</given-names> </name><name name-style="western"><surname>Belkin</surname><given-names>M</given-names> </name><name name-style="western"><surname>Maor</surname><given-names>E</given-names> </name><name name-style="western"><surname>Shechter</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fabian</surname><given-names>ID</given-names> </name></person-group><article-title>Retinal microvascular signs as screening and prognostic factors for cardiac disease: a systematic review of current evidence</article-title><source>Am J Med</source><year>2021</year><month>01</month><volume>134</volume><issue>1</issue><fpage>36</fpage><lpage>47</lpage><pub-id pub-id-type="doi">10.1016/j.amjmed.2020.07.013</pub-id><pub-id pub-id-type="medline">32861624</pub-id></nlm-citation></ref><ref id="ref80"><label>80</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chua</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chin</surname><given-names>CWL</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Impact of hypertension on retinal capillary microvasculature using optical coherence tomographic angiography</article-title><source>J Hypertens</source><year>2019</year><month>03</month><volume>37</volume><issue>3</issue><fpage>572</fpage><lpage>580</lpage><pub-id pub-id-type="doi">10.1097/HJH.0000000000001916</pub-id><pub-id pub-id-type="medline">30113530</pub-id></nlm-citation></ref><ref id="ref81"><label>81</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Al-Absi</surname><given-names>HRH</given-names> </name><name name-style="western"><surname>Islam</surname><given-names>MT</given-names> </name><name name-style="western"><surname>Refaee</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Chowdhury</surname><given-names>MEH</given-names> </name><name name-style="western"><surname>Alam</surname><given-names>T</given-names> </name></person-group><article-title>Cardiovascular disease diagnosis from DXA scan and retinal images using deep learning</article-title><source>Sensors (Basel)</source><year>2022</year><month>06</month><day>7</day><volume>22</volume><issue>12</issue><fpage>4310</fpage><pub-id pub-id-type="doi">10.3390/s22124310</pub-id><pub-id pub-id-type="medline">35746092</pub-id></nlm-citation></ref><ref id="ref82"><label>82</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>YC</given-names> </name><name name-style="western"><surname>Cha</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shim</surname><given-names>I</given-names> </name><etal/></person-group><article-title>Multimodal deep learning of fundus abnormalities and traditional risk factors for cardiovascular risk prediction</article-title><source>NPJ Digit Med</source><year>2023</year><month>02</month><volume>6</volume><issue>1</issue><fpage>36732671</fpage><pub-id pub-id-type="doi">10.1038/s41746-023-00748-4</pub-id></nlm-citation></ref><ref id="ref83"><label>83</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sedlakova</surname><given-names>J</given-names> </name><name name-style="western"><surname>Daniore</surname><given-names>P</given-names> </name><name name-style="western"><surname>Horn Wintsch</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Challenges and best practices for digital unstructured data enrichment in health research: a systematic narrative review</article-title><source>PLOS Digit Health</source><year>2023</year><month>10</month><volume>2</volume><issue>10</issue><fpage>e0000347</fpage><pub-id pub-id-type="doi">10.1371/journal.pdig.0000347</pub-id><pub-id pub-id-type="medline">37819910</pub-id></nlm-citation></ref><ref id="ref84"><label>84</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flores</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Claborne</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Weller</surname><given-names>ZD</given-names> </name><name name-style="western"><surname>Webb-Robertson</surname><given-names>BJM</given-names> </name><name name-style="western"><surname>Waters</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Bramer</surname><given-names>LM</given-names> </name></person-group><article-title>Missing data in multi-omics integration: recent advances through artificial intelligence</article-title><source>Front Artif Intell</source><year>2023</year><volume>6</volume><issue>1098308</issue><fpage>1098308</fpage><pub-id pub-id-type="doi">10.3389/frai.2023.1098308</pub-id><pub-id pub-id-type="medline">36844425</pub-id></nlm-citation></ref><ref id="ref85"><label>85</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Theodos</surname><given-names>K</given-names> </name><name name-style="western"><surname>Sittig</surname><given-names>S</given-names> </name></person-group><article-title>Health information privacy laws in the digital age: HIPAA doesn&#x2019;t apply</article-title><source>Perspect Health Inf Manag</source><year>2021</year><volume>18</volume><issue>Winter</issue><fpage>1l</fpage><pub-id pub-id-type="medline">33633522</pub-id></nlm-citation></ref><ref id="ref86"><label>86</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schwartz</surname><given-names>PH</given-names> </name><name name-style="western"><surname>Caine</surname><given-names>K</given-names> </name><name name-style="western"><surname>Alpert</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Meslin</surname><given-names>EM</given-names> </name><name name-style="western"><surname>Carroll</surname><given-names>AE</given-names> </name><name name-style="western"><surname>Tierney</surname><given-names>WM</given-names> </name></person-group><article-title>Patient preferences in controlling access to their electronic health records: a prospective cohort study in primary care</article-title><source>J Gen Intern Med</source><year>2015</year><month>01</month><volume>30 Suppl 1</volume><issue>Suppl 1</issue><fpage>S25</fpage><lpage>30</lpage><pub-id pub-id-type="doi">10.1007/s11606-014-3054-z</pub-id><pub-id pub-id-type="medline">25480721</pub-id></nlm-citation></ref><ref id="ref87"><label>87</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Amal</surname><given-names>S</given-names> </name><name name-style="western"><surname>Safarnejad</surname><given-names>L</given-names> </name><name name-style="western"><surname>Omiye</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Ghanzouri</surname><given-names>I</given-names> </name><name name-style="western"><surname>Cabot</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Ross</surname><given-names>EG</given-names> </name></person-group><article-title>Use of multi-modal data and machine learning to improve cardiovascular disease care</article-title><source>Front Cardiovasc Med</source><year>2022</year><volume>9</volume><issue>840262</issue><fpage>840262</fpage><pub-id pub-id-type="doi">10.3389/fcvm.2022.840262</pub-id><pub-id pub-id-type="medline">35571171</pub-id></nlm-citation></ref><ref id="ref88"><label>88</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mittelstadt</surname><given-names>BD</given-names> </name><name name-style="western"><surname>Floridi</surname><given-names>L</given-names> </name></person-group><article-title>The ethics of big data: current and foreseeable issues in biomedical contexts</article-title><source>Sci Eng Ethics</source><year>2016</year><month>04</month><volume>22</volume><issue>2</issue><fpage>303</fpage><lpage>341</lpage><pub-id pub-id-type="doi">10.1007/s11948-015-9652-2</pub-id><pub-id pub-id-type="medline">26002496</pub-id></nlm-citation></ref><ref id="ref89"><label>89</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choudhury</surname><given-names>S</given-names> </name><name name-style="western"><surname>Fishman</surname><given-names>JR</given-names> </name><name name-style="western"><surname>McGowan</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Juengst</surname><given-names>ET</given-names> </name></person-group><article-title>Big data, open science and the brain: lessons learned from genomics</article-title><source>Front Hum Neurosci</source><year>2014</year><volume>8</volume><issue>239</issue><fpage>24904347</fpage><pub-id pub-id-type="doi">10.3389/fnhum.2014.00239</pub-id></nlm-citation></ref><ref id="ref90"><label>90</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shojaei</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vlahu-Gjorgievska</surname><given-names>E</given-names> </name><name name-style="western"><surname>Chow</surname><given-names>YW</given-names> </name></person-group><article-title>Security and privacy of technologies in health information systems: a systematic literature review</article-title><source>Computers</source><year>2024</year><volume>13</volume><issue>2</issue><fpage>41</fpage><pub-id pub-id-type="doi">10.3390/computers13020041</pub-id></nlm-citation></ref><ref id="ref91"><label>91</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelly</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Osorio-Marin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kothari</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hague</surname><given-names>S</given-names> </name><name name-style="western"><surname>Dever</surname><given-names>JK</given-names> </name></person-group><article-title>Genetic improvement in cotton fiber elongation can impact yarn quality</article-title><source>Ind Crops Prod</source><year>2019</year><month>03</month><volume>129</volume><fpage>1</fpage><lpage>9</lpage><pub-id pub-id-type="doi">10.1016/j.indcrop.2018.11.066</pub-id></nlm-citation></ref><ref id="ref92"><label>92</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Greenhalgh</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wherton</surname><given-names>J</given-names> </name><name name-style="western"><surname>Papoutsi</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Beyond adoption: a new framework for theorizing and evaluating nonadoption, abandonment, and challenges to the scale-up, spread, and sustainability of health and care technologies</article-title><source>J Med Internet Res</source><year>2017</year><month>11</month><day>1</day><volume>19</volume><issue>11</issue><fpage>e367</fpage><pub-id pub-id-type="doi">10.2196/jmir.8775</pub-id><pub-id pub-id-type="medline">29092808</pub-id></nlm-citation></ref><ref id="ref93"><label>93</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ahmed</surname><given-names>SF</given-names> </name><name name-style="western"><surname>Alam</surname><given-names>MdSB</given-names> </name><name name-style="western"><surname>Hassan</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Deep learning modelling techniques: current progress, applications, advantages, and challenges</article-title><source>Artif Intell Rev</source><year>2023</year><month>11</month><volume>56</volume><issue>11</issue><fpage>13521</fpage><lpage>13617</lpage><pub-id pub-id-type="doi">10.1007/s10462-023-10466-8</pub-id><pub-id pub-id-type="medline">37362885</pub-id></nlm-citation></ref><ref id="ref94"><label>94</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bornstein</surname><given-names>S</given-names> </name></person-group><article-title>Antidiscriminatory algorithms</article-title><source>Ala L Rev</source><year>2018</year><access-date>2025-08-14</access-date><volume>70</volume><issue>2</issue><fpage>519</fpage><comment><ext-link ext-link-type="uri" xlink:href="https://law.ua.edu/wp-content/uploads/2018/12/4-Bornstein-518-572.pdf">https://law.ua.edu/wp-content/uploads/2018/12/4-Bornstein-518-572.pdf</ext-link></comment></nlm-citation></ref><ref id="ref95"><label>95</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Miasato</surname><given-names>A</given-names> </name><name name-style="western"><surname>Reis Silva</surname><given-names>F</given-names> </name></person-group><article-title>Artificial intelligence as an instrument of discrimination in workforce recruitment</article-title><source>AUSLEG</source><year>2020</year><month>01</month><day>15</day><access-date>2025-08-14</access-date><volume>8</volume><issue>2</issue><fpage>191</fpage><lpage>212</lpage><comment><ext-link ext-link-type="uri" xlink:href="http://acta.sapientia.ro/acta-legal/legal-main.htm">http://acta.sapientia.ro/acta-legal/legal-main.htm</ext-link></comment><pub-id pub-id-type="doi">10.47745/AUSLEG.2019.8.2.04</pub-id></nlm-citation></ref><ref id="ref96"><label>96</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Madan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Henry</surname><given-names>T</given-names> </name><name name-style="western"><surname>Dozier</surname><given-names>J</given-names> </name><etal/></person-group><article-title>When and how convolutional neural networks generalize to out-of-distribution category&#x2013;viewpoint combinations</article-title><source>Nat Mach Intell</source><year>2022</year><volume>4</volume><issue>2</issue><fpage>146</fpage><lpage>153</lpage><pub-id pub-id-type="doi">10.1038/s42256-021-00437-5</pub-id></nlm-citation></ref><ref id="ref97"><label>97</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sadeghi</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Alizadehsani</surname><given-names>R</given-names> </name><name name-style="western"><surname>Cifci</surname><given-names>MA</given-names> </name><etal/></person-group><article-title>A review of explainable artificial intelligence in healthcare</article-title><source>Computers and Electrical Engineering</source><year>2024</year><month>08</month><volume>118</volume><fpage>109370</fpage><pub-id pub-id-type="doi">10.1016/j.compeleceng.2024.109370</pub-id></nlm-citation></ref><ref id="ref98"><label>98</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Calaon</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tosello</surname><given-names>G</given-names> </name></person-group><article-title>Integration of multimodal data and explainable artificial intelligence for root cause analysis in manufacturing processes</article-title><source>CIRP Annals</source><year>2024</year><volume>73</volume><issue>1</issue><fpage>365</fpage><lpage>368</lpage><pub-id pub-id-type="doi">10.1016/j.cirp.2024.04.014</pub-id></nlm-citation></ref><ref id="ref99"><label>99</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Rodis</surname><given-names>N</given-names> </name><name name-style="western"><surname>Sardianos</surname><given-names>C</given-names> </name><name name-style="western"><surname>Radoglou-Grammatikis</surname><given-names>P</given-names> </name><name name-style="western"><surname>Sarigiannidis</surname><given-names>P</given-names> </name><name name-style="western"><surname>Varlamis</surname><given-names>I</given-names> </name><name name-style="western"><surname>Papadopoulos</surname><given-names>G</given-names> </name></person-group><article-title>Multimodal explainable artificial intelligence: a comprehensive review of methodological advances and future research directions</article-title><source>arXiv</source><pub-id pub-id-type="doi">10.1109/ACCESS.2024.3467062</pub-id></nlm-citation></ref><ref id="ref100"><label>100</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>C</given-names> </name><name name-style="western"><surname>Yuan</surname><given-names>X</given-names> </name><name name-style="western"><surname>Yan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>W</given-names> </name><etal/></person-group><article-title>From redundancy to relevance: enhancing explainability in multimodal large language models</article-title><source>arXiv</source><comment>Preprint posted online on 2024</comment></nlm-citation></ref><ref id="ref101"><label>101</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>P</given-names> </name><name name-style="western"><surname>Dong</surname><given-names>W</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Kaymak</surname><given-names>U</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>Z</given-names> </name></person-group><article-title>Interpretable clinical prediction via attention-based neural network</article-title><source>BMC Med Inform Decis Mak</source><year>2020</year><month>07</month><day>9</day><volume>20</volume><issue>Suppl 3</issue><fpage>131</fpage><pub-id pub-id-type="doi">10.1186/s12911-020-1110-7</pub-id><pub-id pub-id-type="medline">32646437</pub-id></nlm-citation></ref><ref id="ref102"><label>102</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sharma</surname><given-names>D</given-names> </name><name name-style="western"><surname>Purushotham</surname><given-names>S</given-names> </name><name name-style="western"><surname>Reddy</surname><given-names>CK</given-names> </name></person-group><article-title>MedFuseNet: an attention-based multimodal deep learning model for visual question answering in the medical domain</article-title><source>Sci Rep</source><year>2021</year><month>10</month><day>6</day><volume>11</volume><issue>1</issue><fpage>19826</fpage><pub-id pub-id-type="doi">10.1038/s41598-021-98390-1</pub-id><pub-id pub-id-type="medline">34615894</pub-id></nlm-citation></ref><ref id="ref103"><label>103</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Luo</surname><given-names>B</given-names> </name><name name-style="western"><surname>Teng</surname><given-names>F</given-names> </name><name name-style="western"><surname>Tang</surname><given-names>G</given-names> </name><etal/></person-group><article-title>StereoMM: a graph fusion model for integrating spatial transcriptomic data and pathological images</article-title><source>Brief Bioinform</source><year>2025</year><month>05</month><day>1</day><volume>26</volume><issue>3</issue><fpage>bbaf210</fpage><pub-id pub-id-type="doi">10.1093/bib/bbaf210</pub-id><pub-id pub-id-type="medline">40407386</pub-id></nlm-citation></ref><ref id="ref104"><label>104</label><nlm-citation citation-type="book"><person-group person-group-type="editor"><name name-style="western"><surname>Jain</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wallace</surname><given-names>BC</given-names> </name></person-group><source>Attention Is Not Explanation</source><year>2019</year><publisher-name>North American Chapter of the Association for Computational Linguistics</publisher-name></nlm-citation></ref><ref id="ref105"><label>105</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Niu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhong</surname><given-names>G</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>H</given-names> </name></person-group><article-title>A review on the attention mechanism of deep learning</article-title><source>Neurocomputing</source><year>2021</year><month>09</month><volume>452</volume><fpage>48</fpage><lpage>62</lpage><pub-id pub-id-type="doi">10.1016/j.neucom.2021.03.091</pub-id></nlm-citation></ref><ref id="ref106"><label>106</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Selvaraju</surname><given-names>RR</given-names> </name><name name-style="western"><surname>Cogswell</surname><given-names>M</given-names> </name><name name-style="western"><surname>Das</surname><given-names>A</given-names> </name><name name-style="western"><surname>Vedantam</surname><given-names>R</given-names> </name><name name-style="western"><surname>Parikh</surname><given-names>D</given-names> </name><name name-style="western"><surname>Batra</surname><given-names>D</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Batra</surname><given-names>D</given-names> </name></person-group><article-title>Grad-cam: visual explanations from deep networks via gradient-based localization</article-title><conf-name>2017 IEEE International Conference on Computer Vision (ICCV)</conf-name><conf-date>Sep 10, 2021</conf-date><conf-loc>Venice</conf-loc><pub-id pub-id-type="doi">10.1109/ICCV.2017.74</pub-id></nlm-citation></ref><ref id="ref107"><label>107</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>D</given-names> </name><name name-style="western"><surname>McClement</surname><given-names>D</given-names> </name><name name-style="western"><surname>Oladosu</surname><given-names>O</given-names> </name><name name-style="western"><surname>Pridham</surname><given-names>G</given-names> </name><name name-style="western"><surname>Slaney</surname><given-names>G</given-names> </name></person-group><article-title>Grad-CAM helps interpret the deep learning models trained to classify multiple sclerosis types using clinical brain magnetic resonance imaging</article-title><source>J Neurosci Methods</source><year>2021</year><month>04</month><day>1</day><volume>353</volume><issue>109098</issue><fpage>109098</fpage><pub-id pub-id-type="doi">10.1016/j.jneumeth.2021.109098</pub-id><pub-id pub-id-type="medline">33582174</pub-id></nlm-citation></ref><ref id="ref108"><label>108</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Ogasawara</surname><given-names>K</given-names> </name></person-group><article-title>Grad-CAM-based explainable artificial intelligence related to medical text processing</article-title><source>Bioengineering (Basel)</source><year>2023</year><month>09</month><day>10</day><volume>10</volume><issue>9</issue><fpage>1070</fpage><pub-id pub-id-type="doi">10.3390/bioengineering10091070</pub-id><pub-id pub-id-type="medline">37760173</pub-id></nlm-citation></ref><ref id="ref109"><label>109</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zambrano Chaves</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Wentland</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Desai</surname><given-names>AD</given-names> </name><etal/></person-group><article-title>Opportunistic assessment of ischemic heart disease risk using abdominopelvic computed tomography and medical record data: a multimodal explainable artificial intelligence approach</article-title><source>Sci Rep</source><year>2023</year><month>11</month><day>29</day><volume>13</volume><issue>1</issue><fpage>21034</fpage><pub-id pub-id-type="doi">10.1038/s41598-023-47895-y</pub-id><pub-id pub-id-type="medline">38030716</pub-id></nlm-citation></ref><ref id="ref110"><label>110</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Feng</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Learning from longitudinal data in electronic health record and genetic data to improve cardiovascular event prediction</article-title><source>Sci Rep</source><year>2019</year><month>01</month><day>24</day><volume>9</volume><issue>1</issue><fpage>30679510</fpage><pub-id pub-id-type="doi">10.1038/s41598-018-36745-x</pub-id></nlm-citation></ref><ref id="ref111"><label>111</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Detection of coronary artery disease using multi-modal feature fusion and hybrid feature selection</article-title><source>Physiol Meas</source><year>2020</year><month>11</month><day>1</day><volume>41</volume><issue>11</issue><fpage>115007</fpage><pub-id pub-id-type="doi">10.1088/1361-6579/abc323</pub-id></nlm-citation></ref><ref id="ref112"><label>112</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>von Spiczak</surname><given-names>J</given-names> </name><name name-style="western"><surname>Mannil</surname><given-names>M</given-names> </name><name name-style="western"><surname>Model</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Multimodal multiparametric three-dimensional image fusion in coronary artery disease: combining the best of two worlds</article-title><source>Radiol Cardiothorac Imaging</source><year>2020</year><month>04</month><volume>2</volume><issue>2</issue><fpage>e190116</fpage><pub-id pub-id-type="doi">10.1148/ryct.2020190116</pub-id><pub-id pub-id-type="medline">33778554</pub-id></nlm-citation></ref><ref id="ref113"><label>113</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flores</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Schuler</surname><given-names>A</given-names> </name><name name-style="western"><surname>Eberhard</surname><given-names>AV</given-names> </name><etal/></person-group><article-title>Unsupervised learning for automated detection of coronary artery disease subgroups</article-title><source>J Am Heart Assoc</source><year>2021</year><month>12</month><day>7</day><volume>10</volume><issue>23</issue><fpage>e021976</fpage><pub-id pub-id-type="doi">10.1161/JAHA.121.021976</pub-id><pub-id pub-id-type="medline">34845917</pub-id></nlm-citation></ref><ref id="ref114"><label>114</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ali</surname><given-names>F</given-names> </name><name name-style="western"><surname>El-Sappagh</surname><given-names>S</given-names> </name><name name-style="western"><surname>Islam</surname><given-names>SMR</given-names> </name><etal/></person-group><article-title>A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion</article-title><source>Information Fusion</source><year>2020</year><month>11</month><volume>63</volume><fpage>208</fpage><lpage>222</lpage><pub-id pub-id-type="doi">10.1016/j.inffus.2020.06.008</pub-id></nlm-citation></ref><ref id="ref115"><label>115</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qiu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>MI</given-names> </name><name name-style="western"><surname>Joshi</surname><given-names>PS</given-names> </name><etal/></person-group><article-title>Multimodal deep learning for Alzheimer&#x2019;s disease dementia assessment</article-title><source>Nat Commun</source><year>2022</year><month>06</month><day>20</day><volume>13</volume><issue>1</issue><fpage>35725739</fpage><pub-id pub-id-type="doi">10.1038/s41467-022-31037-5</pub-id></nlm-citation></ref><ref id="ref116"><label>116</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gabitto</surname><given-names>MI</given-names> </name><name name-style="western"><surname>Travaglini</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>Rachleff</surname><given-names>VM</given-names> </name><etal/></person-group><article-title>Integrated multimodal cell atlas of Alzheimer&#x2019;s disease</article-title><source>Res Sq</source><year>2023</year><month>05</month><day>23</day><fpage>37292694</fpage><pub-id pub-id-type="doi">10.21203/rs.3.rs-2921860/v1</pub-id><pub-id pub-id-type="medline">37292694</pub-id></nlm-citation></ref><ref id="ref117"><label>117</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Makarious</surname><given-names>MB</given-names> </name><name name-style="western"><surname>Leonard</surname><given-names>HL</given-names> </name><name name-style="western"><surname>Vitale</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Multi-modality machine learning predicting Parkinson&#x2019;s disease</article-title><source>NPJ Parkinsons Dis</source><year>2022</year><month>04</month><day>1</day><volume>8</volume><issue>1</issue><fpage>35</fpage><pub-id pub-id-type="doi">10.1038/s41531-022-00288-w</pub-id><pub-id pub-id-type="medline">35365675</pub-id></nlm-citation></ref><ref id="ref118"><label>118</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lincoln</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Bernstam</surname><given-names>EV</given-names> </name><name name-style="western"><surname>Shams</surname><given-names>S</given-names> </name></person-group><article-title>Predicting multiple sclerosis severity with multimodal deep neural networks</article-title><source>BMC Med Inform Decis Mak</source><year>2023</year><month>11</month><day>9</day><volume>23</volume><issue>1</issue><fpage>255</fpage><pub-id pub-id-type="doi">10.1186/s12911-023-02354-6</pub-id><pub-id pub-id-type="medline">37946182</pub-id></nlm-citation></ref><ref id="ref119"><label>119</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ding</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Thao</surname><given-names>PNM</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>WC</given-names> </name><etal/></person-group><article-title>Large language multimodal models for new-onset type 2 diabetes prediction using five-year cohort electronic health records</article-title><source>Sci Rep</source><year>2024</year><month>09</month><day>6</day><volume>14</volume><issue>1</issue><fpage>20774</fpage><pub-id pub-id-type="doi">10.1038/s41598-024-71020-2</pub-id><pub-id pub-id-type="medline">39237580</pub-id></nlm-citation></ref><ref id="ref120"><label>120</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bhatt</surname><given-names>RR</given-names> </name><name name-style="western"><surname>Todorov</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sood</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Integrated multi-modal brain signatures predict sex-specific obesity status</article-title><source>Brain Commun</source><year>2023</year><volume>5</volume><issue>2</issue><fpage>fcad098</fpage><pub-id pub-id-type="doi">10.1093/braincomms/fcad098</pub-id><pub-id pub-id-type="medline">37091587</pub-id></nlm-citation></ref><ref id="ref121"><label>121</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lafci</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hadjihambi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Determann</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Multimodal assessment of non-alcoholic fatty liver disease with transmission-reflection optoacoustic ultrasound</article-title><source>Theranostics</source><year>2023</year><volume>13</volume><issue>12</issue><fpage>4217</fpage><lpage>4228</lpage><pub-id pub-id-type="doi">10.7150/thno.78548</pub-id><pub-id pub-id-type="medline">37554280</pub-id></nlm-citation></ref><ref id="ref122"><label>122</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Pan</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>A deep learning model for classification of parotid neoplasms based on multimodal magnetic resonance image sequences</article-title><source>Laryngoscope</source><year>2023</year><month>02</month><volume>133</volume><issue>2</issue><fpage>327</fpage><lpage>335</lpage><pub-id pub-id-type="doi">10.1002/lary.30154</pub-id><pub-id pub-id-type="medline">35575610</pub-id></nlm-citation></ref><ref id="ref123"><label>123</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Bang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Seo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jang</surname><given-names>J</given-names> </name></person-group><article-title>Deep learning-based multimodal segmentation of oropharyngeal squamous cell carcinoma on CT and MRI using self-configuring nnU-Net</article-title><source>Eur Radiol</source><year>2024</year><month>08</month><volume>34</volume><issue>8</issue><fpage>5389</fpage><lpage>5400</lpage><pub-id pub-id-type="doi">10.1007/s00330-024-10585-y</pub-id><pub-id pub-id-type="medline">38243135</pub-id></nlm-citation></ref><ref id="ref124"><label>124</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sundgaard</surname><given-names>JV</given-names> </name><name name-style="western"><surname>Hannemose</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Laugesen</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Multi-modal deep learning for joint prediction of otitis media and diagnostic difficulty</article-title><source>Laryngoscope Investig Otolaryngol</source><year>2024</year><month>02</month><volume>9</volume><issue>1</issue><fpage>e1199</fpage><pub-id pub-id-type="doi">10.1002/lio2.1199</pub-id><pub-id pub-id-type="medline">38362190</pub-id></nlm-citation></ref><ref id="ref125"><label>125</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Callej&#x00F3;n-Leblic</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Blanco-Trejo</surname><given-names>S</given-names> </name><name name-style="western"><surname>Villarreal-Garza</surname><given-names>B</given-names> </name><etal/></person-group><article-title>A multimodal database for the collection of interdisciplinary audiological research data in Spain</article-title><source>Auditio</source><year>2024</year><month>09</month><volume>8</volume><fpage>e109</fpage><pub-id pub-id-type="doi">10.51445/sja.auditio.vol8.2024.109</pub-id></nlm-citation></ref><ref id="ref126"><label>126</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Singhal</surname><given-names>K</given-names> </name><name name-style="western"><surname>Azizi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Large language models encode clinical knowledge</article-title><source>Nature New Biol</source><year>2023</year><month>08</month><day>3</day><volume>620</volume><issue>7972</issue><fpage>172</fpage><lpage>180</lpage><pub-id pub-id-type="doi">10.1038/s41586-023-06291-2</pub-id></nlm-citation></ref><ref id="ref127"><label>127</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Yan</surname><given-names>C</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>X</given-names> </name></person-group><article-title>From large language models to large multimodal models: a literature review</article-title><source>Appl Sci (Basel)</source><year>2024</year><volume>14</volume><issue>12</issue><fpage>5068</fpage><pub-id pub-id-type="doi">10.3390/app14125068</pub-id></nlm-citation></ref><ref id="ref128"><label>128</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Rao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xiao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name></person-group><article-title>What is the limitation of multimodal LLMs? A deeper look into multimodal LLMs through prompt probing</article-title><source>Inf Process Manag</source><year>2023</year><month>11</month><volume>60</volume><issue>6</issue><fpage>103510</fpage><pub-id pub-id-type="doi">10.1016/j.ipm.2023.103510</pub-id></nlm-citation></ref><ref id="ref129"><label>129</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>X</given-names> </name><etal/></person-group><article-title>A medical multimodal large language model for future pandemics</article-title><source>NPJ Digit Med</source><year>2023</year><month>12</month><day>2</day><volume>6</volume><issue>1</issue><fpage>38042919</fpage><pub-id pub-id-type="doi">10.1038/s41746-023-00952-2</pub-id></nlm-citation></ref><ref id="ref130"><label>130</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>P</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Clifton</surname><given-names>DA</given-names> </name></person-group><article-title>Multimodal learning with transformers: a survey</article-title><source>IEEE Trans Pattern Anal Mach Intell</source><year>2023</year><month>10</month><volume>45</volume><issue>10</issue><fpage>12113</fpage><lpage>12132</lpage><pub-id pub-id-type="doi">10.1109/TPAMI.2023.3275156</pub-id><pub-id pub-id-type="medline">37167049</pub-id></nlm-citation></ref><ref id="ref131"><label>131</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>HY</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><etal/></person-group><article-title>A transformer-based representation-learning model with unified processing of multimodal input for clinical diagnostics</article-title><source>Nat Biomed Eng</source><year>2023</year><month>06</month><volume>7</volume><issue>6</issue><fpage>743</fpage><lpage>755</lpage><pub-id pub-id-type="doi">10.1038/s41551-023-01045-x</pub-id></nlm-citation></ref><ref id="ref132"><label>132</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Steurer</surname><given-names>B</given-names> </name><name name-style="western"><surname>Vanhaelen</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Zhavoronkov</surname><given-names>A</given-names> </name></person-group><article-title>Multimodal transformers and their applications in drug target discovery for aging and age-related diseases</article-title><source>J Gerontol A Biol Sci Med Sci</source><year>2024</year><month>09</month><day>1</day><volume>79</volume><issue>9</issue><fpage>39126345</fpage><pub-id pub-id-type="doi">10.1093/gerona/glae006</pub-id><pub-id pub-id-type="medline">39126345</pub-id></nlm-citation></ref><ref id="ref133"><label>133</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Takagi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hashimoto</surname><given-names>N</given-names> </name><name name-style="western"><surname>Masuda</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Transformer-based personalized attention mechanism for medical images with clinical records</article-title><source>J Pathol Inform</source><year>2023</year><volume>14</volume><issue>100185</issue><fpage>100185</fpage><pub-id pub-id-type="doi">10.1016/j.jpi.2022.100185</pub-id></nlm-citation></ref><ref id="ref134"><label>134</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Narhi-Martinez</surname><given-names>W</given-names> </name><name name-style="western"><surname>Dube</surname><given-names>B</given-names> </name><name name-style="western"><surname>Golomb</surname><given-names>JD</given-names> </name></person-group><article-title>Attention as a multi-level system of weights and balances</article-title><source>Wiley Interdiscip Rev Cogn Sci</source><year>2023</year><month>01</month><volume>14</volume><issue>1</issue><fpage>e1633</fpage><pub-id pub-id-type="doi">10.1002/wcs.1633</pub-id><pub-id pub-id-type="medline">36317275</pub-id></nlm-citation></ref><ref id="ref135"><label>135</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sha</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>MD</given-names> </name></person-group><article-title>Interpretable predictions of clinical outcomes with an attention-based recurrent neural network</article-title><source>ACM BCB</source><year>2017</year><month>08</month><volume>2017</volume><fpage>233</fpage><lpage>240</lpage><pub-id pub-id-type="doi">10.1145/3107411.3107445</pub-id><pub-id pub-id-type="medline">32577628</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Additional material.</p><media xlink:href="jmir_v27i1e76557_app1.docx" xlink:title="DOCX File, 17 KB"/></supplementary-material></app-group></back></article>