<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v28i1e85663</article-id><article-id pub-id-type="doi">10.2196/85663</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>Patient Concerns Regarding Artificial Intelligence Applications in Health Care: Systematic Review and Meta-Synthesis Based on Social Ecological Theory</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Hou</surname><given-names>Jiayu</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Zhang</surname><given-names>Zhiqiao</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Cheng</surname><given-names>Xuan</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Wang</surname><given-names>Weihong</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff3">3</xref></contrib></contrib-group><aff id="aff1"><institution>School of Nursing, Hunan Normal University</institution><addr-line>No. 371, Tongzipo Road, Yuelu District</addr-line><addr-line>Changsha City</addr-line><addr-line>Hunan Province</addr-line><country>China</country></aff><aff id="aff2"><institution>Kiang Wu Nursing College of Macau</institution><addr-line>Macao</addr-line><country>China</country></aff><aff id="aff3"><institution>Medical Humanities Research Center, Hunan Normal University</institution><addr-line>Changsha</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Brini</surname><given-names>Stefano</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Heyen</surname><given-names>Nils</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Jackson</surname><given-names>Sara</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Weihong Wang, PhD, School of Nursing, Hunan Normal University, No. 371, Tongzipo Road, Yuelu District, Changsha City, Hunan Province, China, 86 13548968918; <email>wwh270025@hunnu.edu.cn</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>28</day><month>4</month><year>2026</year></pub-date><volume>28</volume><elocation-id>e85663</elocation-id><history><date date-type="received"><day>11</day><month>10</month><year>2025</year></date><date date-type="rev-recd"><day>12</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>12</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Jiayu Hou, Zhiqiao Zhang, Xuan Cheng, Weihong Wang. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 28.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2026/1/e85663"/><abstract><sec><title>Background</title><p>The use of artificial intelligence (AI) in health care is growing quickly, but there is not enough research that looks at patient concerns from a multilevel perspective. Existing reviews predominantly summarize patient attitudes descriptively, lacking theoretical frameworks to explain the underlying mechanisms of these concerns.</p></sec><sec><title>Objective</title><p>This systematic review and meta-synthesis aimed to identify and analyze patient concerns regarding health care AI applications, using social ecological theory to reveal the multilevel interactive mechanisms of concern at the individual, interpersonal, organizational, and societal levels.</p></sec><sec sec-type="methods"><title>Methods</title><p>Following the PRISMA-S (Preferred Reporting Items for Systematic Reviews and Meta-Analyses literature search extension) guidelines, databases including PubMed, Embase, Web of Science, CINAHL, and Scopus were searched on March 1, 2026. Qualitative studies exploring patient perceptions of clinical AI applications were included, excluding those involving only healthy populations, technical performance, or nonclinical settings. Two researchers independently screened the literature and assessed methodological quality using the JBI-QARI (Joanna Briggs Institute Qualitative Assessment and Review Instrument) checklist. Confidence in synthesized findings was assessed using the GRADE-CERQual (Confidence in the Evidence from Reviews of Qualitative Research) approach.</p></sec><sec sec-type="results"><title>Results</title><p>A total of 25 qualitative studies involving 528 participants from diverse patient groups across multiple countries were included. Six themes emerged: (1) microlevel worries about privacy and data security, including data breaches and loss of control over personal health information; (2) worries about the limits and reliability of technology, especially AI diagnostic accuracy and &#x201C;black box&#x201D; decision-making; (3) mesolevel effects on physician-patient relationships, including reduced face-to-face interaction and empathy; (4) trust and accountability issues, including unclear responsibility attribution and institutional oversight problems; (5) macrolevel ethical and equity issues, including algorithmic bias and health care access inequalities; and (6) worries about technology diffusion and possible replacement of health care workers.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This review represents the first meta-synthesis applying social ecological theory to construct patient concerns regarding medical AI. Unlike previous descriptive reviews, it reveals the interconnected &#x201C;ecological imbalance&#x201D; mechanisms at micro-, meso-, and macrolevels when AI is embedded in health care systems. The findings suggest that many patient concerns are based on facts rather than just misunderstandings, indicating that systemic rather than isolated interventions are needed. Practical implications include explainable algorithm design at the microlevel, improved physician-patient communication, and institutional accountability at the mesolevel, and coordinated global ethical norms and equity-promoting policies at the macrolevel. Limitations include the inclusion of studies primarily from developed regions, significant heterogeneity in AI application scenarios, and constraints inherent to secondary research. Nevertheless, addressing these multilevel concerns remains crucial for balancing technological advancement with patient-centered care and enabling sustainable AI integration.</p></sec><sec><title>Trial Registration</title><p>Trial Registration: PROSPERO CRD420251156502; https://www.crd.york.ac.uk/PROSPERO/view/CRD420251156502</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>medical ethics</kwd><kwd>patient concerns</kwd><kwd>data privacy</kwd><kwd>physician-patient relationship</kwd><kwd>health care equity</kwd><kwd>meta-integration</kwd><kwd>social ecological theory</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The application of artificial intelligence (AI) technology in health care is rapidly advancing, particularly across multiple domains, including disease diagnosis, treatment decision support, personalized health management, image analysis, and drug discovery [<xref ref-type="bibr" rid="ref1">1</xref>]. Through techniques such as deep learning and natural language processing, AI can extract critical insights from vast datasets, empowering clinicians to make more precise decisions. With breakthroughs in generative AI and large language models, AI capabilities have expanded into complex clinical decision support and physician-patient interactions, enhancing communication and improving patient outcomes through more tailored and responsive care [<xref ref-type="bibr" rid="ref2">2</xref>]. As of 2024, the US Food and Drug Administration has approved over 950 AI or machine learning medical devices, with radiology accounting for 76% of these approvals [<xref ref-type="bibr" rid="ref3">3</xref>]. In 2024, AI adoption among US physicians reached 66%, nearly doubling from 38% in 2023 [<xref ref-type="bibr" rid="ref4">4</xref>]. AI implementation has significantly enhanced health care efficiency, optimized resource allocation, and delivered more personalized and precise treatment plans for patients [<xref ref-type="bibr" rid="ref5">5</xref>]. For instance, AI can identify minute lesions during early screening and match genes with therapies in precision medicine, substantially improving cure rates and patient quality of life [<xref ref-type="bibr" rid="ref6">6</xref>]. However, as AI technology becomes more widespread and deeply integrated into health care, patient concerns regarding ethics, privacy, and security have increasingly come to the fore [<xref ref-type="bibr" rid="ref7">7</xref>]. These concerns not only impact patient acceptance of AI technology but also directly affect the smooth promotion and implementation of AI in medical practice [<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>When encountering AI-based medical applications, one of the patients&#x2019; primary concerns is privacy protection and data security [<xref ref-type="bibr" rid="ref9">9</xref>]. With the digitization of medical data and widespread AI adoption, patient health information is being collected, stored, and analyzed at an unprecedented scale. Many patients fear potential misuse or leakage of these sensitive data, particularly when data protection measures remain inadequate [<xref ref-type="bibr" rid="ref10">10</xref>]. A 2023 national survey revealed that nearly half of US adults express low trust in the health care system&#x2019;s responsible use of AI [<xref ref-type="bibr" rid="ref11">11</xref>]. Additionally, the transparency and explainability of AI decisions have become major patient concerns. Due to the &#x201C;black box&#x201D; nature of AI algorithms, patients often cannot understand how AI arrives at diagnostic and treatment recommendations. This opacity fuels skepticism about the reliability of AI systems [<xref ref-type="bibr" rid="ref12">12</xref>] and hinders clinicians&#x2019; ability to explain treatment decisions to patients [<xref ref-type="bibr" rid="ref13">13</xref>]. Another core issue is the potential loss of patient decision autonomy, particularly in urgent situations such as emergency rooms. AI&#x2019;s &#x201C;automated&#x201D; decisions may deprive patients of choice and diminish their active role in the treatment process [<xref ref-type="bibr" rid="ref14">14</xref>]. Furthermore, as AI gradually assumes certain medical tasks, patient concerns about the erosion of humanistic care are growing. Many patients believe that while AI can provide efficient diagnostic support, it cannot replace a physician&#x2019;s empathy and care. The resulting distancing in the physician-patient relationship may impact treatment outcomes and the overall patient experience [<xref ref-type="bibr" rid="ref15">15</xref>]. A mixed methods survey of 600 US adults revealed that 30% of respondents expressed concern about AI&#x2019;s lack of a physician&#x2019;s &#x201C;human touch,&#x201D; while 84.2% preferred AI for tasks unrelated to the physician-patient relationship, such as appointment scheduling [<xref ref-type="bibr" rid="ref16">16</xref>].</p><p>Therefore, the application of AI in health care involves not only technological advancement but also profound shifts in physician-patient relationships, ethical principles, and cultural values. During the development of new technologies, it is important to fully consider and address patients&#x2019; ethical concerns, need for privacy protection, and expectations for more humanized medical care. Existing qualitative syntheses primarily adopt a broad public perspective, empirically summarizing the advantages, risks, and recommendations of AI in health care [<xref ref-type="bibr" rid="ref17">17</xref>], yet theoretical frameworks remain relatively underdeveloped. Given this context, this study aimed to synthesize existing literature systematically on patients&#x2019; ethical concerns regarding AI through the Social Ecological Model framework. Specific research objectives are as follows:</p><list list-type="order"><list-item><p>To identify fundamental ethical issues (eg, privacy, trust, and humanistic care) that patients encounter with AI health care applications;</p></list-item><list-item><p>To analyze how these concerns influence patient acceptance of AI technology;</p></list-item><list-item><p>To reveal tension mechanisms at 3 levels&#x2014;individual cognition, physician-patient interaction, and macrolevel institutional systems&#x2014;to address potential &#x201C;ecological imbalance risks.&#x201D;</p></list-item></list></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Registration and Reporting Framework</title><p>This study strictly followed the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines (<xref ref-type="supplementary-material" rid="app2">Checklist 1</xref>) [<xref ref-type="bibr" rid="ref18">18</xref>] and the ENTREQ (Enhancing Transparency in Reporting the Synthesis of Qualitative Research; <xref ref-type="supplementary-material" rid="app3">Checklist 2</xref>) [<xref ref-type="bibr" rid="ref19">19</xref>]. This review aimed to explore patients&#x2019; values, attitudes, and experiences regarding the application of AI in health care. The Cochrane Handbook recognizes qualitative evidence synthesis as the appropriate methodology for such questions [<xref ref-type="bibr" rid="ref20">20</xref>]. As all 25 included studies used qualitative designs yielding narrative findings rather than quantitative effect estimates amenable to statistical pooling [<xref ref-type="bibr" rid="ref21">21</xref>], statistical meta-analysis was not applicable to this review. We therefore used the Joanna Briggs Institute meta-aggregation approach [<xref ref-type="bibr" rid="ref22">22</xref>]. The review was prospectively registered in PROSPERO (CRD420251156502). The registered protocol was subsequently updated to clarify the qualitative study design and the use of JBI-QARI (Joanna Briggs Institute Qualitative Assessment and Review Instrument) for quality assessment. In addition, the following methodological enhancements were made during the conduct of this review: (1) CINAHL was added as a fifth database to improve coverage of nursing and allied health literature, (2) the GRADE-CERQual (Confidence in the Evidence from Reviews of Qualitative Research) approach was adopted to assess confidence in the synthesized findings, and (3) social ecological theory (SET) was used as a theoretical lens to interpret and discuss the findings.</p></sec><sec id="s2-2"><title>Search Strategy</title><p>This study&#x2019;s search strategy followed the PRISMA-S (Preferred Reporting Items for Systematic Reviews and Meta-Analyses literature search extension) guideline (<xref ref-type="supplementary-material" rid="app4">Checklist 3</xref>) [<xref ref-type="bibr" rid="ref23">23</xref>]. The initial search was conducted on September 28, 2025. Following iterative refinement of the search strategy, updated searches were performed on January 4, 2026, and March 1, 2026, to capture the most recent literature. Five databases were searched from inception to the date of each search: PubMed (via National Library of Medicine), Embase (via Embase.com), Web of Science (via Clarivate, encompassing the Core Collection, KCI-Korean Journal Database, MEDLINE, ProQuest Dissertations &#x0026; Theses Citation Index, SciELO Citation Index, and the Grants Index), CINAHL (via EBSCOhost), and Scopus (via Scopus.com); all other databases were searched independently through their respective platforms. The detailed search strategies for all databases are provided in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. A 3-step search strategy was used. First, an initial search was conducted in PubMed, and the titles, abstracts, and index terms of relevant records were analyzed to identify key search terms. Second, a comprehensive search using all identified keywords and index terms was undertaken across all databases. Third, the reference lists of included studies were hand-searched to identify any additional relevant studies. The search strategy was developed de novo for this review, and its effectiveness was validated by confirming the retrieval of known relevant studies. No formal peer review of the search strategy was conducted using standardized appraisal tools. This review focused on published qualitative research; clinical trial registries were not searched. Beyond the systematic database searches and reference list screening, no supplementary search methods were used, such as contacting authors, browsing conference proceedings, or setting up citation alerts. Only studies published in English or Chinese were included. The PubMed search strategy is presented below:</p><list list-type="bullet"><list-item><p>#1 (&#x201C;Artificial Intelligence&#x201D;[MeSH]) OR (&#x201C;Machine Learning&#x201D;[MeSH]) OR (&#x201C;Decision Support Systems, Clinical&#x201D;[MeSH]) OR (&#x201C;artificial intelligence&#x201D;[tiab]) OR (&#x201C;machine learning&#x201D;[tiab]) OR (&#x201C;deep learning&#x201D;[tiab]) OR (&#x201C;AI-based&#x201D;[tiab]) OR (&#x201C;AI-assisted&#x201D;[tiab]) OR (&#x201C;AI-driven&#x201D;[tiab]) OR (&#x201C;ChatGPT&#x201D;[tiab]) OR (&#x201C;large language model&#x201D;[tiab]) OR (&#x201C;clinical decision support&#x201D;[tiab]) OR (&#x201C;CDSS&#x201D;[tiab]) OR (&#x201C;generative AI&#x201D;[tiab]) OR (&#x201C;algorithm*"[tiab])</p></list-item><list-item><p>#2 (&#x201C;Patients&#x201D;[MeSH]) OR (patient*[tiab]) OR (stakeholder*[tiab])</p></list-item><list-item><p>#3 (healthcare[tiab]) OR (&#x201C;health care&#x201D;[tiab]) OR (medical[tiab]) OR (clinical[tiab])</p></list-item><list-item><p>#4 (&#x201C;Attitude to Health&#x201D;[MeSH]) OR (concern[tiab]) OR (perception[tiab]) OR (perspective[tiab]) OR (attitude[tiab]) OR (trust[tiab]) OR (acceptance[tiab]) OR (barrier[tiab]) OR (ethical[tiab]) OR (privacy[tiab]) OR (view[tiab]) OR (opinion[tiab]) OR (experience[tiab]) OR (feeling[tiab]) OR (worry[tiab])</p></list-item><list-item><p>#5 (&#x201C;Qualitative Research&#x201D;[MeSH]) OR (qualitative[tiab]) OR (&#x201C;focus group&#x201D;[tiab]) OR (interview[tiab]) OR (&#x201C;thematic analysis&#x201D;[tiab]) OR (&#x201C;content analysis&#x201D;[tiab]) OR (&#x201C;grounded theory&#x201D;[tiab]) OR (phenomenolog*[tiab]) OR (ethnograph*[tiab]) OR (&#x201C;lived experience&#x201D;[tiab]) OR (&#x201C;narrative analysis&#x201D;[tiab]) OR (&#x201C;in-depth interview&#x201D;[tiab]) OR (&#x201C;qualitative study&#x201D;[tiab]) OR (&#x201C;qualitative research&#x201D;[tiab])</p></list-item><list-item><p>#6 #1 AND #2 AND #3 AND #4 AND #5</p></list-item></list></sec><sec id="s2-3"><title>Inclusion and Exclusion Criteria</title><p>This review focused on patient concerns arising from the use of AI in clinical practice. The inclusion and exclusion criteria are presented in <xref ref-type="table" rid="table1">Table 1</xref>. A total of 19,090 records were retrieved and imported into EndNote (version 21; Clarivate Analytics). Through both automated and manual deduplication in EndNote 21, a total of 7132 (37.4%) duplicate records were identified, leaving 11,958 (62.6%) records for assessment based on the relevance of titles and abstracts. At the title and abstract screening stage, 11,874 (99.3%) records were excluded for the following reasons: not addressing AI applications in clinical health care settings (n=5818, 49%), not involving patients as participants (n=3871, 32.6%), ineligible study design (n=2030, 17.1%), and not published in English or Chinese (n=155, 1.3%). A total of 84 (0.7%) records met the selection criteria, and their full texts were retrieved for further evaluation. Following full-text assessment, 25 studies met the criteria for quality appraisal. Two authors independently conducted the screening process, and any disagreements regarding inclusion were resolved through consultation with a third author.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Inclusion and exclusion criteria for qualitative studies on patient concerns regarding artificial intelligence (AI) applications in health care based on the PICOS framework.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Components</td><td align="left" valign="bottom">Inclusion criteria</td><td align="left" valign="bottom">Exclusion criteria</td></tr></thead><tbody><tr><td align="left" valign="top">Participants</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Patients who receive medical services supported by AI technology and stakeholders.</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Stakeholders without patients and healthy people.</p></list-item></list></td></tr><tr><td align="left" valign="top">Phenomenon of interest</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>The focus of the research is on the application of AI technology in health care, such as AI-assisted diagnosis, AI-assisted surgeries, AI health monitoring, and virtual nursing assistants.</p></list-item><list-item><p>The research focuses on patients&#x2019; perception of AI applications, their psychological reactions, concerns, and emotions, such as concerns about privacy, fears of physician substitution, and trust in AI.</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>It does not involve the application of AI technology in medical fields, or does not focus on the research of patients&#x2019; psychological and emotional responses.</p></list-item><list-item><p>Only focusing on the technical aspects such as the performance and algorithm optimization of AI technology, without considering the patients&#x2019; perception and emotions.</p></list-item></list></td></tr><tr><td align="left" valign="top">Context</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>The research background focuses on AI application scenarios in the medical environment, including hospitals, clinics, telemedicine platforms, and public health monitoring.</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Research that is not conducted in actual medical settings or that is limited to technological development research in laboratory environments.</p></list-item><list-item><p>The research environment has no relation to the actual medical experience of patients and thus cannot provide in-depth information about patients&#x2019; perceptions and reactions.</p></list-item></list></td></tr><tr><td align="left" valign="top">Study design</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>The data collection method focuses on the subjective experiences of patients, paying attention to their emotional, cognitive, and worrying psychological reactions during their medical intervention with AI technology.</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Not a qualitative study.</p></list-item><list-item><p>The overly simplistic design, such as relying solely on short questionnaires or tool tests, fails to thoroughly explore the psychological and emotional responses of patients.</p></list-item></list></td></tr></tbody></table></table-wrap></sec><sec id="s2-4"><title>Quality Assessment</title><p>The JBI-QARI was used to evaluate the methodological rigor of each published study [<xref ref-type="bibr" rid="ref24">24</xref>]. Questions answered &#x201C;yes&#x201D; received 1 point, and studies scoring 5 points or lower were deemed low quality and excluded from the synthesis. Two reviewers independently conducted rigorous assessments of the selected research reviews. Disagreements were resolved through discussion or consultation with a third reviewer within the team. <xref ref-type="table" rid="table2">Table 2</xref> presents the quality assessment results for the studies included in this review.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Methodological quality assessment of 25 included qualitative studies using the Joanna Briggs Institute Qualitative Assessment and Review Instrument critical appraisal checklist.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Studies</td><td align="left" valign="bottom">Q1<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td><td align="left" valign="bottom">Q2<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="bottom">Q3<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup></td><td align="left" valign="bottom">Q4<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="bottom">Q5<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td><td align="left" valign="bottom">Q6<sup><xref ref-type="table-fn" rid="table2fn6">f</xref></sup></td><td align="left" valign="bottom">Q7<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="bottom">Q8<sup><xref ref-type="table-fn" rid="table2fn8">h</xref></sup></td><td align="left" valign="bottom">Q9<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="bottom">Q10<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="bottom">Total</td></tr></thead><tbody><tr><td align="left" valign="top">Hurley et al [<xref ref-type="bibr" rid="ref25">25</xref>]</td><td align="left" valign="top">Y<sup><xref ref-type="table-fn" rid="table2fn11">k</xref></sup></td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">9</td></tr><tr><td align="left" valign="top">Annamalai [<xref ref-type="bibr" rid="ref26">26</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">N<sup><xref ref-type="table-fn" rid="table2fn12">l</xref></sup></td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">&#x010C;artolovni et al [<xref ref-type="bibr" rid="ref27">27</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U<sup><xref ref-type="table-fn" rid="table2fn13">m</xref></sup></td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Sujan et al [<xref ref-type="bibr" rid="ref28">28</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Kostick-Quenet et al [<xref ref-type="bibr" rid="ref29">29</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">9</td></tr><tr><td align="left" valign="top">Hesjedal et al [<xref ref-type="bibr" rid="ref30">30</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">9</td></tr><tr><td align="left" valign="top">McCradden et al [<xref ref-type="bibr" rid="ref31">31</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Al-Anezi [<xref ref-type="bibr" rid="ref32">32</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Freeman et al [<xref ref-type="bibr" rid="ref33">33</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Viberg Johansson et al [<xref ref-type="bibr" rid="ref34">34</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Jeyakumar et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Haan et al [<xref ref-type="bibr" rid="ref36">36</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Berger et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Khairat et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Litchfield et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Omori et al [<xref ref-type="bibr" rid="ref40">40</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Funer et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Giebel et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">9</td></tr><tr><td align="left" valign="top">Ly et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">10</td></tr><tr><td align="left" valign="top">Richardson et al [<xref ref-type="bibr" rid="ref44">44</xref>]</td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Foresman et al [<xref ref-type="bibr" rid="ref45">45</xref>]</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr><tr><td align="left" valign="top">Schneider et al [<xref ref-type="bibr" rid="ref46">46</xref>]</td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">N</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Zhang et al [<xref ref-type="bibr" rid="ref47">47</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Gundlack et al [<xref ref-type="bibr" rid="ref48">48</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">N</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Steerling et al [<xref ref-type="bibr" rid="ref49">49</xref>]</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">U</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">Y</td><td align="left" valign="top">8</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>Q1. Is there congruity between the stated philosophical perspective and the research methodology?</p></fn><fn id="table2fn2"><p><sup>b</sup>Q2. Is there congruity between the research methodology and the research question or objectives?</p></fn><fn id="table2fn3"><p><sup>c</sup>Q3. Is there congruity between the research methodology and the methods used to collect data?</p></fn><fn id="table2fn4"><p><sup>d</sup>Q4. Is there congruity between the research methodology and the representation and analysis of data?</p></fn><fn id="table2fn5"><p><sup>e</sup>Q5. Is there congruity between the research methodology and the interpretation of results?</p></fn><fn id="table2fn6"><p><sup>f</sup>Q6. Is there a statement locating the researcher culturally or theoretically?</p></fn><fn id="table2fn7"><p><sup>g</sup>Q7. Is the influence of the researcher on the research, and vice-versa, addressed?</p></fn><fn id="table2fn8"><p><sup>h</sup>Q8. Are participants, and their voices, adequately represented?</p></fn><fn id="table2fn9"><p><sup>i</sup>Q9. Is the research ethical according to current criteria or, for recent studies, and is there evidence of ethical approval by an appropriate body?</p></fn><fn id="table2fn10"><p><sup>j</sup>Q10. Do the conclusions drawn in the research report flow from the analysis, or interpretation, of the data? the conclusions drawn in the research report flow from the analysis, or interpretation, of the data?</p></fn><fn id="table2fn11"><p><sup>k</sup>Y: yes.</p></fn><fn id="table2fn12"><p><sup>l</sup>N: no.</p></fn><fn id="table2fn13"><p><sup>m</sup>U: unclear.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s2-5"><title>Data Extraction and Synthesis</title><p>Data were extracted from the studies included in this review using the JBI-QARI standardized data extraction tool. The first author extracted data from these 25 studies, including the first author and publication year, study population and sample size, research topic, study type, and primary outcomes, as detailed in <xref ref-type="table" rid="table3">Table 3</xref>. This study used a meta-synthesis approach [<xref ref-type="bibr" rid="ref22">22</xref>]. Two researchers repeatedly read and interpreted the original studies, analyzed and interpreted the implications of the findings, grouped similar results into new categories, and then synthesized these categories into integrated outcomes to form new perspectives or interpretations. When disagreements arose between the 2 coders, a third coder was consulted.</p><p>The confidence in each synthesized finding was assessed using the GRADE-CERQual approach [<xref ref-type="bibr" rid="ref50">50</xref>]. CERQual assesses confidence based on four components: (1) methodological limitations of the included studies, informed by the JBI-QARI critical appraisal results; (2) coherence of the finding across contributing studies; (3) adequacy of data supporting the finding, considering both the number of studies and the richness of data; and (4) relevance, defined as the extent to which the contexts of contributing studies are applicable to the review question. Each finding was assigned an overall confidence level of high, moderate, low, or very low. Two reviewers independently assessed each component and resolved disagreements through discussion. The CERQual assessment results are presented in <xref ref-type="table" rid="table4">Table 4</xref>. Heterogeneity across included studies was explored narratively by examining differences in AI application types, patient populations, and geographic contexts, as reported in <xref ref-type="table" rid="table3">Table 3</xref> and discussed in the <italic>Limitations</italic> section.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Characteristics of 25 included qualitative studies on patient concerns regarding artificial intelligence (AI) in health care (2019&#x2010;2025).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Studies</td><td align="left" valign="bottom">Year</td><td align="left" valign="bottom">Country</td><td align="left" valign="bottom">Setting</td><td align="left" valign="bottom">Data collection</td><td align="left" valign="bottom">Research method</td><td align="left" valign="bottom">Participants</td><td align="left" valign="bottom">Interested topics</td><td align="left" valign="bottom">Main results</td></tr></thead><tbody><tr><td align="left" valign="top">Hurley et al [<xref ref-type="bibr" rid="ref25">25</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">United States</td><td align="left" valign="top">Not specified</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 20 patients)</td><td align="left" valign="top">Ethical considerations in integrating multimodal computer perception and neurotechnology in clinical care</td><td align="left" valign="top">Three main themes:<list list-type="order"><list-item><p>Perceived invasiveness of passive and continuous data collection</p></list-item><list-item><p>Data protection and security concerns</p></list-item><list-item><p>Ethical issues related to patients&#x2019; awareness of data collection</p></list-item></list></td></tr><tr><td align="left" valign="top">Annamalai [<xref ref-type="bibr" rid="ref26">26</xref>]</td><td align="left" valign="top">2020</td><td align="left" valign="top">India</td><td align="left" valign="top">Telepsychiatry</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Grounded theory approach</td><td align="left" valign="top">Stakeholders (including 14 patients)</td><td align="left" valign="top">Exploring challenges of AI-enabled telepsychiatry for clinical practice among urban Indian stakeholders</td><td align="left" valign="top">Four themes:<list list-type="order"><list-item><p>Ethical, legal, accountability, and regulatory issues</p></list-item><list-item><p>Challenges related to data</p></list-item><list-item><p>Health system infrastructure</p></list-item><list-item><p>AI related</p></list-item></list></td></tr><tr><td align="left" valign="top">&#x010C;artolovni et al [<xref ref-type="bibr" rid="ref27">27</xref>]</td><td align="left" valign="top">2023</td><td align="left" valign="top">Croatia</td><td align="left" valign="top">Hospital</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 15 patients)</td><td align="left" valign="top">Exploring multistakeholder (patients, physicians, and health care managers) insights into AI&#x2019;s impact on the patient-physician relationship</td><td align="left" valign="top">Four themes:<list list-type="order"><list-item><p>The current state of health care and the patient-physician relationship</p></list-item><list-item><p>Expectations of AI</p><p>a synergetic effect between physicians</p></list-item><list-item><p>AI the future of health care</p></list-item><list-item><p>The patient-physician relationship</p></list-item></list></td></tr><tr><td align="left" valign="top">Sujan et al [<xref ref-type="bibr" rid="ref28">28</xref>]</td><td align="left" valign="top">2022</td><td align="left" valign="top">United Kingdom</td><td align="left" valign="top">Hospital</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 4 patients)</td><td align="left" valign="top">Exploring perceptions of safety and safety assurance of health care AI (using AI-based ICU<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup> infusion pumps as a case) among patients, hospital staff, technology developers, and regulators in the United Kingdom</td><td align="left" valign="top">Four themes:<list list-type="order"><list-item><p>The potential impact of health care AI</p></list-item><list-item><p>Requirements for human-AI interaction</p></list-item><list-item><p>Safety assurance practices and regulatory frameworks for AI and the gaps that exist</p></list-item><list-item><p>How incidents involving AI should be managed</p></list-item></list></td></tr><tr><td align="left" valign="top">Kostick-Quenet et al [<xref ref-type="bibr" rid="ref29">29</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">United States</td><td align="left" valign="top">Cardiology</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 18 patients)</td><td align="left" valign="top">Exploring trust criteria for an AI/ML<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup>-based survival prediction algorithm (for LVAD<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup> therapy in advanced heart failure) among patients, nurse coordinators, and physicians, focusing on normative and epistemic considerations</td><td align="left" valign="top">Three themes:<list list-type="order"><list-item><p>Epistemic trust considerations</p></list-item><list-item><p>Relational trust considerations</p></list-item><list-item><p>Personal belief-based trust considerations</p></list-item></list></td></tr><tr><td align="left" valign="top">Hesjedal et al [<xref ref-type="bibr" rid="ref30">30</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">Norway</td><td align="left" valign="top">Prostate cancer diagnostics</td><td align="left" valign="top">Participant observation, semistructured interviews, and focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 38 patients)</td><td align="left" valign="top">Investigating how scientists, MDs, and patients with PCa<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup> relate ethical challenges of AI decision-making tools in PCa diagnostics to their understanding of &#x201C;good health care,&#x201D; focusing on registers of valuing</td><td align="left" valign="top"><list list-type="simple"><list-item><p>The main finding of the study is that medical physicians, scientists, and patients perceive ethical challenges with AI in prostate cancer diagnostics differently, shaped by their roles and experiences, which influence their understanding of good health care.</p></list-item></list></td></tr><tr><td align="left" valign="top">McCradden et al [<xref ref-type="bibr" rid="ref31">31</xref>]</td><td align="left" valign="top">2020</td><td align="left" valign="top">Canada</td><td align="left" valign="top">Hospital</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">30 patients and caregivers</td><td align="left" valign="top">Investigate perspectives on ethical issues surrounding AI in health care research among Canadian patients with meningioma, their caregivers, and health care providers</td><td align="left" valign="top">Eight themes:<list list-type="order"><list-item><p>Consent</p></list-item><list-item><p>Privacy</p></list-item><list-item><p>Confidentiality</p></list-item><list-item><p>Responsibility</p></list-item><list-item><p>Accountability</p></list-item><list-item><p>Unintended consequences/harms</p></list-item><list-item><p>Trust</p></list-item><list-item><p>Public engagement</p></list-item></list></td></tr><tr><td align="left" valign="top">Al-Anezi [<xref ref-type="bibr" rid="ref32">32</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">Saudi Arabia</td><td align="left" valign="top">Chronic disease management</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">29 patients</td><td align="left" valign="top">Analyze the feasibility of ChatGPT (free version 3.5) as a virtual health coach for chronic disease management, focusing on its ability to promote health literacy and support patients&#x2019; self-management</td><td align="left" valign="top"><list list-type="simple"><list-item><p>The main finding of this study is that ChatGPT shows potential as a virtual health coach by providing accessible lifestyle advice and motivation, but it faces limitations in accuracy, personalization, and trust that must be addressed before clinical use.</p></list-item></list></td></tr><tr><td align="left" valign="top">Freeman et al [<xref ref-type="bibr" rid="ref33">33</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">Australia</td><td align="left" valign="top">Emergency department</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Grounded theory approach</td><td align="left" valign="top">28 patients</td><td align="left" valign="top">Health consumers&#x2019; ethical concerns toward the use of AI in Australian emergency departments.</td><td align="left" valign="top">Six themes:<list list-type="order"><list-item><p>Health consumer autonomy and consent</p></list-item><list-item><p>Physician decision-making and autonomy</p></list-item><list-item><p>Automation bias and reliance</p></list-item><list-item><p>Data use and privacy</p></list-item><list-item><p>Bias and discrimination</p></list-item><list-item><p>Regulation</p></list-item></list></td></tr><tr><td align="left" valign="top">Viberg Johansson et al [<xref ref-type="bibr" rid="ref34">34</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">Sweden</td><td align="left" valign="top">Breast cancer screening</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">16 patients</td><td align="left" valign="top">Swedish women&#x2019;s perceptions and attitudes toward the use of AI in mammography (as part of the national breast cancer screening program)</td><td align="left" valign="top">Three themes:<list list-type="order"><list-item><p>Perceived differences between AI and human assessment</p></list-item><list-item><p>Attitudes when implementing AI in mammography</p></list-item><list-item><p>Requirements when using AI in mammography</p></list-item></list></td></tr><tr><td align="left" valign="top">Jeyakumar et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td><td align="left" valign="top">2023</td><td align="left" valign="top">Canada</td><td align="left" valign="top">Acute or long-term medical centers</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">12 patients</td><td align="left" valign="top">How patients view AI in health care and emphasize the need for trust, engagement, and strong data governance to ensure ethical and effective integration</td><td align="left" valign="top">Three themes:<list list-type="order"><list-item><p>Cultivating patients&#x2019; trust</p><p>fostering</p></list-item><list-item><p>Patient engagement</p><p>establishing data governance</p></list-item><list-item><p>Validation of AI technologies</p></list-item></list></td></tr><tr><td align="left" valign="top">Haan et al [<xref ref-type="bibr" rid="ref36">36</xref>]</td><td align="left" valign="top">2019</td><td align="left" valign="top">Netherlands</td><td align="left" valign="top">Department of Radiology at a tertiary care academic institution</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Grounded theory approach</td><td align="left" valign="top">20 patients</td><td align="left" valign="top">Patient perspective on the use of AI in radiology, including awareness, uncertainties, and expectations</td><td align="left" valign="top">Six key domains:<list list-type="order"><list-item><p>Proof of technology (need for evidence on efficacy and reliability of AI)</p></list-item><list-item><p>Procedural knowledge (understanding how AI is implemented in practice)</p></list-item><list-item><p>Competence (capability of AI to produce reliable results)</p></list-item><list-item><p>Efficiency (faster scanning and reduced waiting times)</p></list-item><list-item><p>Personal interaction (importance of human communication for results)</p></list-item><list-item><p>Accountability (responsibility for errors made by AI)</p></list-item></list></td></tr><tr><td align="left" valign="top">Berger et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Norway</td><td align="left" valign="top">Prostate cancer diagnostics (AI supported)</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">18 patients</td><td align="left" valign="top">Patient perspectives on trust in AI-powered tools in prostate cancer diagnostics</td><td align="left" valign="top">Three main dimensions of trust were identified:<list list-type="order"><list-item><p>Foundational trust</p></list-item><list-item><p>Interpersonal trust as mediator</p></list-item><list-item><p>Need for human oversight additional finding: participants were more forgiving of human errors than those made by AI, highlighting the relational and moral dimensions of trust in health care.</p></list-item></list></td></tr><tr><td align="left" valign="top">Khairat et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">United States</td><td align="left" valign="top">Cancer survivor organizations (recruitment); virtual/zoom (interviews)</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">21 patients</td><td align="left" valign="top">Cancer survivors&#x2019; experiences, facilitators, and barriers regarding the use of AI-based conversational tools (chatbots)</td><td align="left" valign="top">Three overarching themes:<list list-type="order"><list-item><p>Preference for humans: participants strongly preferred interacting with health care professionals over chatbots.</p></list-item><list-item><p>Lack of empathy: chatbots were perceived as lacking the necessary empathy and emotional support.</p></list-item><list-item><p>Information and privacy concerns: concerns about information overload (generic and nonspecific responses) and data privacy.</p></list-item></list></td></tr><tr><td align="left" valign="top">Litchfield et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">United Kingdom</td><td align="left" valign="top">Primary care</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">7 patients</td><td align="left" valign="top">Acceptability of &#x201C;AmarDoctor,&#x201D; an AI-enabled translation and symptom-checking tool, among underserved Bangladeshi populations</td><td align="left" valign="top">Three overarching themes:<list list-type="order"><list-item><p>Enhanced accessibility and inclusivity</p></list-item><list-item><p>Anonymity for sensitive issues</p></list-item><list-item><p>Trust and safety concerns</p></list-item></list></td></tr><tr><td align="left" valign="top">Omori et al [<xref ref-type="bibr" rid="ref40">40</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Australia</td><td align="left" valign="top">Screening program</td><td align="left" valign="top">Semistructured interviews and focus groups</td><td align="left" valign="top">Grounded theory approach</td><td align="left" valign="top">26 patients</td><td align="left" valign="top">Development of a typology of women&#x2019;s attitudes toward the use of AI in breast cancer screening</td><td align="left" valign="top">Four attitude types identified:<list list-type="order"><list-item><p>Enthusiasts</p></list-item><list-item><p>Practicalists</p></list-item><list-item><p>Traditionalists</p></list-item><list-item><p>Guardians</p></list-item></list></td></tr><tr><td align="left" valign="top">Funer et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td><td align="left" valign="top">2024</td><td align="left" valign="top">Germany</td><td align="left" valign="top">Surgery, nephrology, intensive home care</td><td align="left" valign="top">Semistructured interviews and focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 18 patients)</td><td align="left" valign="top">Impacts of CDSSs<sup><xref ref-type="table-fn" rid="table3fn5">e</xref></sup> on the relationship, communication, and shared decision-making</td><td align="left" valign="top">Three overarching themes:<list list-type="order"><list-item><p>Impact on professional roles</p></list-item><list-item><p>Impact on relationship</p></list-item><list-item><p>Transparency needs</p></list-item></list></td></tr><tr><td align="left" valign="top">Giebel et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Germany</td><td align="left" valign="top">Hospital</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 4 patients)</td><td align="left" valign="top">Opportunities to optimize AI-based CDSS and their integration into health care</td><td align="left" valign="top">Three overarching themes:<list list-type="order"><list-item><p>System optimization</p></list-item><list-item><p>User competence and support</p></list-item><list-item><p>Environmental framework</p></list-item></list></td></tr><tr><td align="left" valign="top">Ly et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Australia</td><td align="left" valign="top">Ophthalmology</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">Stakeholders (including 8 patients)</td><td align="left" valign="top">Stakeholder experiences, attitudes, enablers, barriers, and possible futures of digital diagnosis using AI for age-related macular degeneration in Australia</td><td align="left" valign="top">Three overarching themes:<list list-type="order"><list-item><p>Technological preferences</p></list-item><list-item><p>Divergent stakeholder priorities</p></list-item><list-item><p>Systemic and ethical considerations</p></list-item></list></td></tr><tr><td align="left" valign="top">Richardson et al [<xref ref-type="bibr" rid="ref44">44</xref>]</td><td align="left" valign="top">2021</td><td align="left" valign="top">United States</td><td align="left" valign="top">Primary care</td><td align="left" valign="top">Focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">87 patients</td><td align="left" valign="top">Patient apprehensions and perspectives regarding the use of AI in health care</td><td align="left" valign="top">Major themes identified:<list list-type="order"><list-item><p>Safety and human oversight</p></list-item><list-item><p>Preservation of choice</p></list-item><list-item><p>Cost, equity, and data concerns: cost, bias, security</p></list-item></list></td></tr><tr><td align="left" valign="top">Foresman et al [<xref ref-type="bibr" rid="ref45">45</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">United States</td><td align="left" valign="top">Primary care, radiology, telehealth</td><td align="left" valign="top">Focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">17 patients</td><td align="left" valign="top">Patient perspectives on AI use in health care, specifically in diagnostic processes and communication</td><td align="left" valign="top">Five cross-cutting themes emerged:<list list-type="order"><list-item><p>Validation</p></list-item><list-item><p>Usability</p></list-item><list-item><p>Transparency</p></list-item><list-item><p>Opportunities, privacy, additional finding: comfort levels varied by interaction type; participants were most comfortable with &#x201C;Digital Scribe&#x201D; (low interaction) and least comfortable with &#x201C;virtual human&#x201D; (high interaction).</p></list-item></list></td></tr><tr><td align="left" valign="top">Schneider et al [<xref ref-type="bibr" rid="ref46">46</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Germany</td><td align="left" valign="top">Nephrology, surgery, home-ventilated care</td><td align="left" valign="top">Focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">18 patients</td><td align="left" valign="top">Patient perspectives on AI-based AI-CDSS, specifically focusing on trust, responsibility, and self-determination</td><td align="left" valign="top">Three overarching observations:<list list-type="order"><list-item><p>Indecision and uncertainty</p></list-item><list-item><p>Shift in trust and responsibility</p></list-item><list-item><p>AI as support, not replacement</p></list-item></list></td></tr><tr><td align="left" valign="top">Zhang et al [<xref ref-type="bibr" rid="ref47">47</xref>]</td><td align="left" valign="top">2021</td><td align="left" valign="top">United States</td><td align="left" valign="top">Radiology</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">13 patients</td><td align="left" valign="top">Patients&#x2019; perceptions and acceptance of using AI-based technology to interpret and comprehend radiology reports or imaging data</td><td align="left" valign="top">Four main findings:<list list-type="order"><list-item><p>Positive attitude and utility</p></list-item><list-item><p>Concerns</p></list-item><list-item><p>Trust and design requirements</p></list-item><list-item><p>Designers must ensure systems deliver concerning health results in an empathetic manner to optimize user experience.</p></list-item></list></td></tr><tr><td align="left" valign="top">Gundlack et al [<xref ref-type="bibr" rid="ref48">48</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Germany</td><td align="left" valign="top">Primary care and psychiatry</td><td align="left" valign="top">Semistructured focus groups</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">35 patients</td><td align="left" valign="top">Patients&#x2019; perceptions of AI in medical care regarding caregiving relationships and ethics</td><td align="left" valign="top">Four main themes:<list list-type="order"><list-item><p>AI perceived as beneficial for efficiency, data processing, and patient safety.</p></list-item><list-item><p>Key concerns include impersonality, data security, and overreliance on AI by medical staff.</p></list-item><list-item><p>Human interaction and emotional understanding deemed irreplaceable by AI.</p></list-item><list-item><p>Physicians considered primarily responsible for AI-related decisions; data transparency and privacy protection are essential for implementation.</p></list-item></list></td></tr><tr><td align="left" valign="top">Steerling et al [<xref ref-type="bibr" rid="ref49">49</xref>]</td><td align="left" valign="top">2025</td><td align="left" valign="top">Sweden</td><td align="left" valign="top">Primary care</td><td align="left" valign="top">Semistructured interviews</td><td align="left" valign="top">Phenomenological research</td><td align="left" valign="top">14 health care professionals and 12 patients</td><td align="left" valign="top">Influences on trust in the use of AI-based triage in primary care</td><td align="left" valign="top">Three main influences on trust:<list list-type="order"><list-item><p>Provision of accurate patient information (patients&#x2019; capability and willingness).</p></list-item><list-item><p>Alignment with clinical expertise (standardized reasoning and experience-based knowledge).</p></list-item><list-item><p>Supervision of patients&#x2019; health and safety (professionalism and guidance in information use). Both groups emphasized constructive dialogue and clear instructions for the use and storage of information.</p></list-item></list></td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>ICU: intensive care unit.</p></fn><fn id="table3fn2"><p><sup>b</sup>ML: machine language.</p></fn><fn id="table3fn3"><p><sup>c</sup>LVAD: left ventricular assist device. </p></fn><fn id="table3fn4"><p><sup>d</sup>PCa: prostate cancer. </p></fn><fn id="table3fn5"><p><sup>e</sup>CDSS: clinical decision support system.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>GRADE-CERQual summary of qualitative findings on patient concerns regarding artificial intelligence in health care.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Review finding</td><td align="left" valign="bottom">Contributing studies</td><td align="left" valign="bottom">Methodological limitations</td><td align="left" valign="bottom">Coherence</td><td align="left" valign="bottom">Adequacy</td><td align="left" valign="bottom">Relevance</td><td align="left" valign="bottom">Overall confidence</td></tr></thead><tbody><tr><td align="left" valign="top">Theme 1 (3.1): Privacy and data security</td><td align="left" valign="top">13 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">High</td></tr><tr><td align="left" valign="top">Theme 2 (3.2): Technical limitations and reliability</td><td align="left" valign="top">11 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">High</td></tr><tr><td align="left" valign="top">Theme 3 (3.3): Impact on physician-patient relationship</td><td align="left" valign="top">11 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">High</td></tr><tr><td align="left" valign="top">Theme 4 (3.4): Trust and accountability</td><td align="left" valign="top">13 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">High</td></tr><tr><td align="left" valign="top">Theme 5 (3.5): Ethics and equity</td><td align="left" valign="top">8 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">Moderate concerns</td><td align="left" valign="top">Moderate</td></tr><tr><td align="left" valign="top">Theme 6 (3.6): Future outlook</td><td align="left" valign="top">10 studies</td><td align="left" valign="top">Minor concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">No concerns</td><td align="left" valign="top">High</td></tr></tbody></table></table-wrap></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>This paper provides a systematic review and qualitative synthesis of ethical issues surrounding the application of AI in health care. A total of 25 qualitative studies involving 528 patients were included. Details of the search and screening process are presented in the PRISMA flow diagram (<xref ref-type="fig" rid="figure1">Figure 1</xref>). The findings were categorized into 14 themes and further consolidated into 6 key themes (<xref ref-type="fig" rid="figure2">Figure 2</xref>), elaborated as follows.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram of study selection process (5 databases, inception to March 2026).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v28i1e85663_fig01.png"/></fig><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Social ecological theory of patient concerns regarding artificial intelligence (AI) in health care: thematic structure and multilevel interactions (25 studies, n=528).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v28i1e85663_fig02.png"/></fig></sec><sec id="s3-2"><title>Privacy and Data Security</title><sec id="s3-2-1"><title>Concerns Over Data Protection and Privacy Breaches</title><p>Patients express concerns about the potential misuse of collected medical data, particularly in the absence of effective safeguards. Many believe that once data are uploaded to the internet, they cannot be completely erased and may even be exploited for commercial purposes. Some patients state they consent to data use only for medical purposes, provided they are explicitly informed of its scope:</p><disp-quote><p>They're going to use that [data] for whatever, and [it&#x2019;s like] the Internet, when you [post] something [and] it&#x2019;s stuck forever [there].</p><attrib>P_05 25</attrib></disp-quote><disp-quote><p>I am not confident enough in sharing my personal health data with ChatGPT, as it is based on AI, and there is a high possibility that the data could be misused.</p><attrib>Unspecified 32</attrib></disp-quote></sec><sec id="s3-2-2"><title>Concerns Regarding Data Sharing and Ownership</title><p>Patients expressed confusion about data ownership, particularly when they were unclear about how their data were shared or used. They desire explicit prior notification before data use and fear unauthorized third-party access. Some patients demanded assurances that data use would not exceed the scope of their informed consent:</p><disp-quote><p>I would like to know beforehand so that they don't just send it off without informing me. It doesn't feel entirely right.</p><attrib>P_08 34</attrib></disp-quote><disp-quote><p>I'm worried that the data might be used by others without my permission, especially when I don't know the full extent of how it will be used.</p><attrib>P_14 25</attrib></disp-quote></sec></sec><sec id="s3-3"><title>Limitations and Reliability of AI Technology</title><sec id="s3-3-1"><title>AI Accuracy and Individual Variability</title><p>Patients generally believe that while AI algorithms can analyze vast amounts of data, they cannot fully adapt to each individual&#x2019;s unique health condition. A patient&#x2019;s health status is influenced by factors such as genetics, lifestyle, and environment&#x2014;variables that algorithms often cannot fully predict. Particularly in the early stages of disease or complex scenarios, AI may fail to capture these individual differences, resulting in assessments that do not precisely align with each patient&#x2019;s specific needs. Therefore, patients view AI primarily as a reference tool, asserting that final diagnostic and treatment decisions should remain with experienced physicians:</p><disp-quote><p>The accuracy would be kind of iffy, because everybody has their own risk factor...</p><attrib>P_09 29</attrib></disp-quote><disp-quote><p>I'm fully aware that everybody&#x2019;s situation is unique to their own individual set of circumstances...</p><attrib>P_03 29</attrib></disp-quote></sec><sec id="s3-3-2"><title>Limitations of AI Assessment</title><p>Although AI can process vast amounts of information and provide data-driven analysis, patients believe it cannot comprehensively account for all factors potentially affecting health. For instance, AI struggles to anticipate the impact of nonroutine variables such as unexpected events, environmental changes, or sudden health issues. Patients noted that AI assessments are typically built on historical data, failing to dynamically reflect current health status or psychological state. This limitation frequently constrains AI assessments to broad contexts. Consequently, patients view AI assessments as supplementary tools rather than complete replacements for traditional physician judgment:</p><disp-quote><p>There are so many variables that you don't understand or can't predict...</p><attrib>P_11 29</attrib></disp-quote><disp-quote><p>A lot of these assessments aren't taking things into consideration like diabetes, exercise levels...</p><attrib>P_04 29</attrib></disp-quote></sec><sec id="s3-3-3"><title>Concerns Over Training Data Quality</title><p>Patients&#x2019; doubts about AI reliability extend beyond the algorithms themselves to the quality of their training data. They worry that medical AI systems&#x2019; performance heavily depends on the data they learn from, yet real-world electronic health records often suffer from incomplete documentation, outdated information, or human input errors. If AI is trained on such flawed data, even the most sophisticated algorithms struggle to produce accurate, reliable outputs. Patients are further unsettled by the invisible risk of &#x201C;garbage in, garbage out&#x201D;&#x2014;they cannot know what data the AI has learned from, nor whether it accurately reflects patients&#x2019; actual health conditions:</p><disp-quote><p>So I've had a lot of different things in my medical chart that are inaccurate, very inaccurate, so if they're training [AI on that]...</p><attrib>Unspecified 44</attrib></disp-quote><disp-quote><p>I don't understand how it learns. If it learns wrong, who corrects it?</p><attrib>Participant 11 37</attrib></disp-quote></sec></sec><sec id="s3-4"><title>The Impact of Physician-Patient Relationships</title><sec id="s3-4-1"><title>The Lack of Emotional Support and Human Interaction</title><p>While AI can provide more precise diagnoses and treatment recommendations on a technical level, patients emphasize that a physician&#x2019;s emotional support and human interaction remain indispensable parts of the medical process. Especially when facing serious illnesses, patients require not only medical treatment but also emotional support from their physicians. AI cannot comprehend patients&#x2019; emotional fluctuations, anxieties, or unease. This lack of emotional resonance unsettles many patients. They believe that while AI can assist in treatment, it cannot replace the emotional care provided by health care professionals&#x2014;a crucial element in the therapeutic process.</p><disp-quote><p>People will still want human contact. Machines might be able to do a damned good job, but I think people still need that little human contact...</p><attrib>P_01 28</attrib></disp-quote><disp-quote><p>Excessive digitalization reduces personal contact between people, and that, in turn, reduces communication and connection...</p><p>[P_04] [<xref ref-type="bibr" rid="ref27">27</xref>]</p></disp-quote></sec><sec id="s3-4-2"><title>Challenges in Physician-Patient Communication</title><p>Patients widely agree that effective communication is fundamental to building trust between physicians and patients. Face-to-face interaction helps physicians better understand patients&#x2019; conditions while making patients feel respected and cared for. With AI integration, patients worry physicians may overrely on technology, reducing face-to-face interactions. They emphasize that while AI improves efficiency, it risks neglecting deep patient engagement, which is essential for building trust and understanding in the physician-patient relationship. Especially during complex disease treatments, patients hope physicians will listen more to their needs and feelings rather than solely relying on algorithmic decisions.</p><disp-quote><p>It is crucial that he listen to me and that I listen to him. That&#x2019;s really the most important thing... [<xref ref-type="bibr" rid="ref27">27</xref>]</p></disp-quote><disp-quote><p>I wouldn't only rely on Cronko [AI]; I would use Cronko alongside a physician...</p><attrib>P_13 27</attrib></disp-quote></sec></sec><sec id="s3-5"><title>Trust and Accountability</title><sec id="s3-5-1"><title>Responsibility for AI Decision-Making</title><p>Patients widely express concern about unclear accountability when AI errors occur. AI decisions often lack the experience and emotional judgment of human physicians, making it difficult for patients to determine responsibility in the event of medical incidents. While acknowledging AI&#x2019;s potential to enhance efficiency and accuracy, patients indicate that the absence of a clear accountability mechanism fosters significant distrust toward AI. Particularly when AI systems malfunction, patients are uncertain whether responsibility lies with developers, hospitals, or health care providers, creating unease about adopting AI-driven medical technologies. Additionally, patients emphasize the right to question and correct AI judgments. They argue that when AI makes decisions based on flawed data or assumptions, patients should have avenues to raise objections and demand corrections, rather than passively accepting machine conclusions:</p><disp-quote><p>I believe the doctor always has the responsibility to be checking for you, and you're his responsibility, you know? The AI is not responsible; that&#x2019;s just a tool.</p><attrib>Unspecified 44</attrib></disp-quote><disp-quote><p>So I'd rather know what they're observing, and if it&#x2019;s [AI] wrong, I would [want to be able to] correct it rather than have them just collect data and make assumptions.</p><attrib>Unspecified 44</attrib></disp-quote></sec><sec id="s3-5-2"><title>Acceptance Based on Professional Trust</title><p>Despite reservations about AI, patients accept it when recommended by their medical team due to trust in their physician or health care institution. Patients believe that the medical team&#x2019;s professional endorsement and experience enhance trust in AI systems. Many patients indicated that they would be more receptive if the medical team confirmed that the AI technology is thoroughly validated and offers benefits. Patients view physician recommendations and oversight as the primary basis for trusting AI technology. The professional team&#x2019;s opinion serves as a safety net for patients, particularly when emerging technologies remain unproven.</p><disp-quote><p>I trust it pretty much because I did ask how long they [the clinical team] have been doing LVAD, and I was surprised to find out that this had been going on longer than I thought it was... [<xref ref-type="bibr" rid="ref29">29</xref>]</p></disp-quote></sec></sec><sec id="s3-6"><title>Ethical Challenges of AI and Health Care Equity</title><sec id="s3-6-1"><title>Requirement of Ethical Standards</title><p>Patients believe that the application of AI in health care requires strict ethical standards, particularly when handling sensitive data. The absence of reasonable regulations may compromise patient rights. Many patients worry that AI research may prioritize speed over ethical standards, leading to data misuse or privacy violations. Additionally, patients emphasize the right to informed consent regarding AI interventions. Some patients believe health care institutions should disclose AI use before consultations, not during them, allowing time to process information and prepare mentally. Therefore, AI technology must operate under rigorous ethical review and legal frameworks to safeguard patients&#x2019; fundamental rights:</p><disp-quote><p>I am afraid that [...] those who do research would like to reach a result...[but] the result should be so well grounded that you can actually vouch for it all the way.</p><attrib>M30 30</attrib></disp-quote><disp-quote><p>I want to know about it [AI involvement] before I get to the doctor... I&#x2019;d rather have the opportunity to think about it and review it beforehand.</p><attrib>Participant 45</attrib></disp-quote></sec><sec id="s3-6-2"><title>Threats to Health Care Equity</title><p>AI&#x2019;s impact on health care equity presents a complex duality. Some patients fear AI may exacerbate existing inequalities&#x2014;low-income groups and the technologically disadvantaged may struggle to access AI-enabled health care equitably, thereby widening health disparities. However, evidence from marginalized communities reveals another possibility: minority patients view AI as a &#x201C;safe space&#x201D; enabling them to discuss sensitive issues without cultural shame, while multilingual capabilities compensate for dialect service gaps and bypass human &#x201C;gatekeepers&#x201D; perceived as biased.</p><disp-quote><p>At that moment, I didn't want to speak to anyone... If I had this tool, I would use it first before speaking to anyone else.</p><attrib>P01, male, aged 28 y 39</attrib></disp-quote><disp-quote><p>I feel that patients with fewer resources or access to technology will fall behind in terms of benefiting from AI in health care.</p><attrib>Unspecified 26</attrib></disp-quote></sec></sec><sec id="s3-7"><title>The Irreplaceability of Human Care</title><p>Patients believe that while AI can provide technical support, it cannot replace the human care and ethical judgment physicians offer in treatment. AI cannot comprehend patients&#x2019; emotional needs or life contexts, factors crucial to treatment decisions. Patients emphasized that AI should only serve as an auxiliary tool, unable to substitute for physicians&#x2019; roles in complex medical decisions&#x2014;particularly regarding emotional support and personalized treatment.</p><disp-quote><p>Humans can express emotions, empathy, help, and give hope for a better tomorrow better than any machine.</p><attrib>P_04 27</attrib></disp-quote></sec><sec id="s3-8"><title>Future Outlook for AI Applications</title><sec id="s3-8-1"><title>AI Technology Diffusion</title><p>Patients believe that the advancement of AI in health care is inevitable. Despite concerns, they acknowledge AI&#x2019;s potential to enhance efficiency and reduce errors, particularly in areas such as diagnostics, treatment planning, and patient management. As AI becomes widely adopted across global industries, patients recognize that it will become a vital component of future health care&#x2014;although privacy and ethical issues arising from the technology require further regulation to ensure that patient data are protected and that AI systems are used responsibly in clinical settings.</p><disp-quote><p>It&#x2019;s becoming more and more the case that robots are controlling much of everyday life in various professions, and it&#x2019;s happening everywhere.</p><attrib>M12 30</attrib></disp-quote></sec><sec id="s3-8-2"><title>Potential for AI-Physician Collaboration</title><p>Patients hold a positive view of AI collaborating with physicians, believing it can alleviate physicians&#x2019; workload and improve efficiency, but it cannot replace the physician&#x2019;s decision-making role. Patients hope AI serves as an auxiliary tool to assist physicians rather than dominate the treatment process. Especially in complex conditions, the physician&#x2019;s clinical judgment and humanistic care remain crucial.</p><disp-quote><p>I think the whole system makes sense, but it can&#x2019;t function independently.</p><attrib>P_13 27</attrib></disp-quote></sec></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study aimed to conduct an in-depth analysis of patients&#x2019; ethical concerns regarding AI medical applications, explore their impact on patient acceptance and trust, and use SET to reveal multilevel tension mechanisms. SET, which is based on Bronfenbrenner&#x2019;s ecological systems theory [<xref ref-type="bibr" rid="ref51">51</xref>], was later improved by researchers such as McLeroy and has become a well-known way to study public health and health behavior [<xref ref-type="bibr" rid="ref52">52</xref>,<xref ref-type="bibr" rid="ref53">53</xref>]. The theory&#x2019;s core assumption posits that individual behavior does not exist in isolation but is embedded within a multilevel, interacting environmental system. The microlevel focuses on individual knowledge, attitudes, and psychological characteristics; the mesolevel involves interpersonal relationships and organizational environments; and the macrolevel encompasses sociocultural norms and policy systems [<xref ref-type="bibr" rid="ref52">52</xref>,<xref ref-type="bibr" rid="ref54">54</xref>]. These 3 levels form a complex ecosystem through dynamic bidirectional influences, where changes in any one level may propagate to others.</p><p>Through a meta-synthesis of 25 qualitative studies, we identified 6 core themes: privacy and data security, technological reliability, impacts on physician-patient relationships, trust and accountability, ethical challenges and health care equity, and future perspectives. Socioecological analysis reveals that these concerns create mutually reinforcing &#x201C;ecological imbalances&#x201D; across micro (technological cognitive biases and data control anxieties), meso (broken physician-patient trust and institutional accountability gaps), and macro (health inequities and lagging ethical standards) levels, thereby hindering patient acceptance and trust in AI technologies. Using this framework, the following sections will examine these findings in detail.</p></sec><sec id="s4-2"><title>Microlevel: The Vicious Cycle of Technological Cognitive Bias and Data Control Anxiety</title><p>At the individual level, patients&#x2019; concerns about AI technology primarily manifest as cognitive uncertainty about the technology and anxiety over data control. SET emphasizes the interaction between individual behavior and multilevel environments, where individuals&#x2019; cognitive and emotional attitudes are influenced not only by internal factors but also closely tied to their social and cultural contexts [<xref ref-type="bibr" rid="ref55">55</xref>]. During the application of AI medical technologies, individuals&#x2019; cognitive biases about the technology are often accompanied by intense concerns about its transparency and controllability, which can lead to hesitance in adopting these technologies for their health care needs.</p><p>Research indicates that patients&#x2019; understanding of AI technology is closely linked to their acceptance of it [<xref ref-type="bibr" rid="ref56">56</xref>]. However, the complexity of AI technology and the black box nature of algorithms make it difficult for patients to comprehend its decision-making processes. This uncertainty directly leads to patients questioning the technology&#x2019;s accuracy [<xref ref-type="bibr" rid="ref57">57</xref>]. For instance, patients worry that AI cannot account for their individualized health variations&#x2014;such as differences in underlying conditions or lifestyle habits&#x2014;potentially compromising the effectiveness of medical interventions [<xref ref-type="bibr" rid="ref58">58</xref>]. Furthermore, patients&#x2019; doubts about AI accuracy have extended from the algorithmic level to the data level. Drawing from experiences reviewing their own electronic health records, they have identified numerous errors in the documentation. Consequently, they fear that even well-designed algorithms trained on such flawed data may struggle to produce reliable outputs. This technological cognitive bias does not exist in isolation; it is intertwined with individuals&#x2019; anxieties over data control. Patients&#x2019; concerns about data control cannot be reduced to instrumental worries about privacy breaches but should be understood as existential threats to their self-integrity and narrative sovereignty [<xref ref-type="bibr" rid="ref59">59</xref>]. Patients&#x2019; anxiety over data control extends beyond privacy leaks to a profound fear of losing self-determination [<xref ref-type="bibr" rid="ref60">60</xref>]. Medical data encapsulate patients&#x2019; life experiences, health statuses, and identity information. When processed by algorithms for unforeseen purposes, patients lose not only informational control but also dominion over their own health narratives [<xref ref-type="bibr" rid="ref61">61</xref>-<xref ref-type="bibr" rid="ref63">63</xref>].</p><p>SET posits that individual behavior is shaped by multiple interrelated factors [<xref ref-type="bibr" rid="ref55">55</xref>]. At the microlevel, patients&#x2019; technological cognitive biases and data control anxieties do not exist in isolation but form a vicious cycle through the &#x201C;cognition-anxiety&#x201D; interaction [<xref ref-type="bibr" rid="ref64">64</xref>]. Distrust of AI technology not only heightens patients&#x2019; concerns about data security but also deepens their skepticism regarding technological transparency and controllability [<xref ref-type="bibr" rid="ref65">65</xref>]. This vicious cycle may lead to patient resistance toward AI technology, subsequently affecting their overall attitude toward health care services. Similar perspectives are reflected in other studies. Research indicates that uncertainty about information technology and privacy concerns often amplify individual resistance, thereby influencing their acceptance of technology [<xref ref-type="bibr" rid="ref66">66</xref>].</p></sec><sec id="s4-3"><title>Mesolevel: Resonance Between Fractured Physician-Patient Trust and Institutional Accountability Deficits</title><p>At the mesolevel, patient concerns primarily manifest as fractures in physician-patient trust and inadequacies in health care organizational governance. Interpersonal relationships and organizational culture often constrain individual behavioral changes, according to SET [<xref ref-type="bibr" rid="ref55">55</xref>]. Patients&#x2019; apprehensions toward AI medical technologies stem not only from individual perceptions of the technology but are also closely intertwined with the quality of physician-patient relationships and the accountability of health care organizations.</p><p>Extensive research indicates that emotional empathy and interpersonal interaction between physicians and patients are crucial for building trust [<xref ref-type="bibr" rid="ref67">67</xref>-<xref ref-type="bibr" rid="ref69">69</xref>]. However, the introduction of AI has undermined this foundation to some extent. Studies reveal that when physicians overrely on AI technology, patients experience reduced interaction time and emotional support from their physicians, leading to diminished trust [<xref ref-type="bibr" rid="ref70">70</xref>]. The core of the physician-patient relationship lies in physicians viewing patients as whole individuals, not machines requiring repair [<xref ref-type="bibr" rid="ref71">71</xref>]. This relationship is based on listening, understanding, and being there for each other. Physicians know how much their patients are hurting and help them by talking to them [<xref ref-type="bibr" rid="ref72">72</xref>-<xref ref-type="bibr" rid="ref74">74</xref>]. AI intervention threatens the quality of this interpersonal engagement. When physicians&#x2019; attention shifts to screens, when diagnoses rely on algorithmic outputs, and when communication is replaced by standardized processes, patients transform from &#x201C;people receiving care&#x201D; into &#x201C;objects undergoing testing&#x201D; [<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref75">75</xref>].</p><p>Conversely, the absence of accountability mechanisms within health care organizations is another significant source of patient concern. Existing research indicates that many institutions lack clear AI liability frameworks, leaving patients unable to identify responsible parties when AI systems malfunction [<xref ref-type="bibr" rid="ref76">76</xref>]. This lack of accountability not only makes patients less trusting of AI technology but also makes health care organizations look less credible [<xref ref-type="bibr" rid="ref77">77</xref>]. SET emphasizes that deficiencies in internal organizational governance structures amplify individual-level concerns by eroding patients&#x2019; trust in the health care system [<xref ref-type="bibr" rid="ref78">78</xref>]. The absence of accountability in AI decision-making within health care organizations leaves patients without effective avenues for redress when encountering medical issues, thereby creating a mutually reinforcing negative cycle of &#x201C;accountability-trust&#x201D; [<xref ref-type="bibr" rid="ref79">79</xref>]. Similar &#x201C;accountability vacuum&#x201D; issues have been explored in other studies, with research indicating that patients&#x2019; trust in medical decisions is often severely compromised when health care organizations lack clear responsibility frameworks [<xref ref-type="bibr" rid="ref80">80</xref>].</p></sec><sec id="s4-4"><title>Macrolevel: Intensifying Issues of Health Care Equity and Lagging Ethical Standards</title><p>At the macrolevel, patient concerns regarding AI technology primarily center on health care equity and the lag in ethical regulations. SET asserts that sociocultural inequities and the lack of ethical standards directly impact individual behavioral choices [<xref ref-type="bibr" rid="ref81">81</xref>]. During the implementation of AI health care technologies, widespread adoption encounters obstacles arising from socioeconomic disparities. Low-income groups and patients with limited technological literacy often struggle to access AI medical services equitably. This gap exacerbates health care inequalities, fueling patient resistance toward AI technologies [<xref ref-type="bibr" rid="ref82">82</xref>,<xref ref-type="bibr" rid="ref83">83</xref>].</p><p>Not all socioeconomic strata equally benefit from the application of AI health care technologies, especially in regions with slower economic development or scarce resources [<xref ref-type="bibr" rid="ref84">84</xref>]. High-end medical facilities and advanced technological resources are concentrated in major cities and economically developed regions, while low-income communities and remote areas still face significant gaps in AI health care adoption [<xref ref-type="bibr" rid="ref85">85</xref>]. Furthermore, the high technical support and maintenance requirements of AI health care technologies make them unaffordable for many resource-constrained medical institutions, further exacerbating the unequal distribution of health care resources [<xref ref-type="bibr" rid="ref86">86</xref>]. In this context, the &#x201C;technological access barriers&#x201D; experienced by patients are not merely technical difficulties but deep-seated social problems stemming from unequal socioeconomic structures and resource allocation [<xref ref-type="bibr" rid="ref87">87</xref>].</p><p>Although the World Health Organization (WHO) has issued global AI ethics principles (WHO, 2021), implementation and regulatory rigor vary significantly across countries [<xref ref-type="bibr" rid="ref88">88</xref>]. The lag in ethical standards manifests not only at the policy level but also creates gaps in practical implementation. For instance, some countries may lack cross-regional ethical collaboration mechanisms, leading certain health care institutions to prioritize AI services for economically advantaged groups over low-income populations due to cost considerations [<xref ref-type="bibr" rid="ref89">89</xref>]. Such practices exacerbate health inequalities, deepening patients&#x2019; concerns that AI serves only select groups rather than benefiting the broader public [<xref ref-type="bibr" rid="ref90">90</xref>].</p></sec><sec id="s4-5"><title>Interaction Mechanisms and Mutual Influence Across Levels</title><p>According to SET, concerns at the microlevel, mesolevel, and macrolevel do not exist in isolation but form mutually reinforcing chain reactions through &#x201C;risk perception transmission&#x201D; [<xref ref-type="bibr" rid="ref55">55</xref>]. Research indicates that individual cognition and anxiety at the microlevel generate reverse effects at interpersonal and organizational levels, subsequently impacting broader social structures [<xref ref-type="bibr" rid="ref91">91</xref>]. At the microlevel, individuals&#x2019; cognitive biases regarding AI technology and anxieties over data control rights are transmitted to the mesolevel through interactions between individuals and health care organizations, increasing communication pressures for these organizations when deploying AI technologies [<xref ref-type="bibr" rid="ref92">92</xref>,<xref ref-type="bibr" rid="ref93">93</xref>]. Patients&#x2019; concerns about data privacy and security influence individual behavioral decisions and prompt health care institutions to reevaluate the boundaries of AI applications, thereby driving societal-level attention to AI ethical norms [<xref ref-type="bibr" rid="ref94">94</xref>].</p><p>The absence of accountability mechanisms for medical organizations at the mesolevel exacerbates microlevel cognitive biases and anxieties, leading to increased distrust among patients and further complicating the integration of AI technologies in health care [<xref ref-type="bibr" rid="ref93">93</xref>,<xref ref-type="bibr" rid="ref95">95</xref>]. Extensive research indicates that the breakdown of social support networks directly impacts individual health decisions and attitudes, with trust deficits further intensifying emotional distress and technological apprehension [<xref ref-type="bibr" rid="ref96">96</xref>-<xref ref-type="bibr" rid="ref98">98</xref>]. The breakdown of physician-patient trust not only reduces patient acceptance of AI technology but also leads patients to rely more on personal emotional judgments when facing medical decisions, overlooking the potential of AI technology [<xref ref-type="bibr" rid="ref99">99</xref>].</p><p>Social equity issues at the macrolevel further permeate the microlevel, particularly the rejection of AI health care services by low-income groups. This affects their acceptance of AI technology and reinforces fears of technological displacement through sociocultural perceptions [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref100">100</xref>]. SET indicates that sociocultural beliefs not only shape patients&#x2019; perceptions of AI technology through individual behavior but also influence organizational behavior via mesolevel cultural diffusion, thereby amplifying implementation barriers in unequal societies [<xref ref-type="bibr" rid="ref101">101</xref>,<xref ref-type="bibr" rid="ref102">102</xref>].</p></sec><sec id="s4-6"><title>Policy Recommendations: Specific Pathways and Response Strategies</title><p>Reflecting on the findings of this study, patients&#x2019; concerns are far from unfounded. Across the 25 studies included in this review, patients from diverse clinical contexts and cultural backgrounds consistently expressed anxieties about data security, skepticism toward algorithmic opacity, and questions about accountability attribution. These concerns point to objective limitations of AI medical technologies at their current stage of development and institutional gaps that remain unaddressed. Recent studies have independently corroborated the reality of these issues from various perspectives: data bias and privacy risks in clinical AI systems have been extensively documented [<xref ref-type="bibr" rid="ref103">103</xref>], the erosion of physician-patient trust caused by algorithmic inexplicability has attracted sustained attention [<xref ref-type="bibr" rid="ref104">104</xref>], and accountability attribution in AI-assisted medical decision-making still lacks clear legal delineation [<xref ref-type="bibr" rid="ref105">105</xref>]. The WHO, in its 2021 guidance, also identified transparency, accountability, and equity as core principles for the ethical governance of AI in health care [<xref ref-type="bibr" rid="ref106">106</xref>]. In other words, the concerns articulated by patients based on their lived experiences correspond precisely with the risks identified through systematic analyses by the academic community and international organizations. This implies that the goal of policy intervention should not be to &#x201C;correct patient misconceptions,&#x201D; but rather to substantively address these well-founded and evidence-based concerns. The following recommendations are organized across the microlevel, mesolevel, and macrolevel while acknowledging the implementation challenges inherent in each pathway.</p></sec><sec id="s4-7"><title>Microlevel: Incremental Transparency and Patient Empowerment</title><p>Given that the field of Explainable Artificial Intelligence remains in its developmental stages, fully visualizing algorithmic decision-making logic is not realistic in the short term; an incremental transparency strategy should therefore be adopted [<xref ref-type="bibr" rid="ref107">107</xref>]. Specifically, drawing on the access rights for data subjects under the European Union&#x2019;s General Data Protection Regulation [<xref ref-type="bibr" rid="ref108">108</xref>], health care institutions should be required to establish data access logging systems that enable patients to query the access records and intended purposes of their health data [<xref ref-type="bibr" rid="ref109">109</xref>]. To address patients&#x2019; concerns regarding the quality of training data, accessible mechanisms for medical record review and correction should be established, safeguarding patients&#x2019; rights to audit and amend their own medical records [<xref ref-type="bibr" rid="ref110">110</xref>]. Furthermore, informed consent regarding AI involvement should be front-loaded&#x2014;patients should be notified before their clinical encounter&#x2014;and channels for questioning AI judgments and requesting human review should be established [<xref ref-type="bibr" rid="ref111">111</xref>].</p></sec><sec id="s4-8"><title>Mesolevel: Rebuilding Trust and Clarifying Accountability</title><p>The principle of &#x201C;human oversight, AI assistance&#x201D; should be institutionally safeguarded. Rather than imposing rigid communication time benchmarks, mandatory &#x201C;physician confirmation checkpoints&#x201D; should be embedded within AI-assisted diagnostic and treatment workflows, ensuring that critical decisions are subject to physician review and explained to patients before implementation [<xref ref-type="bibr" rid="ref112">112</xref>]. Concurrently, performance evaluation systems should be adjusted to prevent efficiency-driven metrics from encroaching upon the space for physician-patient communication [<xref ref-type="bibr" rid="ref113">113</xref>]. Regarding accountability attribution, a tiered liability framework should be constructed that differentiates the responsible parties for data errors, algorithmic defects, and clinical misjudgments [<xref ref-type="bibr" rid="ref114">114</xref>], while acknowledging that existing legal frameworks contain gaps in the attribution of responsibility for AI-driven decisions, necessitating legislative follow-up [<xref ref-type="bibr" rid="ref115">115</xref>]. Additionally, clinical scenarios in which AI cannot substitute for human practitioners should be explicitly delineated, particularly those highly dependent on emotional support and clinical judgment, such as end-of-life care, mental health treatment, and complex decision-making situations that require empathy and nuanced understanding of patient needs [<xref ref-type="bibr" rid="ref116">116</xref>].</p></sec><sec id="s4-9"><title>Macrolevel: Advancing Equitable Access and Ethical Governance</title><p>The WHO has set out global ethical standards for AI in health care, which can be used to help create an international ethical framework. National governments should build upon this foundation to develop regulatory frameworks aligned with their respective health care systems and cultural traditions, rather than pursuing an unrealistic goal of globally uniform standards. At the level of resource allocation, investment should be increased for low-income populations and remote areas, and age-friendly and low-cost AI health care tools should be developed to narrow the gap in technological accessibility [<xref ref-type="bibr" rid="ref117">117</xref>]. However, it must be acknowledged that the barriers facing these regions extend beyond equipment scarcity to include insufficient technical maintenance capacity and digital health literacy, necessitating complementary capacity building and educational support [<xref ref-type="bibr" rid="ref118">118</xref>]. Furthermore, equity assessments should be incorporated into the market approval review of AI health care products, requiring developers to submit impact assessment reports for diverse population groups [<xref ref-type="bibr" rid="ref119">119</xref>].</p><p>Finally, it must be acknowledged that structural tensions exist in the advancement of AI in health care: technology developers&#x2019; pursuit of algorithmic efficiency and commercial returns may conflict with patients&#x2019; safety needs; health care institutions under cost pressures may find it difficult to reconcile these with adequate physician-patient communication. Confronting rather than evading these tensions is a prerequisite for formulating pragmatic policies [<xref ref-type="bibr" rid="ref120">120</xref>].</p></sec><sec id="s4-10"><title>Research Limitations</title><p>This study has the following limitations. First, the 25 included studies were primarily conducted in developed regions, such as North America, Europe, and Australia, with insufficient representation of patients from developing countries. This limits the generalizability of findings to resource-constrained settings. Second, the included studies covered a wide range of AI applications&#x2014;including imaging diagnostics, clinical decision support, and virtual health assistants&#x2014;where patient concerns may vary across different AI types. The thematic synthesis in this study may obscure such context-specific differences. Third, as a secondary analysis, the quality of the meta-synthesis depends on the reporting depth of the original studies. Some studies provided insufficient contextual information for patient quotes, hindering more nuanced interpretation. Finally, AI medical technology is undergoing rapid iteration. The literature included in this study reflects patient perceptions within a specific time window. As technological transparency and regulatory frameworks evolve, patient attitudes may change, potentially leading to increased trust in AI medical technology and greater acceptance of its use in health care. Future research should continuously track this dynamic process.</p></sec><sec id="s4-11"><title>Conclusions</title><p>This research, through a systematic synthesis of 25 qualitative studies, identified 6 primary patient concerns regarding AI health care applications: privacy and data security, technological reliability, the impact on physician-patient relationships, trust and accountability, ethics and fairness, and ambivalent attitudes toward future developments. Unlike previous reviews focusing on the general public, this study centers on patients as core stakeholders. It pioneers the application of SET to this field, revealing a &#x201C;disrupted ecological equilibrium&#x201D; mechanism that propagates across microlevel, mesolevel, and macrolevel. This provides an explanatory framework&#x2014;transcending descriptive induction&#x2014;for understanding the deep-seated reasons behind patient resistance to AI medical technologies. The findings offer direct implications for practice: clinical institutions should establish a &#x201C;human-led, AI-assisted&#x201D; diagnostic model; policymakers should accelerate liability legislation and prioritize equitable technology access; and developers should adopt incremental transparency strategies while providing patients with avenues for questioning and review. Future research may explore the following directions: incorporating perspectives from patients in more developing countries and resource-constrained regions to test cross-cultural applicability; conducting comparative studies across different AI application scenarios to explore the context-specific nature of concerns; and using longitudinal designs to track the dynamic evolution of patient attitudes as technology advances and regulatory frameworks mature.</p></sec></sec></body><back><ack><p>Generative artificial intelligence was not used in any portion of the manuscript writing.</p></ack><notes><sec><title>Funding</title><p>This study was supported by the Hunan Provincial Department of Education's Postgraduate Education Innovation and Professional Ability Improvement Project (grant Xiangjiaotong [2020] 172-26). The funder had no involvement in the study design, data collection, analysis, interpretation, or the writing of the manuscript.</p></sec><sec><title>Data Availability</title><p>Data sharing is not applicable to this study as no datasets were generated or analyzed during this study.</p></sec></notes><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">ENTREQ</term><def><p>Enhancing Transparency in Reporting the Synthesis of Qualitative Research</p></def></def-item><def-item><term id="abb3">GRADE-CERQual</term><def><p>Confidence in the Evidence from Reviews of Qualitative Research</p></def></def-item><def-item><term id="abb4">JBI-QARI</term><def><p>Joanna Briggs Institute Qualitative Assessment and Review Instrument</p></def></def-item><def-item><term id="abb5">PRISMA</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses</p></def></def-item><def-item><term id="abb6">PRISMA-S</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses literature search extension</p></def></def-item><def-item><term id="abb7">SET</term><def><p>social ecological theory</p></def></def-item><def-item><term id="abb8">WHO</term><def><p>World Health Organization</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zeb</surname><given-names>S</given-names> </name><name name-style="western"><surname>Fnu</surname><given-names>N</given-names> </name><name name-style="western"><surname>Abbasi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Fahad</surname><given-names>M</given-names> </name></person-group><article-title>AI in healthcare: revolutionizing diagnosis and therapy</article-title><source>ijmdsa</source><year>2024</year><volume>3</volume><issue>3</issue><fpage>118</fpage><lpage>128</lpage><pub-id pub-id-type="doi">10.47709/ijmdsa.v3i3.4546</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thirunavukarasu</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSJ</given-names> </name><name name-style="western"><surname>Elangovan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gutierrez</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>TF</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSW</given-names> </name></person-group><article-title>Large language models in medicine</article-title><source>Nat Med</source><year>2023</year><month>08</month><volume>29</volume><issue>8</issue><fpage>1930</fpage><lpage>1940</lpage><pub-id pub-id-type="doi">10.1038/s41591-023-02448-8</pub-id><pub-id pub-id-type="medline">37460753</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Joshi</surname><given-names>G</given-names> </name><name name-style="western"><surname>Jain</surname><given-names>A</given-names> </name><name name-style="western"><surname>Araveeti</surname><given-names>SR</given-names> </name><name name-style="western"><surname>Adhikari</surname><given-names>S</given-names> </name><name name-style="western"><surname>Garg</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bhandari</surname><given-names>M</given-names> </name></person-group><article-title>FDA-approved artificial intelligence and machine learning (AI/ML)-enabled medical devices: an updated landscape</article-title><source>Electronics (Basel)</source><year>2024</year><volume>13</volume><issue>3</issue><fpage>498</fpage><pub-id pub-id-type="doi">10.3390/electronics13030498</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Maslej</surname><given-names>N</given-names> </name><name name-style="western"><surname>Fattorini</surname><given-names>L</given-names> </name><name name-style="western"><surname>Perrault</surname><given-names>R</given-names> </name><name name-style="western"><surname>Gil</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Parli</surname><given-names>V</given-names> </name><name name-style="western"><surname>Kariuki</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Artificial Intelligence Index Report 2025</article-title><source>arXiv</source><year>2025</year><month>04</month><day>8</day><pub-id pub-id-type="doi">10.48550/arXiv.2504.07139</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Juang</surname><given-names>WC</given-names> </name><name name-style="western"><surname>Hsu</surname><given-names>MH</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>ZX</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>CM</given-names> </name></person-group><article-title>Developing an AI-assisted clinical decision support system to enhance in-patient holistic health care</article-title><source>PLOS ONE</source><year>2022</year><volume>17</volume><issue>10</issue><fpage>e0276501</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0276501</pub-id><pub-id pub-id-type="medline">36315554</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Spencer</surname><given-names>KR</given-names> </name><name name-style="western"><surname>Ulrickson</surname><given-names>M</given-names> </name><name name-style="western"><surname>Grierson</surname><given-names>P</given-names> </name><etal/></person-group><article-title>BASECAMP-1: an approach to improve patient screening efficiency and to provide large datasets for precision medicine clinical trials</article-title><source>JCO</source><year>2025</year><month>06</month><volume>43</volume><issue>16_suppl</issue><pub-id pub-id-type="doi">10.1200/JCO.2025.43.16_suppl.e23020</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Prakash</surname><given-names>S</given-names> </name><name name-style="western"><surname>Balaji</surname><given-names>JN</given-names> </name><name name-style="western"><surname>Joshi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Surapaneni</surname><given-names>KM</given-names> </name></person-group><article-title>Ethical conundrums in the application of artificial intelligence (AI) in healthcare-a scoping review of reviews</article-title><source>J Pers Med</source><year>2022</year><month>11</month><day>16</day><volume>12</volume><issue>11</issue><fpage>1914</fpage><pub-id pub-id-type="doi">10.3390/jpm12111914</pub-id><pub-id pub-id-type="medline">36422090</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Albashrawi</surname><given-names>MA</given-names> </name></person-group><article-title>AI adoption in healthcare: trust and privacy concerns</article-title><access-date>2026-04-13</access-date><conf-name>Proceedings of the International Conferences on Applied Computing 2022</conf-name><conf-date>Nov 8-10, 2022</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.computing-conf.org/wp-content/uploads/2022/11/3_ICWI2022_R_017.pdf">https://www.computing-conf.org/wp-content/uploads/2022/11/3_ICWI2022_R_017.pdf</ext-link></comment></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Gawankar</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nair</surname><given-names>S</given-names> </name><name name-style="western"><surname>Pawar</surname><given-names>V</given-names> </name><name name-style="western"><surname>Vhatkar</surname><given-names>A</given-names> </name><name name-style="western"><surname>Chavan</surname><given-names>P</given-names> </name></person-group><article-title>Patient privacy and data security in the era of AI-driven healthcare</article-title><conf-name>2024 8th International Conference on Computing, Communication, Control and Automation (ICCUBEA)</conf-name><conf-date>Aug 23, 2024</conf-date><conf-loc>Pune, India</conf-loc><fpage>1</fpage><lpage>6</lpage><pub-id pub-id-type="doi">10.1109/ICCUBEA61740.2024.10775004</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Williamson</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Prybutok</surname><given-names>VR</given-names> </name></person-group><article-title>Balancing privacy and progress: a review of privacy challenges, systemic oversight, and patient perceptions in AI-driven healthcare</article-title><source>Appl Sci (Basel)</source><year>2024</year><volume>14</volume><issue>2</issue><fpage>675</fpage><pub-id pub-id-type="doi">10.3390/app14020675</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nong</surname><given-names>P</given-names> </name><name name-style="western"><surname>Platt</surname><given-names>J</given-names> </name></person-group><article-title>Patients&#x2019; trust in health systems to use artificial intelligence</article-title><source>JAMA Netw Open</source><year>2025</year><month>02</month><day>3</day><volume>8</volume><issue>2</issue><fpage>e2460628</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2024.60628</pub-id><pub-id pub-id-type="medline">39951270</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dur&#x00E1;n</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Jongsma</surname><given-names>KR</given-names> </name></person-group><article-title>Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical AI</article-title><source>J Med Ethics</source><year>2021</year><month>03</month><day>18</day><volume>47</volume><issue>5</issue><fpage>329</fpage><lpage>335</lpage><pub-id pub-id-type="doi">10.1136/medethics-2020-106820</pub-id><pub-id pub-id-type="medline">33737318</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Byrnes</surname><given-names>J</given-names> </name><name name-style="western"><surname>Robinson</surname><given-names>M</given-names> </name></person-group><article-title>Transparency and authority concerns with using AI to make ethical recommendations in clinical settings</article-title><source>Nurs Ethics</source><year>2025</year><month>09</month><volume>32</volume><issue>6</issue><fpage>1749</fpage><lpage>1760</lpage><pub-id pub-id-type="doi">10.1177/09697330241307317</pub-id><pub-id pub-id-type="medline">39710586</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>D</given-names> </name><name name-style="western"><surname>Vegt</surname><given-names>N</given-names> </name><name name-style="western"><surname>Visch</surname><given-names>V</given-names> </name><name name-style="western"><surname>Bos-De Vos</surname><given-names>M</given-names> </name></person-group><article-title>How much decision power should (A)I have?: investigating patients&#x2019; preferences towards AI autonomy in healthcare decision making</article-title><conf-name>Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems</conf-name><conf-date>May 11, 2024</conf-date><pub-id pub-id-type="doi">10.1145/3613904.3642883</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Montemayor</surname><given-names>C</given-names> </name><name name-style="western"><surname>Halpern</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fairweather</surname><given-names>A</given-names> </name></person-group><article-title>In principle obstacles for empathic AI: why we can&#x2019;t replace human empathy in healthcare</article-title><source>AI Soc</source><year>2022</year><volume>37</volume><issue>4</issue><fpage>1353</fpage><lpage>1359</lpage><pub-id pub-id-type="doi">10.1007/s00146-021-01230-z</pub-id><pub-id pub-id-type="medline">34054228</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Witkowski</surname><given-names>K</given-names> </name><name name-style="western"><surname>Dougherty</surname><given-names>RB</given-names> </name><name name-style="western"><surname>Neely</surname><given-names>SR</given-names> </name></person-group><article-title>Public perceptions of artificial intelligence in healthcare: ethical concerns and opportunities for patient-centered care</article-title><source>BMC Med Ethics</source><year>2024</year><month>06</month><day>22</day><volume>25</volume><issue>1</issue><fpage>74</fpage><pub-id pub-id-type="doi">10.1186/s12910-024-01066-4</pub-id><pub-id pub-id-type="medline">38909180</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gao</surname><given-names>S</given-names> </name><name name-style="western"><surname>He</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>D</given-names> </name><name name-style="western"><surname>Lai</surname><given-names>K</given-names> </name></person-group><article-title>Public perception of artificial intelligence in medical care: content analysis of social media</article-title><source>J Med Internet Res</source><year>2020</year><month>07</month><day>13</day><volume>22</volume><issue>7</issue><fpage>e16649</fpage><pub-id pub-id-type="doi">10.2196/16649</pub-id><pub-id pub-id-type="medline">32673231</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Page</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>McKenzie</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Bossuyt</surname><given-names>PM</given-names> </name><etal/></person-group><article-title>The PRISMA 2020 statement: an updated guideline for reporting systematic reviews</article-title><source>BMJ</source><year>2021</year><month>03</month><day>29</day><volume>372</volume><fpage>n71</fpage><pub-id pub-id-type="doi">10.1136/bmj.n71</pub-id><pub-id pub-id-type="medline">33782057</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tong</surname><given-names>A</given-names> </name><name name-style="western"><surname>Flemming</surname><given-names>K</given-names> </name><name name-style="western"><surname>McInnes</surname><given-names>E</given-names> </name><name name-style="western"><surname>Oliver</surname><given-names>S</given-names> </name><name name-style="western"><surname>Craig</surname><given-names>J</given-names> </name></person-group><article-title>Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ</article-title><source>BMC Med Res Methodol</source><year>2012</year><month>11</month><day>27</day><volume>12</volume><issue>1</issue><fpage>181</fpage><pub-id pub-id-type="doi">10.1186/1471-2288-12-181</pub-id><pub-id pub-id-type="medline">23185978</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Noyes</surname><given-names>J</given-names> </name><name name-style="western"><surname>Booth</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cargo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Flemming</surname><given-names>K</given-names> </name><name name-style="western"><surname>Harden</surname><given-names>A</given-names> </name><name name-style="western"><surname>Harris</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Qualitative evidence</article-title><source>Cochrane Handbook for Systematic Reviews of Interventions</source><year>2019</year><publisher-name>John Wiley &#x0026; Sons</publisher-name><fpage>525</fpage><lpage>545</lpage><pub-id pub-id-type="doi">10.1002/9781119536604</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Borenstein</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hedges</surname><given-names>LV</given-names> </name><name name-style="western"><surname>Higgins</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Rothstein</surname><given-names>HR</given-names> </name></person-group><source>Introduction to Meta-Analysis</source><year>2021</year><publisher-name>John Wiley &#x0026;Amp; Sons</publisher-name><pub-id pub-id-type="other">1119558387</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lockwood</surname><given-names>C</given-names> </name><name name-style="western"><surname>Munn</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Porritt</surname><given-names>K</given-names> </name></person-group><article-title>Qualitative research synthesis: methodological guidance for systematic reviewers utilizing meta-aggregation</article-title><source>Int J Evid Based Healthc</source><year>2015</year><month>09</month><volume>13</volume><issue>3</issue><fpage>179</fpage><lpage>187</lpage><pub-id pub-id-type="doi">10.1097/XEB.0000000000000062</pub-id><pub-id pub-id-type="medline">26262565</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rethlefsen</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Kirtley</surname><given-names>S</given-names> </name><name name-style="western"><surname>Waffenschmidt</surname><given-names>S</given-names> </name><etal/></person-group><article-title>PRISMA-S: an extension to the PRISMA statement for reporting literature searches in systematic reviews</article-title><source>Syst Rev</source><year>2021</year><month>01</month><day>26</day><volume>10</volume><issue>1</issue><fpage>39</fpage><pub-id pub-id-type="doi">10.1186/s13643-020-01542-z</pub-id><pub-id pub-id-type="medline">33499930</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hannes</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lockwood</surname><given-names>CS</given-names> </name><name name-style="western"><surname>Pearson</surname><given-names>AS</given-names> </name></person-group><article-title>A comparative analysis of three online appraisal instruments&#x2019; ability to assess validity in qualitative research</article-title><source>Qual Health Res</source><year>2010</year><month>12</month><volume>20</volume><issue>12</issue><fpage>1736</fpage><lpage>1743</lpage><pub-id pub-id-type="doi">10.1177/1049732310378656</pub-id><pub-id pub-id-type="medline">20671302</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hurley</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Sonig</surname><given-names>A</given-names> </name><name name-style="western"><surname>Herrington</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Ethical considerations for integrating multimodal computer perception and neurotechnology</article-title><source>Front Hum Neurosci</source><year>2024</year><volume>18</volume><fpage>1332451</fpage><pub-id pub-id-type="doi">10.3389/fnhum.2024.1332451</pub-id><pub-id pub-id-type="medline">38435745</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Annamalai</surname><given-names>A</given-names> </name></person-group><article-title>Challenges of building, deploying, and using AI-enabled telepsychiatry platforms for clinical practice among urban Indians: a qualitative study</article-title><source>Indian J Psychol Med</source><year>2021</year><month>07</month><volume>43</volume><issue>4</issue><fpage>336</fpage><lpage>342</lpage><pub-id pub-id-type="doi">10.1177/0253717620973414</pub-id><pub-id pub-id-type="medline">34385728</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>&#x010C;artolovni</surname><given-names>A</given-names> </name><name name-style="western"><surname>Male&#x0161;evi&#x0107;</surname><given-names>A</given-names> </name><name name-style="western"><surname>Poslon</surname><given-names>L</given-names> </name></person-group><article-title>Critical analysis of the AI impact on the patient-physician relationship: a multi-stakeholder qualitative study</article-title><source>Digit Health</source><year>2023</year><volume>9</volume><fpage>20552076231220833</fpage><pub-id pub-id-type="doi">10.1177/20552076231220833</pub-id><pub-id pub-id-type="medline">38130798</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sujan</surname><given-names>MA</given-names> </name><name name-style="western"><surname>White</surname><given-names>S</given-names> </name><name name-style="western"><surname>Habli</surname><given-names>I</given-names> </name><name name-style="western"><surname>Reynolds</surname><given-names>N</given-names> </name></person-group><article-title>Stakeholder perceptions of the safety and assurance of artificial intelligence in healthcare</article-title><source>SSRN Journal</source><year>2022</year><pub-id pub-id-type="doi">10.2139/ssrn.4000675</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kostick-Quenet</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Lang</surname><given-names>BH</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>JN</given-names> </name><name name-style="western"><surname>Hurley</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Blumenthal-Barby</surname><given-names>J</given-names> </name></person-group><article-title>Trust criteria for artificial intelligence in health: normative and epistemic considerations</article-title><source>J Med Ethics</source><year>2024</year><month>08</month><volume>50</volume><issue>8</issue><fpage>544</fpage><lpage>551</lpage><pub-id pub-id-type="doi">10.1136/jme-2023-109338</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hesjedal</surname><given-names>MB</given-names> </name><name name-style="western"><surname>Lys&#x00F8;</surname><given-names>EH</given-names> </name><name name-style="western"><surname>Solbj&#x00F8;r</surname><given-names>M</given-names> </name><name name-style="western"><surname>Skolbekken</surname><given-names>JA</given-names> </name></person-group><article-title>Valuing good health care: how medical doctors, scientists and patients relate ethical challenges with artificial intelligence decision-making support tools in prostate cancer diagnostics to good health care</article-title><source>Sociol Health Illn</source><year>2024</year><month>11</month><volume>46</volume><issue>8</issue><fpage>1808</fpage><lpage>1827</lpage><pub-id pub-id-type="doi">10.1111/1467-9566.13818</pub-id><pub-id pub-id-type="medline">39037701</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McCradden</surname><given-names>MD</given-names> </name><name name-style="western"><surname>Baba</surname><given-names>A</given-names> </name><name name-style="western"><surname>Saha</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Ethical concerns around use of artificial intelligence in health care research from the perspective of patients with meningioma, caregivers and health care providers: a qualitative study</article-title><source>CMAJ Open</source><year>2020</year><volume>8</volume><issue>1</issue><fpage>E90</fpage><lpage>E95</lpage><pub-id pub-id-type="doi">10.9778/cmajo.20190151</pub-id><pub-id pub-id-type="medline">32071143</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Al-Anezi</surname><given-names>FM</given-names> </name></person-group><article-title>Exploring the use of ChatGPT as a virtual health coach for chronic disease management</article-title><source>Learn Health Syst</source><year>2024</year><month>07</month><volume>8</volume><issue>3</issue><fpage>e10406</fpage><pub-id pub-id-type="doi">10.1002/lrh2.10406</pub-id><pub-id pub-id-type="medline">39036525</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Freeman</surname><given-names>S</given-names> </name><name name-style="western"><surname>Stewart</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kaard</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Health consumers&#x2019; ethical concerns towards artificial intelligence in Australian emergency departments</article-title><source>Emerg Medicine Australasia</source><year>2024</year><month>10</month><volume>36</volume><issue>5</issue><fpage>768</fpage><lpage>776</lpage><pub-id pub-id-type="doi">10.1111/1742-6723.14449</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Viberg Johansson</surname><given-names>J</given-names> </name><name name-style="western"><surname>Dembrower</surname><given-names>K</given-names> </name><name name-style="western"><surname>Strand</surname><given-names>F</given-names> </name><name name-style="western"><surname>Grauman</surname><given-names>&#x00C5;</given-names> </name></person-group><article-title>Women&#x2019;s perceptions and attitudes towards the use of AI in mammography in Sweden: a qualitative interview study</article-title><source>BMJ Open</source><year>2024</year><month>02</month><day>14</day><volume>14</volume><issue>2</issue><fpage>e084014</fpage><pub-id pub-id-type="doi">10.1136/bmjopen-2024-084014</pub-id><pub-id pub-id-type="medline">38355190</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jeyakumar</surname><given-names>T</given-names> </name><name name-style="western"><surname>Younus</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Preparing for an artificial intelligence-enabled future: patient perspectives on engagement and health care professional training for adopting artificial intelligence technologies in health care settings</article-title><source>JMIR AI</source><year>2023</year><month>03</month><day>2</day><volume>2</volume><fpage>e40973</fpage><pub-id pub-id-type="doi">10.2196/40973</pub-id><pub-id pub-id-type="medline">38875561</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Haan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ongena</surname><given-names>YP</given-names> </name><name name-style="western"><surname>Hommes</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kwee</surname><given-names>TC</given-names> </name><name name-style="western"><surname>Yakar</surname><given-names>D</given-names> </name></person-group><article-title>A qualitative study to understand patient perspective on the use of artificial intelligence in radiology</article-title><source>J Am Coll Radiol</source><year>2019</year><month>10</month><volume>16</volume><issue>10</issue><fpage>1416</fpage><lpage>1419</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2018.12.043</pub-id><pub-id pub-id-type="medline">30878311</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Berger</surname><given-names>SA</given-names> </name><name name-style="western"><surname>H&#x00E5;land</surname><given-names>E</given-names> </name><name name-style="western"><surname>Solbj&#x00F8;r</surname><given-names>M</given-names> </name></person-group><article-title>Patient perspectives on trust in artificial intelligence-powered tools in prostate cancer diagnostics</article-title><source>Qual Health Res</source><year>2026</year><month>03</month><volume>36</volume><issue>2-3</issue><fpage>276</fpage><lpage>288</lpage><pub-id pub-id-type="doi">10.1177/10497323251387545</pub-id><pub-id pub-id-type="medline">41251145</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Khairat</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mehraby</surname><given-names>H</given-names> </name><name name-style="western"><surname>Masoumi</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Evaluation of cancer survivors&#x2019; experience of using AI-based conversational tools: qualitative study</article-title><source>JMIR Cancer</source><year>2025</year><month>11</month><day>14</day><volume>11</volume><fpage>e77390</fpage><pub-id pub-id-type="doi">10.2196/77390</pub-id><pub-id pub-id-type="medline">41237331</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Litchfield</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Delanerolle</surname><given-names>G</given-names> </name><name name-style="western"><surname>Harper</surname><given-names>L</given-names> </name><name name-style="western"><surname>Dunning</surname><given-names>S</given-names> </name></person-group><article-title>Understanding acceptability of AI triage tools amongst underserved populations: lessons from the early phases of co-production wIth Bangladeshi communities in Birmingham, UK</article-title><source>Health Expect</source><year>2025</year><month>12</month><volume>28</volume><issue>6</issue><fpage>e70523</fpage><pub-id pub-id-type="doi">10.1111/hex.70523</pub-id><pub-id pub-id-type="medline">41342429</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Omori</surname><given-names>M</given-names> </name><name name-style="western"><surname>Basnayake</surname><given-names>P</given-names> </name><name name-style="western"><surname>Keogh</surname><given-names>L</given-names> </name><name name-style="western"><surname>Frazer</surname><given-names>HML</given-names> </name><name name-style="western"><surname>Kunicki</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Lippey</surname><given-names>JF</given-names> </name></person-group><article-title>Developing a typology of women&#x2019;s attitudes towards AI use in the BreastScreen Programme-a qualitative study with BreastScreen Victoria clients</article-title><source>Health Expect</source><year>2025</year><month>10</month><volume>28</volume><issue>5</issue><fpage>e70415</fpage><pub-id pub-id-type="doi">10.1111/hex.70415</pub-id><pub-id pub-id-type="medline">40884416</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Funer</surname><given-names>F</given-names> </name><name name-style="western"><surname>Schneider</surname><given-names>D</given-names> </name><name name-style="western"><surname>Heyen</surname><given-names>NB</given-names> </name><etal/></person-group><article-title>Impacts of clinical decision support systems on the relationship, communication, and shared decision-making between health care professionals and patients: multistakeholder interview study</article-title><source>J Med Internet Res</source><year>2023</year><pub-id pub-id-type="doi">10.2196/55717</pub-id><pub-id pub-id-type="medline">39178023</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Giebel</surname><given-names>GD</given-names> </name><name name-style="western"><surname>Raszke</surname><given-names>P</given-names> </name><name name-style="western"><surname>Nowak</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Improving AI-based clinical decision support systems and their integration into care from the perspective of experts: interview study among different stakeholders</article-title><source>JMIR Med Inform</source><year>2025</year><month>07</month><day>7</day><volume>13</volume><fpage>e69688</fpage><pub-id pub-id-type="doi">10.2196/69688</pub-id><pub-id pub-id-type="medline">40623684</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ly</surname><given-names>A</given-names> </name><name name-style="western"><surname>Herse</surname><given-names>S</given-names> </name><name name-style="western"><surname>Williams</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Stapleton</surname><given-names>F</given-names> </name></person-group><article-title>Artificial intelligence for age&#x2010;related macular degeneration diagnosis in Australia: a novel qualitative interview study</article-title><source>Ophthalmic Physiologic Optic</source><year>2025</year><month>09</month><volume>45</volume><issue>6</issue><fpage>1282</fpage><lpage>1292</lpage><pub-id pub-id-type="doi">10.1111/opo.13542</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Richardson</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>C</given-names> </name><name name-style="western"><surname>Curtis</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Patient apprehensions about the use of artificial intelligence in healthcare</article-title><source>NPJ Digit Med</source><year>2021</year><month>09</month><day>21</day><volume>4</volume><issue>1</issue><fpage>140</fpage><pub-id pub-id-type="doi">10.1038/s41746-021-00509-1</pub-id><pub-id pub-id-type="medline">34548621</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Foresman</surname><given-names>G</given-names> </name><name name-style="western"><surname>Biro</surname><given-names>J</given-names> </name><name name-style="western"><surname>Tran</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Patient perspectives on artificial intelligence in health care: focus group study for diagnostic communication and tool implementation</article-title><source>J Particip Med</source><year>2024</year><volume>17</volume><fpage>e69564</fpage><lpage>e69564</lpage><pub-id pub-id-type="doi">10.2196/69564</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schneider</surname><given-names>D</given-names> </name><name name-style="western"><surname>Liedtke</surname><given-names>W</given-names> </name><name name-style="western"><surname>Klausen</surname><given-names>AD</given-names> </name><etal/></person-group><article-title>Indecision on the use of artificial intelligence in healthcare-a qualitative study of patient perspectives on trust, responsibility and self-determination using AI-CDSS</article-title><source>Digit Health</source><year>2025</year><volume>11</volume><fpage>20552076251339522</fpage><pub-id pub-id-type="doi">10.1177/20552076251339522</pub-id><pub-id pub-id-type="medline">40469779</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Citardi</surname><given-names>D</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Genc</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Shan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>X</given-names> </name></person-group><article-title>Patients&#x2019; perceptions of using artificial intelligence (AI)-based technology to comprehend radiology imaging data</article-title><source>Health Informatics J</source><year>2021</year><volume>27</volume><issue>2</issue><fpage>14604582211011215</fpage><pub-id pub-id-type="doi">10.1177/14604582211011215</pub-id><pub-id pub-id-type="medline">33913359</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gundlack</surname><given-names>J</given-names> </name><name name-style="western"><surname>Negash</surname><given-names>S</given-names> </name><name name-style="western"><surname>Thiel</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Artificial intelligence in medical care - patients&#x2019; perceptions on caregiving relationships and ethics: a qualitative study</article-title><source>Health Expect</source><year>2025</year><month>04</month><volume>28</volume><issue>2</issue><fpage>e70216</fpage><pub-id pub-id-type="doi">10.1111/hex.70216</pub-id><pub-id pub-id-type="medline">40094179</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Steerling</surname><given-names>E</given-names> </name><name name-style="western"><surname>Svedberg</surname><given-names>P</given-names> </name><name name-style="western"><surname>Nilsen</surname><given-names>P</given-names> </name><name name-style="western"><surname>Siira</surname><given-names>E</given-names> </name><name name-style="western"><surname>Nygren</surname><given-names>J</given-names> </name></person-group><article-title>Influences on trust in the use of AI-based triage-an interview study with primary healthcare professionals and patients in Sweden</article-title><source>Front Digit Health</source><year>2025</year><volume>7</volume><fpage>1565080</fpage><pub-id pub-id-type="doi">10.3389/fdgth.2025.1565080</pub-id><pub-id pub-id-type="medline">40463579</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lewin</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bohren</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rashidian</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Applying GRADE-CERQual to qualitative evidence synthesis findings-paper 2: how to make an overall CERQual assessment of confidence and create a Summary of Qualitative Findings table</article-title><source>Implement Sci</source><year>2018</year><month>01</month><day>25</day><volume>13</volume><issue>Suppl 1</issue><fpage>10</fpage><pub-id pub-id-type="doi">10.1186/s13012-017-0689-2</pub-id><pub-id pub-id-type="medline">29384082</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bronfenbrenner</surname><given-names>U</given-names> </name></person-group><article-title>Toward an experimental ecology of human development</article-title><source>American Psychologist</source><year>1977</year><volume>32</volume><issue>7</issue><fpage>513</fpage><lpage>531</lpage><pub-id pub-id-type="doi">10.1037/0003-066X.32.7.513</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McLeroy</surname><given-names>KR</given-names> </name><name name-style="western"><surname>Bibeau</surname><given-names>D</given-names> </name><name name-style="western"><surname>Steckler</surname><given-names>A</given-names> </name><name name-style="western"><surname>Glanz</surname><given-names>K</given-names> </name></person-group><article-title>An ecological perspective on health promotion programs</article-title><source>Health Educ Q</source><year>1988</year><month>12</month><volume>15</volume><issue>4</issue><fpage>351</fpage><lpage>377</lpage><pub-id pub-id-type="doi">10.1177/109019818801500401</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thaivalappil</surname><given-names>A</given-names> </name><name name-style="western"><surname>Young</surname><given-names>I</given-names> </name><name name-style="western"><surname>Papadopoulos</surname><given-names>A</given-names> </name></person-group><article-title>More than a framework? Applying socioecological models within mental health and community psychology inquiries</article-title><source>Canadian Journal of Community Mental Health</source><year>2024</year><month>09</month><day>1</day><volume>43</volume><issue>3</issue><fpage>37</fpage><lpage>45</lpage><pub-id pub-id-type="doi">10.7870/cjcmh-2024-020</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="book"><person-group person-group-type="editor"><name name-style="western"><surname>Bronfenbrenner</surname><given-names>U</given-names> </name></person-group><source>The Ecology of Human Development: Experiments by Nature and Design</source><year>1979</year><publisher-name>Harvard University Press</publisher-name><pub-id pub-id-type="doi">10.4159/9780674028845</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roura</surname><given-names>M</given-names> </name></person-group><article-title>The social ecology of power in participatory health research</article-title><source>Qual Health Res</source><year>2021</year><month>03</month><volume>31</volume><issue>4</issue><fpage>778</fpage><lpage>788</lpage><pub-id pub-id-type="doi">10.1177/1049732320979187</pub-id><pub-id pub-id-type="medline">33349170</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>K</given-names> </name></person-group><article-title>Acceptance of medical treatment regimens provided by AI vs. human</article-title><source>Appl Sci (Basel)</source><year>2021</year><volume>12</volume><issue>1</issue><fpage>110</fpage><pub-id pub-id-type="doi">10.3390/app12010110</pub-id><pub-id pub-id-type="medline">34386269</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hennrich</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ritz</surname><given-names>E</given-names> </name><name name-style="western"><surname>Hofmann</surname><given-names>P</given-names> </name><name name-style="western"><surname>Urbach</surname><given-names>N</given-names> </name></person-group><article-title>Capturing artificial intelligence applications&#x2019; value proposition in healthcare - a qualitative research study</article-title><source>BMC Health Serv Res</source><year>2024</year><month>04</month><day>3</day><volume>24</volume><issue>1</issue><fpage>420</fpage><pub-id pub-id-type="doi">10.1186/s12913-024-10894-4</pub-id><pub-id pub-id-type="medline">38570809</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Babel</surname><given-names>A</given-names> </name><name name-style="western"><surname>Taneja</surname><given-names>R</given-names> </name><name name-style="western"><surname>Mondello Malvestiti</surname><given-names>F</given-names> </name><name name-style="western"><surname>Monaco</surname><given-names>A</given-names> </name><name name-style="western"><surname>Donde</surname><given-names>S</given-names> </name></person-group><article-title>Artificial intelligence solutions to increase medication adherence in patients with non-communicable diseases</article-title><source>Front Digit Health</source><year>2021</year><volume>3</volume><fpage>669869</fpage><pub-id pub-id-type="doi">10.3389/fdgth.2021.669869</pub-id><pub-id pub-id-type="medline">34713142</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lalova-Spinks</surname><given-names>T</given-names> </name><name name-style="western"><surname>Saesen</surname><given-names>R</given-names> </name><name name-style="western"><surname>Silva</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Patients&#x2019; knowledge, preferences, and perspectives about data protection and data control: an exploratory survey</article-title><source>Front Pharmacol</source><year>2023</year><volume>14</volume><fpage>1280173</fpage><pub-id pub-id-type="doi">10.3389/fphar.2023.1280173</pub-id><pub-id pub-id-type="medline">38445168</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Narkhede</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Wankhede</surname><given-names>NI</given-names> </name><name name-style="western"><surname>Kamble</surname><given-names>AM</given-names> </name></person-group><article-title>Enhancing patient autonomy in data ownership: privacy models and consent frameworks for healthcare</article-title><source>JDH</source><year>2025</year><fpage>1</fpage><lpage>23</lpage><pub-id pub-id-type="doi">10.55976/jdh.4202513361-23</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hahn</surname><given-names>E</given-names> </name></person-group><article-title>Das &#x201C;Recht auf Nichtwissen&#x201D; des Patienten bei algorithmengesteuerter Auswertung von Big Data</article-title><source>MedR</source><year>2019</year><month>03</month><volume>37</volume><issue>3</issue><fpage>197</fpage><lpage>202</lpage><pub-id pub-id-type="doi">10.1007/s00350-019-5176-0</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kalkman</surname><given-names>S</given-names> </name><name name-style="western"><surname>van Delden</surname><given-names>J</given-names> </name><name name-style="western"><surname>Banerjee</surname><given-names>A</given-names> </name><name name-style="western"><surname>Tyl</surname><given-names>B</given-names> </name><name name-style="western"><surname>Mostert</surname><given-names>M</given-names> </name><name name-style="western"><surname>van Thiel</surname><given-names>G</given-names> </name></person-group><article-title>Patients&#x2019; and public views and attitudes towards the sharing of health data for research: a narrative review of the empirical evidence</article-title><source>J Med Ethics</source><year>2022</year><month>01</month><volume>48</volume><issue>1</issue><fpage>3</fpage><lpage>13</lpage><pub-id pub-id-type="doi">10.1136/medethics-2019-105651</pub-id><pub-id pub-id-type="medline">31719155</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zolnoori</surname><given-names>M</given-names> </name><name name-style="western"><surname>Balls-Berry</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Brockman</surname><given-names>TA</given-names> </name><name name-style="western"><surname>Patten</surname><given-names>CA</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>M</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>L</given-names> </name></person-group><article-title>A systematic framework for analyzing patient-generated narrative data: protocol for a content analysis</article-title><source>JMIR Res Protoc</source><year>2019</year><month>08</month><day>26</day><volume>8</volume><issue>8</issue><fpage>13914</fpage><pub-id pub-id-type="doi">10.2196/13914</pub-id><pub-id pub-id-type="medline">31452524</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Norris</surname><given-names>S</given-names> </name><name name-style="western"><surname>Salgado</surname><given-names>F</given-names> </name><name name-style="western"><surname>Murray</surname><given-names>S</given-names> </name><name name-style="western"><surname>Amen</surname><given-names>D</given-names> </name><name name-style="western"><surname>Keator</surname><given-names>DB</given-names> </name></person-group><article-title>The role of negativity bias in emotional and cognitive dysregulation: a neuroimaging study in anxiety disorders</article-title><source>Depress Anxiety</source><year>2025</year><month>01</month><volume>2025</volume><issue>1</issue><pub-id pub-id-type="doi">10.1155/da/2739947</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Clements</surname><given-names>W</given-names> </name><name name-style="western"><surname>Thong</surname><given-names>LP</given-names> </name><name name-style="western"><surname>Zia</surname><given-names>A</given-names> </name><name name-style="western"><surname>Moriarty</surname><given-names>HK</given-names> </name><name name-style="western"><surname>Goh</surname><given-names>GS</given-names> </name></person-group><article-title>A prospective study assessing patient perception of the use of artificial intelligence in radiology</article-title><source>APJHM</source><year>2022</year><pub-id pub-id-type="doi">10.24083/apjhm.v17i1.861</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dhagarra</surname><given-names>D</given-names> </name><name name-style="western"><surname>Goswami</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>G</given-names> </name></person-group><article-title>Impact of trust and privacy concerns on technology acceptance in healthcare: an Indian perspective</article-title><source>Int J Med Inform</source><year>2020</year><month>09</month><volume>141</volume><issue>104164</issue><fpage>104164</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2020.104164</pub-id><pub-id pub-id-type="medline">32593847</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arshad</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sriram</surname><given-names>S</given-names> </name><name name-style="western"><surname>Khan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Gollapalli</surname><given-names>PK</given-names> </name><name name-style="western"><surname>Albadrani</surname><given-names>MS</given-names> </name></person-group><article-title>Mediating role of physician&#x2019;s empathy between physician&#x2019;s communication and patient&#x2019;s satisfaction</article-title><source>J Family Med Prim Care</source><year>2024</year><month>04</month><volume>13</volume><issue>4</issue><fpage>1530</fpage><lpage>1534</lpage><pub-id pub-id-type="doi">10.4103/jfmpc.jfmpc_1615_23</pub-id><pub-id pub-id-type="medline">38827677</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Karakaya</surname><given-names>&#x00C7;</given-names> </name><name name-style="western"><surname>S&#x00FC;mer</surname><given-names>FE</given-names> </name><name name-style="western"><surname>Demir</surname><given-names>ZG</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Durmaz</surname><given-names>e</given-names> </name></person-group><source>Is Communication A Receipt For Doctors? The Relationship Between Communication Competence And Trust: A Research On Breast Patients</source><year>2021</year><publisher-name>Marmara University</publisher-name><pub-id pub-id-type="doi">10.17829/turcom.861934</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>L</given-names> </name><name name-style="western"><surname>Zhai</surname><given-names>F</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wan</surname><given-names>C</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zeng</surname><given-names>L</given-names> </name></person-group><article-title>Construction of trust relationship between doctors and patients: a social psychological analysis</article-title><source>Environ Soc Psychol</source><year>2022</year><volume>6</volume><issue>2</issue><pub-id pub-id-type="doi">10.18063/esp.v6.i2.1395</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Esmaeilzadeh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Mirzaei</surname><given-names>T</given-names> </name><name name-style="western"><surname>Dharanikota</surname><given-names>S</given-names> </name></person-group><article-title>Patients&#x2019; perceptions toward human-artificial intelligence interaction in health care: experimental study</article-title><source>J Med Internet Res</source><year>2021</year><month>11</month><day>25</day><volume>23</volume><issue>11</issue><fpage>e25856</fpage><pub-id pub-id-type="doi">10.2196/25856</pub-id><pub-id pub-id-type="medline">34842535</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ventres</surname><given-names>WB</given-names> </name></person-group><article-title>Looking within: intentions of practice for person-centered care</article-title><source>Ann Fam Med</source><year>2017</year><month>03</month><volume>15</volume><issue>2</issue><fpage>171</fpage><lpage>174</lpage><pub-id pub-id-type="doi">10.1370/afm.2037</pub-id><pub-id pub-id-type="medline">28289119</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fuehrer</surname><given-names>S</given-names> </name><name name-style="western"><surname>Weil</surname><given-names>AB</given-names> </name><name name-style="western"><surname>Osterberg</surname><given-names>LG</given-names> </name><name name-style="western"><surname>Zulman</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Meunier</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Schwartz</surname><given-names>R</given-names> </name></person-group><article-title>Building authentic connection in the patient-physician relationship</article-title><source>J Prim Care Community Health</source><year>2024</year><volume>15</volume><fpage>21501319231225996</fpage><pub-id pub-id-type="doi">10.1177/21501319231225996</pub-id><pub-id pub-id-type="medline">38281122</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kishton</surname><given-names>R</given-names> </name><name name-style="western"><surname>Patel</surname><given-names>H</given-names> </name><name name-style="western"><surname>Saini</surname><given-names>D</given-names> </name><name name-style="western"><surname>Millstein</surname><given-names>J</given-names> </name><name name-style="western"><surname>Levy</surname><given-names>A</given-names> </name></person-group><article-title>Listening as medicine: a thematic analysis</article-title><source>Patient Exp J</source><year>2023</year><volume>10</volume><issue>1</issue><fpage>64</fpage><lpage>71</lpage><pub-id pub-id-type="doi">10.35680/2372-0247.1728</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sakumoto, MD</surname><given-names>M</given-names> </name><name name-style="western"><surname>Joshi, MD</surname><given-names>A</given-names> </name></person-group><article-title>Digital Empathy 2.0: connecting with patients using the written word</article-title><source>THMT</source><year>2023</year><volume>8</volume><issue>5</issue><pub-id pub-id-type="doi">10.30953/thmt.v8.433</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Perivolaris</surname><given-names>A</given-names> </name><name name-style="western"><surname>Adams-McGavin</surname><given-names>C</given-names> </name><name name-style="western"><surname>Madan</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Quality of interaction between clinicians and artificial intelligence systems. A systematic review</article-title><source>Future Healthc J</source><year>2024</year><month>09</month><volume>11</volume><issue>3</issue><fpage>100172</fpage><pub-id pub-id-type="doi">10.1016/j.fhj.2024.100172</pub-id><pub-id pub-id-type="medline">39281326</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cestonaro</surname><given-names>C</given-names> </name><name name-style="western"><surname>Delicati</surname><given-names>A</given-names> </name><name name-style="western"><surname>Marcante</surname><given-names>B</given-names> </name><name name-style="western"><surname>Caenazzo</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tozzo</surname><given-names>P</given-names> </name></person-group><article-title>Defining medical liability when artificial intelligence is applied on diagnostic algorithms: a systematic review</article-title><source>Front Med (Lausanne)</source><year>2023</year><volume>10</volume><fpage>1305756</fpage><pub-id pub-id-type="doi">10.3389/fmed.2023.1305756</pub-id><pub-id pub-id-type="medline">38089864</pub-id></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>ZM</given-names> </name></person-group><article-title>Ethics and governance of trustworthy medical artificial intelligence</article-title><source>BMC Med Inform Decis Mak</source><year>2023</year><month>01</month><day>13</day><volume>23</volume><issue>1</issue><fpage>7</fpage><pub-id pub-id-type="doi">10.1186/s12911-023-02103-9</pub-id><pub-id pub-id-type="medline">36639799</pub-id></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Platt</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Nong</surname><given-names>P</given-names> </name></person-group><article-title>An ecosystem approach to earning and sustaining trust in health care-too big to care</article-title><source>JAMA Health Forum</source><year>2023</year><month>01</month><day>6</day><volume>4</volume><issue>1</issue><fpage>e224882</fpage><pub-id pub-id-type="doi">10.1001/jamahealthforum.2022.4882</pub-id><pub-id pub-id-type="medline">36637813</pub-id></nlm-citation></ref><ref id="ref79"><label>79</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lang</surname><given-names>BH</given-names> </name><name name-style="western"><surname>Nyholm</surname><given-names>S</given-names> </name><name name-style="western"><surname>Blumenthal-Barby</surname><given-names>J</given-names> </name></person-group><article-title>Responsibility gaps and black box healthcare AI: shared responsibilization as a solution</article-title><source>Digit Soc</source><year>2023</year><month>12</month><volume>2</volume><issue>3</issue><fpage>52</fpage><pub-id pub-id-type="doi">10.1007/s44206-023-00073-z</pub-id><pub-id pub-id-type="medline">38596344</pub-id></nlm-citation></ref><ref id="ref80"><label>80</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>B</given-names> </name><name name-style="western"><surname>Asan</surname><given-names>O</given-names> </name><name name-style="western"><surname>Mansouri</surname><given-names>M</given-names> </name></person-group><article-title>Perspectives of patients with chronic diseases on future acceptance of AI-based home care systems: cross-sectional web-based survey study</article-title><source>JMIR Hum Factors</source><year>2023</year><month>11</month><day>6</day><volume>10</volume><fpage>e49788</fpage><pub-id pub-id-type="doi">10.2196/49788</pub-id><pub-id pub-id-type="medline">37930780</pub-id></nlm-citation></ref><ref id="ref81"><label>81</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nirmalasary</surname><given-names>TN</given-names> </name><name name-style="western"><surname>Rejo</surname><given-names>A</given-names> </name><name name-style="western"><surname>Windusari</surname><given-names>Y</given-names> </name></person-group><article-title>Analysis model of socio-cultural environment on criminality rate: a rural case study</article-title><source>J of Law and Sust Develop</source><year>2024</year><volume>12</volume><issue>5</issue><fpage>e3670</fpage><pub-id pub-id-type="doi">10.55908/sdgs.v12i5.3670</pub-id></nlm-citation></ref><ref id="ref82"><label>82</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kambala</surname><given-names>M</given-names> </name><collab>Elevance Health</collab></person-group><article-title>AI-powered healthcare: transforming patient outcomes with machine learning</article-title><source>JMSCR</source><year>2024</year><volume>12</volume><issue>8</issue><fpage>34</fpage><lpage>47</lpage><pub-id pub-id-type="doi">10.18535/jmscr/v12i08.07</pub-id></nlm-citation></ref><ref id="ref83"><label>83</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Veinot</surname><given-names>TC</given-names> </name><name name-style="western"><surname>Mitchell</surname><given-names>H</given-names> </name><name name-style="western"><surname>Ancker</surname><given-names>JS</given-names> </name></person-group><article-title>Good intentions are not enough: how informatics interventions can worsen inequality</article-title><source>J Am Med Inform Assoc</source><year>2018</year><month>08</month><day>1</day><volume>25</volume><issue>8</issue><fpage>1080</fpage><lpage>1088</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocy052</pub-id><pub-id pub-id-type="medline">29788380</pub-id></nlm-citation></ref><ref id="ref84"><label>84</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wahl</surname><given-names>B</given-names> </name><name name-style="western"><surname>Cossy-Gantner</surname><given-names>A</given-names> </name><name name-style="western"><surname>Germann</surname><given-names>S</given-names> </name><name name-style="western"><surname>Schwalbe</surname><given-names>NR</given-names> </name></person-group><article-title>Artificial intelligence (AI) and global health: how can AI contribute to health in resource-poor settings?</article-title><source>BMJ Glob Health</source><year>2018</year><volume>3</volume><issue>4</issue><fpage>e000798</fpage><pub-id pub-id-type="doi">10.1136/bmjgh-2018-000798</pub-id><pub-id pub-id-type="medline">30233828</pub-id></nlm-citation></ref><ref id="ref85"><label>85</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bu</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Qi</surname><given-names>L</given-names> </name><name name-style="western"><surname>He</surname><given-names>C</given-names> </name><name name-style="western"><surname>Liren</surname><given-names>S</given-names> </name><name name-style="western"><surname>Xin</surname><given-names>M</given-names> </name></person-group><article-title>Supervision and application analysis of AI medical devices</article-title><source>Social Medicine and Health Management</source><year>2023</year><volume>4</volume><issue>2</issue><pub-id pub-id-type="doi">10.23977/socmhm.2023.040209</pub-id></nlm-citation></ref><ref id="ref86"><label>86</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Parikh</surname><given-names>S</given-names> </name><name name-style="western"><surname>Costa</surname><given-names>A</given-names> </name></person-group><article-title>A critical look into artificial intelligence and healthcare disparities</article-title><source>Front Artif Intell</source><year>2025</year><volume>8</volume><fpage>1545869</fpage><pub-id pub-id-type="doi">10.3389/frai.2025.1545869</pub-id><pub-id pub-id-type="medline">40115119</pub-id></nlm-citation></ref><ref id="ref87"><label>87</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Weiss</surname><given-names>D</given-names> </name><name name-style="western"><surname>Rydland</surname><given-names>HT</given-names> </name><name name-style="western"><surname>&#x00D8;versveen</surname><given-names>E</given-names> </name><name name-style="western"><surname>Jensen</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Solhaug</surname><given-names>S</given-names> </name><name name-style="western"><surname>Krokstad</surname><given-names>S</given-names> </name></person-group><article-title>Innovative technologies and social inequalities in health: a scoping review of the literature</article-title><source>PLoS ONE</source><year>2018</year><volume>13</volume><issue>4</issue><fpage>e0195447</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0195447</pub-id><pub-id pub-id-type="medline">29614114</pub-id></nlm-citation></ref><ref id="ref88"><label>88</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gattinger</surname><given-names>G</given-names> </name></person-group><article-title>The evolving regulatory paradigm of AI in MedTech: a review of perspectives and where we are today</article-title><source>Ther Innov Regul Sci</source><year>2024</year><month>05</month><volume>58</volume><issue>3</issue><fpage>456</fpage><lpage>464</lpage><pub-id pub-id-type="doi">10.1007/s43441-024-00628-3</pub-id><pub-id pub-id-type="medline">38528278</pub-id></nlm-citation></ref><ref id="ref89"><label>89</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jobin</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ienca</surname><given-names>M</given-names> </name><name name-style="western"><surname>Vayena</surname><given-names>E</given-names> </name></person-group><article-title>Artificial intelligence: the global landscape of ethics guidelines</article-title><source>Nat Mach Intell</source><year>2019</year><volume>1</volume><fpage>389</fpage><lpage>399</lpage><pub-id pub-id-type="doi">10.1038/s42256-019-0088-2</pub-id></nlm-citation></ref><ref id="ref90"><label>90</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>d&#x2019;Elia</surname><given-names>A</given-names> </name><name name-style="western"><surname>Gabbay</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rodgers</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Artificial intelligence and health inequities in primary care: a systematic scoping review and framework</article-title><source>Fam Med Com Health</source><year>2022</year><month>11</month><volume>10</volume><issue>Suppl 1</issue><fpage>e001670</fpage><pub-id pub-id-type="doi">10.1136/fmch-2022-001670</pub-id></nlm-citation></ref><ref id="ref91"><label>91</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>The mediation of trust on artificial intelligence anxiety and continuous adoption of artificial intelligence technology among primacy nurses: a cross-sectional study</article-title><source>BMC Nurs</source><year>2025</year><month>07</month><day>1</day><volume>24</volume><issue>1</issue><fpage>724</fpage><pub-id pub-id-type="doi">10.1186/s12912-025-03406-0</pub-id><pub-id pub-id-type="medline">40597194</pub-id></nlm-citation></ref><ref id="ref92"><label>92</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Frenkenberg</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hochman</surname><given-names>G</given-names> </name></person-group><article-title>It&#x2019;s scary to use it, it&#x2019;s scary to refuse it: the psychological dimensions of AI Adoption&#x2014;anxiety, motives, and dependency</article-title><source>Systems</source><year>2025</year><volume>13</volume><issue>2</issue><fpage>82</fpage><pub-id pub-id-type="doi">10.3390/systems13020082</pub-id></nlm-citation></ref><ref id="ref93"><label>93</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huo</surname><given-names>W</given-names> </name><name name-style="western"><surname>Yuan</surname><given-names>X</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>W</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>B</given-names> </name></person-group><article-title>Increasing acceptance of medical AI: the role of medical staff participation in AI development</article-title><source>Int J Med Inform</source><year>2023</year><month>07</month><volume>175</volume><fpage>105073</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2023.105073</pub-id><pub-id pub-id-type="medline">37119693</pub-id></nlm-citation></ref><ref id="ref94"><label>94</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rony</surname><given-names>MKK</given-names> </name><name name-style="western"><surname>Numan</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Akter</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Nurses&#x2019; perspectives on privacy and ethical concerns regarding artificial intelligence adoption in healthcare</article-title><source>Heliyon</source><year>2024</year><month>09</month><day>15</day><volume>10</volume><issue>17</issue><fpage>e36702</fpage><pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e36702</pub-id><pub-id pub-id-type="medline">39281626</pub-id></nlm-citation></ref><ref id="ref95"><label>95</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flugelman</surname><given-names>MY</given-names> </name><name name-style="western"><surname>Jaffe</surname><given-names>R</given-names> </name><name name-style="western"><surname>Luria</surname><given-names>G</given-names> </name><name name-style="western"><surname>Yagil</surname><given-names>D</given-names> </name></person-group><article-title>Trust in the referring physician reduces anxiety in an integrated community-to-hospital care system</article-title><source>Isr J Health Policy Res</source><year>2020</year><month>05</month><day>11</day><volume>9</volume><issue>1</issue><fpage>7</fpage><pub-id pub-id-type="doi">10.1186/s13584-020-00365-6</pub-id><pub-id pub-id-type="medline">32393391</pub-id></nlm-citation></ref><ref id="ref96"><label>96</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Coffey</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hannigan</surname><given-names>B</given-names> </name><name name-style="western"><surname>Meudell</surname><given-names>A</given-names> </name><name name-style="western"><surname>Jones</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hunt</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fitzsimmons</surname><given-names>D</given-names> </name></person-group><article-title>Quality of life, recovery and decision-making: a mixed methods study of mental health recovery in social care</article-title><source>Soc Psychiatry Psychiatr Epidemiol</source><year>2019</year><month>06</month><volume>54</volume><issue>6</issue><fpage>715</fpage><lpage>723</lpage><pub-id pub-id-type="doi">10.1007/s00127-018-1635-6</pub-id><pub-id pub-id-type="medline">30470882</pub-id></nlm-citation></ref><ref id="ref97"><label>97</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name></person-group><article-title>Seeking health information on social media: a perspective of trust, self-determination, and social support</article-title><source>J Organ End User Comput</source><year>2018</year><volume>30</volume><fpage>1</fpage><lpage>22</lpage><pub-id pub-id-type="doi">10.4018/JOEUC.2018010101</pub-id></nlm-citation></ref><ref id="ref98"><label>98</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zeb</surname><given-names>R</given-names> </name><name name-style="western"><surname>Arzeen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mumtaz</surname><given-names>M</given-names> </name><name name-style="western"><surname>Haroon</surname><given-names>M</given-names> </name></person-group><article-title>The perils of technology: social networking addiction and anxiety</article-title><source>PJMHS</source><year>2023</year><volume>17</volume><issue>5</issue><fpage>507</fpage><lpage>512</lpage><pub-id pub-id-type="doi">10.53350/pjmhs2023175507</pub-id></nlm-citation></ref><ref id="ref99"><label>99</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yokoi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Eguchi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Fujita</surname><given-names>T</given-names> </name><name name-style="western"><surname>Nakayachi</surname><given-names>K</given-names> </name></person-group><article-title>Artificial intelligence is trusted less than a doctor in medical treatment decisions: influence of perceived care and value similarity</article-title><source>International Journal of Human&#x2013;Computer Interaction</source><year>2021</year><month>06</month><day>15</day><volume>37</volume><issue>10</issue><fpage>981</fpage><lpage>990</lpage><pub-id pub-id-type="doi">10.1080/10447318.2020.1861763</pub-id></nlm-citation></ref><ref id="ref100"><label>100</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>D</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>X</given-names> </name></person-group><article-title>Public perceptions on the application of artificial intelligence in healthcare: a qualitative meta-synthesis</article-title><source>BMJ Open</source><year>2023</year><month>01</month><day>4</day><volume>13</volume><issue>1</issue><fpage>e066322</fpage><pub-id pub-id-type="doi">10.1136/bmjopen-2022-066322</pub-id><pub-id pub-id-type="medline">36599634</pub-id></nlm-citation></ref><ref id="ref101"><label>101</label><nlm-citation citation-type="confproc"><person-group person-group-type="editor"><name name-style="western"><surname>Buolamwini</surname><given-names>J</given-names> </name><name name-style="western"><surname>Gebru</surname><given-names>T</given-names> </name></person-group><article-title>Gender shades: intersectional accuracy disparities in commercial gender classification</article-title><year>2018</year><access-date>2026-04-13</access-date><conf-name>Proceedings of the 1st Conference on Fairness, Accountability and Transparency</conf-name><conf-loc>New York, NY</conf-loc><publisher-name>Proceedings of Machine Learning Research</publisher-name><fpage>77</fpage><lpage>91</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://proceedings.mlr.press/v81/buolamwini18a.html">https://proceedings.mlr.press/v81/buolamwini18a.html</ext-link></comment></nlm-citation></ref><ref id="ref102"><label>102</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Obermeyer</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Powers</surname><given-names>B</given-names> </name><name name-style="western"><surname>Vogeli</surname><given-names>C</given-names> </name><name name-style="western"><surname>Mullainathan</surname><given-names>S</given-names> </name></person-group><article-title>Dissecting racial bias in an algorithm used to manage the health of populations</article-title><source>Science</source><year>2019</year><month>10</month><day>25</day><volume>366</volume><issue>6464</issue><fpage>447</fpage><lpage>453</lpage><pub-id pub-id-type="doi">10.1126/science.aax2342</pub-id><pub-id pub-id-type="medline">31649194</pub-id></nlm-citation></ref><ref id="ref103"><label>103</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>E</given-names> </name><name name-style="western"><surname>Banerjee</surname><given-names>O</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>AI in health and medicine</article-title><source>Nat Med</source><year>2022</year><month>01</month><volume>28</volume><issue>1</issue><fpage>31</fpage><lpage>38</lpage><pub-id pub-id-type="doi">10.1038/s41591-021-01614-0</pub-id><pub-id pub-id-type="medline">35058619</pub-id></nlm-citation></ref><ref id="ref104"><label>104</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Markus</surname><given-names>AF</given-names> </name><name name-style="western"><surname>Kors</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Rijnbeek</surname><given-names>PR</given-names> </name></person-group><article-title>The role of explainability in creating trustworthy artificial intelligence for health care: a comprehensive survey of the terminology, design choices, and evaluation strategies</article-title><source>J Biomed Inform</source><year>2021</year><month>01</month><volume>113</volume><fpage>103655</fpage><pub-id pub-id-type="doi">10.1016/j.jbi.2020.103655</pub-id><pub-id pub-id-type="medline">33309898</pub-id></nlm-citation></ref><ref id="ref105"><label>105</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Smith</surname><given-names>H</given-names> </name><name name-style="western"><surname>Fotheringham</surname><given-names>K</given-names> </name></person-group><article-title>Artificial intelligence in clinical decision-making: rethinking liability</article-title><source>Med Law Int</source><year>2020</year><month>06</month><volume>20</volume><issue>2</issue><fpage>131</fpage><lpage>154</lpage><pub-id pub-id-type="doi">10.1177/0968533220945766</pub-id></nlm-citation></ref><ref id="ref106"><label>106</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Guidance</surname><given-names>W</given-names> </name></person-group><source>Ethics and Governance of Artificial Intelligence for Health</source><year>2021</year><publisher-name>World Health Organization</publisher-name><fpage>1</fpage><lpage>165</lpage></nlm-citation></ref><ref id="ref107"><label>107</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hulsen</surname><given-names>T</given-names> </name></person-group><article-title>Explainable artificial intelligence (XAI): concepts and challenges in healthcare</article-title><source>AI</source><year>2023</year><volume>4</volume><issue>3</issue><fpage>652</fpage><lpage>666</lpage><pub-id pub-id-type="doi">10.3390/ai4030034</pub-id></nlm-citation></ref><ref id="ref108"><label>108</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bovenberg</surname><given-names>J</given-names> </name><name name-style="western"><surname>Peloquin</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bierer</surname><given-names>B</given-names> </name><name name-style="western"><surname>Barnes</surname><given-names>M</given-names> </name><name name-style="western"><surname>Knoppers</surname><given-names>BM</given-names> </name></person-group><article-title>How to fix the GDPR&#x2019;s frustration of global biomedical research</article-title><source>Science</source><year>2020</year><month>10</month><day>2</day><volume>370</volume><issue>6512</issue><fpage>40</fpage><lpage>42</lpage><pub-id pub-id-type="doi">10.1126/science.abd2499</pub-id><pub-id pub-id-type="medline">33004505</pub-id></nlm-citation></ref><ref id="ref109"><label>109</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kiseleva</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kotzinos</surname><given-names>D</given-names> </name><name name-style="western"><surname>De Hert</surname><given-names>P</given-names> </name></person-group><article-title>Transparency of AI in healthcare as a multilayered system of accountabilities: between legal requirements and technical limitations</article-title><source>Front Artif Intell</source><year>2022</year><volume>5</volume><fpage>879603</fpage><pub-id pub-id-type="doi">10.3389/frai.2022.879603</pub-id><pub-id pub-id-type="medline">35707765</pub-id></nlm-citation></ref><ref id="ref110"><label>110</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hanauer</surname><given-names>DA</given-names> </name><name name-style="western"><surname>Preib</surname><given-names>R</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>K</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>SW</given-names> </name></person-group><article-title>Patient-initiated electronic health record amendment requests</article-title><source>J Am Med Inform Assoc</source><year>2014</year><volume>21</volume><issue>6</issue><fpage>992</fpage><lpage>1000</lpage><pub-id pub-id-type="doi">10.1136/amiajnl-2013-002574</pub-id><pub-id pub-id-type="medline">24863430</pub-id></nlm-citation></ref><ref id="ref111"><label>111</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>HJ</given-names> </name></person-group><article-title>Patient perspectives on informed consent for medical AI: a web-based experiment</article-title><source>Digit HEALTH</source><year>2024</year><volume>10</volume><fpage>20552076241247938</fpage><pub-id pub-id-type="doi">10.1177/20552076241247938</pub-id><pub-id pub-id-type="medline">38698829</pub-id></nlm-citation></ref><ref id="ref112"><label>112</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bakken</surname><given-names>S</given-names> </name></person-group><article-title>AI in health: keeping the human in the loop</article-title><source>J Am Med Inform Assoc</source><year>2023</year><month>06</month><day>20</day><volume>30</volume><issue>7</issue><fpage>1225</fpage><lpage>1226</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocad091</pub-id><pub-id pub-id-type="medline">37337923</pub-id></nlm-citation></ref><ref id="ref113"><label>113</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nouis</surname><given-names>SCE</given-names> </name><name name-style="western"><surname>Uren</surname><given-names>V</given-names> </name><name name-style="western"><surname>Jariwala</surname><given-names>S</given-names> </name></person-group><article-title>Evaluating accountability, transparency, and bias in AI-assisted healthcare decision- making: a qualitative study of healthcare professionals&#x2019; perspectives in the UK</article-title><source>BMC Med Ethics</source><year>2025</year><month>07</month><day>8</day><volume>26</volume><issue>1</issue><fpage>89</fpage><pub-id pub-id-type="doi">10.1186/s12910-025-01243-z</pub-id><pub-id pub-id-type="medline">40629303</pub-id></nlm-citation></ref><ref id="ref114"><label>114</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Maliha</surname><given-names>G</given-names> </name><name name-style="western"><surname>Gerke</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cohen</surname><given-names>IG</given-names> </name><name name-style="western"><surname>Parikh</surname><given-names>RB</given-names> </name></person-group><article-title>Artificial intelligence and liability in medicine: balancing safety and innovation</article-title><source>Milbank Q</source><year>2021</year><month>09</month><volume>99</volume><issue>3</issue><fpage>629</fpage><lpage>647</lpage><pub-id pub-id-type="doi">10.1111/1468-0009.12504</pub-id><pub-id pub-id-type="medline">33822422</pub-id></nlm-citation></ref><ref id="ref115"><label>115</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fotheringham</surname><given-names>K</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>H</given-names> </name></person-group><article-title>Accidental injustice: healthcare AI legal responsibility must be prospectively planned prior to its adoption</article-title><source>Future Healthc J</source><year>2024</year><month>09</month><volume>11</volume><issue>3</issue><fpage>100181</fpage><pub-id pub-id-type="doi">10.1016/j.fhj.2024.100181</pub-id><pub-id pub-id-type="medline">39371533</pub-id></nlm-citation></ref><ref id="ref116"><label>116</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Iserson</surname><given-names>KV</given-names> </name></person-group><article-title>Informed consent for artificial intelligence in emergency medicine: a practical guide</article-title><source>Am J Emerg Med</source><year>2024</year><month>02</month><volume>76</volume><fpage>225</fpage><lpage>230</lpage><pub-id pub-id-type="doi">10.1016/j.ajem.2023.11.022</pub-id><pub-id pub-id-type="medline">38128163</pub-id></nlm-citation></ref><ref id="ref117"><label>117</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Osonuga</surname><given-names>A</given-names> </name><name name-style="western"><surname>Osonuga</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Fidelis</surname><given-names>SC</given-names> </name><name name-style="western"><surname>Osonuga</surname><given-names>GC</given-names> </name><name name-style="western"><surname>Juckes</surname><given-names>J</given-names> </name><name name-style="western"><surname>Olawade</surname><given-names>DB</given-names> </name></person-group><article-title>Bridging the digital divide: artificial intelligence as a catalyst for health equity in primary care settings</article-title><source>Int J Med Inform</source><year>2025</year><month>12</month><volume>204</volume><fpage>106051</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2025.106051</pub-id><pub-id pub-id-type="medline">40706198</pub-id></nlm-citation></ref><ref id="ref118"><label>118</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>R</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>Y</given-names> </name></person-group><article-title>Digital divide as a determinant of health in the U.S. older adults: prevalence, trends, and risk factors</article-title><source>BMC Geriatr</source><year>2024</year><volume>24</volume><issue>1</issue><fpage>1027</fpage><pub-id pub-id-type="doi">10.1186/s12877-024-05612-y</pub-id></nlm-citation></ref><ref id="ref119"><label>119</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pham</surname><given-names>T</given-names> </name></person-group><article-title>Ethical and legal considerations in healthcare AI: innovation and policy for safe and fair use</article-title><source>R Soc Open Sci</source><year>2025</year><month>05</month><volume>12</volume><issue>5</issue><pub-id pub-id-type="doi">10.1098/rsos.241873</pub-id></nlm-citation></ref><ref id="ref120"><label>120</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cestonaro</surname><given-names>C</given-names> </name><name name-style="western"><surname>Delicati</surname><given-names>A</given-names> </name><name name-style="western"><surname>Marcante</surname><given-names>B</given-names> </name><name name-style="western"><surname>Caenazzo</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tozzo</surname><given-names>P</given-names> </name></person-group><article-title>Defining medical liability when artificial intelligence is applied on diagnostic algorithms: a systematic review</article-title><source>Front Med</source><year>2023</year><month>11</month><day>27</day><volume>10</volume><fpage>2023</fpage><pub-id pub-id-type="doi">10.3389/fmed.2023.1305756</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Search stage.</p><media xlink:href="jmir_v28i1e85663_app1.pdf" xlink:title="PDF File, 1500 KB"/></supplementary-material><supplementary-material id="app2"><label>Checklist 1</label><p>PRISMA 2020 checklist.</p><media xlink:href="jmir_v28i1e85663_app2.pdf" xlink:title="PDF File, 163 KB"/></supplementary-material><supplementary-material id="app3"><label>Checklist 2</label><p>ENTREQ checklist.</p><media xlink:href="jmir_v28i1e85663_app3.pdf" xlink:title="PDF File, 95 KB"/></supplementary-material><supplementary-material id="app4"><label>Checklist 3</label><p>PRISMA-S checklist.</p><media xlink:href="jmir_v28i1e85663_app4.pdf" xlink:title="PDF File, 73 KB"/></supplementary-material></app-group></back></article>