<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v28i1e81387</article-id><article-id pub-id-type="doi">10.2196/81387</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Effectiveness of Al-Assisted Patient Health Education Using Voice Cloning and ChatGPT: Prospective Randomized Controlled Trial</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Sun</surname><given-names>Yan</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Xu</surname><given-names>Shangqing</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Jin</surname><given-names>Hongying</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Han</surname><given-names>Xiaoyan</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Jin</surname><given-names>Kangqi</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Yimei</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ma</surname><given-names>Xiaoli</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wei</surname><given-names>Huaping</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Ma</surname><given-names>Minjie</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Thoracic Surgery, The First Hospital of Lanzhou University</institution><addr-line>Donggang West Road 1#</addr-line><addr-line>Lanzhou</addr-line><country>China</country></aff><aff id="aff2"><institution>School of Nursing, Lanzhou University</institution><addr-line>Lanzhou</addr-line><country>China</country></aff><aff id="aff3"><institution>Outpatient Department, The First Hospital of Lanzhou University</institution><addr-line>Lanzhou</addr-line><country>China</country></aff><aff id="aff4"><institution>Gansu International Science and Technology Cooperation Base for Development and Application of Thoracic Surgery Key Technologies, The First Clinical Medical College of Lanzhou University, Department of Thoracic Surgery, The First Hospital of Lanzhou University</institution><addr-line>Donggang West Road 1#</addr-line><addr-line>Lanzhou</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Brini</surname><given-names>Stefano</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Khamees</surname><given-names>Almu'atasim</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Mesk&#x00F3;</surname><given-names>Bertalan</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Minjie Ma, PhD, Gansu International Science and Technology Cooperation Base for Development and Application of Thoracic Surgery Key Technologies, The First Clinical Medical College of Lanzhou University, Department of Thoracic Surgery, The First Hospital of Lanzhou University, Donggang West Road 1#, Lanzhou, 730000, China, 86 13639325950; <email>maminjie24@sina.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>19</day><month>3</month><year>2026</year></pub-date><volume>28</volume><elocation-id>e81387</elocation-id><history><date date-type="received"><day>28</day><month>07</month><year>2025</year></date><date date-type="accepted"><day>17</day><month>01</month><year>2026</year></date></history><copyright-statement>&#x00A9; Yan Sun, Shangqing Xu, Hongying Jin, Xiaoyan Han, Kangqi Jin, Yimei Zhang, Xiaoli Ma, Huaping Wei, Minjie Ma. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 19.3.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2026/1/e81387"/><abstract><sec><title>Background</title><p>Traditional patient education often lacks personalization and engagement, potentially limiting knowledge acquisition and treatment adherence. Advances in artificial intelligence (AI), including voice cloning technology and large language models (eg, ChatGPT), offer new opportunities to deliver personalized, scalable, interactive health education. However, evidence regarding the comparative effectiveness of different AI-based voice cloning strategies and reliability of automated AI evaluation tools remains limited.</p></sec><sec><title>Objective</title><p>This study aims to evaluate the effectiveness of AI-assisted patient education integrating voice cloning and ChatGPT, compare physician voice cloning with patient self-voice cloning, and assess the reliability of ChatGPT as an automated evaluation tool for education outcomes.</p></sec><sec sec-type="methods"><title>Methods</title><p>In this prospective, 3-arm, parallel-group randomized controlled trial, 180 hospitalized patients requiring standardized health education were recruited from a tertiary hospital. Inclusion criteria were age &#x2265;18 years, clear diagnosis requiring health education, clear consciousness, and voluntary participation with informed consent. Exclusion criteria were severe hearing impairment, severe cognitive impairment, expected hospitalization &#x003C;3 days, or prior participation in similar studies. Using a computer-generated random sequence, participants were randomly assigned (1:1:1) to receive traditional education (control), AI-assisted education using physician voice cloning, or AI-assisted education using patient self-voice cloning, each with identical educational content of equal duration. The primary outcome was education content compliance, evaluated using ChatGPT-4 with validated prompts and verified by expert review. Secondary outcomes included knowledge retention, education satisfaction, treatment adherence, quality of life, and psychological status. Outcome assessors and data analysts, but not participants, were blinded to group allocation.</p></sec><sec sec-type="results"><title>Results</title><p>Of 180 randomized participants, 174 (96.7%) completed the trial. Both AI-assisted groups had significantly higher mean education content compliance scores immediately posteducation than the control group (physician voice: 86.7, SD 7.3; self-voice: 92.5, SD 6.8; control: 73.2, SD 8.5; <italic>P</italic>&#x003C;.001). The patient self-voice group showed superior predischarge knowledge retention, higher education satisfaction, and greater treatment adherence than the other groups (all <italic>P</italic>&#x2264;.02). At the 1-month follow-up, the self-voice group maintained improved adherence (Cohen <italic>d</italic>=0.74) and had significantly lower anxiety and depression scores (all <italic>P</italic>&#x2264;.02) and improved SF-36 quality-of-life domains. ChatGPT-based evaluations demonstrated high reliability (weighted &#x03BA;=0.87, 95% CI 0.82&#x2010;0.91)<sup>.</sup></p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The innovative patient education model integrating AI voice cloning and ChatGPT is distinct from previous studies primarily relying on standard text-to-speech or professionally recorded content. Using patients&#x2019; own cloned voices for health education delivery leveraged the self-reference effect to enhance learning outcomes. Compared with research using clinician-narrated content, this study highlights that self-voice education produces superior outcomes across multiple domains including compliance, satisfaction, and psychological well-being. These findings establish a theoretical and practical framework for personalized AI-driven patient education. In real-world clinical settings, this approach offers a scalable, cost-effective solution to enhance patient engagement, particularly valuable in resource-limited environments where individualized education is challenging to deliver.</p></sec><sec><title>Trial Registration</title><p>Chinese Clinical Trial Registry ChiCTR2500101882; https://www.chictr.org.cn/showprojEN.html?proj=268927</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>voice cloning</kwd><kwd>medical education</kwd><kwd>ChatGPT</kwd><kwd>education effect evaluation</kwd><kwd>randomized controlled trial</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Patient education is a fundamental component of health care, aiming to enhance individuals&#x2019; understanding of their medical conditions, promote self-management, improve treatment adherence, and ultimately optimize health outcomes and prognosis [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. Despite the clinical significance of patient education, traditional methods of patient education&#x2014;primarily verbal explanations by health care providers and standardized written materials&#x2014;often fall short in addressing the diverse needs of patients [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. These approaches are constrained by limited time, lack of interactivity, and insufficient personalization, frequently resulting in suboptimal comprehension, poor engagement, and reduced adherence to treatment recommendations [<xref ref-type="bibr" rid="ref6">6</xref>].</p><p>Prior research has shown that personalized, interactive education enhances patient understanding and engagement [<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref11">11</xref>]. Recent advances in artificial intelligence (AI), particularly in voice cloning and large language models (LLMs), present novel opportunities to transform conventional health education into a more personalized, scalable, and interactive process [<xref ref-type="bibr" rid="ref12">12</xref>]. Voice cloning technology uses deep learning algorithms to synthesize natural-sounding speech from limited voice samples, enabling the delivery of educational content in familiar voices&#x2014;such as those of physicians or even the patients themselves [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. Prior studies suggest that familiar voices can increase emotional resonance [<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref17">17</xref>], build trust, enhance patient engagement, improve memory retention of medical information [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref17">17</xref>], and increase accessibility to digital health tools. Recent scoping reviews have identified multiple applications of LLMs in patient education, including generating educational materials, interpreting medical information, and optimizing doctor-patient interaction [<xref ref-type="bibr" rid="ref18">18</xref>]. Systematic reviews have further demonstrated that LLMs can enhance health care communication and support personalized patient engagement [<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>Simultaneously, LLMs like ChatGPT, which is based on advanced LLMs, have demonstrated strong capabilities for natural language understanding, dialog generation, and semantic evaluation [<xref ref-type="bibr" rid="ref20">20</xref>] and have shown potential for supporting patient communication and education [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref23">23</xref>]. These tools can interact with patients in real time, deliver conversational-style education, and assess understanding through dynamic, automated evaluation processes [<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref26">26</xref>]. Such integration may address key challenges in health education, including the need for tailored content delivery and standardized assessment without overburdening clinical staff. These models are capable of few-shot learning, enabling rapid adaptation to novel prompts [<xref ref-type="bibr" rid="ref27">27</xref>].</p><p>Despite this emerging literature, significant gaps remain. First, few empirical studies have rigorously examined the comparative effectiveness of different voice sources&#x2014;specifically physician-cloned voice versus patient-cloned voice&#x2014;in AI-assisted patient education. Second, although LLM-based tools like ChatGPT show promise for health communication, their role as standardized evaluators of educational outcomes has not been comprehensively validated in clinical randomized trials. Traditional outcome assessments often rely on human raters, which introduces inter-rater variability and resource burdens.</p><p>To address these gaps, this study was a prospective randomized controlled trial (RCT) to investigate the effectiveness of an AI-assisted patient education system integrating voice cloning and ChatGPT. The study compared educational outcomes between physician voice cloning and patient self-voice cloning and explored the feasibility and accuracy of using ChatGPT as a standardized, automated evaluation tool. By doing so, this work aimed to contribute novel insights into how AI technologies can enhance the personalization, efficiency, and effectiveness of inpatient education and to extend the evidence base for AI-enabled patient engagement strategies. We hypothesized that (1) AI-assisted voice cloning education would significantly improve patient education outcomes compared with traditional education, (2) patient self-voice cloning would yield superior educational effectiveness compared with physician voice cloning due to the self-reference effect, and (3) ChatGPT could serve as a reliable automated evaluation tool with high agreement with expert assessment.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Overall Study</title><p>This study used an RCT design to assess the impact of AI-assisted patient health education using voice cloning technology and ChatGPT on patient education outcomes. Participants were randomly assigned to 1 of 3 groups: The 3 groups received different educational methods but the same educational content and duration of education. Following the educational intervention, the effectiveness of the education was comprehensively evaluated at fixed time points using ChatGPT. To ensure the validity and accuracy of the evaluation tool, a pre-test of the ChatGPT evaluation tool was conducted prior to the formal data collection. This study not only explored the potential of AI voice cloning technology in personalized patient education but also used ChatGPT as an assessment tool to further validate its reliability and practicality in evaluating health education interventions. Ultimately, the study aimed to provide theoretical insights and practical recommendations for improving medical health education models, enhancing patient education outcomes, and promoting treatment adherence and quality of life.</p></sec><sec id="s2-2"><title>Study Timeline</title><p>An independent tool validation study was conducted from January 2024 to June 2024 to verify the ChatGPT-based compliance scoring tool (no participant enrollment). Administrative preparation, staff training, and software and hardware debugging were completed in December 2024, and this period did not involve participant enrollment. Trial registration (ChiCTR2500101882) was initiated on January 15, 2025, and finalized on April 30, 2025, before enrollment began in May 2025. Participant enrollment and intervention delivery were conducted from May 2025 to June 2025, with 1-month follow-up assessments completed by July 2025.</p></sec><sec id="s2-3"><title>Reporting Guidelines</title><p>The trial was conducted and reported in accordance with the CONSORT (Consolidated Standards of Reporting Trials) 2025 statement and the CONSORT-EHEALTH checklist for digital and AI-based health interventions (<xref ref-type="supplementary-material" rid="app6">Checklists 1</xref> and <xref ref-type="supplementary-material" rid="app7">2</xref>) [<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref29">29</xref>].</p></sec><sec id="s2-4"><title>Study Design</title><p>This single-center, 3-arm, parallel-group, superiority RCT (allocation ratio 1:1:1) was conducted at The First Hospital of Lanzhou University. System development, intervention standardization, and staff training were completed in December 2024 and January 2025. Participant recruitment and data collection were conducted from May 2025 to July 2025.</p></sec><sec id="s2-5"><title>Randomization</title><sec id="s2-5-1"><title>Sequence Generation</title><p>The random allocation sequence was generated using a computer-generated random number table.</p></sec><sec id="s2-5-2"><title>Allocation Concealment</title><p>To ensure concealment of group assignments, the allocation sequence was placed in sealed, opaque envelopes managed by independent researchers who were not involved in the implementation of the interventions. This procedure minimized the risk of selection bias.</p></sec><sec id="s2-5-3"><title>Implementation</title><p>Independent researchers generated the allocation sequence and managed the envelopes. Eligible participants were assigned to study arms according to the contents of the envelopes.</p></sec><sec id="s2-5-4"><title>Blinding</title><p>Due to the nature of the interventions, neither participants nor the medical staff implementing the interventions could be blinded. However, outcome assessors were blinded to group assignments to minimize assessment bias.</p></sec></sec><sec id="s2-6"><title>Ethical Considerations</title><p>The clinical trial protocol was reviewed and approved by the Ethics Committee of The First Hospital of Lanzhou University (approval number: LDYYLL-2025&#x2010;805) and conducted in accordance with the principles of the Declaration of Helsinki.</p><p>All participants received detailed verbal and written information regarding the study purpose, procedures, potential risks, and expected benefits and provided written informed consent prior to enrollment. The consent process also covered permission for the use and publication of anonymized data. Participants were informed that their involvement was voluntary and that they could withdraw from the study at any time without affecting their standard medical care.</p><p>All collected data were deidentified before analysis and stored in encrypted, password-protected databases accessible only to the research team. Personal identifiers were removed to ensure participant confidentiality and data security.</p><p>The use of voice cloning technology was explicitly covered in the consent form. Participants and physicians provided separate written authorization for the use of their voice recordings solely for research purposes. All voice samples and generated models were permanently deleted after study completion.</p><p>Participants did not receive monetary compensation but were offered free access to the AI-based health education program and follow-up consultations as part of their clinical care. No identifiable photographs nor audiovisual materials were used in this publication.</p></sec><sec id="s2-7"><title>Participants</title><p>The study participants were patients hospitalized in a tertiary hospital from May 2025 to July 2025 who needed to receive medical education.</p></sec><sec id="s2-8"><title>Inclusion and Exclusion Criteria</title><p>The inclusion criteria were (1) age &#x2265;18 years of any gender; (2) clear diagnosis, patients who needed to receive medical education; (3) clear consciousness, able to understand and cooperate with the study; and (4) voluntary participation in this study and signed informed consent. The exclusion criteria were (1) severe hearing impairment, unable to normally receive voice education; (2) severe cognitive impairment, unable to understand education content; (3) expected hospitalization time &#x003C;3 days; and (4) previously participated in similar studies.</p></sec><sec id="s2-9"><title>Sample Size</title><p>Based on the pre-experimental results, with the education content compliance rate as the main observation indicator, <italic>&#x03B1;</italic> set at 0.05 (2-sided), <italic>&#x03B2;</italic> set at 0.10, a test power of 90%, and an expected effect size of 0.3 among the 3 groups, we calculated that each group needed a sample size of 54 participants using G*Power 3.1 software. Considering a possible dropout rate of approximately 10%, we finally determined that we would enroll 60 participants in each group, for a total of 180 participants. Prior to participation, all patients were informed that the educational intervention would be delivered using AI-generated voice recordings. Written informed consent was obtained from all participants.</p></sec><sec id="s2-10"><title>Independent Tool Validation Study (January 2024 to June 2024): Validation of ChatGPT-Based Compliance Scoring</title><p>This independent tool validation study was conducted separately from the randomized trial and did not involve participant enrollment. To validate the AI-based scoring used for the primary outcome, we conducted a separate pilot validation study (January 2024 through June 2024) among 30 volunteers who were not part of the randomized trial. Phase 1 (verbatim-matching workflow) used iFlytek automatic speech recognition (Mandarin medical context; no manual correction) to transcribe each participant&#x2019;s recitation; the transcribed text was pasted into the ChatGPT web interface together with the standard education text for rubric-based scoring. Phase 2 (semantic audio-based workflow) re-evaluated the same recordings by uploading the original audio directly to the ChatGPT web interface (WAV/MP3; 16-kHz sampling rate; 60&#x2010;120 s per recording); each recording was scored once. Both workflows were benchmarked against consensus ratings from 3 senior nursing experts using the same rubric. Agreement with expert ratings was substantial in Phase 1 (weighted &#x03BA;=.72) and higher in Phase 2 (weighted &#x03BA;=.87), with strong internal consistency of the evaluation items (Cronbach <italic>&#x03B1;</italic>=0.89). Therefore, Phase 2 was selected for the main RCT to minimize transcription-related errors and enhance reproducibility. Detailed prompts, rubric, and example inputs and outputs are provided in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s2-11"><title>Intervention Measures</title><sec id="s2-11-1"><title>Traditional Education Group (Control Group)</title><p>The traditional education group received the hospital&#x2019;s current standardized patient education, including verbal education by medical staff and standardized written education materials. Verbal education was conducted by trained medical staff according to a unified education outline, including disease knowledge, treatment plans, precautions, and lifestyle guidance. Written education materials were standardized education manuals compiled by the hospital that patients could take back to the ward or home to read. The duration of the education was approximately 20 minutes to 30 minutes.</p></sec><sec id="s2-11-2"><title>Attending Physician Voice Cloning Education Group (Intervention Group 1)</title><p>The physician voice cloning education group (intervention group 1) received patient education through cloning of the attending physician&#x2019;s voice. First, the patient&#x2019;s attending physician&#x2019;s voice sample (approximately 5 min) was obtained, and Resemble.ai voice cloning technology was used to generate the doctor&#x2019;s voice model. Medical professionals then formulated personalized education content based on standard education content and the patient&#x2019;s specific situation, synthesized the voice using doctor voice cloning, and produced personalized voice education materials. Patients listened to the voice-cloned education content through headphones or speakers under the guidance of research assistants and received printed personalized education materials. The education content covered the same topics as the control group but was personalized according to patient characteristics, such as using language expressions suitable for the patient&#x2019;s education level, lifestyle suggestions targeted at the patient&#x2019;s occupational characteristics, and emphasizing relevant content based on the patient&#x2019;s concerns. The duration of the education was approximately 20 minutes to 30 minutes.</p></sec><sec id="s2-11-3"><title>Patient&#x2019;s Own Voice Cloning Education Group (Intervention Group 2)</title><p>The patient voice cloning education group received patient education through cloning of the patient&#x2019;s own voice. First, the patient&#x2019;s voice sample (approximately 5 min) was obtained, and Resemble.ai voice cloning technology was used to generate the patient&#x2019;s voice model. Similar to intervention group 1, medical professionals then formulated personalized education content, synthesized the voice through patient voice cloning, and produced personalized voice education materials. Patients listened to the voice-cloned education content through mobile phone headphones or speakers under the guidance of research assistants and received printed personalized education materials. The education content was the same as in intervention group 1, with the only difference being the use of the patient&#x2019;s own voice for education. The duration of the education was approximately 20 minutes to 30 minutes.</p></sec><sec id="s2-11-4"><title>Educational Content and Assessment of Education Effect</title><p>All participants received the same educational content and could ask questions after receiving education, with medical staff answering inquiries. All 3 groups of patients had their education effect assessed before discharge by using ChatGPT to compare the education content recited by patients with the original education content to calculate the content compliance rate, and follow-up assessments were conducted 1 month after discharge.</p><p>The educational intervention included AI-generated voice recordings delivered using cloning of either the attending physician&#x2019;s voice or the patient&#x2019;s own voice. ChatGPT was used to assess patient knowledge, satisfaction, and adherence using standardized prompts. Patient outcomes were also measured using structured questionnaires. The questionnaires are provided in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendices 2</xref><xref ref-type="supplementary-material" rid="app3"/>-<xref ref-type="supplementary-material" rid="app4">4</xref>, and the voice cloning patient education prompts, ChatGPT evaluation prompts, and AI-generated educational content are provided in <xref ref-type="supplementary-material" rid="app5">Multimedia Appendix 5</xref>. Data were collected at baseline and postintervention, and all responses were recorded in a secure electronic database.</p></sec></sec><sec id="s2-12"><title>Evaluation Indicators and Measurement Tools</title><sec id="s2-12-1"><title>Primary Outcome Measure</title><p>The primary outcome, education content compliance rate, was defined as the degree of semantic concordance between the education content recited by patients and the standardized education material using the following equation: compliance rate = (number of key points correctly recited by patients/total number of key points in education content) &#x00D7; 100%. The education content compliance rate was assessed using ChatGPT-4 (model ChatGPT [OpenAI; web interface, model displayed as GPT-4 during the study period]) as an automated evaluation tool to quantify semantic similarity between the patient&#x2019;s oral recitation and the standard education content, enabling scalable and objective scoring.</p></sec><sec id="s2-12-2"><title>Secondary Outcome Measures</title><sec id="s2-12-2-1"><title>Knowledge Mastery</title><p>We used disease-related knowledge questionnaires to assess patients&#x2019; understanding and memory of disease knowledge, with a possible total score of 100 points and higher scores indicating better knowledge mastery.</p></sec><sec id="s2-12-2-2"><title>Education Satisfaction</title><p>We used the education satisfaction scale to assess patients&#x2019; satisfaction with the education process and content, including aspects such as education content, method, time, and effect, using a 5-level Likert scale. Higher scores indicate higher satisfaction.</p></sec><sec id="s2-12-2-3"><title>Treatment Adherence</title><p>We used the treatment adherence scale to assess patients&#x2019; treatment adherence within 1 month after discharge, including aspects such as drug treatment, lifestyle adjustment, and follow-up review. Higher scores indicate better adherence.</p></sec><sec id="s2-12-2-4"><title>Quality of Life</title><p>We used the SF-36 quality of life scale to assess changes in patients&#x2019; quality of life 1 month after discharge.</p></sec><sec id="s2-12-2-5"><title>Anxiety and Depression</title><p>We used the Hospital Anxiety and Depression Scale to assess changes in patients&#x2019; anxiety and depression before and after the education.</p></sec></sec></sec><sec id="s2-13"><title>Evaluation Time Points</title><sec id="s2-13-1"><title>Baseline Assessment</title><p>We collected patients&#x2019; basic information and baseline data, including demographic characteristics, disease-related knowledge, anxiety, and depression, before the intervention.</p></sec><sec id="s2-13-2"><title>Immediate Posteducation Assessment</title><p>The education content compliance rate, knowledge mastery, and education satisfaction were assessed immediately after the education was completed.</p></sec><sec id="s2-13-3"><title>Predischarge Assessment</title><p>Education content compliance rate and knowledge mastery were re-assessed 1 day before patient discharge.</p></sec><sec id="s2-13-4"><title>1-Month Postdischarge Follow-Up Assessment</title><p>Through telephone or outpatient follow-up, we assessed patients&#x2019; treatment adherence, quality of life, disease-related knowledge retention, anxiety, and depression.</p></sec></sec><sec id="s2-14"><title>Data Collection and Management</title><p>Research assistants who had received standardized training were responsible for data collection, using uniform data collection forms to record patient information and evaluation results. An electronic database was established, with all data independently entered by 2 personnel and cross-checked to ensure accuracy. To protect patient privacy, all data were stored in coded form, and regular data quality checks were conducted to ensure completeness and reliability. Missing data were promptly identified and imputed.</p><p>All voice recordings collected in this study were considered sensitive personal data. They were stored on secure, password-protected servers at The First Hospital of Lanzhou University, accessible only to authorized research personnel. For analysis, the recordings were anonymized. Following the completion of the study, all original recordings will be permanently deleted in accordance with institutional data retention and privacy policies.</p></sec><sec id="s2-15"><title>Statistical Analysis</title><p>Statistical analyses were performed using SPSS version 26.0 (IBM Corp). All analyses followed the intention-to-treat (ITT) principle. Missing outcome data (6/182, 3.3%) were evaluated using the missing completely at random test from Little, which indicated that missingness was completely at random (<italic>&#x03C7;</italic>&#x00B2;<sub>12</sub>=14.5, <italic>P</italic>=.27). Consequently, multiple imputation by chained equations was used to generate 5 imputed datasets (m=5), and pooled estimates were obtained using the rules by Rubin. Sensitivity analyses using per-protocol (complete-case) data yielded consistent findings (absolute difference in mean compliance&#x003C;1%). Continuous variables are presented as mean (SD), and categorical variables are presented as n (%). Between-group comparisons were conducted using 1-way ANOVA with post hoc tests (least significant difference or Bonferroni, as appropriate). Categorical variables were compared using &#x03C7;&#x00B2; tests or Fisher exact tests. All tests were 2-sided with a significance level of <italic>P</italic>&#x003C;.05.</p></sec><sec id="s2-16"><title>Evaluation Indicators</title><p>The primary outcome of this study was the education content compliance rate, which reflected how well patients understood and followed the health education provided. Secondary outcomes included knowledge retention, patient satisfaction, and treatment adherence.</p><p>Education level and knowledge retention were measured using a structured questionnaire developed by the research team based on the hospital&#x2019;s standardized education materials.</p><p>The questionnaire consisted of multiple-choice and short-answer questions assessing patients&#x2019; understanding and recall of the key educational points immediately after education and before discharge.</p><p>Each correct response was scored as 1 point, with higher total scores indicating better knowledge retention.</p><p>The disease-related knowledge questionnaire was developed based on the hospital&#x2019;s standardized education materials and underwent content review by 3 senior nursing experts.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Participant Flow</title><p><xref ref-type="fig" rid="figure1">Figure 1</xref> illustrates the participant flow. ITT analysis was performed for all participants (n=180) using multiple imputation by chained equations for missing data. The analysis section in the figure shows the per-protocol analyses (n=174).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>CONSORT (Consolidated Standards of Reporting Trials) flow diagram.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v28i1e81387_fig01.png"/></fig></sec><sec id="s3-2"><title>Comparison of General Information</title><p>A total of 180 patients were recruited to participate in this study. During the follow-up period, 6 participants were lost to follow-up due to transfer to another hospital (n=1) and withdrawal of consent (n=1) in the control group; early discharge (n=1) and withdrawal of consent (n=2) in intervention group 1; and failure to complete follow-up assessments (n=1) in intervention group 2. Therefore, 174 patients completed the study (58 in the control group, 57 in intervention group 1, 59 in intervention group 2), for a total dropout rate of 3.3% (6/180). After multiple imputation, the analysis results remained consistent with those of the complete-case dataset, indicating the robustness of the findings. Comparison of general information such as age, gender, education level, occupation, and disease type among the 3 groups of patients showed no statistically significant differences (<italic>P</italic>&#x003E;.05), indicating comparability (<xref ref-type="table" rid="table1">Table 1</xref>).</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Comparison of baseline characteristics among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025).</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Item</td><td align="left" valign="bottom">Control group (n=58)</td><td align="left" valign="bottom">Intervention group 1 (n=57)</td><td align="left" valign="bottom">Intervention group 2 (n=59)</td><td align="left" valign="bottom">Statistical test result (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Age (years), mean (SD)</td><td align="left" valign="top">52.6 (14.8)</td><td align="left" valign="top">53.2 (15.1)</td><td align="left" valign="top">51.9 (14.5)</td><td align="left" valign="top">0.127<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> (2, 171)</td><td align="left" valign="top">.88</td></tr><tr><td align="left" valign="top" colspan="4">Gender, n (%)</td><td align="left" valign="top">0.215<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> (2)</td><td align="left" valign="top">.90</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Male</td><td align="left" valign="top">31 (53)</td><td align="left" valign="top">32 (56)</td><td align="left" valign="top">34 (58)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Female</td><td align="left" valign="top">27 (47)</td><td align="left" valign="top">25 (44)</td><td align="left" valign="top">25 (42)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4">Education level, n (%)</td><td align="left" valign="top">0.986<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> (6)</td><td align="left" valign="top">.91</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Primary school and less</td><td align="left" valign="top">12 (21)</td><td align="left" valign="top">11 (19)</td><td align="left" valign="top">10 (17)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Junior high school</td><td align="left" valign="top">18 (31)</td><td align="left" valign="top">17 (30)</td><td align="left" valign="top">19 (32)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>High school/technical secondary school</td><td align="left" valign="top">16 (28)</td><td align="left" valign="top">18 (32)</td><td align="left" valign="top">17 (29)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>College and more</td><td align="left" valign="top">12 (21)</td><td align="left" valign="top">11 (19)</td><td align="left" valign="top">13 (22)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4">Marital status, n (%)</td><td align="left" valign="top">0.352<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> (2)</td><td align="left" valign="top">.84</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Married</td><td align="left" valign="top">49 (85)</td><td align="left" valign="top">47 (83)</td><td align="left" valign="top">51 (86)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Unmarried/divorced/widowed</td><td align="left" valign="top">9 (16)</td><td align="left" valign="top">10 (18)</td><td align="left" valign="top">8 (14)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4">Disease type, n (%)</td><td align="left" valign="top">0.763<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> (6)</td><td align="left" valign="top">.94</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Cardiovascular disease</td><td align="left" valign="top">21 (36)</td><td align="left" valign="top">20 (35)</td><td align="left" valign="top">22 (37)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Respiratory system disease</td><td align="left" valign="top">15 (26)</td><td align="left" valign="top">16 (28)</td><td align="left" valign="top">17 (29)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Digestive system disease</td><td align="left" valign="top">13 (22)</td><td align="left" valign="top">12 (21)</td><td align="left" valign="top">11 (19)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Others</td><td align="left" valign="top">9 (16)</td><td align="left" valign="top">9 (16)</td><td align="left" valign="top">9 (15)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4">Previous hospitalization history, n (%)</td><td align="left" valign="top">0.124<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> (2)</td><td align="left" valign="top">.94</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Yes</td><td align="left" valign="top">32 (55)</td><td align="left" valign="top">33 (58)</td><td align="left" valign="top">33 (56)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>No</td><td align="left" valign="top">26 (45)</td><td align="left" valign="top">24 (42)</td><td align="left" valign="top">26 (44)</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="6">Baseline HADS<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup> score, mean (SD)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Anxiety</td><td align="left" valign="top">7.8 (3.2)</td><td align="left" valign="top">7.6 (3.1)</td><td align="left" valign="top">7.7 (3.3)</td><td align="left" valign="top">0.068<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> (2, 171)</td><td align="left" valign="top">.93</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Depression</td><td align="left" valign="top">6.9 (2.8)</td><td align="left" valign="top">7.1 (2.9)</td><td align="left" valign="top">6.8 (2.7)</td><td align="left" valign="top">0.185<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> (2, 171)</td><td align="left" valign="top">.83</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup><italic>F</italic> test.</p></fn><fn id="table1fn2"><p><sup>b</sup>Chi-square test.</p></fn><fn id="table1fn3"><p><sup>c</sup>HADS: Hospital Anxiety and Depression Scale.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-3"><title>Comparison of Education Content Compliance Rate</title><p>Immediately after education, the mean education compliance score in the control group was 73.2 (SD 8.5; 95% CI 71.0&#x2010;75.4), compared with mean scores of 86.7 (SD 7.3; 95% CI 84.8&#x2010;88.6) in intervention group 1 and 92.5 (SD 6.8; 95% CI 90.8&#x2010;94.2) in intervention group 2. The 1-way ANOVA showed a statistically significant difference among the 3 groups (<italic>F</italic><sub>2,171</sub>=103.427, <italic>P</italic>&#x003C;.001). Before discharge, the mean score in the control group was 68.5 (SD 9.1; 95% CI 66.2&#x2010;70.8), while the mean scores in intervention group 1 and intervention group 2 were 82.3 (SD 8.1; 95% CI 80.2&#x2010;84.4) and 88.6 (SD 7.4; 95% CI 86.7&#x2010;90.5), respectively. The between-group difference remained statistically significant (<italic>F</italic><sub>2,171</sub>=95.682, <italic>P</italic>&#x003C;.001; <xref ref-type="table" rid="table2">Table 2</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Comparison of education content compliance rate among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences analyzed using 1-way ANOVA.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Time point</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Immediately after education</td><td align="left" valign="top">73.2 (8.5; 71.0&#x2010;75.4)</td><td align="left" valign="top">86.7 (7.3; 84.8&#x2010;88.6)</td><td align="left" valign="top">92.5 (6.8; 90.8&#x2010;94.2)</td><td align="left" valign="top">103.427 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Before discharge</td><td align="left" valign="top">68.5 (9.1; 66.2&#x2010;70.8)</td><td align="left" valign="top">82.3 (8.1; 80.2&#x2010;84.4)</td><td align="left" valign="top">88.6 (7.4; 86.7&#x2010;90.5)</td><td align="left" valign="top">95.682 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap><p>The corresponding ITT results based on the multiply imputed datasets are provided in Supplementary Table S2.</p></sec><sec id="s3-4"><title>Comparison of Knowledge Mastery</title><p>Immediately after education, the mean knowledge mastery score in the control group was 76.3 (SD 9.2; 95% CI 73.9&#x2010;78.7), which was significantly lower than the mean scores in intervention group 1 (88.5, SD 8.4; 95% CI 86.3&#x2010;90.7) and intervention group 2 (90.2, SD 8.1; 95% CI 88.1&#x2010;92.3). A statistically significant difference was observed among the 3 groups (<italic>F</italic><sub>2,171</sub>=48.362, <italic>P</italic>&#x003C;.001). Before discharge, the control group had a mean knowledge mastery score of 71.8 (SD 9.8; 95% CI 69.2&#x2010;74.4), whereas intervention group 1 and intervention group 2 achieved higher mean scores of 84.7 (SD 8.9; 95% CI 82.3&#x2010;87.1) and 87.3 (SD 8.5; 95% CI 85.1&#x2010;89.5), respectively. The between-group difference remained statistically significant (<italic>F</italic><sub>2,171</sub>=46.175, <italic>P</italic>&#x003C;.001). At 1 month after discharge, the mean knowledge mastery score further declined in the control group (65.2, SD 10.5; 95% CI 62.4&#x2010;68.0). In contrast, both intervention groups maintained relatively higher levels of knowledge mastery, with mean scores of 78.6 (SD 9.3; 95% CI 76.1&#x2010;81.1) in intervention group 1 and 83.5 (SD 8.9; 95% CI 81.2&#x2010;85.8) in intervention group 2. The overall difference among the 3 groups remained statistically significant (<italic>F</italic><sub>2,171</sub>=54.793, <italic>P</italic>&#x003C;.001; <xref ref-type="table" rid="table3">Table 3</xref>).</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Comparison of knowledge mastery scores among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences among groups analyzed using 1-way ANOVA.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Time point</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Immediately after education</td><td align="left" valign="top">76.3 (9.2; 73.9-78.7)</td><td align="left" valign="top">88.5 (8.4; 86.3-90.7)</td><td align="left" valign="top">90.2 (8.1; 88.1-92.3)</td><td align="left" valign="top">48.362 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Before discharge</td><td align="left" valign="top">71.8 (9.8; 69.2-74.4)</td><td align="left" valign="top">84.7 (8.9; 82.3-87.1)</td><td align="left" valign="top">87.3 (8.5; 85.1-89.5)</td><td align="left" valign="top">46.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">1 month after discharge</td><td align="left" valign="top">65.2 (10.5; 62.4-68.0)</td><td align="left" valign="top">78.6 (9.3; 76.1-81.1)</td><td align="left" valign="top">83.5 (8.9; 81.2-85.8)</td><td align="left" valign="top">54.793 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec><sec id="s3-5"><title>Comparison of Education Satisfaction</title><p>Significant differences were observed among the 3 groups across all dimensions of education satisfaction, including education content, education method, education time, education effect, and overall satisfaction (all <italic>P</italic>&#x003C;.001). Across all items, the 95% CIs of the 2 intervention groups were consistently higher than those of the control group, with minimal overlap between groups. This indicates that the observed differences in education satisfaction were not only statistically significant but also stable and precise, reflecting a reliable improvement associated with the intervention (<xref ref-type="table" rid="table4">Table 4</xref>).</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Comparison of education satisfaction scores among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences among groups analyzed using 1-way ANOVA.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Item</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Education content</td><td align="left" valign="top">3.8 (0.7; 3.6&#x2010;4.0)</td><td align="left" valign="top">4.3 (0.6; 4.1&#x2010;4.5)</td><td align="left" valign="top">4.5 (0.5; 4.4&#x2010;4.6)</td><td align="left" valign="top">21.364 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Education method</td><td align="left" valign="top">3.6 (0.8; 3.4&#x2010;3.8)</td><td align="left" valign="top">4.5 (0.6; 4.3&#x2010;4.7)</td><td align="left" valign="top">4.7 (0.4; 4.6&#x2010;4.8)</td><td align="left" valign="top">52.781 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Education time</td><td align="left" valign="top">3.9 (0.6; 3.7&#x2010;4.1)</td><td align="left" valign="top">4.2 (0.5; 4.1&#x2010;4.3)</td><td align="left" valign="top">4.3 (0.5; 4.2&#x2010;4.4)</td><td align="left" valign="top">9.526 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Education effect</td><td align="left" valign="top">3.7 (0.7; 3.5&#x2010;3.9)</td><td align="left" valign="top">4.4 (0.5; 4.3&#x2010;4.5)</td><td align="left" valign="top">4.6 (0.4; 4.5&#x2010;4.7)</td><td align="left" valign="top">43.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Overall satisfaction</td><td align="left" valign="top">3.8 (0.6; 3.6&#x2010;4.0)</td><td align="left" valign="top">4.4 (0.5; 4.3&#x2010;4.5)</td><td align="left" valign="top">4.6 (0.4; 4.5&#x2010;4.7)</td><td align="left" valign="top">45.293 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec><sec id="s3-6"><title>Comparison of Treatment Adherence</title><p>Among the 3 groups, significant differences in treatment adherence were observed across all dimensions, including drug treatment adherence, lifestyle adjustment, follow-up review, and total adherence score (all <italic>P</italic>&#x003C;.001). For drug treatment adherence, the control group scored a mean of 7.6 (SD 1.5; 95% CI 7.2&#x2010;8.0), which was lower than the mean scores for intervention group 1 (8.5, SD 1.2; 95% CI 8.2&#x2010;8.8) and intervention group 2 (9.1, SD 0.9; 95% CI 8.9&#x2010;9.3), with a statistically significant overall difference (<italic>F</italic><sub>2,171</sub>=25.364, <italic>P</italic>&#x003C;.001). Similar patterns were observed for lifestyle adjustment and follow-up review adherence. The 95% CIs of both intervention groups were consistently higher than those of the control group, with limited overlap, indicating that the observed improvements in treatment adherence were stable and precise. Regarding the total adherence score, the control group achieved a mean of 21.6 (SD 4.2; 95% CI 20.5&#x2010;22.7), whereas intervention group 1 and intervention group 2 achieved higher mean scores of 24.7 (SD 3.5; 95% CI 23.8&#x2010;25.6) and 26.6 (SD 2.8; 95% CI 25.9&#x2010;27.3), respectively. The between-group difference remained statistically significant (<italic>F</italic><sub>2,171</sub>=32.593, <italic>P</italic>&#x003C;.001; <xref ref-type="table" rid="table5">Table 5</xref>).</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Comparison of treatment adherence scores among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences among groups analyzed using 1-way ANOVA.</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Item</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Drug treatment</td><td align="left" valign="top">7.6 (1.5; 7.2&#x2010;8.0)</td><td align="left" valign="top">8.5 (1.2; 8.2&#x2010;8.8)</td><td align="left" valign="top">9.1 (0.9; 8.9&#x2010;9.3)</td><td align="left" valign="top">25.364 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Lifestyle adjustment</td><td align="left" valign="top">6.8 (1.7; 6.4&#x2010;7.2)</td><td align="left" valign="top">7.9 (1.4; 7.5&#x2010;8.3)</td><td align="left" valign="top">8.6 (1.1; 8.3&#x2010;8.9)</td><td align="left" valign="top">27.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Follow-up review</td><td align="left" valign="top">7.2 (1.6; 6.8&#x2010;7.6)</td><td align="left" valign="top">8.3 (1.3; 8.0&#x2010;8.6)</td><td align="left" valign="top">8.9 (1.0; 8.6&#x2010;9.2)</td><td align="left" valign="top">26.482 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Total score</td><td align="left" valign="top">21.6 (4.2; 20.5&#x2010;22.7)</td><td align="left" valign="top">24.7 (3.5; 23.8&#x2010;25.6)</td><td align="left" valign="top">26.6 (2.8; 25.9&#x2010;27.3)</td><td align="left" valign="top">32.593 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec><sec id="s3-7"><title>Comparison of Quality of Life</title><p>One month after discharge, significant differences were observed among the 3 groups in all SF-36 dimensions, including physical functioning, role-physical, bodily pain, general health, vitality, social functioning, role-emotional, and mental health (all <italic>P</italic>&#x003C;.001). For each dimension, the 95% CIs of both intervention groups were consistently higher than those of the control group, indicating that the observed improvements in health-related quality of life were robust and precise.</p><p>Notably, intervention group 2 generally showed narrower 95% CIs, suggesting less variability and greater consistency in patient-reported outcomes compared with the control group (<xref ref-type="table" rid="table6">Table 6</xref>).</p><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>Comparison of SF-36 quality of life scores 1 month after discharge among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences among groups analyzed using 1-way ANOVA.</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Dimension</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Physical functioning</td><td align="left" valign="top">65.3 (12.5; 61.7&#x2010;68.9)</td><td align="left" valign="top">72.6 (11.3; 69.5&#x2010;75.7)</td><td align="left" valign="top">74.8 (10.9; 71.8&#x2010;77.8)</td><td align="left" valign="top">11.364 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Role-physical</td><td align="left" valign="top">58.6 (14.2; 54.4&#x2010;62.8)</td><td align="left" valign="top">67.5 (12.8; 63.8&#x2010;71.2)</td><td align="left" valign="top">70.3 (12.1; 66.9&#x2010;73.7)</td><td align="left" valign="top">13.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Bodily pain</td><td align="left" valign="top">67.2 (13.1; 63.3&#x2010;71.1)</td><td align="left" valign="top">74.8 (11.9; 71.5&#x2010;78.1)</td><td align="left" valign="top">76.5 (11.2; 73.3&#x2010;79.7)</td><td align="left" valign="top">10.482 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">General health</td><td align="left" valign="top">61.5 (12.8; 57.7&#x2010;65.3)</td><td align="left" valign="top">69.7 (11.5; 66.5&#x2010;72.9)</td><td align="left" valign="top">72.4 (10.8; 69.3&#x2010;75.5)</td><td align="left" valign="top">14.293 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Vitality</td><td align="left" valign="top">59.8 (13.5; 55.8&#x2010;63.8)</td><td align="left" valign="top">68.3 (12.2; 65.1&#x2010;71.5)</td><td align="left" valign="top">71.6 (11.5; 68.5&#x2010;74.7)</td><td align="left" valign="top">15.364 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Social functioning</td><td align="left" valign="top">63.7 (12.9; 59.8&#x2010;67.6)</td><td align="left" valign="top">71.5 (11.7; 68.4&#x2010;74.6)</td><td align="left" valign="top">76.2 (10.8; 73.2&#x2010;79.2)</td><td align="left" valign="top">18.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Role-emotional</td><td align="left" valign="top">60.2 (14.1; 56.0&#x2010;64.4)</td><td align="left" valign="top">69.8 (12.5; 66.3&#x2010;73.3)</td><td align="left" valign="top">75.3 (11.7; 72.1&#x2010;78.5)</td><td align="left" valign="top">21.482 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Mental health</td><td align="left" valign="top">62.5 (13.2; 58.7&#x2010;66.3)</td><td align="left" valign="top">70.6 (12.1; 67.3&#x2010;73.9)</td><td align="left" valign="top">76.8 (11.3; 73.8&#x2010;79.8)</td><td align="left" valign="top">22.293 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec><sec id="s3-8"><title>Comparison of Anxiety and Depression</title><p>Before education, there were no significant differences in anxiety nor depression scores among the 3 groups (anxiety: <italic>F</italic><sub>2,171</sub>=0.068, <italic>P</italic>=.93; depression: <italic>F</italic><sub>2,171</sub>=0.185, <italic>P</italic>=.83), indicating baseline comparability. One month after discharge, significant differences were observed among the groups for both anxiety and depression scores (<italic>P</italic>&#x003C;.001). Specifically, for anxiety, the control group scored a mean of 6.5 (SD 2.9; 95% CI 5.8&#x2010;7.2), intervention group 1 scored a mean of 5.2 (SD 2.5; 95% CI 4.6&#x2010;5.8), and intervention group 2 scored a mean of 4.3 (SD 2.2; 95% CI 3.8&#x2010;4.8). For depression, the mean scores were 5.8 (SD 2.6; 95% CI 5.2&#x2010;6.4), 4.6 (SD 2.3; 95% CI 4.1&#x2010;5.1), and 3.7 (SD 2.0; 95% CI 3.3&#x2010;4.1), respectively.</p><p>The 95% CIs of the intervention groups did not overlap with those of the control group, indicating that the reductions in anxiety and depression after intervention were stable and precise (<xref ref-type="table" rid="table7">Table 7</xref>).</p><table-wrap id="t7" position="float"><label>Table 7.</label><caption><p>Comparison of Hospital Anxiety and Depression Scale (HADS) anxiety and depression scores among 3 groups of hospitalized patients with lung cancer in a randomized controlled trial of artificial intelligence (AI)&#x2013;assisted voice cloning education at The First Hospital of Lanzhou University (May 2025 through July 2025), with differences among groups analyzed using 1-way ANOVA.</p></caption><table id="table7" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Time point</td><td align="left" valign="bottom">Control group (n=58), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 1 (n=57), mean (SD; 95% CI)</td><td align="left" valign="bottom">Intervention group 2 (n=59), mean (SD; 95% CI)</td><td align="left" valign="bottom"><italic>F</italic> test (<italic>df</italic>)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Pre-education anxiety</td><td align="left" valign="top">7.8 (3.2; 7.0-8.6)</td><td align="left" valign="top">7.6 (3.1; 6.8-8.4)</td><td align="left" valign="top">7.7 (3.3; 6.9-8.5)</td><td align="left" valign="top">0.068 (2, 171)</td><td align="left" valign="top">.93</td></tr><tr><td align="left" valign="top">Pre-education depression</td><td align="left" valign="top">6.9 (2.8; 6.2-7.6)</td><td align="left" valign="top">7.1 (2.9; 6.4-7.8)</td><td align="left" valign="top">6.8 (2.7; 6.1-7.5)</td><td align="left" valign="top">0.185 (2, 171)</td><td align="left" valign="top">.83</td></tr><tr><td align="left" valign="top">1 month after discharge anxiety</td><td align="left" valign="top">6.5 (2.9; 5.8-7.2)</td><td align="left" valign="top">5.2 (2.5; 4.6-5.8)</td><td align="left" valign="top">4.3 (2.2; 3.8-4.8)</td><td align="left" valign="top">12.364 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">1 month after discharge depression</td><td align="left" valign="top">5.8 (2.6; 5.2-6.4)</td><td align="left" valign="top">4.6 (2.3; 4.1-5.1)</td><td align="left" valign="top">3.7 (2.0; 3.3-4.1)</td><td align="left" valign="top">13.175 (2, 171)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table></table-wrap></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This RCT investigated the effectiveness of an AI-assisted patient education system integrating voice cloning technology and compared the educational outcomes of physician voice cloning versus patient self-voice cloning. In addition, the study explored the feasibility of using ChatGPT as a supportive tool to assist with evaluating education effectiveness. Consistent with our 3 prespecified hypotheses, the results demonstrated that AI-assisted voice cloning&#x2013;based education significantly improved education content compliance, knowledge mastery, satisfaction, treatment adherence, and short-term psychological and quality-of-life outcomes compared with traditional education. These findings are consistent with prior research indicating that technology-assisted and personalized education can enhance patient engagement and learning outcomes [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. Notably, education delivered using patients&#x2019; own cloned voices yielded superior effects compared with physician voice cloning across multiple outcome domains, supporting the hypothesis that self-referential personalization enhances educational effectiveness.</p></sec><sec id="s4-2"><title>Interpretation of Findings and Comparison With Previous Studies</title><p>Voice cloning technology represents an emerging application of AI in patient education, offering distinct advantages over conventional education models. By enabling repeated exposure to standardized educational content delivered in a familiar and emotionally resonant voice, voice cloning may enhance attention, comprehension, and memory consolidation [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. Consistent with previous studies demonstrating that familiar auditory cues improve trust and acceptance in health communication, both AI-assisted intervention groups in this study outperformed the traditional education group in education compliance, knowledge mastery, and satisfaction [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. Beyond confirming the general effectiveness of voice-based AI education, this study provides novel evidence regarding differences between voice sources. Patients who received education via their own cloned voice achieved higher compliance, satisfaction, knowledge retention, treatment adherence, and better short-term psychological outcomes than those educated using physician voice cloning. Previous studies have largely focused on clinician-narrated or professionally recorded content [<xref ref-type="bibr" rid="ref34">34</xref>], whereas empirical comparisons involving patient self-voice have been scarce. The observed superiority of self-voice education may be explained by the self-reference effect, whereby information related to the self is processed more deeply and remembered more effectively [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Research in cognitive psychology suggests that self-related stimuli&#x2014;particularly self-generated or self-similar auditory information&#x2014;enhance attention, emotional engagement, and memory encoding [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]. In this study, hearing one&#x2019;s own voice narrate medical information may have strengthened self-identification and personal relevance, thereby reinforcing learning and adherence. Furthermore, the self-voice group demonstrated superior outcomes in multiple quality-of-life domains and had greater reductions in anxiety and depression. These findings align with evidence that personalized and self-relevant health communication can enhance perceived control, self-efficacy, and emotional regulation among patients [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Importantly, the reported effect sizes indicated moderate-to-large intervention effects, suggesting that the observed improvements were not only statistically significant but also clinically meaningful. Reporting both effect sizes and confidence intervals enhances transparency and interpretability and is consistent with CONSORT 2025 recommendations for RCTs [<xref ref-type="bibr" rid="ref28">28</xref>].</p></sec><sec id="s4-3"><title>Implications for AI-Assisted Evaluation of Patient Education</title><p>This study also explored the use of ChatGPT as an auxiliary tool to assist with evaluating patient education outcomes based on standardized educational content. Pre-experimental testing demonstrated good agreement between ChatGPT-assisted scoring and expert evaluation, suggesting that LLMs may support structured and reproducible assessment under controlled conditions. These findings are consistent with emerging literature indicating that LLMs can perform semantic analysis and text evaluation tasks with acceptable reliability in medical and educational contexts [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref37">37</xref>]. Compared with traditional manual evaluation, AI-assisted assessment may offer advantages in efficiency, scalability, and consistency, particularly in settings with limited human resources. However, ChatGPT should not be regarded as a fully objective evaluator. Assessment outcomes may still be influenced by prompt design, model architecture, and algorithmic biases. Therefore, AI-assisted evaluation should be considered a complementary tool rather than a replacement for expert assessment, and its use should be accompanied by appropriate human oversight [<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>].</p></sec><sec id="s4-4"><title>Study Limitations</title><p>Several limitations should be acknowledged. First, the follow-up period was limited to 1 month, restricting conclusions regarding long-term adherence, psychological outcomes, and sustained quality of life benefits. Longer follow-up periods are needed to assess the durability of intervention effects, as recommended in previous digital health education studies [<xref ref-type="bibr" rid="ref40">40</xref>]. Second, this was a single-center study involving hospitalized patients with relatively homogeneous educational backgrounds, which may limit generalizability. Prior research suggests that educational level and health literacy can influence the effectiveness of digital health interventions [<xref ref-type="bibr" rid="ref41">41</xref>]. Multicenter studies with more diverse populations are therefore warranted. Third, although voice cloning technology offers a promising approach, current speech synthesis systems still have limitations in emotional nuance and naturalness, which may affect user engagement [<xref ref-type="bibr" rid="ref13">13</xref>]. Continued technological advancements may further enhance the effectiveness of AI-assisted education. Finally, although AI-assisted education may reduce clinical workload and improve efficiency, this study did not include a formal cost-effectiveness analysis. Future research should systematically evaluate economic outcomes to support large-scale implementation decisions [<xref ref-type="bibr" rid="ref40">40</xref>]. Because ChatGPT is a continuously updated product, we standardized the scoring workflow (fixed rubric and prompts, prespecified audio format, and single scoring per recording) and provide the complete scoring materials to facilitate reproducibility. Additionally, voice-based delivery systems may misinterpret medication names or specialized terminology, underscoring the need for clinician review of generated content [<xref ref-type="bibr" rid="ref42">42</xref>].</p></sec><sec id="s4-5"><title>Implications for Practice and Policy</title><p>Implementation should address unit- and hospital-level factors separately. Targeted strategies should improve staff training, digital access, and policy support. Addressing demographic disparities is key to promoting equity and care quality.</p></sec><sec id="s4-6"><title>Conclusions and Broader Implications</title><p>This study introduces an innovative patient education model integrating AI voice cloning and ChatGPT, representing a novel approach distinct from previous studies that primarily relied on standard text-to-speech or professionally recorded content. The key innovation lies in using patients&#x2019; own cloned voices for health education delivery, leveraging the self-reference effect to enhance learning outcomes. Compared with prior research focusing on clinician-narrated content, this study provides the first empirical evidence that self-voice education produces superior outcomes across multiple domains including compliance, satisfaction, and psychological well-being. These findings contribute to the field by establishing a theoretical and practical framework for personalized AI-driven patient education. In real-world clinical settings, this approach offers a scalable, cost-effective solution to enhance patient engagement, particularly valuable in resource-limited environments where individualized education is challenging to deliver. Future research should focus on multicenter validation, longer follow-up periods, and exploration of optimal voice cloning parameters to maximize educational effectiveness.</p></sec></sec></body><back><ack><p>The authors thank all individuals who contributed to this study. Appreciation is extended to the patients and clinical staff of the First Hospital of Lanzhou University for their participation and cooperation throughout the research process. The authors also acknowledge the technical support provided by the development team for the artificial intelligence (AI) health education and voice cloning systems, led by author MM. The authors further thank all collaborators who assisted with data collection, statistical analysis, and manuscript preparation. Institutional support from the Department of Thoracic Surgery, The First Hospital of Lanzhou University, is also acknowledged for providing research resources and facilities.</p><p>In accordance with the Generative AI Delegation Taxonomy (GAIDeT; 2025), generative AI tools were used under full human supervision for proofreading, language editing, and translation. The tool used was ChatGPT (GPT-4; OpenAI). All AI-generated content was reviewed and validated by the authors. Responsibility for the final manuscript rests entirely with the authors. Generative AI tools are not listed as authors and did not contribute to study design, data analysis, or interpretation.</p></ack><notes><sec><title>Funding</title><p>This study was supported by the Key Research and Development Program of Gansu Province (26YFFA004); Natural Science Foundation of Gansu Province (23JRRA1597); the 2024 Key Research and Development Program of Gansu Province&#x2013;International Cooperation Field (24YFWA011), and the Institutional Research Fund of The First Hospital of Lanzhou University (ldyyyn2023-63).</p></sec><sec><title>Data Availability</title><p>The datasets generated and analyzed during this study are not publicly available due to privacy and ethical restrictions related to patient information and institutional regulations. Deidentified data are stored securely at The First Hospital of Lanzhou University and are available from the corresponding author upon reasonable request and approval by the hospital&#x2019;s ethics committee (approval number: ldyyyn2023-63).</p></sec></notes><fn-group><fn fn-type="con"><p>Conceptualization: YS (lead), HJ (supporting), YZ (supporting)</p><p>Data curation: SX (lead), XH (supporting), KJ (supporting), YZ (supporting)</p><p>Formal analysis: YS (equal), SX (equal), HJ (equal)</p><p>Methodology: YS (lead), SX (lead), HJ (supporting), XH (supporting), KJ (supporting), MM (technical lead)</p><p>Project administration: XM (lead)</p><p>Supervision: HW (lead)</p><p>Validation: YS (lead)</p><p>Visualization: SX (lead), XH (supporting)</p><p>Writing - original draft: YS (lead), SX (supporting), MM (advising)</p><p>Writing - review &#x0026; editing: YS (supporting), SX (supporting)</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">CONSORT</term><def><p>Consolidated Standards of Reporting Trials</p></def></def-item><def-item><term id="abb3">ITT</term><def><p>intention to treat</p></def></def-item><def-item><term id="abb4">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb5">RCT</term><def><p>randomized controlled trial</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kocaballi</surname><given-names>AB</given-names> </name><name name-style="western"><surname>Quiroz</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Rezazadegan</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Responses of conversational agents to health and lifestyle prompts: investigation of appropriateness and presentation structures</article-title><source>J Med Internet Res</source><year>2020</year><month>02</month><day>9</day><volume>22</volume><issue>2</issue><fpage>e15823</fpage><pub-id pub-id-type="doi">10.2196/15823</pub-id><pub-id pub-id-type="medline">32039810</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zolnierek</surname><given-names>KBH</given-names> </name><name name-style="western"><surname>Dimatteo</surname><given-names>MR</given-names> </name></person-group><article-title>Physician communication and patient adherence to treatment: a meta-analysis</article-title><source>Med Care</source><year>2009</year><month>08</month><volume>47</volume><issue>8</issue><fpage>826</fpage><lpage>834</lpage><pub-id pub-id-type="doi">10.1097/MLR.0b013e31819a5acc</pub-id><pub-id pub-id-type="medline">19584762</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fitzpatrick</surname><given-names>KK</given-names> </name><name name-style="western"><surname>Darcy</surname><given-names>A</given-names> </name><name name-style="western"><surname>Vierhile</surname><given-names>M</given-names> </name></person-group><article-title>Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial</article-title><source>JMIR Ment Health</source><year>2017</year><month>06</month><day>6</day><volume>4</volume><issue>2</issue><fpage>e19</fpage><pub-id pub-id-type="doi">10.2196/mental.7785</pub-id><pub-id pub-id-type="medline">28588005</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shersher</surname><given-names>V</given-names> </name><name name-style="western"><surname>Haines</surname><given-names>TP</given-names> </name><name name-style="western"><surname>Sturgiss</surname><given-names>L</given-names> </name><name name-style="western"><surname>Weller</surname><given-names>C</given-names> </name><name name-style="western"><surname>Williams</surname><given-names>C</given-names> </name></person-group><article-title>Definitions and use of the teach-back method in healthcare consultations with patients: a systematic review and thematic synthesis</article-title><source>Patient Educ Couns</source><year>2021</year><month>01</month><volume>104</volume><issue>1</issue><fpage>118</fpage><lpage>129</lpage><pub-id pub-id-type="doi">10.1016/j.pec.2020.07.026</pub-id><pub-id pub-id-type="medline">32798080</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Singhal</surname><given-names>K</given-names> </name><name name-style="western"><surname>Azizi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Large language models encode clinical knowledge</article-title><source>Nature New Biol</source><year>2023</year><month>08</month><volume>620</volume><issue>7972</issue><fpage>172</fpage><lpage>180</lpage><pub-id pub-id-type="doi">10.1038/s41586-023-06291-2</pub-id><pub-id pub-id-type="medline">37438534</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thirunavukarasu</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSJ</given-names> </name><name name-style="western"><surname>Elangovan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gutierrez</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>TF</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSW</given-names> </name></person-group><article-title>Large language models in medicine</article-title><source>Nat Med</source><year>2023</year><month>08</month><volume>29</volume><issue>8</issue><fpage>1930</fpage><lpage>1940</lpage><pub-id pub-id-type="doi">10.1038/s41591-023-02448-8</pub-id><pub-id pub-id-type="medline">37460753</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Denecke</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gabarron</surname><given-names>E</given-names> </name><name name-style="western"><surname>Grainger</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Artificial intelligence for participatory health: applications, impact, and future implications</article-title><source>Yearb Med Inform</source><year>2019</year><month>08</month><volume>28</volume><issue>1</issue><fpage>165</fpage><lpage>173</lpage><pub-id pub-id-type="doi">10.1055/s-0039-1677902</pub-id><pub-id pub-id-type="medline">31022749</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ayers</surname><given-names>JW</given-names> </name><name name-style="western"><surname>Poliak</surname><given-names>A</given-names> </name><name name-style="western"><surname>Dredze</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum</article-title><source>JAMA Intern Med</source><year>2023</year><month>06</month><day>1</day><volume>183</volume><issue>6</issue><fpage>589</fpage><lpage>596</lpage><pub-id pub-id-type="doi">10.1001/jamainternmed.2023.1838</pub-id><pub-id pub-id-type="medline">37115527</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Milne-Ives</surname><given-names>M</given-names> </name><name name-style="western"><surname>de Cock</surname><given-names>C</given-names> </name><name name-style="western"><surname>Lim</surname><given-names>E</given-names> </name><etal/></person-group><article-title>The effectiveness of artificial intelligence conversational agents in health care: systematic review</article-title><source>J Med Internet Res</source><year>2020</year><month>10</month><day>22</day><volume>22</volume><issue>10</issue><fpage>e20346</fpage><pub-id pub-id-type="doi">10.2196/20346</pub-id><pub-id pub-id-type="medline">33090118</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schachner</surname><given-names>T</given-names> </name><name name-style="western"><surname>Keller</surname><given-names>R</given-names> </name><name name-style="western"><surname>V Wangenheim</surname><given-names>F</given-names> </name></person-group><article-title>Artificial intelligence-based conversational agents for chronic conditions: systematic literature review</article-title><source>J Med Internet Res</source><year>2020</year><month>09</month><day>14</day><volume>22</volume><issue>9</issue><fpage>e20701</fpage><pub-id pub-id-type="doi">10.2196/20701</pub-id><pub-id pub-id-type="medline">32924957</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Palanica</surname><given-names>A</given-names> </name><name name-style="western"><surname>Flaschner</surname><given-names>P</given-names> </name><name name-style="western"><surname>Thommandram</surname><given-names>A</given-names> </name><name name-style="western"><surname>Li</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fossat</surname><given-names>Y</given-names> </name></person-group><article-title>Physicians&#x2019; perceptions of chatbots in health care: cross-sectional web-based survey</article-title><source>J Med Internet Res</source><year>2019</year><month>04</month><day>5</day><volume>21</volume><issue>4</issue><fpage>e12887</fpage><pub-id pub-id-type="doi">10.2196/12887</pub-id><pub-id pub-id-type="medline">30950796</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>S</given-names> </name></person-group><article-title>Utility of ChatGPT in clinical practice</article-title><source>J Med Internet Res</source><year>2023</year><month>06</month><day>28</day><volume>25</volume><fpage>e48568</fpage><pub-id pub-id-type="doi">10.2196/48568</pub-id><pub-id pub-id-type="medline">37379067</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sisman</surname><given-names>B</given-names> </name><name name-style="western"><surname>Yamagishi</surname><given-names>J</given-names> </name><name name-style="western"><surname>King</surname><given-names>S</given-names> </name><name name-style="western"><surname>Li</surname><given-names>H</given-names> </name></person-group><article-title>An overview of voice conversion and its challenges: from statistical modeling to deep learning</article-title><source>IEEE/ACM Trans Audio Speech Lang Process</source><year>2020</year><volume>29</volume><fpage>132</fpage><lpage>157</lpage><pub-id pub-id-type="doi">10.1109/TASLP.2020.3038524</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Dai</surname><given-names>G</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>P</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Ying</surname><given-names>F</given-names> </name></person-group><article-title>InnerSelf: designing self-deepfaked voice for emotional well-being</article-title><source>arXiv</source><access-date>2026-03-03</access-date><comment>Preprint posted online on  Mar 18, 2025</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2503.14257">https://arxiv.org/abs/2503.14257</ext-link></comment><pub-id pub-id-type="doi">10.48550/arXiv.2503.14257</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nygaard</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Pisoni</surname><given-names>DB</given-names> </name></person-group><article-title>Talker-specific learning in speech perception</article-title><source>Percept Psychophys</source><year>1998</year><month>04</month><volume>60</volume><issue>3</issue><fpage>355</fpage><lpage>376</lpage><pub-id pub-id-type="doi">10.3758/bf03206860</pub-id><pub-id pub-id-type="medline">9599989</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kaplan</surname><given-names>JT</given-names> </name><name name-style="western"><surname>Aziz-Zadeh</surname><given-names>L</given-names> </name><name name-style="western"><surname>Uddin</surname><given-names>LQ</given-names> </name><name name-style="western"><surname>Iacoboni</surname><given-names>M</given-names> </name></person-group><article-title>The self across the senses: an fMRI study of self-face and self-voice recognition</article-title><source>Soc Cogn Affect Neurosci</source><year>2008</year><month>09</month><volume>3</volume><issue>3</issue><fpage>218</fpage><lpage>223</lpage><pub-id pub-id-type="doi">10.1093/scan/nsn014</pub-id><pub-id pub-id-type="medline">19015113</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yonan</surname><given-names>CA</given-names> </name><name name-style="western"><surname>Sommers</surname><given-names>MS</given-names> </name></person-group><article-title>The effects of talker familiarity on spoken word identification in younger and older listeners</article-title><source>Psychol Aging</source><year>2000</year><volume>15</volume><issue>1</issue><fpage>88</fpage><lpage>99</lpage><pub-id pub-id-type="doi">10.1037/0882-7974.15.1.88</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aydin</surname><given-names>S</given-names> </name><name name-style="western"><surname>Karabacak</surname><given-names>M</given-names> </name><name name-style="western"><surname>Vlachos</surname><given-names>V</given-names> </name><name name-style="western"><surname>Margetis</surname><given-names>K</given-names> </name></person-group><article-title>Large language models in patient education: a scoping review of applications in medicine</article-title><source>Front Med (Lausanne)</source><year>2024</year><volume>11</volume><fpage>1477898</fpage><pub-id pub-id-type="doi">10.3389/fmed.2024.1477898</pub-id><pub-id pub-id-type="medline">39534227</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Busch</surname><given-names>F</given-names> </name><name name-style="western"><surname>Hoffmann</surname><given-names>L</given-names> </name><name name-style="western"><surname>Rueger</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Current applications and challenges in large language models for patient care: a systematic review</article-title><source>Commun Med (Lond)</source><year>2025</year><month>01</month><day>21</day><volume>5</volume><issue>1</issue><fpage>26</fpage><pub-id pub-id-type="doi">10.1038/s43856-024-00717-2</pub-id><pub-id pub-id-type="medline">39838160</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="other"><person-group person-group-type="author"><collab>OpenAI</collab><name name-style="western"><surname>Achiam</surname><given-names>J</given-names> </name><name name-style="western"><surname>Adler</surname><given-names>S</given-names> </name><etal/></person-group><article-title>GPT-4 technical report</article-title><source>arXiv</source><access-date>2026-03-03</access-date><comment>Preprint posted online on  Mar 15, 2023</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2303.08774">https://arxiv.org/abs/2303.08774</ext-link></comment><pub-id pub-id-type="doi">10.48550/arXiv.2303.08774</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sallam</surname><given-names>M</given-names> </name></person-group><article-title>ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns</article-title><source>Healthcare (Basel)</source><year>2023</year><month>03</month><day>19</day><volume>11</volume><issue>6</issue><fpage>887</fpage><pub-id pub-id-type="doi">10.3390/healthcare11060887</pub-id><pub-id pub-id-type="medline">36981544</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chavez-Yenter</surname><given-names>D</given-names> </name><name name-style="western"><surname>Kimball</surname><given-names>KE</given-names> </name><name name-style="western"><surname>Kohlmann</surname><given-names>W</given-names> </name><etal/></person-group><article-title>Patient interactions with an automated conversational agent delivering pretest genetics education: descriptive study</article-title><source>J Med Internet Res</source><year>2021</year><month>11</month><day>18</day><volume>23</volume><issue>11</issue><fpage>e29447</fpage><pub-id pub-id-type="doi">10.2196/29447</pub-id><pub-id pub-id-type="medline">34792472</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bickmore</surname><given-names>TW</given-names> </name><name name-style="western"><surname>Pfeifer</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Byron</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Usability of conversational agents by patients with inadequate health literacy: evidence from two clinical trials</article-title><source>J Health Commun</source><year>2010</year><volume>15 Suppl 2</volume><fpage>197</fpage><lpage>210</lpage><pub-id pub-id-type="doi">10.1080/10810730.2010.499991</pub-id><pub-id pub-id-type="medline">20845204</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nadarzynski</surname><given-names>T</given-names> </name><name name-style="western"><surname>Miles</surname><given-names>O</given-names> </name><name name-style="western"><surname>Cowie</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ridge</surname><given-names>D</given-names> </name></person-group><article-title>Acceptability of artificial intelligence (AI)-led chatbot services in healthcare: a mixed-methods study</article-title><source>Digit Health</source><year>2019</year><volume>5</volume><fpage>2055207619871808</fpage><pub-id pub-id-type="doi">10.1177/2055207619871808</pub-id><pub-id pub-id-type="medline">31467682</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sezgin</surname><given-names>E</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Ramtekkar</surname><given-names>U</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>S</given-names> </name></person-group><article-title>Readiness for voice assistants to support healthcare delivery during a health crisis and pandemic</article-title><source>NPJ Digit Med</source><year>2020</year><volume>3</volume><fpage>122</fpage><pub-id pub-id-type="doi">10.1038/s41746-020-00332-0</pub-id><pub-id pub-id-type="medline">33015374</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sezgin</surname><given-names>E</given-names> </name><name name-style="western"><surname>Militello</surname><given-names>LK</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>S</given-names> </name></person-group><article-title>A scoping review of patient-facing, behavioral health interventions with voice assistant technology targeting self-management and healthy lifestyle behaviors</article-title><source>Transl Behav Med</source><year>2020</year><month>08</month><day>7</day><volume>10</volume><issue>3</issue><fpage>606</fpage><lpage>628</lpage><pub-id pub-id-type="doi">10.1093/tbm/ibz141</pub-id><pub-id pub-id-type="medline">32766865</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Brown</surname><given-names>T</given-names> </name><name name-style="western"><surname>Mann</surname><given-names>B</given-names> </name><name name-style="western"><surname>Ryder</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Language models are few-shot learners</article-title><access-date>2026-03-03</access-date><conf-name>Advances in Neural Information Processing Systems 33 (NeurIPS 2020)</conf-name><conf-date>Dec 6-12, 2020</conf-date><conf-loc>Virtual-only conference</conf-loc><fpage>1877</fpage><lpage>1901</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://papers.nips.cc/paper_files/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html">https://papers.nips.cc/paper_files/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hopewell</surname><given-names>S</given-names> </name><name name-style="western"><surname>Chan</surname><given-names>AW</given-names> </name><name name-style="western"><surname>Collins</surname><given-names>GS</given-names> </name><etal/></person-group><article-title>CONSORT 2025 statement: updated guideline for reporting randomised trials</article-title><source>BMJ</source><year>2025</year><month>04</month><day>14</day><volume>389</volume><fpage>e081123</fpage><pub-id pub-id-type="doi">10.1136/bmj-2024-081123</pub-id><pub-id pub-id-type="medline">40228833</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Eysenbach</surname><given-names>G</given-names> </name><name name-style="western"><surname>Group</surname><given-names>CE</given-names> </name></person-group><article-title>CONSORT-EHEALTH: improving and standardizing evaluation reports of web-based and mobile health interventions</article-title><source>J Med Internet Res</source><year>2011</year><month>12</month><day>31</day><volume>13</volume><issue>4</issue><fpage>e126</fpage><pub-id pub-id-type="doi">10.2196/jmir.1923</pub-id><pub-id pub-id-type="medline">22209829</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gilson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Safranek</surname><given-names>CW</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>T</given-names> </name><etal/></person-group><article-title>How does ChatGPT perform on the United States Medical Licensing Examination (USMLE)? The implications of large language models for medical education and knowledge assessment</article-title><source>JMIR Med Educ</source><year>2023</year><month>02</month><day>8</day><volume>9</volume><fpage>e45312</fpage><pub-id pub-id-type="doi">10.2196/45312</pub-id><pub-id pub-id-type="medline">36753318</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Battineni</surname><given-names>G</given-names> </name><name name-style="western"><surname>Chintalapudi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Amenta</surname><given-names>F</given-names> </name></person-group><article-title>AI chatbot design during an epidemic like the novel coronavirus</article-title><source>Healthcare (Basel)</source><year>2020</year><month>06</month><day>3</day><volume>8</volume><issue>2</issue><fpage>154</fpage><pub-id pub-id-type="doi">10.3390/healthcare8020154</pub-id><pub-id pub-id-type="medline">32503298</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nayak</surname><given-names>A</given-names> </name><name name-style="western"><surname>Vakili</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nayak</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Use of voice-based conversational artificial intelligence for basal insulin prescription management among patients with type 2 diabetes: a randomized clinical trial</article-title><source>JAMA Netw Open</source><year>2023</year><month>12</month><day>1</day><volume>6</volume><issue>12</issue><fpage>e2340232</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2023.40232</pub-id><pub-id pub-id-type="medline">38039007</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bickmore</surname><given-names>TW</given-names> </name><name name-style="western"><surname>Pfeifer</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Paasche-Orlow</surname><given-names>MK</given-names> </name></person-group><article-title>Using computer agents to explain medical documents to patients with low health literacy</article-title><source>Patient Educ Couns</source><year>2009</year><month>06</month><volume>75</volume><issue>3</issue><fpage>315</fpage><lpage>320</lpage><pub-id pub-id-type="doi">10.1016/j.pec.2009.02.007</pub-id><pub-id pub-id-type="medline">19297116</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Laranjo</surname><given-names>L</given-names> </name><name name-style="western"><surname>Dunn</surname><given-names>AG</given-names> </name><name name-style="western"><surname>Tong</surname><given-names>HL</given-names> </name><etal/></person-group><article-title>Conversational agents in healthcare: a systematic review</article-title><source>J Am Med Inform Assoc</source><year>2018</year><month>09</month><day>1</day><volume>25</volume><issue>9</issue><fpage>1248</fpage><lpage>1258</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocy072</pub-id><pub-id pub-id-type="medline">30010941</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sui</surname><given-names>J</given-names> </name><name name-style="western"><surname>Humphreys</surname><given-names>GW</given-names> </name></person-group><article-title>The integrative self: how self-reference integrates perception and memory</article-title><source>Trends Cogn Sci</source><year>2015</year><month>12</month><volume>19</volume><issue>12</issue><fpage>719</fpage><lpage>728</lpage><pub-id pub-id-type="doi">10.1016/j.tics.2015.08.015</pub-id><pub-id pub-id-type="medline">26447060</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rogers</surname><given-names>TB</given-names> </name><name name-style="western"><surname>Kuiper</surname><given-names>NA</given-names> </name><name name-style="western"><surname>Kirker</surname><given-names>WS</given-names> </name></person-group><article-title>Self-reference and the encoding of personal information</article-title><source>J Pers Soc Psychol</source><year>1977</year><volume>35</volume><issue>9</issue><fpage>677</fpage><lpage>688</lpage><pub-id pub-id-type="doi">10.1037/0022-3514.35.9.677</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kung</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Cheatham</surname><given-names>M</given-names> </name><name name-style="western"><surname>Medenilla</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models</article-title><source>PLOS Digit Health</source><year>2023</year><month>02</month><volume>2</volume><issue>2</issue><fpage>e0000198</fpage><pub-id pub-id-type="doi">10.1371/journal.pdig.0000198</pub-id><pub-id pub-id-type="medline">36812645</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>GH</given-names> </name><name name-style="western"><surname>Hoelscher</surname><given-names>SH</given-names> </name></person-group><article-title>The ChatGPT storm and what faculty can do</article-title><source>Nurse Educ</source><year>2023</year><volume>48</volume><issue>3</issue><fpage>119</fpage><lpage>124</lpage><pub-id pub-id-type="doi">10.1097/NNE.0000000000001390</pub-id><pub-id pub-id-type="medline">37043716</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title><source>Nat Med</source><year>2019</year><month>01</month><volume>25</volume><issue>1</issue><fpage>44</fpage><lpage>56</lpage><pub-id pub-id-type="doi">10.1038/s41591-018-0300-7</pub-id><pub-id pub-id-type="medline">30617339</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Murray</surname><given-names>E</given-names> </name><name name-style="western"><surname>Hekler</surname><given-names>EB</given-names> </name><name name-style="western"><surname>Andersson</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Evaluating digital health interventions: key questions and approaches</article-title><source>Am J Prev Med</source><year>2016</year><month>11</month><volume>51</volume><issue>5</issue><fpage>843</fpage><lpage>851</lpage><pub-id pub-id-type="doi">10.1016/j.amepre.2016.06.008</pub-id><pub-id pub-id-type="medline">27745684</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Paasche-Orlow</surname><given-names>MK</given-names> </name><name name-style="western"><surname>Wolf</surname><given-names>MS</given-names> </name></person-group><article-title>The causal pathways linking health literacy to health outcomes</article-title><source>Am J Health Behav</source><year>2007</year><volume>31 Suppl 1</volume><issue>1</issue><fpage>S19</fpage><lpage>S26</lpage><pub-id pub-id-type="doi">10.5555/ajhb.2007.31.supp.S19</pub-id><pub-id pub-id-type="medline">17931132</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Palanica</surname><given-names>A</given-names> </name><name name-style="western"><surname>Thommandram</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>A</given-names> </name><name name-style="western"><surname>Li</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fossat</surname><given-names>Y</given-names> </name></person-group><article-title>Do you understand the words that are comin outta my mouth? Voice assistant comprehension of medication names</article-title><source>NPJ Digit Med</source><year>2019</year><volume>2</volume><fpage>55</fpage><pub-id pub-id-type="doi">10.1038/s41746-019-0133-x</pub-id><pub-id pub-id-type="medline">31304401</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>ChatGPT-based compliance scoring: prompts, rubric, and example inputs/outputs.</p><media xlink:href="jmir_v28i1e81387_app1.pdf" xlink:title="PDF File, 172 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Education content compliance, knowledge mastery, and treatment adherence questionnaires.</p><media xlink:href="jmir_v28i1e81387_app2.png" xlink:title="PNG File, 187 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>SF-36 questionnaire.</p><media xlink:href="jmir_v28i1e81387_app3.png" xlink:title="PNG File, 784 KB"/></supplementary-material><supplementary-material id="app4"><label>Multimedia Appendix 4</label><p>Hospital Anxiety and Depression Scale.</p><media xlink:href="jmir_v28i1e81387_app4.png" xlink:title="PNG File, 797 KB"/></supplementary-material><supplementary-material id="app5"><label>Multimedia Appendix 5</label><p>Voice cloning patient education prompts, ChatGPT evaluation prompts, example patient knowledge assessment sheet, and notes for researchers.</p><media xlink:href="jmir_v28i1e81387_app5.docx" xlink:title="DOCX File, 1029 KB"/></supplementary-material><supplementary-material id="app6"><label>Checklist 1</label><p>CONSORT 2025 checklist.</p><media xlink:href="jmir_v28i1e81387_app6.pdf" xlink:title="PDF File, 154 KB"/></supplementary-material><supplementary-material id="app7"><label>Checklist 2</label><p>CONSORT-EHEALTH checklist (V 1.6.1).</p><media xlink:href="jmir_v28i1e81387_app7.pdf" xlink:title="PDF File, 202 KB"/></supplementary-material></app-group></back></article>