Original Paper
Abstract
Background: Patients often struggle to understand standard hospital discharge letters, increasing the risk of medication errors and misunderstandings. According to cognitive load theory (CLT), complex, information-dense texts can overload working memory and impair comprehension. Artificial intelligence tools that generate patient-centered versions could help reduce extraneous cognitive load and bridge this gap. However, evidence for their effectiveness remains limited.
Objective: This study aimed to evaluate whether GPT-4 (OpenAI)–generated patient-centered letters improve standardized patients’ retention and understanding of safety-relevant medical information compared with standard hospital discharge letters, and to explore potential effects on cognitive load as described by CLT.
Methods: In this prospective, randomized, crossover study, 48 trained standardized patients received a conventional discharge letter for an assigned disease (out of 3) and its matching GPT-4–generated patient-centered letter. Participants read one version first, identified predefined safety-relevant “learning objectives,” and then repeated the task with the alternate version. The primary outcome was the proportion of learning objectives fully, partially, or not reported. In a secondary analysis, results were stratified by content field (Medication, Organization, Prevention of Complications, Lifestyle/Disease Management) and Bloom taxonomy level (“Remember,” “Understand”).
Results: The letter type significantly influenced comprehension (odds ratio [OR] 1.74, 95% CI 1.45-2.08; P<.001). Patient letters, compared with discharge letters, led to higher rates of fully (490/1073, 45.7% vs 413/1073, 38.5%) or partially (322/1073, 30% vs 287/1073, 26.7%) stated learning objectives and fewer omissions (261/1073, 24.3% vs 373/1073, 34.8%). Participants performed better on “Remember” than on “Understand” learning objectives, regardless of letter type (OR 3.33, 95% CI 1.96-5.88; P<.001). Compared with standard hospital discharge letters, patient letters consistently improved results at both cognitive levels (“Remember”: 278/545, 51% vs 242/545, 44.4%; “Understand”: 212/528, 40.2% vs 171/528, 32.4% fully stated). The effect of patient letters varied by content field (P<.001). The greatest improvements were observed for “Medication” (170/254, 66.9% vs 129/254, 50.8% fully stated) and “Organization” (78/158, 49.4% vs 62/158, 39.2% fully stated). Improvements in the content field “Prevention of Complications” were modest, and those for “Lifestyle/Disease Management” were even smaller across all conditions. A total of 24.3% (261/1073) of key information remained unrecognized.
Conclusions: In this explanatory study, GPT-4–generated patient letters improved comprehension of safety-relevant discharge information among standardized patients, particularly regarding medication and organizational aspects. However, they were less effective in supporting higher-order understanding, such as risk prevention or lifestyle management. These hypothesis-driven findings can be interpreted within a CLT framework and may motivate prospective evaluation of multimodal, iterative supports.
doi:10.2196/81243
Keywords
Introduction
The transition from inpatient to outpatient care represents a critical juncture for information exchange and patient safety, often marked by communication errors and incomplete understanding []. Nearly half (49%) of patients experience at least 1 postdischarge medical error, most frequently concerning medication management, diagnostic workup, or test result follow-up []. Effective patient understanding during this phase is crucial, as misinterpretation or omission of medical instructions can directly impact adherence, follow-up care, and safety outcomes [,]. Studies estimate that up to 78% of patients misunderstand key elements of their hospital care and discharge instructions [], and approximately half (50.8%) of patients receiving polypharmacy misinterpret at least 1 dosage instruction [].
Patients’ comprehension of discharge information is influenced by multiple factors, including health literacy, educational background, prior knowledge, and emotional stress during this vulnerable phase [,]. Considering that 36% of US adults have limited health literacy [], a substantial proportion of patients may struggle to fully understand and follow postdischarge instructions, which are typically written for communication between health care professionals and often fail to meet patients’ informational needs []. Therefore, there is a clear need for discharge documents that are patient-centered, clearly structured, and accessible; these will be referred to as patient letters throughout this manuscript.
Recent advances in artificial intelligence (AI), particularly large language models (LLMs) such as GPT-4 (OpenAI), offer a promising approach to address this need []. By transforming complex medical documents into plain-language patient letters, AI has the potential to improve patients’ comprehension and recall of important medical information. Several feasibility studies have demonstrated that LLMs can generate readable and accurate patient letters from standard discharge documents, highlighting their technical capability to simplify language, structure information, and provide consistent formatting [-]. To date, only a small number of studies have investigated the impact of GPT-generated patient letters on patient understanding [-]. These studies have primarily relied on semistructured interviews or multiple-choice assessments, offering only quantitative evaluations of patient knowledge. A more nuanced understanding of patient comprehension is still lacking, particularly regarding which types of information are retained, how the cognitive complexity of content influences understanding, and whether comprehension differs across medical content domains. Accordingly, evidence on the educational impact and cognitive benefits of AI-generated patient communications remains limited.
This study is based on our previous work, in which we explored the feasibility of transforming discharge letters into patient letters using GPT-4 []. That study primarily focused on the accuracy of the generated letters and measures of patient-centeredness, demonstrating that overall, GPT-4 can produce readable and structured patient-directed documents, with some limitations including omissions. Building on these findings, this study shifts the focus to an application-oriented perspective. It investigates whether GPT-4–generated patient letters improve comprehension, using standardized patients (SPs) to simulate real-world understanding and follow-up tasks.
To systematically evaluate comprehension, we framed patient understanding as a learning process informed by cognitive load theory (CLT) []. According to CLT, learning efficiency depends on three types of cognitive load:
- Intrinsic load refers to the inherent complexity of the information and is dependent on the learner’s level of expertise. In the context of discharge letters, intrinsic load is generally high due to complex medical content and procedural details, and it can only be modified to a limited extent.
- Extraneous load reflects mental resources allocated to elements that do not contribute to learning or schema acquisition. High extraneous load can arise from poorly structured information, excessive jargon, or frequent use of abbreviations []. In our prior study, we demonstrated that GPT-4–generated patient letters mitigated these contributing factors by providing increased repetition of key points, fewer abbreviations, and reduced use of medical terminology [].
- Germane load represents the mental resources devoted to acquiring and automating schemata in long-term memory. In our context, the goal is to maximize germane load so that patients allocate as much cognitive energy as possible to truly understand, retain, and integrate the medical information.
Guided by this theoretical framework, the study addressed the following research questions:
- Does the reduction of extraneous load through AI-generated patient letters translate into improved comprehension of safety-relevant information?
- Are comprehension gains consistent across different content domains (eg, Medication, Organization, Prevention of Complications, and Lifestyle)?
- Do comprehension improvements differ between lower-order (“Remember”) and higher-order (“Understand”) cognitive processes as defined by Bloom taxonomy []?
By combining a patient-centered perspective with a learning-theoretical framework, this study examines the educational potential and limitations of AI-generated patient letters and considers how they might contribute to patient safety during vulnerable care transitions.
Methods
Study Design and Setting
In our previous study, we developed a methodology for generating GPT-4–based patient letters, demonstrating the feasibility of transforming discharge letters into patient-friendly formats while maintaining medical accuracy and patient-centeredness. The prompting strategy and human review process were described in detail in that publication [].
Building directly on this work, this prospective, randomized, crossover study with SPs used 3 discharge letters and the corresponding GPT-4–generated patient letters from our prior study, each representing a common chronic condition: arterial hypertension, type 2 diabetes mellitus, and diabetic kidney disease. These diagnoses were selected because they are highly prevalent and frequently occur in both inpatient and outpatient care. The letters are provided in in German (original) and English (translated).
Both letter formats were presented to 48 SPs using a crossover design (). We compared perception and retention of key medical information—defined as “learning objectives”—presented in the letters.

Participants and Study Procedures
A total of 48 SPs participated in our study (). SPs are individuals who are specifically trained as patient actors to portray medical patients in a consistent and realistic manner, and they participated in the study in their specific role [,]. The sample size was determined by the number of participants available for recruitment and was not calculated prior to the study. The SPs were recruited from the SP programs of the Universities of Tübingen and Erlangen, Germany. Data were collected between May and September 2024. After reading a brief case vignette covering basic medical and socioeconomic information, each participant was randomized to the discharge letter (group A) or its patient-friendly version (group B). Participants were not informed about the sequence in which they would receive the letters. However, the differences between the discharge letter and the GPT-4–generated patient letter were apparent, making full blinding impossible. Therefore, no formal assessment of blinding adequacy was performed.
At both sites, letters were presented digitally through a web-based interface. Participants were instructed to read each letter with care and were permitted to scroll back to the text while answering the open-ended questions. This setting mirrors the real-world access to written discharge information. The SPs extracted key information from 4 content fields and answered predefined open questions (). This approach was chosen because continuous access limits short-term memory bias and mirrors real practice []. To foster active knowledge construction, participants were instructed to answer in their own words without copying. Letters were then crossed over, and the task was repeated with the alternate version.
Learning Objectives and Content Fields
For each of the 3 clinical cases (arterial hypertension, diabetes mellitus, and diabetic kidney disease), we defined 24 safety-related learning objectives that cover information considered essential by the treating physicians for patient self-management after discharge (eg, knowledge of medication changes, warning symptoms, and follow-up appointments). The learning objectives were designed and revised by an interdisciplinary committee of 4 physicians (specialized in internal medicine, nephrology, diabetology, general medicine, and medical education) based on current clinical guidelines [-] and their professional experience. They were formulated as short, observable statements (eg, “The patient can ... list” and “The patient understands that ...”) and checked for comprehensibility and redundancy.
For each case, the 24 learning objectives were evenly distributed across the levels of the revised Bloom taxonomy, with 12 learning objectives in the “Remember” category and 12 in the “Understand” category. In Bloom taxonomy, the “Remember” category refers to the ability to recall facts and basic concepts, while the “Understand” category goes beyond recall and involves interpreting, explaining, and making sense of the information [,]. Each learning objective was further mapped to 1 of 4 content fields: “Organization” (follow-up tasks), “Medication” (drug changes), “Prevention of Complications” (early detection and management of disease-related complications), and “Lifestyle/Disease Management” (self-care advice/lifestyle modifications; ). This structure ensured comparable cognitive demands and content coverage across all 3 cases and enabled stratified analyses by content area and Bloom category. The process of defining learning objectives and creating the original discharge letters has been described in detail in our previous publication [].
Across the 3 letters used in this study, 72 learning objectives were assessed in total (24 per letter). A total of 5 of these 72 learning objectives were not represented in the GPT-4–generated patient letters because the model did not carry them over from the original discharge letters. These learning objectives were therefore excluded from the analysis for both letter types (). The quality and completeness of the translation from discharge letter to patient letter were examined in detail in our previous publication and were therefore not reassessed within the scope of this study []. After this exclusion, 34 of the 67 remaining learning objectives were classified as “Remember” and 33 as “Understand” according to Bloom taxonomy ().
Assessment Approach
Reading the discharge or patient letter and then answering the case questions was designed as a structured learning assessment sequence []: the letters provided the information to be learned, and the SPs’ responses reflected what they had actually understood. To capture constructed understanding rather than simple recognition, we used open-ended questions that required participants to formulate their answers in their own words rather than choosing from predefined options.
Open-ended response formats are less prone to guessing than multiple-choice questions and are better suited for assessing higher levels of processing for learning objectives at the “Understand” level. This approach closely aligns the assessment with the predefined learning objectives and the intended depth of understanding. This choice is supported by assessment literature [].
Outcome Measures
The primary outcome was the understanding of safety-related information at the level of individual learning objectives. Each participant response (participant × learning objective × letter) was rated on an ordinal scale with 3 categories reflecting the degree of goal achievement in the open-ended response (0=not specified, 1=partially specified, 2=fully specified). The main comparison of interest was the difference between the patient letters generated by GPT-4 and the standard discharge letters with respect to this ordinal measure of SPs’ understanding.
Secondary end points were the distribution of this ordinal comprehension metric across the 4 predefined content areas (Organization, Medication, Prevention of Complications, and Lifestyle/Disease Management) and the distribution across the levels of Bloom taxonomy (“Remember” vs “Understand”). As an additional secondary end point, the frequency of medically incorrect answers was examined.
Randomization
Participants at the University Hospital Tübingen were allocated to the 3 clinical cases and the order of letter presentation (discharge letter first vs patient letter first) according to a pregenerated randomization list. This list was created prior to participant recruitment using a randomly generated allocation sequence that specified both case assignment and presentation order, with approximately balanced group sizes. Participants were enrolled consecutively and assigned to the next available entry on the randomization list. Allocation was not known in advance to either participants or study personnel conducting the sessions, thereby preserving allocation concealment at the participant level.
Participants from the University of Erlangen, who took part remotely via a web-based study platform, were randomized using a computer-generated 1:1 allocation algorithm implemented within the platform. Upon enrollment, the platform automatically assigned participants to the letter-order sequence and the clinical case.
The use of 2 different randomization methods resulted from the different study sites: on-site participation in Tübingen required an offline procedure, whereas the Erlangen cohort participated exclusively online. Randomization was conducted separately within each site and not across sites. Although randomization procedures differed between sites, both approaches ensured allocation concealment at the individual level. Importantly, the substantive data collection procedure (including exposure to the letters and assessment of learning objectives) was identical at both locations.
Data Analysis
Rating Process
The answers of the SPs were independently rated by 2 experienced clinicians (FH and AS) to determine whether the predefined learning objectives were correctly and completely identified and reported. The rating process followed a predefined, ordinally scaled rating structure, ranging from “not listed” over “partially listed” to “fully listed,” and disagreements were solved by discussion. A specific definition was established in advance for each learning objective (). Raters were blinded to both the letter type (discharge letter or patient letter) and the order in which it was presented. No formal assessment of blinding adequacy was performed. The raters achieved a Cohen κ of 0.97. Due to the high level of agreement, one of the two ratings was selected at random for use in the statistical analysis.
Statistical Analysis
Due to the ordinal nature of the ratings of the learning objectives (comprehensively reported, partially reported, and not reported), cumulative link mixed models were fitted to the data. Random intercepts for participant and learning objective were used to account for participant and learning objective random effects (). The fixed effects of letter type, Bloom category, and content field were investigated as variables of interest with model comparisons. The fixed effects disease type and time point of measurement were included as controls. Relevant models were assessed, and alternatives were considered. P values for differences in individual content fields were calculated by subsetting the data to each content field. Models were compared according to the Akaike Information Criterion and likelihood ratio tests. No missing data were present. The analyses were conducted using R software (version 4.4; R Foundation for Statistical Computing) and the ordinal package version 2023.12-4.1 [,].
Reporting Standards
The reporting of this study followed the STROBE (Strengthening the Reporting of Observational Studies in Epidemiology) guideline for cohort studies []. A reporting checklist can be found in the .
Ethical Considerations
This study was approved by the local ethics committee (778/2023BO2) of Tübingen University Hospital, Germany. Written informed consent was obtained from the SPs before inclusion in the study.
No direct identifiers (eg, names or dates of birth) were collected in the study. A short demographic questionnaire (age, gender, educational background, and prior experience as an SP) was completed by the SPs; these results were stored separately and not linked to individual response data. During data collection, response data were temporarily saved on secured cloud infrastructure; evaluation data and the separate demographic file were stored in different locations with a role-based access control. Reidentification was not possible at any time. Due to data loss at one site, demographics were available for 38 out of 48 SPs. As demographics were not used in the analysis of learning objective performance, and there is no evidence of systematic bias between sites, this missing information does not affect the validity of the study outcomes. The study was conducted in accordance with the Declaration of Helsinki.
Results
Demographic Data
The median age of the participants was 46.7 (IQR 28.3-65.0) years, with a female-to-male ratio of approximately 1.92:1. The majority (24/38, 63.2%) had a university degree. Task order did not influence performance (P=.80), although participants performed slightly better on their second attempt (odds ratio [OR] 1.23, 95% CI 1.03-1.47; P=.02). The type of disease did not have a significant main effect on the retention rate of learning objectives (P=.96; ). Therefore, all patient data were pooled for the subsequent comparison of the discharge letter and the patient letter.
Patient Letters vs Discharge Letters
Overall, patient letters improved information uptake: SPs retained more learning objectives after reading the patient letter compared with the discharge letter (OR 1.74, 95% CI 1.45-2.08; P<.001; ). Comprehensive reporting rose to 45.7% (490/1073) with the patient letter compared to 38.5% (413/1073) with the discharge letter. Partial reporting also improved (322/1073, 30% vs 287/1073, 26.7%), while omissions decreased from 34.8% (373/1073) to 24.3% (261/1073; ; ). Remarkably, despite overall gains, those 24.3% (261/1073) of all ratings remained “not reported” after reading the patient letter. Because omissions varied by participant, we conducted an item-level analysis to identify objectives consistently missed (predefined as “not reported” in ≥50% of the cases). A total of 6 (9%) of 67 objectives met this criterion. By Bloom level, 5 out of 6 were “Understand” and 1/6 “Remember”; by content field, 3 out of 6 fell under “Prevention of Complications,” 2 out of 6 under “Lifestyle/Disease Management,” and 1 out of 6 under “Organization” (). On the other hand, the overall proportion of medically incorrect reports by the participants was low, with no significant difference between the letter types (P=.55; ).


Impact of Cognitive Demand (Bloom Taxonomy) on Information Uptake
A subgroup analysis of learning objectives based on Bloom taxonomy revealed that, regardless of letter type, participants performed better on learning objectives classified under the “Remember” category compared to those classified under the “Understand” category (OR 3.33, 95% CI 1.96-5.88; P<.001; ). In both categories, the proportion of comprehensively reported learning objectives was higher when presented in the patient letter format compared to the discharge letter format (278/545, 51% vs 242/545, 44.4% for “Remember” and 212/528, 40.2% vs 171/528, 32.4% for “Understand”; ; ). A subsequent analysis showed no significant interaction between Bloom taxonomy level and letter type (P=.13), indicating that the observed benefits of the patient letter were consistent across both cognitive levels.
Impact of Content Field on Information Uptake
In a subgroup analysis based on content fields, we observed a statistically significant interaction effect between the type of letter and the content field (P<.001), indicating that the benefit of the patient letter depends on the content field. Additional values for effect size, OR, and CI for the results in this paragraph are illustrated in . A substantial increase in comprehension associated with the patient letter was observed for the domain of “Medication” (P<.001), with a rate of accurately reported learning objectives of 66.9% (170/254) compared to 50.8% (129/254) for the discharge letter. In the GPT-4–generated patient letter, the changes and purpose of each drug were explicitly stated in plain language (eg, “You were also given two other medications, metformin and empagliflozin, which help control blood sugar”), whereas the standard discharge letter contained this information in a condensed table format without further explanation. For further details on the translation process from discharge to patient letters, we refer to our previous work [].
The second largest improvement was seen in the domain of “Organization” (78/158, 49.4% vs 62/158, 39.2%; P<.001; ; ). In contrast, improvements observed with the patient letter in the content fields “Lifestyle/Disease Management” (P=.38) and “Prevention of Complications” (P<.01) were small and medium. In the content field “Prevention of Complications,” the patient letter resulted in a minor increase in the proportion of partially addressed learning objectives compared to the discharge letter (102/311, 32.8% vs 70/311, 22.5%). However, the rate of completely reported learning objectives remained the lowest compared to the other content fields (; ). Regarding the content field “Lifestyle/Disease Management,” there was even a small decrease observed for the partially reported learning objectives (100/350, 28.6% in the patient letter vs 107/350, 30.6% in the discharge letter). However, the amount of correctly stated learning objectives still showed a small increase (149/350, 42.6% in the patient letter vs 139/350, 39.7% in the discharge letter; ; ).

Discussion
Principal Findings
This prospective, explanatory study with SPs showed that AI-generated patient letters increased the retrieval of medication-related learning objectives to 66.9% (170/254) and organizational objectives to 49.4% (78/158), compared with 50.8% (129/254) and 39.2% (62/158), respectively, after reading the standard hospital discharge letters. These gains are larger than those reported in earlier single-arm or pre-post studies of human or AI-generated patient letters that lacked counterbalancing [,,]. By scoring open, free-text answers, we also captured a broader slice of comprehension than recent GPT-based studies that relied on recognition formats [-]. However, the reliance on predefined learning objectives, evaluated by clinicians, may still miss nuances of SP understanding. In particular, misconceptions and emotional interpretations may be underreported, and these remain important aspects for future qualitative work.
Content- and Bloom-Specific Insights
Although overall performance improved, SPs still failed to report, on average, nearly a quarter of the key information, highlighting a persistent patient safety gap for patients. Gains clustered in the areas of Medication and Organization, which are long-recognized weak spots: <50% of patients understand their discharge medicines [], and medication errors account for roughly half of preventable harm []. By increasing the retention of medication-related information, we hypothesize that patient letters have the potential to improve medication management and reduce medication errors, thereby potentially contributing to enhanced patient safety. Similarly, improving awareness of necessary follow-up appointments may also support patient safety by reducing missed visits and associated complications. Previous research supports these mechanisms: written patient information material has been shown to increase medication adherence [], missed follow-up appointments predict readmissions and emergency visits [-], and low health literacy is associated with higher readmission rates []. Moreover, communication interventions at hospital discharge are significantly linked to fewer readmissions and higher treatment adherence []. While these findings provide a rationale for the use of AI-generated patient letters, the effects on actual patient safety outcomes remain to be empirically evaluated in future research.
In contrast, improvements in the content field “Prevention of Complications” were modest, and those for “Lifestyle/Disease Management” were even smaller. Notably, 37.3% (116/311) of the key information regarding “Prevention of Complications” remained unknown. While higher cognitive demand likely contributed, other factors such as complex sentence structures, domain-specific phrasing, multistep reasoning, and the absence of visual aids may have further limited comprehension. As shown in our previous study, GPT-4 also faced greater challenges translating “Prevention of Complications” content compared with other domains []. Interestingly, although GPT-4 more successfully simplified “Lifestyle/Disease Management” information, SP performance remained low, suggesting that simplification alone does not guarantee comprehension when integration into practical self-management is required, and suggesting a distinction between content that is challenging for the AI to translate versus content that is difficult for SPs to understand and apply. Stratifying learning objectives by Bloom taxonomy showed higher performance for “Remember” than for “Understand” items (278/545, 51% vs 212/528, 40.2% correct with the patient letter), and 5 out of 6 objectives omitted by ≥50% of participants fell into the “Understand” category. This difference is unsurprising given that “Remember” tasks primarily involve locating explicit information, whereas “Understand” tasks require higher-order cognitive processes. Importantly, the patient letter improved performance in both categories compared with the discharge letter (“Remember”: 278/545, 51% vs 242/545, 44.4%; Understand: 212/528, 40.2% vs 171/528, 32.4%), and there was no significant interaction between letter type and Bloom level (P=.13), indicating that the observed benefit of the patient letter was approximately equal for both cognitive levels.
These findings suggest that GPT-4–generated patient letters can support not only the retrieval of factual information but may also modestly enhance comprehension-oriented learning objectives, though higher-level understanding remains more challenging. This is consistent with prior longitudinal research demonstrating that discharge instructions decay rapidly without reinforcement [-].
The clustering of consistently missed objectives in “Prevention of Complications” and, secondarily, “Lifestyle/Disease Management,” together with the predominance of “Understand”-level items, suggests that a static, text-only letter may not be sufficient in contexts where patients need to link symptoms to actions, reason about conditional risks (eg, sleep apnea as source of high blood pressure), or maintain self-management routines (eg, home blood pressure documentation or dietary change).
Didactic Implications (CLT)
CLT provides a helpful framework for interpreting our findings []. Recent work applying CLT to clinical practice shows that stress, emotion, and uncertainty increase intrinsic cognitive load and that working memory depletion can impair clinical performance []. Against this background, it seems plausible that linguistic simplification and better structure primarily reduce extraneous cognitive load. Our prior work demonstrated that GPT-4 can reorganize discharge letters into patient letters mainly by removing jargon and chunking information [], which likely supports this reduction. While discharge letters are generally already well-structured, AI-generated patient letters further optimize content organization in a patient-centered manner.
We did not directly measure intrinsic, extraneous, or germane cognitive load in this study; CLT is therefore used as an interpretive framework rather than an empirically tested mechanism. Our learning objectives were deliberately designed so that each content field contained both “Remember” and “Understand” objectives. However, objectives that remained most difficult (defined as those omitted by ≥50% of participants) were predominantly from the “Understand” category in the area of “Prevention of Complications” and, to a lesser extent, in the area of “Lifestyle/Disease Management.” These learning objectives (eg, hypoglycemia symptom-action mapping, recognition of statin-associated muscle symptoms, dehydration-related kidney risk, or the contribution of home blood pressure monitoring and Mediterranean diet to long-term disease control) require the integration of multiple information elements, conditional if-then thinking, and sustained self-regulation. Within the framework of CLT, they can therefore be interpreted as tasks with a comparatively high intrinsic cognitive load.
In contrast, the largest gains from patient letters were observed predominantly in factual learning objectives in the areas of “Medication” or “Organization,” which required participants to find and reproduce explicitly stated information (eg, medication changes or appointments). Such learning objectives are likely to be particularly sensitive to a reduction in extrinsic load through simpler language and clearer structure.
Based on these findings, we conclude that AI-based translation into patient letters may reduce extraneous cognitive load, thereby possibly freeing working-memory resources for the processing of concrete, action-oriented tasks (eg, tablet strength and appointment date). However, for learning objectives with high intrinsic cognitive load, such as symptom-to-action coupling (eg, tremors-hypoglycemia compensation required), conditional risk thinking (eg, allergy present notification to the care team), or permanent self-regulation (eg, home blood pressure monitoring and dietary changes) reducing extrinsic cognitive load alone may be not sufficient to free up enough relevant resources (germane cognitive load) for deeper understanding and behavioral change.
Taken together, these CLT-based interpretations are hypothesis-generating and should be tested in future studies that directly measure cognitive load. They further suggest that patient letters may be most effective for immediate, concrete instructions, whereas higher-order learning objectives may require complementary, multimodal supports (eg, infographics, microvideos, interactive prompts, and teach-back). These, however, need prospective evaluation. Evidence for each adjunct already exists in discharge-education and adherence research [,,].
Beyond CLT, factors such as stress-induced cognitive depletion, low health literacy, and limited self-efficacy may also reduce patients’ ability to process complex medical content [-]. Tailoring follow-up communication to these variables will be essential [].
Clinical Implications
The implementation of AI-generated patient letters requires careful attention to ethical, legal, and social aspects []. In general, human oversight (“human-in-the-loop”) remains essential when using AI to translate clinician-written discharge letters into patient letters. Our study indicates that the patient letter format alone does not yield sufficient comprehension gains for cognitively demanding learning objectives, particularly those classified as “Understand” by Bloom taxonomy and thematically related to prevention of complications or lifestyle/disease management. It is further underscored by omissions and hallucinations observed in the prior translation process [], which required the exclusion of 5 learning objectives in this study. Therefore, establishing practical governance structures is essential for clinical practice. By identifying the specific content domains most prone to errors or omissions, our results can help to target quality-assurance efforts efficiently, making oversight both feasible and focused on the areas critical for patient safety.
From a legal perspective, robust data protection is critical, as personal patient information constitutes a central component of these letters [,]. Socially, simplified language may improve accessibility for patients with low health literacy and reduce language barriers []. While the format does not accommodate individual literacy needs or reading difficulties, it may still help reduce social barriers in discharge communication. Finally, practical and feasibility considerations include technical integration into existing clinical workflows, usability for health care staff, and costs associated with development, implementation, and maintenance. However, manual letter production is time-consuming and costly, whereas AI offers a potentially scalable solution. Effective measures without AI can be found in the literature, including patient letters written in simple language and structured summaries that reduce technical jargon and organize content for easy recall [], portal access (facilitating review but not guaranteeing understanding) [,], supplementary written materials (linked with better medication adherence) [], audio recordings of consultations (supporting recall) [], and interactive elements such as brief teach-back or app-based prompts (which enhance encoding and memory) [,]. However, implementation is frequently constrained by time and workload, limited organizational support and training, workflow barriers, and gaps in digital access and use [,]. These problems are documented across teach-back implementation (time/sustainability barriers), discharge letter production (time pressure, authorship level, and workload effects), and portal-based delivery (digital divide: differences in access, skills, and use) [-].
Our study isolates the contribution of AI-generated formulations and was not designed as a head-to-head comparison. In this context, AI-generated patient letters, with clinician oversight, may offer a practical, scalable way to produce patient-centered text for domains where our data indicate clear gains (eg, medication and organization), and future work should address workflow integration, usability for clinicians, maintenance costs, and robust data protection.
Limitations
This study has several limitations. First, we used only 1 GPT-4 version and focused on 3 chronic diseases (hypertension, diabetes, and diabetic kidney disease), which limits generalizability to other LLM versions and medical conditions. Although participants and clinician raters were blinded to letter type and order as far as possible, full blinding was not achievable, and the success of blinding was not formally assessed for either group. Because the stylistic differences between discharge letters and patient letters are recognizable, which was unavoidable given the study content, participants and raters may have inferred the format, introducing potential expectancy effects. Such effects could operate in both directions: some participants might assume the patient letter to be easier, whereas others may approach AI-generated text with skepticism, potentially reducing trust or attention. These bidirectional biases could either amplify or attenuate format-related differences. While standardized instructions, structured scoring procedures, and uniform training were used to minimize subjective influence, residual bias cannot be excluded and should be considered when interpreting the magnitude of effects. Additionally, we did not formally assess the feasibility or acceptability of patient letters through validated usability or satisfaction scales. These aspects should be addressed in future implementation studies. Furthermore, our study used standardized rather than real-world patients to examine the effects of AI-generated patient letters. This approach ensured high procedural consistency and enabled a controlled comparison between AI-generated and human-written letters, as reflected in recent high-impact AI studies using SPs []. However, it limits external validity; SPs are not able to fully represent real patients’ stress levels, emotional states, cognitive vulnerability, and health literacy diversity, especially in the context of acute illness or hospitalization. In clinical practice, such factors may negatively affect patient comprehension due to reduced germane cognitive load. Consequently, although the direction of the observed effects (improved understanding through AI-generated patient letters) is expected to remain, the overall level of understanding and recall would likely be lower among real patient populations for both letter types. In addition, letters were written in German and later translated for publication, limiting the generalizability of our findings to other languages. Another potential limitation of this study is the exclusion of 5 predefined learning objectives that were not present in the GPT-4–generated patient letters due to omissions, which could introduce bias in favor of the AI-generated format. However, a statistical analysis including all 72 learning objectives showed that the overall pattern of results remained unchanged despite the absence of these 5 items in the AI-generated patient letters. This suggests that the exclusion did not materially affect the observed effects. Additionally, the sample size in our study was determined by availability rather than by formal power calculation, which may limit the ability to detect subtle letter-type differences, particularly in subgroup analyses. Our findings should therefore be interpreted with caution. Finally, given the potential for hallucinations or omissions in AI-generated texts, as observed in the translation process [], clinical review and governance are essential if AI-based tools are deployed in practice, and attention to AI transparency and the potential for unexpected outputs remains critical.
Conclusions
This hypothesis-driven, explanatory study showed that AI-generated patient letters improved the uptake of basic, action-oriented information in SPs. This effect might reflect a reduction in extraneous cognitive load. We hypothesize that, in clinical practice, this could reduce common sources of postdischarge harm and improve patient safety. However, patient letters alone did not close gaps in higher-order understanding, particularly in risk prevention and lifestyle change. Addressing such complex learning goals may require multimodal, interactive support (eg, explanatory graphics, short videos, or app-based follow-ups [,,]) that reinforce the text-based patient letter.
Future studies should therefore evaluate integrative models that layer such tools onto patient letters and tailor content to health literacy, self-efficacy, and situational resources [], ideally within a didactic framework (eg, Observing Patient Involvement Scale or Patient Education Materials Assessment Tool []) to enhance actionability. Studies including more diverse medical conditions, real patient populations with varying health literacy levels, different LLMs, and multiple languages are needed to improve generalizability. We also suggest incorporating qualitative approaches in future studies (eg, think-aloud protocols or teach-back methods) to better capture nuances in patient understanding and potential contributors to medical error.
Acknowledgments
We wish to thank the University of Tübingen and Eric Nazarenus for supporting our study.
The authors declare the use of generative artificial intelligence (GenAI) in the research and writing process. According to the GAIDeT taxonomy (2025), the following tasks were delegated to GenAI tools under full human supervision: proofreading, editing, and translation.
The GenAI tools used were: ChatGPT by OpenAI, DeepL SE.
Responsibility for the final manuscript lies entirely with the authors.
GenAI tools are not listed as authors and do not bear responsibility for the final outcomes.
Data Availability
The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.
Funding
This work was supported by the Open Access Publishing Fund of the University of Tübingen.
Authors' Contributions
FE had full access to all of the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.
Concept and design: FE, FH, AS
Acquisition, analysis, or interpretation of data: all authors
Drafting of the manuscript: FE, AS, FH
Critical review of the manuscript for important intellectual content: all authors
Statistical analysis: KS
Administrative, technical, or material support: CSP, KS, PS
Supervision: AHW, MH
Conflicts of Interest
None declared.
GPT-4–generated patient letters (German original and English translation).
PDF File (Adobe PDF File), 319 KBStudy flowchart. Overview of participant enrollment, randomization, crossover procedure, and analysis population.
PDF File (Adobe PDF File), 86 KBOverview of the learning objectives, standardized questions, and the corresponding rating scheme applied to evaluate information retention and comprehension across study conditions.
XLSX File (Microsoft Excel File), 18 KBList of the learning objectives that were not represented in the GPT-4–generated patient letters because the model did not carry them over from the original discharge letters. These objectives were excluded from the analysis for both letter types.
DOCX File , 22 KBLikelihood ratio tests.
DOCX File , 27 KBSTROBE checklist detailing how each reporting item was fulfilled and where it can be found in the manuscript.
DOCX File , 33 KBNumber of learning outcomes achieved, stratified by letter type (patient letter vs discharge letter).
DOCX File , 25 KBItem-level analysis identifying learning objectives that were consistently missed, predefined as those “not reported” by ≥50% of participants.
DOCX File , 22 KBOdds ratios from the best-fitting cumulative link mixed model. Odds ratios represent the interaction between letter type and content field, estimated relative to an arbitrary reference category. The reference category is defined as letter type = discharge letter (DL) and content field = Lifestyle / Disease Management. P values for the odds ratios are calculated in comparison with this reference category.
PNG File , 106 KBReferences
- Marsall M, Hornung T, Bäuerle A, Weigl M. Quality of care transition, patient safety incidents, and patients' health status: a structural equation model on the complexity of the discharge process. BMC Health Serv Res. 2024;24(1):576. [FREE Full text] [CrossRef] [Medline]
- Moore C, Wisnivesky J, Williams S, McGinn T. Medical errors related to discontinuity of care from an inpatient to an outpatient setting. J Gen Intern Med. 2003;18(8):646-651. [FREE Full text] [CrossRef] [Medline]
- Steele EM, Burney HN, Freije SL, Zellars RC, Lautenschlaeger T, Holmes JA. Missed follow-up is associated with worse survival in stage I lung cancer: results from a large multi-site academic hospital system. Sci Rep. 2024;14(1):17710. [FREE Full text] [CrossRef] [Medline]
- Becker C, Zumbrunn S, Beck K, Vincent A, Loretz N, Müller J, et al. Interventions to improve communication at hospital discharge and rates of readmission: a systematic review and meta-analysis. JAMA Netw Open. 2021;4(8):e2119346. [FREE Full text] [CrossRef] [Medline]
- Engel KG, Heisler M, Smith DM, Robinson CH, Forman JH, Ubel PA. Patient comprehension of emergency department care and instructions: are patients aware of when they do not understand? Ann Emerg Med. 2009;53(4):454-461.e15. [CrossRef] [Medline]
- Belachew EA, Sendekie AK, Wondm SA, Ayele EM, Netere AK. Misunderstanding of dosing regimen instructions among patients with chronic diseases receiving polypharmacy at the University of Gondar comprehensive specialized hospital. PLoS One. 2023;18(1):e0280204. [FREE Full text] [CrossRef] [Medline]
- Kreps GL. Promoting patient comprehension of relevant health information. Isr J Health Policy Res. 2018;7(1):56. [FREE Full text] [CrossRef] [Medline]
- Toole J, Kohansieh M, Khan U, Romero S, Ghali M, Zeltser R, et al. Does your patient understand their treatment plan? Factors affecting patient understanding of their medical care treatment plan in the inpatient setting. J Patient Exp. 2020;7(6):1151-1157. [FREE Full text] [CrossRef] [Medline]
- Magnani JW, Mujahid MS, Aronow HD, Cené CW, Dickson VV, Havranek E, et al. American Heart Association Council on EpidemiologyPrevention; Council on Cardiovascular Disease in the Young; Council on CardiovascularStroke Nursing; Council on Peripheral Vascular Disease; Council on Quality of CareOutcomes Research;Stroke Council. Health literacy and cardiovascular disease: fundamental relevance to primary and secondary prevention: a scientific statement from the American Heart Association. Circulation. 2018;138(2):e48-e74. [FREE Full text] [CrossRef] [Medline]
- Choudhry AJ, Baghdadi YMK, Wagie AE, Habermann EB, Heller SF, Jenkins DH, et al. Readability of discharge summaries: with what level of information are we dismissing our patients? Am J Surg. 2016;211(3):631-636. [FREE Full text] [CrossRef] [Medline]
- Ali SR, Dobbs TD, Hutchings HA, Whitaker IS. Using ChatGPT to write patient clinic letters. Lancet Digit Health. 2023;5(4):e179-e181. [FREE Full text] [CrossRef] [Medline]
- Eisinger F, Holderried F, Mahling M, Stegemann-Philipps C, Herrmann-Werner A, Nazarenus E, et al. What's going on with me and how can I better manage my health? The potential of GPT-4 to transform discharge letters into patient-centered letters to enhance patient safety: prospective, exploratory study. J Med Internet Res. 2025;27:e67143. [FREE Full text] [CrossRef] [Medline]
- Zaretsky J, Kim JM, Baskharoun S, Zhao Y, Austrian J, Aphinyanaphongs Y, et al. Generative artificial intelligence to transform inpatient discharge summaries to patient-friendly language and format. JAMA Netw Open. 2024;7(3):e240357. [FREE Full text] [CrossRef] [Medline]
- Shahnam A, Nindra U, Hitchen N, Tang J, Hong M, Hong JH, et al. Application of generative artificial intelligence for physician and patient oncology letters—AI-OncLetters. JCO Clin Cancer Inform. 2025;(9):e2400323. [CrossRef]
- Kumar A, Wang H, Muir KW, Mishra V, Engelhard M. A cross-sectional study of GPT-4–based plain language translation of clinical notes to improve patient comprehension of disease course and management. NEJM AI. 2025;2(2):AIoa2400402. [CrossRef]
- van Driel MHE, Blok N, van den Brand JAJG, van de Sande D, de Vries M, Eijlers B, et al. Leveraging GPT-4 enables patient comprehension of radiology reports. Eur J Radiol. 2025;187:112111. [FREE Full text] [CrossRef] [Medline]
- Kim SI, Park J, Kim T, Seo W, Kim T, Cha WC, et al. Enhancing patient participation in emergency department through patient-friendly clinical notes generated by large language models. Sci Rep. 2025;16(1):1409. [FREE Full text] [CrossRef] [Medline]
- Sweller J. Cognitive load during problem solving: effects on learning. Cogn Sci. 2010;12(2):257-285. [FREE Full text] [CrossRef]
- Adams NE. Bloom's taxonomy of cognitive learning objectives. J Med Libr Assoc. 2015;103(3):152-153. [FREE Full text] [CrossRef] [Medline]
- Felix HM, Simon LV. Types of Standardized Patients and Recruitment in Medical Simulation. Treasure Island (FL). StatPearls; 2025.
- Tu T, Schaekermann M, Palepu A, Saab K, Freyberg J, Tanno R, et al. Towards conversational diagnostic artificial intelligence. Nature. 2025;642(8067):442-450. [CrossRef] [Medline]
- Hellyer S. Supplementary report: frequency of stimulus presentation and short-term decrement in recall. J Exp Psychol. 1962;64(6):650-650. [CrossRef] [Medline]
- McEvoy JW, McCarthy CP, Bruno RM, Brouwers S, Canavan MD, Ceconi C, et al. ESC Scientific Document Group. 2024 ESC guidelines for the management of elevated blood pressure and hypertension. Eur Heart J. 2024;45(38):3912-4018. [CrossRef] [Medline]
- American Diabetes Association Professional Practice Committee. 6. Glycemic goals and hypoglycemia: standards of care in diabetes-2024. Diabetes Care. 2024;47(Suppl 1):S111-S125. [CrossRef] [Medline]
- American Diabetes Association Professional Practice Committee. 2. Diagnosis and classification of diabetes: standards of care in diabetes-2024. Diabetes Care. 2024;47(Suppl 1):S20-S42. [CrossRef] [Medline]
- American Diabetes Association Professional Practice Committee. 3. Prevention or delay of diabetes and associated comorbidities: standards of care in diabetes-2024. Diabetes Care. 2024;47(Suppl 1):S43-S51. [CrossRef] [Medline]
- American Diabetes Association Professional Practice Committee. 11. Chronic kidney disease and risk management: standards of care in diabetes-2024. Diabetes Care. 2024;47(Suppl 1):S219-S230. [CrossRef] [Medline]
- Krathwohl DR. A revision of Bloom's taxonomy: an overview. Theory Pract. 2010;41(4):212-218. [CrossRef]
- Biggs JB, Tang CSK. Teaching for Quality Learning at University What the Student Does. 4th ed. Maidenhead. Society for Research into Higher Education; 2011.
- Abdolreza GZ, Mansourzadeh A, Montazeri KA, Saeidi M. Reflections on using open-ended questions. Med Educ Bull. 2022;3(2):475-482. [CrossRef]
- Haubo R, Christensen B, editors. Cumulative Link Models for Ordinal Regression With the R Package ordinal. Denmark. Christensen Statistics; 2018.
- R: A Language and Environment for Statistical Computing. Vienna, Austria. R Core Team, R Foundation for Statistical Computing; 2025.
- Von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, Vandenbroucke JP, et al. STROBE Initiative. The strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. PLoS Med. 2007;4(10):e296. [FREE Full text] [CrossRef] [Medline]
- Hoffmann H, Jonietz A, Gräfe W, Zenker R, Voigt K, Riemenschneider H. Associations of an easy-to-understand patient letter on the health literacy of patients after discharge from hospital: results of a randomized controlled intervention study. Gesundheitswesen. 2023;85(S 03):S183-S188. [FREE Full text] [CrossRef] [Medline]
- Lin R, Gallagher R, Spinaze M, Najoumian H, Dennis C, Clifton-Bligh R, et al. Effect of a patient-directed discharge letter on patient understanding of their hospitalisation. Intern Med J. 2014;44(9):851-857. [CrossRef] [Medline]
- Makaryus AN, Friedman EA. Patients' understanding of their treatment plans and diagnosis at discharge. Mayo Clin Proc. 2005;80(8):991-994. [CrossRef] [Medline]
- Hodkinson A, Tyler N, Ashcroft DM, Keers RN, Khan K, Phipps D, et al. Preventable medication harm across health care settings: a systematic review and meta-analysis. BMC Med. 2020;18(1):313. [FREE Full text] [CrossRef] [Medline]
- Al Jeraisy M, Alshammari H, Albassam M, Al Aamer K, Abolfotouh MA. Utility of patient information leaflet and perceived impact of its use on medication adherence. BMC Public Health. 2023;23(1):488. [FREE Full text] [CrossRef] [Medline]
- Coppa K, Kim EJ, Oppenheim MI, Bock KR, Conigliaro J, Hirsch JS. Examination of post-discharge follow-up appointment status and 30-day readmission. J Gen Intern Med. 2021;36(5):1214-1221. [FREE Full text] [CrossRef] [Medline]
- McAlister FA, Youngson E, Bakal JA, Kaul P, Ezekowitz J, van Walraven C. Impact of physician continuity on death or urgent readmission after discharge among patients with heart failure. CMAJ. 2013;185(14):E681-E689. [FREE Full text] [CrossRef] [Medline]
- Baky V, Moran D, Warwick T, George A, Williams T, McWilliams E, et al. Obtaining a follow-up appointment before discharge protects against readmission for patients with acute coronary syndrome and heart failure: a quality improvement project. Int J Cardiol. 2018;257:12-15. [CrossRef] [Medline]
- Shahid R, Shoker M, Chu LM, Frehlick R, Ward H, Pahwa P. Impact of low health literacy on patients' health outcomes: a multicenter cohort study. BMC Health Serv Res. 2022;22(1):1148. [FREE Full text] [CrossRef] [Medline]
- Wolderslund M, Kofoed PE, Holst R, Waidtløw K, Ammentorp J. Outpatients' recall of information when provided with an audio recording: a mixed-methods study. Patient Educ Couns. 2020;103(1):63-70. [CrossRef] [Medline]
- Laws MB, Lee Y, Taubin T, Rogers WH, Wilson IB. Factors associated with patient recall of key information in ambulatory specialty care visits: results of an innovative methodology. PLoS One. 2018;13(2):e0191940. [FREE Full text] [CrossRef] [Medline]
- Kessels RPC. Patients' memory for medical information. J R Soc Med. 2003;96(5):219-222. [FREE Full text] [CrossRef] [Medline]
- Szulewski A, Howes D, van Merriënboer JJG, Sweller J. From theory to practice: the application of cognitive load theory to the practice of medicine. Acad Med. 2021;96(1):24-30. [CrossRef] [Medline]
- Xu Q, Sundar SS. Interactivity and memory: Information processing of interactive versus non-interactive content. Comput Hum Behav. 2016;63:620-629. [FREE Full text] [CrossRef]
- Dada A, Koras OA, Bauer M, Butler A, Smith KE, Kleesiek J. MeDiSumQA: patient-oriented question-answer generation from discharge letters. arXiv. Preprint posted online on February 5, 2025. 2025. [CrossRef]
- Baumeister RF, André N, Southwick DA, Tice DM. Self-control and limited willpower: current status of ego depletion theory and research. Curr Opin Psychol. 2024;60:101882. [FREE Full text] [CrossRef] [Medline]
- Cui Y, Li R, Yang T, Wang H, Jin S, Liu N, et al. Influence of positive and negative affect on self-management among patients with early chronic kidney disease during the COVID-19 pandemic: the mediating and suppressing effect of ego depletion. Front Psychiatry. 2022;13:992404. [FREE Full text] [CrossRef] [Medline]
- Yumiya Y, Goto A, Konta T. Association between health literacy and understanding of doctors' explanations: the Yamagata study. Health Lit Res Pract. 2024;8(3):e175-e183. [FREE Full text] [CrossRef] [Medline]
- Stormacq C, Wosinski J, Boillat E, Van den Broucke S. Effects of health literacy interventions on health-related outcomes in socioeconomically disadvantaged adults living in the community: a systematic review. JBI Evid Synth. 2020;18(7):1389-1469. [CrossRef] [Medline]
- Gilbert S, Harvey H, Melvin T, Vollebregt E, Wicks P. Large language model AI chatbots require approval as medical devices. Nat Med. 2023;29(10):2396-2398. [CrossRef] [Medline]
- Gorelik AJ, Li M, Hahne J, Wang J, Ren Y, Yang L, et al. Ethics of AI in healthcare: a scoping review demonstrating applicability of a foundational framework. Front Digit Health. 2025;7:1662642. [FREE Full text] [CrossRef] [Medline]
- Park MK, Ashwood N, Capes N. Ethics of artificial intelligence in medicine. Cureus. 2025;17(5):e83567. [CrossRef] [Medline]
- Kong M, Fernandez A, Bains J, Milisavljevic A, Brooks KC, Shanmugam A, et al. Evaluation of the accuracy and safety of machine translation of patient-specific discharge instructions: a comparative analysis. BMJ Qual Saf. 2025:bmjqs-2024-018384. [FREE Full text] [CrossRef] [Medline]
- Siegrist V, Mata R, Langewitz W, Gerger H, Furger S, Hertwig R, et al. Does information structuring improve recall of discharge information? A cluster randomized clinical trial. PLoS One. 2021;16(10):e0257656. [FREE Full text] [CrossRef] [Medline]
- Brands MR, Gouw SC, Beestrum M, Cronin RM, Fijnvandraat K, Badawy SM. Patient-centered digital health records and their effects on health outcomes: systematic review. J Med Internet Res. 2022;24(12):e43086. [FREE Full text] [CrossRef] [Medline]
- Neves AL, Freise L, Laranjo L, Carter AW, Darzi A, Mayer E. Impact of providing patients access to electronic health records on quality and safety of care: a systematic review and meta-analysis. BMJ Qual Saf. 2020;29(12):1019-1032. [FREE Full text] [CrossRef] [Medline]
- Talevski J, Wong Shee A, Rasmussen B, Kemp G, Beauchamp A. Teach-back: a systematic review of implementation and impacts. PLoS One. 2020;15(4):e0231350. [FREE Full text] [CrossRef] [Medline]
- Weetman K, Spencer R, Dale J, Scott E, Schnurr S. What makes a "successful" or "unsuccessful" discharge letter? Hospital clinician and general practitioner assessments of the quality of discharge letters. BMC Health Serv Res. 2021;21(1):349. [FREE Full text] [CrossRef] [Medline]
- Coit MH, Katz JT, McMahon GT. The effect of workload reduction on the quality of residents' discharge summaries. J Gen Intern Med. 2011;26(1):28-32. [FREE Full text] [CrossRef] [Medline]
- Carris NW, Bullers K, McKee M, Schanze J, Eubanks T, Epperson C, et al. Automated lifestyle interventions and weight loss: a systematic review of randomized controlled trials. Int J Obes (Lond). 2025;49(6):992-997. [CrossRef] [Medline]
- Sajjadi SL, Ghafourifard M, Khosroshahi HT. The effect of individualized education on learning needs of patients undergoing hemodialysis: a randomized controlled clinical trial. BMC Nephrol. 2024;25(1):452. [FREE Full text] [CrossRef] [Medline]
- Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. 2014;96(3):395-403. [CrossRef] [Medline]
Abbreviations
| AI: artificial intelligence |
| CLT: cognitive load theory |
| LLM: large language model |
| OR: odds ratio |
| SP: standardized patient |
| STROBE: Strengthening the Reporting of Observational Studies in Epidemiology |
Edited by J Sarvestan; submitted 26.Jul.2025; peer-reviewed by F Al Dhabbari, Z Vojvodic, K Fiscella; comments to author 01.Sep.2025; revised version received 23.Dec.2025; accepted 24.Dec.2025; published 26.Feb.2026.
Copyright©Friederike Holderried, Alessandra Sonanini, Christian Stegemann–Philipps, Anne Herrmann–Werner, Philipp Spitzer, Martina Guthoff, Nils Heyne, Konstantin Sering, Martin Holderried, Felix Eisinger. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 26.Feb.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

