<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v28i1e86078</article-id><article-id pub-id-type="doi">10.2196/86078</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Time Savings Through an AI Speech Assistant for Nursing Documentation: Pre-Post Time-Motion Study in German Long-Term Care</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Schwabe</surname><given-names>Katja</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Ferizaj</surname><given-names>Drin</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Neumann</surname><given-names>Susann</given-names></name><degrees>DipNEd</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Strube-Lahmann</surname><given-names>Sandra</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lahmann</surname><given-names>Nils</given-names></name><degrees>Prof Dr</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>voize GmbH</institution><addr-line>Dircksenstra&#x00DF;e 47</addr-line><addr-line>Berlin</addr-line><country>Germany</country></aff><aff id="aff2"><institution>Department of Geriatrics and Medical Gerontology, Charit&#x00E9; &#x2013; Universit&#x00E4;tsmedizin Berlin</institution><addr-line>Berlin</addr-line><country>Germany</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Coristine</surname><given-names>Andrew</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Chen</surname><given-names>Ke</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>He</surname><given-names>Lingxiao</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Mohanadas</surname><given-names>Sadhasivam</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Ye</surname><given-names>Siao</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Katja Schwabe, BSc, voize GmbH, Dircksenstra&#x00DF;e 47, Berlin, 10178, Germany, +49 331 96809186; <email>katja@voize.de</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>8</day><month>4</month><year>2026</year></pub-date><volume>28</volume><elocation-id>e86078</elocation-id><history><date date-type="received"><day>17</day><month>10</month><year>2025</year></date><date date-type="rev-recd"><day>20</day><month>02</month><year>2026</year></date><date date-type="accepted"><day>24</day><month>02</month><year>2026</year></date></history><copyright-statement>&#x00A9; Katja Schwabe, Drin Ferizaj, Susann Neumann, Sandra Strube-Lahmann, Nils Lahmann. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 8.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2026/1/e86078"/><abstract><sec><title>Background</title><p>Nurses in long-term care spend up to one-third of their working time on documentation, contributing to administrative burden and limited time for direct care. Artificial intelligence (AI) speech assistants have shown potential to accelerate documentation, but longitudinal evidence from real-world long-term care settings remains scarce.</p></sec><sec><title>Objective</title><p>This study aimed to evaluate whether implementing a domain-specific, mobile AI speech assistant is associated with reduced documentation time in German long-term care under routine conditions. Secondary objectives included examining usability, perceived documentation effort, interruptions, and workplace satisfaction.</p></sec><sec sec-type="methods"><title>Methods</title><p>A pre-post time-motion study with full-shift observation was conducted. Continuous, event-based observations were performed before (<italic>t</italic><sub>0</sub>) and after (<italic>t</italic><sub>1</sub>) implementation of the mobile speech assistant voize. The primary outcome was total documentation time per morning shift based on direct observations. In addition to observations, questionnaires were administered to assess perceived documentation effort, interruptions, satisfaction with the documentation system, and workplace satisfaction. The primary end point was analyzed using a linear mixed-effects model. Secondary, self-reported outcomes were analyzed exploratorily via paired pre-post differences with pooling across multiple imputations and Holm-Bonferroni correction.</p></sec><sec sec-type="results"><title>Results</title><p>A total of 52 registered nurses from 14 long-term care facilities participated (mean age 42.37, SD 12.37 years; 42/52, 80.8% female). Across 770 observed hours, the observed total documentation time per morning shift decreased significantly by an adjusted mean of 15 (SE 3.36) minutes, <italic>t</italic><sub>46.29</sub>=&#x2212;4.46, <italic>P</italic>&#x003C;.001, with a 95% CI of &#x2212;21.75 to &#x2212;8.23, corresponding to an approximately 28% reduction relative to the baseline mean. Holm-Bonferroni&#x2013;corrected exploratory analyses indicated significant declines in self-reported documentation time and interruptions, and satisfaction with the documentation system improved, while workplace satisfaction showed no significant change. Usability was rated as acceptable.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This study provides real-world evidence from a single-group pre-post design that an AI-based speech assistant is associated with reduced documentation workload in long-term care. In this sample, the integration of a mobile, domain-specific speech system into daily workflows coincided with substantially decreased documentation time and improved perceived efficiency. Beyond these observed time savings, such technology has the potential to alleviate workload, free time for resident care, and enhance working conditions. These findings are also relevant for policy discussions on addressing the nursing workforce shortage, showing that well-integrated, speech-enabled documentation systems can support more sustainable long-term care environments.</p></sec><sec><title>Trial Registration</title><p>German Clinical Trials Register DRKS00035512; https://drks.de/search/en/trial/DRKS00035512</p></sec></abstract><kwd-group><kwd>nursing documentation time</kwd><kwd>automatic speech recognition</kwd><kwd>long-term care</kwd><kwd>artificial intelligence</kwd><kwd>AI speech assistant</kwd><kwd>mobile nursing documentation</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><sec id="s1-1"><title>Background</title><p>In Germany, approximately 1.7 million nursing staff are employed, with nearly 30% working in long-term care facilities, making this the largest sector in the nursing workforce [<xref ref-type="bibr" rid="ref1">1</xref>]. Demand for care continues to rise while shortages of qualified staff worsen [<xref ref-type="bibr" rid="ref2">2</xref>]. Nursing staff, particularly in long-term care, report higher levels of strain compared to other occupational groups [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. They work under considerable time pressure, often lack sufficient time to complete tasks, and frequently perform overtime with irregular or missed breaks [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. Alongside these structural challenges and workload demands, care workers face a high administrative burden: documentation tasks account for about 30% of total working time [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>], leaving less room for direct interaction and care for residents [<xref ref-type="bibr" rid="ref8">8</xref>]. Furthermore, the high documentation workload has been linked to reduced job satisfaction and an elevated level of stress [<xref ref-type="bibr" rid="ref9">9</xref>]. In light of these issues, artificial intelligence (AI)&#x2013;based speech assistants are presented as promising solutions to accelerate the documentation process and improve its accuracy [<xref ref-type="bibr" rid="ref10">10</xref>].</p></sec><sec id="s1-2"><title>Medical Documentation With Speech Recognition</title><p>Research on speech recognition has been most widely conducted in clinical medicine, often involving physicians and mixed professional groups. Previous findings show that speech recognition technologies can significantly streamline documentation workflows compared to traditional methods. For example, Zuchowski and G&#x00F6;ller [<xref ref-type="bibr" rid="ref11">11</xref>] reported notable time savings and lower error rates when using speech recognition versus manual typing. Similarly, Peine et al [<xref ref-type="bibr" rid="ref12">12</xref>] found that a voice-based system enabled faster and more accurate documentation among intensive care professionals compared to both electronic and paper-based approaches.</p><p>These findings are further supported by systematic reviews. Alboksmaty et al [<xref ref-type="bibr" rid="ref13">13</xref>] synthesized findings across several countries, highlighting consistent efficiency gains, reduced administrative burden, and improved provider-patient interaction. Of particular note, Wang et al [<xref ref-type="bibr" rid="ref14">14</xref>] found that a digital scribe could accelerate history-taking and physical examination documentation by more than double compared to typing or traditional dictation.</p><p>However, reviews also point out that most studies have taken place in controlled or experimental settings, which may limit generalizability to routine care [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. In this context, Ng et al [<xref ref-type="bibr" rid="ref15">15</xref>] reviewed 29 research articles and observed that many reported time savings and improvements in documentation completeness, whereas others described challenges such as increased editing workload, variable cost-effectiveness, and issues with error rates for specialized terminology or accented speech.</p></sec><sec id="s1-3"><title>Clinical Nursing Documentation With Speech Recognition</title><p>Compared to physician documentation, evidence on nursing documentation with AI speech assistants is more limited but growing. Findings to date are still heterogeneous. For example, many studies reviewed by Joseph et al [<xref ref-type="bibr" rid="ref10">10</xref>] indicated beneficial effects, such as reduced documentation time and improved workflow. However, a substantial number reported no significant impact, given ongoing challenges integrating AI speech assistants into real-world nursing environments and existing documentation systems.</p><p>However, several recent investigations suggest a promising potential for efficiency gains in real-world nursing environments. In a hospital time-motion study, Ehrler et al [<xref ref-type="bibr" rid="ref16">16</xref>] found that a mobile bedside application meaningfully decreased the duration of electronic health record (EHR) documentation for nurses while also increasing time available for direct patient interaction. Similarly, Mayer et al [<xref ref-type="bibr" rid="ref17">17</xref>] reported statistically significant time savings&#x2014;up to 6.1 minutes per documentation scenario&#x2014;when using a voice recognition system compared to traditional keyboard entry. Mairittha et al [<xref ref-type="bibr" rid="ref18">18</xref>] also demonstrated that a smartphone-based dialogue system allowed for 15 percent faster documentation relative to electronic forms in a controlled setting. Likewise, qualitative findings from German long-term care facilities hint at the potential of AI speech assistants for care documentation, reporting self-reported time savings, fewer interruptions, greater flexibility, and less reliance on stationary computers [<xref ref-type="bibr" rid="ref19">19</xref>].</p></sec><sec id="s1-4"><title>Research Gap</title><p>In summary, evidence across clinical settings suggests that AI speech assistants can shorten documentation time and thus reduce administrative burden. At the same time, findings are mixed and often derive from controlled or pilot evaluations rather than routine practice, which limits generalizability [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref15">15</xref>]. In nursing, several studies report faster entries with mobile or dialogue-based tools, but most were scenario-based, single-unit, or acute-care focused, with heterogeneous outcomes and little full-shift observation. Long-term care therefore remains underrepresented despite the high documentation burden [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref16">16</xref>-<xref ref-type="bibr" rid="ref18">18</xref>].</p></sec><sec id="s1-5"><title>Objectives</title><p>Therefore, this study aims to address these gaps by evaluating a domain-specific mobile AI speech assistant under routine conditions in German long-term care using full-shift time-motion observation at 2 time points, before and after implementation. The primary end point is the adjusted change in total documentation time per observed morning shift, defined as the cumulative time spent on a set of predefined documentation-related tasks. Secondary outcomes are self-reported and include usability, perceived documentation effort and interruptions, workplace satisfaction, and self-reported documentation quality.</p></sec><sec id="s1-6"><title>Hypotheses</title><p>The following hypothesis will be tested for the primary end point:</p><list list-type="simple"><list-item><p>H0: There will be no significant reduction measured by the adjusted mean change in total documentation time from baseline to postimplementation.</p></list-item><list-item><p>H1: There will be a significant reduction measured by the adjusted mean change in total documentation time from baseline to postimplementation.</p></list-item></list><p>The secondary outcomes are evaluated exploratorily by pre-post changes. To adjust for multiple testing, the Holm-Bonferroni correction is applied across all secondary tests.</p></sec></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design and Setting</title><p>The study was conducted as part of the &#x201C;PYSA &#x2013; Nursing Documentation With Hybrid Speech Assistant&#x201D; project, funded by the German Federal Ministry of Education and Research [<xref ref-type="bibr" rid="ref20">20</xref>]. The longitudinal research design follows a pre-post observational time-motion design with 2 measurement points (baseline <italic>t</italic><sub>0</sub> and postassessment <italic>t</italic><sub>1</sub>) and takes place in stationary long-term care facilities in Germany. Facilities were approached nationwide to ensure the representation of different conditions and to include both rural and urban areas as well as different types of providers. All participating facilities used either the same EHR system or paper-based documentation.</p><p>The minimum age for inclusion of participants was 18 years, with no upper age limit. Additional inclusion criteria were a completed 3-year nursing education, regular involvement in nursing documentation, and at least 1 year of professional experience. Participants had to be able to understand instructions in German and provide informed consent. Only staff employed in long-term care facilities in which the speech assistant voize was implemented during the study period were eligible. Exclusion criteria were applied to staff who were not permanently employed in the participating facilities, such as agency or temporary nurses. We report this observational study in accordance with the STROBE (Strengthening the Reporting of Observational Studies in Epidemiology) statement [<xref ref-type="bibr" rid="ref21">21</xref>].</p></sec><sec id="s2-2"><title>Recruitment</title><p>Participants were recruited through long-term care facilities that planned to implement the speech assistant voize and received written information about the study procedures. Following approval from the employee representative bodies of each facility, nursing staff were approached by the facility management. All participants were informed in writing that participation was voluntary, that data would be pseudonymized and analyzed in aggregated form, and that they could withdraw from the study at any time without facing any negative consequences. Participants were offered a voluntary small financial compensation for their time and effort.</p></sec><sec id="s2-3"><title>The Speech Assistant Voize</title><p>The speech assistant voize is a software-based mobile app that enables users to dictate nursing documentation via smartphone. It uses a continuously trained language model that processes voice inputs locally, assigns the content to the appropriate categories, and automatically generates structured documentation entries. These entries are reviewed and approved by nursing staff before being integrated into the existing documentation system. Developed in close collaboration with nursing professionals [<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>], the application is continuously refined through a collaborative, user-centered design process, and its underlying models are regularly trained with voluntarily submitted and anonymized documentation data to ensure ongoing improvements and adaptation to the linguistic characteristics and professional terminology used in nursing.</p><p>Technically, the system converts spoken language into text (speech-to-text) in real time using an end-to-end streaming ASR (automatic speech recognition) model. In conjunction with this, a set of specialized natural language processing models is activated as needed to handle specific tasks&#x2014;such as understanding context, matching spoken language to planned interventions, extracting key information, or structuring data&#x2014;ensuring that all components work together to deliver the supported features. As part of the study, voize was introduced in cooperation with voize GmbH. The implementation included planning and rollout within the facilities.</p></sec><sec id="s2-4"><title>Power Analysis</title><p>In a preliminary anonymous survey with 55 nurses already using voize, about 70% reported subjective time savings of 16 to 30 minutes per shift. It was assumed that about one-third of working time is spent on documentation-related tasks [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>], which is equivalent to about 20 minutes per hour. A 15% reduction was considered meaningful, corresponding to the mean of reported savings (23 minutes per shift), that is, 2.9 minutes less per working hour, reducing documentation from 20 to 17.1 minutes per hour. With an assumed SD of 6 minutes, this equates to an effect size of Cohen <italic>d</italic>=0.5. Power analysis using the R package pwr [<xref ref-type="bibr" rid="ref22">22</xref>] (&#x03B1;=.05, 2-sided; power=0.8) indicated a minimum of 34 participants for paired <italic>t</italic> tests. To account for an expected 15% dropout, the target sample size was set at 40, compensating for likely reasons such as scheduling conflicts, shift changes, sick leave, or vacation.</p><p>Early-stage dropout rates exceeded expectations. Therefore, additional recruitment efforts were employed. Nonetheless, interest from both facilities and individual nurses was high, and incremental participant burden remained low. As a result, a total of 52 nurses were included in this study.</p></sec><sec id="s2-5"><title>Measures</title><sec id="s2-5-1"><title>Primary End Point: Total Documentation Time per Shift</title><p>The primary outcome was total documentation time per observed morning shift in minutes, defined as the sum of all documentation-related categories recorded during the time-motion observation. Observation categories were developed and refined during preparatory nonparticipant observations in a long-term care facility within an unpublished pilot study [<xref ref-type="bibr" rid="ref20">20</xref>]. The identified documentation-related tasks included documentation at the PC or on paper, consulting information at the resident record at the PC, waiting for a PC, booting up and logging into the PC, walking to the ward office for documentation purposes, making notes, and reviewing entries (Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). All facilities applied their usual infection control protocols for mobile devices. Nurses were instructed to disinfect the study smartphones in accordance with local hygiene guidelines. When these hygiene steps occurred in direct connection with documentation, they were included in the observed documentation time. At <italic>t</italic><sub>1</sub>, after implementation of the application, additional categories were included for documentation with voize and consulting information in voize (Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p></sec><sec id="s2-5-2"><title>Nursing Workplace Satisfaction Questionnaire (NWSQ)</title><p>Workplace satisfaction outcomes were assessed using the NWSQ [<xref ref-type="bibr" rid="ref23">23</xref>]. The NWSQ is a brief measure of nursing job satisfaction encompassing 3 domains: intrinsic, extrinsic, and relational satisfaction. Items are rated on a 5-point Likert scale, and domain scores are computed as the sum of their constituent items. Possible ranges are 6&#x2010;30 for intrinsic, 5&#x2010;25 for extrinsic, and 4&#x2010;20 for relational satisfaction, with higher scores indicating greater satisfaction.</p></sec><sec id="s2-5-3"><title>Usability</title><p>Technology-related usability was measured with the system usability scale [<xref ref-type="bibr" rid="ref24">24</xref>] (SUS), a 10-item questionnaire of perceived usability on a 5-point Likert scale and alternating item valence. Responses were scored using the standard SUS procedure, including appropriate reverse-scoring, and converted to a total score from 0 to 100. Scores around 70 indicate acceptable usability, with higher values reflecting better perceived usability [<xref ref-type="bibr" rid="ref24">24</xref>].</p></sec><sec id="s2-5-4"><title>Self-Developed Questionnaire</title><p>Several single-item measures were developed and used to capture several facets of potential impact based on a previous study [<xref ref-type="bibr" rid="ref19">19</xref>]. Items included perceived interruptions during daily work, the perceived time-consuming nature of documentation, satisfaction with the documentation system, expected quality improvement, expected time savings, current documentation quality and completeness, implementation as a good idea, intention-to-use, and self-reported documentation time in minutes per shift. Items used 5-point scales or 10-point scales as shown in the Results section, plus open numerical entries for minutes. Higher ratings indicate more agreement with the item. For example, higher values on &#x201C;
documentation is time-consuming&#x201D;




 indicate greater perceived time burden, and higher values on &#x201C;satisfaction with the documentation system&#x201D;




 indicate greater satisfaction. Item wording and differences across <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub> are provided in Tables S1 and S2 in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p></sec></sec><sec id="s2-6"><title>Procedure</title><p>Data were collected between September 2024 and July 2025 in German long-term care facilities. Trained observers conducted continuous, nonparticipant full-shift observations of morning shifts with a planned duration of 8 hours, and observed shift lengths were very similar between <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub>.</p><p>Time was captured with an event-based smartphone logging app by starting and stopping predefined documentation-related task categories. After completing the shift, participants were asked to complete an online questionnaire that included the measures.</p><p>Approximately 8 weeks after the implementation of voize (<italic>t</italic><sub>1</sub>), participants were observed again during a full morning shift, and documentation times were recorded using the same procedure. At this second measurement point, participants completed an adjusted online survey, including the SUS.</p><p>Observations at both time points took place on typical weekdays outside major holiday periods. According to the participating units, bed capacity and staffing ratios were generally stable between <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub>. Taken together, these conditions suggest that major systematic differences in resident census between <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub> were unlikely, although smaller variations may still have occurred.</p></sec><sec id="s2-7"><title>Observers and Minimization of Bias</title><p>Prior to data collection, all observers were employees of voize GmbH who were specifically hired for the purpose of this study and were not involved in the development or design of the speech assistant. Observers had no performance incentives tied to study outcomes and were instructed to adhere strictly to the predefined task categories and timing rules to minimize the risk of biased classification or timing cessation. They received standardized training from the Charit&#x00E9; research team in the use of the measurement instruments and the predefined task categories. As part of this training, observers first jointly worked through standardized observation scenarios and coded them in parallel. Any discrepancies were then discussed systematically until a shared consensus on the application of each category was reached. In addition, weekly meetings were held throughout the data collection period to address questions, reflect on measurement procedures, and resolve ambiguities in a standardized manner. During the observations, observers also had access to a notes field where they could record questions or uncertainties regarding the assignment to task categories, add comments, and briefly describe the situation. This enabled the study personnel to later compare the selected task category with the descriptive notes and to correct obvious misclassifications where necessary.</p><p>In addition, the direct observation design can lead to behavioral changes among participants. The presence of an external observer can sometimes prompt improved performance, a phenomenon known as the Hawthorne effect [<xref ref-type="bibr" rid="ref25">25</xref>]. To mitigate such influences, repeated briefings and transparent communication of the study objectives were provided to the nursing staff and critical questions were encouraged.</p></sec><sec id="s2-8"><title>Statistical Analysis</title><p>Statistical analyses were performed using Python 3.11.9 and R 4.5.2 in Visual Studio Code 1.103.2 with the packages scikit-learn [<xref ref-type="bibr" rid="ref26">26</xref>], miceforest [<xref ref-type="bibr" rid="ref27">27</xref>], NumPy [<xref ref-type="bibr" rid="ref28">28</xref>], pandas [<xref ref-type="bibr" rid="ref29">29</xref>], SciPy [<xref ref-type="bibr" rid="ref30">30</xref>], Matplotlib [<xref ref-type="bibr" rid="ref31">31</xref>], tidyverse [<xref ref-type="bibr" rid="ref32">32</xref>], lme4 [<xref ref-type="bibr" rid="ref33">33</xref>], and lmerTest [<xref ref-type="bibr" rid="ref34">34</xref>]. All statistical methods used 2-sided tests at &#x03B1;=.05.</p><p>To account for the hierarchical and longitudinal structure of the data with repeated measurements within participants and clustering within facilities, the primary end point was analyzed using a linear mixed-effects model (LMM). The primary estimand was the adjusted mean change in documentation time from baseline to postimplementation (<italic>t</italic><sub>1</sub> &#x2013; <italic>t</italic><sub>0</sub>) in minutes. Time was coded as 0 at baseline and 1 at postimplementation, so that the fixed effect of time directly represents the adjusted mean change. The primary model included fixed effects for time, age, gender, and baseline documentation type, and random intercepts for participants and facilities. The model was estimated with restricted maximum likelihood. Inference for fixed effects used Kenward-Roger degrees of freedom.</p><p>The primary model used all available outcome observations and handled missing primary outcomes via likelihood-based estimation under a missing-at-random assumption. Robustness was assessed in exploratory sensitivity analyses, including a complete-case analysis requiring observations at both time points, models additionally accounting for observer effects, and alternative random-effects specifications. We further repeated analyses using multiple imputation with pooling via Rubin rules and Barnard-Rubin degrees of freedom and conducted missing-not-at-random (MNAR) pattern-mixture sensitivity analyses including jump-to-baseline and delta adjustments. As an additional robustness check, we repeated the analysis using a baseline-adjusted analysis of covariance with change scores. Full model specifications, statistical rationale, imputation settings, and MNAR sensitivity procedures are provided in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>. As the primary and secondary analyses fall under the frequentist paradigm, we additionally conducted an exploratory Bayesian hierarchical repeated-measures analysis (<xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>).</p><p>All secondary, self-reported outcomes were analyzed exploratorily using within-subject pre-post comparisons on multiply imputed datasets pooled according to Rubin rules and Barnard-Rubin small sample adjustment. For each outcome, the estimand was the mean change score (<italic>t</italic><sub>1</sub> &#x2013; <italic>t</italic><sub>0</sub>) with heteroscedasticity-consistent type 3 robust standard errors. To account for multiplicity, a Holm-Bonferroni correction was applied across secondary outcomes.</p></sec><sec id="s2-9"><title>Ethical Considerations</title><p>The study was reviewed and approved by the Ethics Committee of Charit&#x00E9; &#x2013; Universit&#x00E4;tsmedizin Berlin, approval number EA1/209/24. All participants provided written informed consent prior to participation. Data were pseudonymized and analyzed in aggregated form. The study was preregistered in the German Clinical Trials Register (DRKS00035512). Participants received 30 &#x20AC; (US $34.43) for their participation.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Participant Characteristics</title><p>The sample comprised 52 registered nurses from 14 long-term care facilities in Germany. Of the 52 participants, 8 (15.4%) dropped out between baseline and postassessment, primarily due to long-term illness or resignation.</p><p>The mean age was 42.37 (SD 12.37) years, with 17.52 (SD 9.63) years of professional experience and 9.42 (SD 8.13) years of tenure. Most participants were female (42/52, 80.8%). Among completers (44/52, 84.6%), the mean age was 43.23 (SD 12.01) years and mean experience was 18.2 (SD 9.95) years; 84.1% (37/44) were female. Dropouts (8/52, 15.4%) were younger on average (mean age 37.62, SD 14.07 years) and reported less work experience (mean 13.75, SD 6.88 years), and 5/8 (62.5%) of them were female. Detailed counts and percentages for all characteristics are provided in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Participant characteristics of the total sample, completers, and dropouts.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Variable</td><td align="left" valign="top">Total (N=52)</td><td align="left" valign="top">Completers (n=44)</td><td align="left" valign="top">Dropouts (n=8)</td></tr></thead><tbody><tr><td align="left" valign="top">Age (years), mean (SD)</td><td align="left" valign="top">42.37 (12.37)</td><td align="left" valign="top">43.23 (12.01)</td><td align="left" valign="top">37.62 (14.07)</td></tr><tr><td align="left" valign="top">Work experience (years), mean (SD)</td><td align="left" valign="top">17.52 (9.63)</td><td align="left" valign="top">18.20 (9.95)</td><td align="left" valign="top">13.75 (6.88)</td></tr><tr><td align="left" valign="top">Tenure (years), mean (SD)</td><td align="left" valign="top">9.42 (8.13)</td><td align="left" valign="top">9.75 (8.41)</td><td align="left" valign="top">7.62 (6.52)</td></tr><tr><td align="left" valign="top">Gender, n (%)</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Male</td><td align="left" valign="top">10 (19.2)</td><td align="left" valign="top">7 (15.9)</td><td align="left" valign="top">3 (37.5)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Female</td><td align="left" valign="top">42 (80.8)</td><td align="left" valign="top">37 (84.1)</td><td align="left" valign="top">5 (62.5)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Other</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td></tr><tr><td align="left" valign="top">Role, n (%)</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Registered nurse</td><td align="left" valign="top">52 (100)</td><td align="left" valign="top">44 (100)</td><td align="left" valign="top">8 (100)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Leadership role</td><td align="left" valign="top">6 (11.5)</td><td align="left" valign="top">5 (11.4)</td><td align="left" valign="top">1 (12.5)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Practical instructor</td><td align="left" valign="top">6 (11.5)</td><td align="left" valign="top">6 (13.6)</td><td align="left" valign="top">0 (0)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Other</td><td align="left" valign="top">1 (1.9)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">1 (12.5)</td></tr></tbody></table></table-wrap></sec><sec id="s3-2"><title>Primary Outcome</title><sec id="s3-2-1"><title>Descriptive Analyses</title><p>At <italic>t</italic><sub>0</sub> (n=52), documentation was predominantly PC-based (35/52, 67.3%), followed by hybrid use (11/52, 21.2%) and paper (6/52, 11.5%). Among completers at <italic>t</italic><sub>1</sub> (n=44), the speech assistant voize was reported as the primary documentation mode by 18/44 (40.9%), PC for 25/44 (56.8%), and paper by 1/44 (2.3%). Usage frequency of the speech assistant was high: 33/44 (75%) reported daily use, 9/44 (20.5%) used the speech assistant 3&#x2010;4 times per week, and 2/44 (4.5%) 1&#x2010;2 times per week.</p><p>Overall, around 770 hours of work were observed in long-term care during this study. Total documentation-related time decreased from a mean of 53.91 (SD 23.59) minutes at baseline to 39.10 (SD 16.29) minutes postimplementation in the analysis pooled across 40 multiply imputed datasets (&#x0394;=&#x2212;14.81 minutes). For the change in total documentation time (post &#x2013; pre in minutes), complete cases (n=44; see <xref ref-type="fig" rid="figure1">Figure 1</xref>) had Q1 &#x2013;22.48 and Q3 &#x2013;2.90 (min=&#x2212;85.90, max=32.60), and the pooled analysis (n=52) had Q1 &#x2013;23.59 and Q3 &#x2013;2.54 (min=&#x2212;86.95, max=32.91).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Baseline to postassessment changes in total documentation time among participants.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v28i1e86078_fig01.png"/></fig><p>Descriptive reductions were concentrated in active documentation (&#x0394;=&#x2212;13.28) and note-taking (&#x0394;=&#x2212;2.38), while reviewing entries (&#x0394;=+0.64) and information retrieval (&#x0394;=+1.28) showed small descriptive increases, consistent with a shift from producing text to checking/retrieving information.</p></sec><sec id="s3-2-2"><title>Primary Outcome: Changes in Total Documentation Time</title><p>The LMM with random intercepts for participants and facilities and fixed effects for time, age, gender, and baseline documentation type indicated a statistically significant reduction in total documentation time. The adjusted mean change for time (<italic>&#x03B2;</italic>) was &#x2212;15 (SE 3.36) minutes, <italic>t</italic><sub>46.29</sub>=&#x2212;4.46, <italic>P</italic>&#x003C;.001, with a 95% CI &#x2212;21.75 to &#x2212;8.23. Relative to the observed baseline mean of 53.91 minutes, this corresponds to an adjusted 28% reduction in documentation time per morning shift. Variance partitioning indicated an intraclass correlation coefficient of 0.36 at the participant level and 0.08 at the facility level.</p><p>Exploratory analyses including time-by-covariate interactions did not indicate reliable heterogeneity of the time effect by age (<italic>P=</italic>.25), gender (<italic>P=</italic>.87), or baseline documentation type (<italic>P=</italic>.49). The estimated reduction was stable across sensitivity analyses, with effect estimates and confidence intervals closely aligned with the primary model. The reduction also remained highly robust under extreme simulated MNAR departures. Detailed results of the exploratory and sensitivity analyses are provided in <xref ref-type="supplementary-material" rid="app5">Multimedia Appendix 5</xref>.</p></sec></sec><sec id="s3-3"><title>Secondary, Self-Reported Outcomes</title><p>There was a significant reduction in self-reported documentation time (&#x0394;=&#x2212;31.14, SE 6.57 minutes), <italic>t</italic><sub>43.18</sub>=&#x2212;4.88, <italic>P</italic>&#x003C;.001, 95% CI &#x2212;45.31 to &#x2212;18.83. Self-reported daily work interruptions also declined, <italic>t</italic><sub>47.00</sub>=&#x2212;3.95, <italic>P</italic>&#x003C;.001, 95% CI &#x2212;0.95 to &#x2212;0.31. Participants experienced the documentation process as markedly less time-consuming, <italic>t</italic><sub>46.91</sub>=&#x2212;9.32, <italic>P</italic>&#x003C;.001, 95% CI &#x2212;1.97 to &#x2212;1.27, and satisfaction with the documentation system increased after the implementation of voize, <italic>t</italic><sub>46.47</sub>=4.84, <italic>P</italic>&#x003C;.001, 95% CI 1.04&#x2010;2.52. Expectations that the voice assistant would improve documentation quality were also lower after use, <italic>t</italic><sub>46.82</sub>=&#x2212;3.40, <italic>P</italic>=.001, 95% CI &#x2212;0.75 to &#x2212;0.19. All 5 of these effects remained significant after Holm-Bonferroni correction across the 13-test family (Holm-adjusted <italic>P</italic>&#x2264;.013 for each; see <xref ref-type="table" rid="table2">Table 2</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Pooled measures of central tendency for secondary, self-reported outcomes<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Variables (score range: min-max)</td><td align="left" valign="bottom">Pooled baseline, mean (SD)</td><td align="left" valign="bottom">Pooled post, mean (SD)</td><td align="left" valign="bottom">Mean difference (post &#x2013; pre)</td><td align="left" valign="bottom">Median (pre)</td><td align="left" valign="bottom">Median (post)</td></tr></thead><tbody><tr><td align="left" valign="top">Interruptions in daily work (range: 1&#x2010;5)</td><td align="left" valign="top">4.06 (0.78)</td><td align="left" valign="top">3.40 (0.91)</td><td align="left" valign="top">&#x2212;0.66<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>***</td><td align="left" valign="top">4.00</td><td align="left" valign="top">3.97</td></tr><tr><td align="left" valign="top">Self-reported documentation time (in minutes)</td><td align="left" valign="top">67.05 (40.98)</td><td align="left" valign="top">35.91 (24.34)</td><td align="left" valign="top">&#x2212;31.14<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>***</td><td align="left" valign="top">60.00</td><td align="left" valign="top">30.00</td></tr><tr><td align="left" valign="top">Documentation is time-consuming (range: 1&#x2010;5)</td><td align="left" valign="top">4.07 (0.79)</td><td align="left" valign="top">2.46 (0.84)</td><td align="left" valign="top">&#x2212;1.61<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>***</td><td align="left" valign="top">4.00</td><td align="left" valign="top">2.18</td></tr><tr><td align="left" valign="top">Satisfaction with documentation system (range: 1&#x2010;10)</td><td align="left" valign="top">5.07 (2.07)</td><td align="left" valign="top">6.65 (1.82)</td><td align="left" valign="top">+1.58<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>***</td><td align="left" valign="top">5.00</td><td align="left" valign="top">7.00</td></tr><tr><td align="left" valign="top">Expected quality improvement (range: 1&#x2010;5)</td><td align="left" valign="top">3.69 (0.74)</td><td align="left" valign="top">3.18 (0.81)</td><td align="left" valign="top">&#x2212;0.51<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>**</td><td align="left" valign="top">4.00</td><td align="left" valign="top">3.00</td></tr><tr><td align="left" valign="top">Documentation is complete (range: 1&#x2010;5)</td><td align="left" valign="top">3.27 (0.85)</td><td align="left" valign="top">2.83 (0.94)</td><td align="left" valign="top">&#x2212;0.44**</td><td align="left" valign="top">3.00</td><td align="left" valign="top">2.87</td></tr><tr><td align="left" valign="top">Expected time savings (range: 1&#x2010;5)</td><td align="left" valign="top">3.81 (0.69)</td><td align="left" valign="top">3.46 (0.79)</td><td align="left" valign="top">&#x2212;0.35*</td><td align="left" valign="top">4.00</td><td align="left" valign="top">3.57</td></tr><tr><td align="left" valign="top">High documentation quality (range: 1&#x2010;5)</td><td align="left" valign="top">3.38 (0.82)</td><td align="left" valign="top">3.03 (0.75)</td><td align="left" valign="top">&#x2212;0.35*</td><td align="left" valign="top">3.89</td><td align="left" valign="top">3.00</td></tr><tr><td align="left" valign="top">Implementation as a good idea (range: 1&#x2010;10)</td><td align="left" valign="top">7.91 (1.93)</td><td align="left" valign="top">7.35 (2.04)</td><td align="left" valign="top">&#x2212;0.56</td><td align="left" valign="top">8.00</td><td align="left" valign="top">7.73</td></tr><tr><td align="left" valign="top">Intention to use (range: 1&#x2010;10)</td><td align="left" valign="top">8.17 (1.99)</td><td align="left" valign="top">7.88 (1.69)</td><td align="left" valign="top">&#x2212;0.29</td><td align="left" valign="top">8.13</td><td align="left" valign="top">8.00</td></tr><tr><td align="left" valign="top">NWSQ<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup>: extrinsic (range: 5&#x2010;25)</td><td align="left" valign="top">17.60 (2.34)</td><td align="left" valign="top">17.95 (2.94)</td><td align="left" valign="top">+0.35</td><td align="left" valign="top">17.51</td><td align="left" valign="top">18.00</td></tr><tr><td align="left" valign="top">NWSQ: intrinsic (range: 6&#x2010;30)</td><td align="left" valign="top">22.21 (3.94)</td><td align="left" valign="top">22.04 (4.42)</td><td align="left" valign="top">&#x2212;0.17</td><td align="left" valign="top">22.00</td><td align="left" valign="top">21.63</td></tr><tr><td align="left" valign="top">NWSQ: relational (range: 4&#x2010;20)</td><td align="left" valign="top">15.62 (2.62)</td><td align="left" valign="top">15.31 (2.66)</td><td align="left" valign="top">&#x2212;0.31</td><td align="left" valign="top">16.00</td><td align="left" valign="top">16.00</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>All values are pooled across 40 multiply imputed datasets according to Rubin rules. Consequently, some median scores are not restricted to the discrete response categories. The significance of the change is indicated by asterisks based on paired <italic>t</italic> tests. <italic>*P</italic>&#x003C;.05; **<italic>P</italic>&#x003C;.01; ***<italic>P</italic>&#x003C;.001.</p></fn><fn id="table2fn2"><p><sup>b</sup>Effects that remained significant after Holm-Bonferroni correction for 13 tests.</p></fn><fn id="table2fn3"><p><sup>c</sup>NWSQ: Nursing Workplace Satisfaction Questionnaire.</p></fn></table-wrap-foot></table-wrap><p>Perceiving the implementation of voize as a good idea showed no reliable change, <italic>t</italic><sub>46.29</sub>=&#x2212;1.09, <italic>P=</italic>.28, 95% CI &#x2212;1.04 to 0.31, and intention to use remained essentially unchanged at a high level, <italic>t</italic><sub>46.60</sub>=&#x2212;0.45, <italic>P</italic>=.65, 95% CI &#x2212;0.77 to 0.49. Both remain nonsignificant after correction (Holm-adjusted <italic>P</italic>&#x2265;.99). Ratings that the current documentation is qualitatively high decreased, <italic>t</italic><sub>46.97</sub>=&#x2212;2.30, <italic>P</italic>=.03, 95% CI &#x2212;0.61 to &#x2212;0.04, as did ratings that it is complete and comprehensive, <italic>t</italic><sub>47.01</sub>=&#x2013;2.75, <italic>P=</italic>.008, 95% CI &#x2212;0.74 to &#x2013;0.11, and expected time savings with the speech assistant, <italic>t</italic><sub>46.74</sub>=&#x2013;2.60, <italic>P</italic>=.01, 95% CI &#x2212;0.55 to &#x2212;0.07. However, none of these 3 effects survived Holm-Bonferroni correction (Holm-adjusted <italic>P=</italic>.16, .06, and .09, respectively).</p><p>Nursing workplace satisfaction measured by the NWSQ did not change significantly (extrinsic: <italic>t</italic><sub>38.94</sub>=1.48, <italic>P</italic>=.15, 95% CI &#x2212;0.22 to 1.45; intrinsic: <italic>t</italic><sub>39.06</sub>=0.07, <italic>P=</italic>.95, 95% CI &#x2212;1.03 to 1.10; relational: <italic>t</italic><sub>38.46</sub>=&#x2212;0.68, <italic>P</italic>=.5, 95% CI &#x2212;0.84 to 0.42), and these remain nonsignificant after correction (Holm-adjusted <italic>P</italic>=.73, &#x2265;.99, and &#x2265;.99, respectively). The SUS scores indicated acceptable usability [<xref ref-type="bibr" rid="ref24">24</xref>]: the complete-case mean was 70.42 (SD 14.26; n=40).</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study provides real-world evidence from long-term care that a mobile speech assistant is associated with reduced documentation time over full shifts. The primary LMM estimated an adjusted reduction of approximately 15 minutes per morning shift, corresponding to an approximately 28% adjusted decrease relative to the baseline mean. This finding was robust across extensive sensitivity analyses. The interquartile range indicated that most participants experienced time savings. Task-level patterns suggest that time was recaptured by making active documentation more efficient (see <xref ref-type="table" rid="table3">Table 3</xref>). However, net time savings may be overestimated as infection control steps such as disinfection time and adherence to infection control protocols were not systematically recorded in this study. In parallel, nurses reported fewer interruptions, lower perceived time burden, and higher satisfaction with the mobile AI speech assistant. Notably, self-reported retrospective time savings were substantially larger than the observed reduction, and perceived baseline documentation time exceeded our time-motion estimates. This divergence suggests that perceived relief in documentation burden may exceed objective time savings when moving from manual entry to speech-based workflows. Retrospective estimates may apply broader boundaries to what counts as documentation and rely on memory-based summary impressions that overweight salient moments. In this context, cognitively exhaustive, interruption-prone work such as administrative documentation may be remembered as more time-consuming, whereas patient interaction may feel more absorbing and therefore subjectively shorter. By contrast, time-motion observation captures only predefined, observable tasks within the sampled shift, yielding an objective estimate of recorded activities that does not necessarily map onto perceived workload.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Documentation time by task in minutes at baseline and postimplementation.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Task</td><td align="left" valign="bottom">Pre<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup> (minutes), mean (SD)</td><td align="left" valign="bottom">Post-complete<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup> (minutes), mean (SD)</td><td align="left" valign="bottom">Post-pooled<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup> (minutes), mean (SD)</td></tr></thead><tbody><tr><td align="left" valign="top">Computer login</td><td align="left" valign="top">1.45 (1.32)</td><td align="left" valign="top">1.05 (1.13)</td><td align="left" valign="top">1.01 (1.09)</td></tr><tr><td align="left" valign="top">Active documentation</td><td align="left" valign="top">44.23 (19.72)</td><td align="left" valign="top">30.44 (14.71)</td><td align="left" valign="top">30.95 (14.36)</td></tr><tr><td align="left" valign="top">Reviewing entries</td><td align="left" valign="top">0.82 (3.65)</td><td align="left" valign="top">1.47 (3.81)</td><td align="left" valign="top">1.46 (3.68)</td></tr><tr><td align="left" valign="top">Information retrieval</td><td align="left" valign="top">3.48 (4.71)</td><td align="left" valign="top">4.85 (6.69)</td><td align="left" valign="top">4.76 (6.48)</td></tr><tr><td align="left" valign="top">Walking to the PC (for documentation)<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="top">0.56 (1.56)</td><td align="left" valign="top">0.10 (0.24)</td><td align="left" valign="top">0.09 (0.23)</td></tr><tr><td align="left" valign="top">Taking notes</td><td align="left" valign="top">3.15 (2.68)</td><td align="left" valign="top">0.84 (1.67)</td><td align="left" valign="top">0.77 (1.56)</td></tr><tr><td align="left" valign="top">Waiting</td><td align="left" valign="top">0.22 (0.88)</td><td align="left" valign="top">0.07 (0.33)</td><td align="left" valign="top">0.06 (0.31)</td></tr><tr><td align="left" valign="top">Total documentation-related time</td><td align="left" valign="top">53.91 (23.59)</td><td align="left" valign="top">38.82 (16.75)</td><td align="left" valign="top">39.10 (16.29)</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>Pre: observed baseline for all participants (N=52).</p></fn><fn id="table3fn2"><p><sup>b</sup>Post-complete: observed postimplementation among completers (n=44).</p></fn><fn id="table3fn3"><p><sup>c</sup>Post-pooled: estimates pooled across 40 multiply imputed datasets under &#x201C;missing at random&#x201D; assumption (N=52).</p></fn><fn id="table3fn4"><p><sup>d</sup>&#x201C;Walking to the PC (for documentation)&#x201D; refers exclusively to walking carried out to reach stationary ward computers or the ward office for documentation purposes. Walking that occurred while documentation was being performed (eg, dictating with the mobile speech assistant while moving) was not logged as a separate walking task but was counted only as documentation time.</p></fn></table-wrap-foot></table-wrap><p>Overall usability was rated as acceptable. Taken together, these findings suggest that introducing speech-enabled, mobile documentation can reduce documentation time while improving perceived documentation efficiency.</p><p>At the same time, self-reported documentation quality and completeness showed a consistent negative trend after implementation. In the frequentist exploratory analyses, these decreases were statistically significant before multiplicity correction, but not after Holm-Bonferroni adjustment. However, the complementary Bayesian analysis used adaptive regularization for secondary outcomes and assessed practical relevance using a region of practical equivalence. The Bayesian analysis still assigned a high posterior probability of negative changes in perceived documentation completeness and quality, while the 94% highest density intervals slightly overlapped zero (see Multimedia Appendix 4). Together, these findings suggest a potential trade-off in routine use. Faster, more mobile documentation may coincide with lower perceived completeness or quality, possibly because of recognition errors, editing burden, or shorter entries [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>In stationary long-term care, documentation consumes substantial nurse time. A domain-specific mobile speech assistant may reclaim minutes each shift, possibly reduce interruptions, and improve the overall documentation experience. In labor-constrained markets, recovered time can be redeployed to resident care, coordination, supervision, taking necessary breaks, and improving working conditions while lowering overtime [<xref ref-type="bibr" rid="ref19">19</xref>].</p></sec><sec id="s4-2"><title>Relation to Prior Work</title><p>These findings are consistent with experimental and simulation studies showing that speech-based input improves speed and, in some contexts, accuracy. In nursing, Mayer et al [<xref ref-type="bibr" rid="ref17">17</xref>] found time savings of 2.3 to 6.1 minutes per scenario without increased errors and high user preference. Similarly, Mairittha et al [<xref ref-type="bibr" rid="ref18">18</xref>] reported 15% faster entries with a dialogue-based system. In addition, another mobile bedside app reduced EHR charting time and increased uninterrupted patient interaction in a pre-post study [<xref ref-type="bibr" rid="ref16">16</xref>]. In broader clinical samples, speech systems shortened task completion time and reduced error rates versus paper and traditional patient data management systems in intensive care units [<xref ref-type="bibr" rid="ref12">12</xref>]. In hospital-wide deployments, they also achieved high adoption with large transcription-cost reductions and perceived gains in quality and efficiency [<xref ref-type="bibr" rid="ref35">35</xref>]. Observational comparisons have also shown that speech can be substantially faster than typing, even when perceptions may differ [<xref ref-type="bibr" rid="ref11">11</xref>].</p><p>Relative to this body of evidence, this study adds three elements. First, it focuses on long-term care, where evidence has been sparse and documentation burden is high. Second, it evaluates the technology under routine conditions, not only in controlled or simulated scenarios. Third, it uses full-shift observation at both time points, yielding approximately 770 hours of time-motion data and directly aligned with the workflow realities reported in earlier studies [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Together, these features can help bridge the gap highlighted by reviews that call for end-to-end, real-world evaluations beyond transcription accuracy alone [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref15">15</xref>]. The pattern of decreased active entry, waiting, and walking to the stationary ward computers, with slight increases in retrieval and review (see <xref ref-type="table" rid="table3">Table 3</xref>), also may show that mobile tools reduce &#x201C;print-note-retype&#x201D; loops [<xref ref-type="bibr" rid="ref16">16</xref>].</p><p>Several mechanisms may influence the observed association between speech assistant use and reduced documentation time. First, this is in line with previous qualitative findings of Ferizaj and Neumann [<xref ref-type="bibr" rid="ref19">19</xref>] that have shown fewer interruptions, less walking to the ward, shorter information-retrieval times, and higher overall satisfaction with the documentation system. In line with this, the mobile, as-you-go entry may have further reduced travel and queuing at ward PCs, cutting waiting, walking to the ward, and repeated logins. Second, domain-tuned speech and natural language processing (ie, models adapted to the nursing domain) may lower the friction of capturing structured information. This may reduce the need for interim notes and subsequent re-entry while also minimizing information loss and documentation errors. Third, workflow coupling at the bedside may decrease interruptions and handovers. The descriptive rise in information retrieval and review may reflect appropriate verification behavior within the mobile app. Usability in the good range [<xref ref-type="bibr" rid="ref24">24</xref>] and stable acceptance [<xref ref-type="bibr" rid="ref37">37</xref>] suggest the solution was well received with high intention-to-use and satisfaction, albeit with room for refinement.</p></sec><sec id="s4-3"><title>Limitations</title><p>This study has several limitations. The single-group pre-post design without a control condition limits causal inference and leaves room for workload fluctuations or co-occurring process changes as alternative explanations. According to reports from facility management, no other major documentation tools or workflow redesigns were introduced between <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub>, which reduces, but does not eliminate, the risk of confounding by concurrent initiatives. Participation was voluntary, so self-selection may have favored early adopters. Additionally, as facilities reported generally stable conditions, we did not collect shift-level workload indicators such as number of residents cared for or staffing ratios for each observed shift. Therefore, documentation time was not normalized for patient census or workload, and residual confounding by day-to-day fluctuations remains possible. Observations were conducted by multiple observers. Although observers underwent standardized training, joint pilot coding, and ongoing weekly calibration meetings, we did not calculate a formal interrater reliability coefficient (eg,


 Cohen


 &#x03BA;) for the final field observations. This limits the extent to which we can quantify measurement consistency across observers. However, an exploratory sensitivity analysis including observer ID as a covariate did not materially change the estimated reduction in documentation time, suggesting that systematic between-observer differences are unlikely to account for the primary findings. Still, this exploratory analysis cannot rule out shared, potentially unconscious observer bias operating in the same direction across time points. Fourth, direct observation can itself change behavior (Hawthorne effect), which we sought to mitigate with standardized procedures but cannot exclude [<xref ref-type="bibr" rid="ref25">25</xref>]. We did not systematically record time spent on disinfecting mobile devices unless it occurred in direct connection with documentation or monitor adherence to infection control protocols. If device disinfection is performed more consistently in routine use, the net time savings reported here may be overestimated. Documentation quality was captured perceptually rather than via objective measures such as assessing structured completeness or error rates that other work has examined in standardized tasks [<xref ref-type="bibr" rid="ref12">12</xref>]. We did not conduct independent audits of documentation accuracy (eg, chart reviews, error checks). Reduced documentation time therefore cannot be directly linked to unchanged or improved documentation quality. Furthermore, dropouts were descriptively younger and less experienced than completers. This pattern may limit generalizability for adoption, since retention and continued engagement with the tool may differ by career stage, training status, or role-related workload. Finally, results reflect stationary long-term care in Germany, German-language ASR, and a specific EHR integration. Performance may differ with dialects, accents, terminology, and integration quality [<xref ref-type="bibr" rid="ref15">15</xref>], and ongoing advances in speech technology may change effect sizes over time.</p></sec><sec id="s4-4"><title>Future Research</title><p>This study complements prior simulation and laboratory research and provides further evaluation of the potential effects of AI speech assistants in long-term care. To better inform workforce policy in the context of nursing shortages, future implementation research should examine whether documentation automation yields sustained workload relief, frees time for direct resident care, coordination, and recovery, and under which organizational and technical conditions these benefits can be realized at scale.</p><p>To strengthen causal inference and capture secular trends and learning effects, future studies should use cluster-randomized or stepped-wedge designs with extended follow-up. These evaluations should combine time-motion data with independent, objective assessments of documentation quality (eg,

 completeness indices, semantic accuracy ratings, and error rates) to determine whether time savings are achieved without compromising data integrity. In addition, studies should assess how saved time is reinvested and whether this translates into improvements in resident outcomes, psychosocial workload, perceived interruptions, staff satisfaction, and retention.</p><p>Methodologically, linking full-shift observation with EHR telemetry could quantify sustainability, diurnal patterns, and adoption trajectories. Additionally, further subgroup analyses could explore heterogeneity by accent/dialect, professional experience, unit characteristics, and integration context. Finally, economic evaluations should estimate the total cost of ownership and downstream effects on productivity, overtime, sick leave, and turnover to guide procurement and financing decisions for broader implementation.</p></sec><sec id="s4-5"><title>Conclusions</title><p>To the authors&#x2019; knowledge, this is the first study to report real-world time savings observed in association with the use of a mobile AI speech assistant in long-term care. In this longitudinal pre-post time-motion study conducted in 14 German long-term care facilities, implementation of a mobile, domain-specific speech documentation system into routine workflows was associated with substantially lower documentation time. Specifically, the mean documentation-related time decreased from about 54 minutes at baseline to approximately 39 minutes postimplementation. In the primary LMM with random intercepts for participants and facilities and fixed effects for time, age, gender, and baseline documentation type, the adjusted mean change was &#x2212;15 minutes, corresponding to an adjusted 28% reduction relative to the baseline mean. This result remained consistent across sensitivity analyses and MNAR pattern-mixture scenarios. Task-level patterns indicated that time savings were primarily driven by less active documentation and note-taking, while reviewing entries and information retrieval showed small increases. This is consistent with a shift from producing text to verifying and retrieving information [<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>Secondary, exploratory self-report measures converged with the observational findings: participants reported fewer interruptions, lower perceived documentation burden, and higher satisfaction with the documentation system, while nursing workplace satisfaction did not change. Usability was rated as acceptable. Most participants experienced measurable time savings, with approximately one quarter reducing documentation time by more than about 24 minutes per shift. Given the single-group pre-post design and the reliance on self-reported indicators for documentation quality, these results should be interpreted as evidence of an association between implementation and reduced documentation time rather than as causal proof or as objective evidence of changes in documentation quality.</p></sec></sec></body><back><ack><p>We would like to thank the German Federal Ministry of Education and Research as the responsible funding body. In addition, we would like to thank the participating facilities, and especially the nursing staff, who contributed their insights and expertise to this study.</p><p>We used the generative AI tool ChatGPT by OpenAI to assist with language refinement and translation of text originally written by the authors, and to help identify potentially relevant scientific literature. All suggested references were independently reviewed and verified by the authors. No generative AI tools were used to generate scientific content, analyses, results, or interpretations.</p></ack><notes><sec><title>Funding</title><p>This study was conducted as part of the &#x201C;PYSA &#x2013; Nursing Documentation With Hybrid Speech Assistant&#x201D; project, funded by the German Federal Ministry of Education and Research.</p></sec><sec><title>Data Availability</title><p>The datasets generated and analyzed in this study contain highly granular time-motion logs and contextual information from nursing staff in German long-term care facilities. Owing to data protection regulations, institutional data use agreements, and the terms of the ethics approval and informed consent based on a preregistered study protocol that did not foresee open data sharing, these data cannot be made publicly available. Deidentified aggregate data are reported in the manuscript and appendices. Researchers who wish to access additional deidentified data for scientifically sound purposes may contact the corresponding author; data sharing will be considered on a case-by-case basis and may require additional institutional approvals and data use agreements.</p></sec></notes><fn-group><fn fn-type="con"><p>KS contributed to the concept and design of the study, data interpretation, visualization, preparation of the original draft, and review and editing of the manuscript. DF contributed to the study concept and design, data curation, formal analysis, investigation, visualization, preparation of the original draft, and review and editing of the manuscript. SN contributed to the study design, investigation, project administration, interpretation of data, and review and editing of the manuscript. SS-L contributed to project administration, concept and design, funding acquisition, and review and editing of the manuscript. NL contributed to methodological guidance, formal analysis, supervision, funding acquisition, and review and editing of the manuscript. All authors have read and approved the final manuscript.</p></fn><fn fn-type="conflict"><p>KS is employed by voize GmbH, which developed the speech assistant evaluated in this study. She was not involved in the development of the system. All other authors declare no competing interests.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">ASR</term><def><p>automatic speech recognition</p></def></def-item><def-item><term id="abb3">EHR</term><def><p>electronic health record</p></def></def-item><def-item><term id="abb4">LMM</term><def><p>linear mixed-effects model</p></def></def-item><def-item><term id="abb5">MNAR</term><def><p>missing not at random</p></def></def-item><def-item><term id="abb6">NWSQ</term><def><p>Nursing Workplace Satisfaction Questionnaire</p></def></def-item><def-item><term id="abb7">PYSA</term><def><p>&#x201C;Pflegedokumentation mit Hybridem Sprachassistenten&#x201D; (project title)</p></def></def-item><def-item><term id="abb8">STROBE</term><def><p>Strengthening the Reporting of Observational Studies in Epidemiology</p></def></def-item><def-item><term id="abb9">SUS</term><def><p>system usability scale</p></def></def-item><def-item><term id="abb10"><italic>t</italic><sub>0</sub></term><def><p>baseline measurement</p></def></def-item><def-item><term id="abb11"><italic>t</italic><sub>1</sub></term><def><p>postimplementation measurement</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="report"><article-title>Arbeitsmarktsituation im Pflegebereich [Report in German]</article-title><year>2025</year><month>05</month><access-date>2026-03-13</access-date><publisher-name>Statistik der Bundesagentur f&#x00FC;r Arbeit</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://statistik.arbeitsagentur.de/DE/Statischer-Content/Statistiken/Themen-im-Fokus/Berufe/Generische-Publikationen/Altenpflege.html?__blob=publicationFile">https://statistik.arbeitsagentur.de/DE/Statischer-Content/Statistiken/Themen-im-Fokus/Berufe/Generische-Publikationen/Altenpflege.html?__blob=publicationFile</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="web"><article-title>Pflegekr&#x00E4;ftevorausberechnung [Website in German]</article-title><source>Statistisches Bundesamt (Destatis)</source><access-date>2026-3-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Bevoelkerung/Bevoelkerungsvorausberechnung/pflegekraeftevorausberechnung.html">https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Bevoelkerung/Bevoelkerungsvorausberechnung/pflegekraeftevorausberechnung.html</ext-link></comment></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wirth</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ulusoy</surname><given-names>N</given-names> </name><name name-style="western"><surname>Lincke</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Nienhaus</surname><given-names>N</given-names> </name><name name-style="western"><surname>Schablon</surname><given-names>A</given-names> </name></person-group><article-title>Psychosoziale Belastungen und Beanspruchungen von Besch&#x00E4;ftigten in der station&#x00E4;ren und ambulanten Altenpflege [Article in German]</article-title><source>ASU Arbeitsmed Sozialmed Umweltmed</source><year>2017</year><volume>52</volume><issue>9</issue><fpage>662</fpage><lpage>669</lpage><pub-id pub-id-type="doi">10.17147/ASU2017-09-01-01</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hasson</surname><given-names>H</given-names> </name><name name-style="western"><surname>Arnetz</surname><given-names>JE</given-names> </name></person-group><article-title>Nursing staff competence, work strain, stress and satisfaction in elderly care: a comparison of home-based care and nursing homes</article-title><source>J Clin Nurs</source><year>2008</year><month>02</month><volume>17</volume><issue>4</issue><fpage>468</fpage><lpage>481</lpage><pub-id pub-id-type="doi">10.1111/j.1365-2702.2006.01803.x</pub-id><pub-id pub-id-type="medline">17331093</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dall&#x2019;Ora</surname><given-names>C</given-names> </name><name name-style="western"><surname>Ejebu</surname><given-names>OZ</given-names> </name><name name-style="western"><surname>Ball</surname><given-names>J</given-names> </name><name name-style="western"><surname>Griffiths</surname><given-names>P</given-names> </name></person-group><article-title>Shift work characteristics and burnout among nurses: cross-sectional survey</article-title><source>Occup Med</source><year>2023</year><month>05</month><day>18</day><volume>73</volume><issue>4</issue><fpage>199</fpage><lpage>204</lpage><pub-id pub-id-type="doi">10.1093/occmed/kqad046</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schenk</surname><given-names>E</given-names> </name><name name-style="western"><surname>Schleyer</surname><given-names>R</given-names> </name><name name-style="western"><surname>Jones</surname><given-names>CR</given-names> </name><name name-style="western"><surname>Fincham</surname><given-names>S</given-names> </name><name name-style="western"><surname>Daratha</surname><given-names>KB</given-names> </name><name name-style="western"><surname>Monsen</surname><given-names>KA</given-names> </name></person-group><article-title>Time motion analysis of nursing work in ICU, telemetry and medical-surgical units</article-title><source>J Nurs Manag</source><year>2017</year><month>11</month><volume>25</volume><issue>8</issue><fpage>640</fpage><lpage>646</lpage><pub-id pub-id-type="doi">10.1111/jonm.12502</pub-id><pub-id pub-id-type="medline">28853187</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yen</surname><given-names>PY</given-names> </name><name name-style="western"><surname>Kellye</surname><given-names>M</given-names> </name><name name-style="western"><surname>Lopetegui</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Nurses&#x2019; time allocation and multitasking of nursing activities: a time motion study</article-title><source>AMIA Annu Symp Proc</source><year>2018</year><volume>2018</volume><fpage>1137</fpage><lpage>1146</lpage><pub-id pub-id-type="medline">30815156</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Momenipur</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pennathur</surname><given-names>PR</given-names> </name></person-group><article-title>Balancing documentation and direct patient care activities: a study of a mature electronic health record system</article-title><source>Int J Ind Ergon</source><year>2019</year><month>07</month><volume>72</volume><fpage>338</fpage><lpage>346</lpage><pub-id pub-id-type="doi">10.1016/j.ergon.2019.06.012</pub-id><pub-id pub-id-type="medline">32201437</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gesner</surname><given-names>E</given-names> </name><name name-style="western"><surname>Dykes</surname><given-names>PC</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Gazarian</surname><given-names>P</given-names> </name></person-group><article-title>Documentation burden in nursing and its role in clinician burnout syndrome</article-title><source>Appl Clin Inform</source><year>2022</year><month>10</month><volume>13</volume><issue>5</issue><fpage>983</fpage><lpage>990</lpage><pub-id pub-id-type="doi">10.1055/s-0042-1757157</pub-id><pub-id pub-id-type="medline">36261113</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Joseph</surname><given-names>J</given-names> </name><name name-style="western"><surname>Moore</surname><given-names>ZEH</given-names> </name><name name-style="western"><surname>Patton</surname><given-names>D</given-names> </name><name name-style="western"><surname>O&#x2019;Connor</surname><given-names>T</given-names> </name><name name-style="western"><surname>Nugent</surname><given-names>LE</given-names> </name></person-group><article-title>The impact of implementing speech recognition technology on the accuracy and efficiency (time to complete) clinical documentation by nurses: a systematic review</article-title><source>J Clin Nurs</source><year>2020</year><month>07</month><volume>29</volume><issue>13-14</issue><fpage>2125</fpage><lpage>2137</lpage><pub-id pub-id-type="doi">10.1111/jocn.15261</pub-id><pub-id pub-id-type="medline">32243006</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zuchowski</surname><given-names>M</given-names> </name><name name-style="western"><surname>G&#x00F6;ller</surname><given-names>A</given-names> </name></person-group><article-title>Speech recognition for medical documentation: an analysis of time, cost efficiency and acceptance in a clinical setting</article-title><source>Br J Healthc Manag</source><year>2022</year><month>01</month><day>2</day><volume>28</volume><issue>1</issue><fpage>30</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.12968/bjhc.2021.0074</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Peine</surname><given-names>A</given-names> </name><name name-style="western"><surname>Gronholz</surname><given-names>M</given-names> </name><name name-style="western"><surname>Seidl-Rathkopf</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Standardized comparison of voice-based information and documentation systems to established systems in intensive care: crossover study</article-title><source>JMIR Med Inform</source><year>2023</year><month>11</month><day>28</day><volume>11</volume><fpage>e44773</fpage><pub-id pub-id-type="doi">10.2196/44773</pub-id><pub-id pub-id-type="medline">38015593</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alboksmaty</surname><given-names>A</given-names> </name><name name-style="western"><surname>Aldakhil</surname><given-names>R</given-names> </name><name name-style="western"><surname>Hayhoe</surname><given-names>BWJ</given-names> </name><name name-style="western"><surname>Ashrafian</surname><given-names>H</given-names> </name><name name-style="western"><surname>Darzi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Neves</surname><given-names>AL</given-names> </name></person-group><article-title>The impact of using AI-powered voice-to-text technology for clinical documentation on quality of care in primary care and outpatient settings: a systematic review</article-title><source>eBioMedicine</source><year>2025</year><month>08</month><volume>118</volume><fpage>105861</fpage><pub-id pub-id-type="doi">10.1016/j.ebiom.2025.105861</pub-id><pub-id pub-id-type="medline">40694861</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lavender</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hoque</surname><given-names>E</given-names> </name><name name-style="western"><surname>Brophy</surname><given-names>P</given-names> </name><name name-style="western"><surname>Kautz</surname><given-names>H</given-names> </name></person-group><article-title>A patient-centered digital scribe for automatic medical documentation</article-title><source>JAMIA Open</source><year>2021</year><month>01</month><volume>4</volume><issue>1</issue><fpage>ooab003</fpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooab003</pub-id><pub-id pub-id-type="medline">34377960</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ng</surname><given-names>JJW</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Evaluating the performance of artificial intelligence-based speech recognition for clinical documentation: a systematic review</article-title><source>BMC Med Inform Decis Mak</source><year>2025</year><month>07</month><day>1</day><volume>25</volume><issue>1</issue><fpage>236</fpage><pub-id pub-id-type="doi">10.1186/s12911-025-03061-0</pub-id><pub-id pub-id-type="medline">40598136</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ehrler</surname><given-names>F</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>DTY</given-names> </name><name name-style="western"><surname>Ducloux</surname><given-names>P</given-names> </name><name name-style="western"><surname>Blondon</surname><given-names>K</given-names> </name></person-group><article-title>A mobile application to support bedside nurse documentation and care: a time and motion study</article-title><source>JAMIA Open</source><year>2021</year><month>07</month><volume>4</volume><issue>3</issue><fpage>ooab046</fpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooab046</pub-id><pub-id pub-id-type="medline">34345804</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mayer</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Edwards</surname><given-names>N</given-names> </name><name name-style="western"><surname>Bokhart</surname><given-names>G</given-names> </name></person-group><article-title>A comparison of voice recognition program and traditional keyboard charting for nurse documentation</article-title><source>Comput Inform Nurs</source><year>2021</year><month>08</month><day>4</day><volume>40</volume><issue>2</issue><fpage>90</fpage><lpage>94</lpage><pub-id pub-id-type="doi">10.1097/CIN.0000000000000793</pub-id><pub-id pub-id-type="medline">34347642</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mairittha</surname><given-names>T</given-names> </name><name name-style="western"><surname>Mairittha</surname><given-names>N</given-names> </name><name name-style="western"><surname>Inoue</surname><given-names>S</given-names> </name></person-group><article-title>Evaluating a spoken dialogue system for recording systems of nursing care</article-title><source>Sensors (Basel)</source><year>2019</year><month>01</month><volume>19</volume><issue>17</issue><fpage>3736</fpage><pub-id pub-id-type="doi">10.3390/s19173736</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Ferizaj</surname><given-names>D</given-names> </name><name name-style="western"><surname>Neumann</surname><given-names>S</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Kurosu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hashizume</surname><given-names>A</given-names> </name></person-group><article-title>Assessing perceptions and experiences of an AI-driven speech assistant for nursing documentation: a qualitative study in German nursing homes</article-title><source>Human-Computer Interaction, Part V</source><year>2024</year><publisher-name>Springer</publisher-name><fpage>17</fpage><lpage>34</lpage><pub-id pub-id-type="doi">10.1007/978-3-031-60449-2_2</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="web"><article-title>PYSA &#x2013; Miteinander durch Innovation</article-title><source>Bundesministerium f&#x00FC;r Forschung, Technologie und Raumfahrt</source><access-date>2026-3-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.interaktive-technologien.de/projekte/pysa">https://www.interaktive-technologien.de/projekte/pysa</ext-link></comment></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>von Elm</surname><given-names>E</given-names> </name><name name-style="western"><surname>Altman</surname><given-names>DG</given-names> </name><name name-style="western"><surname>Egger</surname><given-names>M</given-names> </name><name name-style="western"><surname>Pocock</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>G&#x00F8;tzsche</surname><given-names>PC</given-names> </name><name name-style="western"><surname>Vandenbroucke</surname><given-names>JP</given-names> </name></person-group><article-title>The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies</article-title><source>J Clin Epidemiol</source><year>2008</year><month>04</month><volume>61</volume><issue>4</issue><fpage>344</fpage><lpage>349</lpage><pub-id pub-id-type="doi">10.1016/j.jclinepi.2007.11.008</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Champely</surname><given-names>S</given-names> </name><name name-style="western"><surname>Ekstrom</surname><given-names>C</given-names> </name><name name-style="western"><surname>Gill</surname><given-names>J</given-names> </name><etal/></person-group><article-title>pwr: basic functions for power analysis</article-title><source>GitHub</source><year>2020</year><access-date>2026-03-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/heliosdrm/pwr">https://github.com/heliosdrm/pwr</ext-link></comment></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fairbrother</surname><given-names>G</given-names> </name><name name-style="western"><surname>Jones</surname><given-names>A</given-names> </name><name name-style="western"><surname>Rivas</surname><given-names>K</given-names> </name></person-group><article-title>Development and validation of the Nursing Workplace Satisfaction Questionnaire (NWSQ)</article-title><source>Contemp Nurse</source><year>2009</year><volume>34</volume><issue>1</issue><fpage>10</fpage><lpage>18</lpage><pub-id pub-id-type="doi">10.5172/conu.2009.34.1.010</pub-id><pub-id pub-id-type="medline">20230168</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bangor</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kortum</surname><given-names>PT</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>JT</given-names> </name></person-group><article-title>An empirical evaluation of the System Usability Scale</article-title><source>Int J Hum-Comput Interact</source><year>2008</year><month>07</month><day>29</day><volume>24</volume><issue>6</issue><fpage>574</fpage><lpage>594</lpage><pub-id pub-id-type="doi">10.1080/10447310802205776</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Franke</surname><given-names>RH</given-names> </name><name name-style="western"><surname>Kaul</surname><given-names>JD</given-names> </name></person-group><article-title>The Hawthorne experiments: first statistical interpretation</article-title><source>Am Sociol Rev</source><year>1978</year><month>10</month><volume>43</volume><issue>5</issue><fpage>623</fpage><pub-id pub-id-type="doi">10.2307/2094540</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pedregosa</surname><given-names>F</given-names> </name><name name-style="western"><surname>Varoquaux</surname><given-names>G</given-names> </name><name name-style="western"><surname>Gramfort</surname><given-names>A</given-names> </name><etal/></person-group><article-title>scikit-learn: Machine learning in Python</article-title><source>J Mach Learn Res</source><year>2011</year><volume>12</volume><issue>85</issue><fpage>2825</fpage><lpage>2830</lpage><pub-id pub-id-type="doi">10.5555/1953048.2078195</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="web"><article-title>miceforest: Multiple imputation by chained equations with LightGBM</article-title><source>GitHub</source><access-date>2026-03-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/AnotherSamWilson/miceforest">https://github.com/AnotherSamWilson/miceforest</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Harris</surname><given-names>CR</given-names> </name><name name-style="western"><surname>Millman</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>van der Walt</surname><given-names>SJ</given-names> </name><etal/></person-group><article-title>Array programming with NumPy</article-title><source>Nature</source><year>2020</year><volume>585</volume><fpage>357</fpage><lpage>362</lpage><pub-id pub-id-type="doi">10.1038/s41586-020-2649-2</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McKinney</surname><given-names>W</given-names> </name></person-group><article-title>Data structures for statistical computing in Python</article-title><source>SciPy</source><year>2010</year><pub-id pub-id-type="doi">10.25080/Majora-92bf1922-00a</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Jones</surname><given-names>E</given-names> </name><name name-style="western"><surname>Oliphant</surname><given-names>T</given-names> </name><name name-style="western"><surname>Peterson</surname><given-names>P</given-names> </name></person-group><source>SciPy: open source scientific tools for Python</source><year>2001</year><access-date>2026-03-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://www.scipy.org">http://www.scipy.org</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hunter</surname><given-names>JD</given-names> </name></person-group><article-title>Matplotlib: A 2D graphics environment</article-title><source>Comput Sci Eng</source><year>2007</year><month>05</month><volume>9</volume><issue>3</issue><fpage>90</fpage><lpage>95</lpage><pub-id pub-id-type="doi">10.1109/MCSE.2007.55</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wickham</surname><given-names>H</given-names> </name><name name-style="western"><surname>Averick</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bryan</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Welcome to the tidyverse</article-title><source>J Open Source Softw</source><year>2019</year><volume>4</volume><issue>43</issue><fpage>1686</fpage><pub-id pub-id-type="doi">10.21105/joss.01686</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bates</surname><given-names>D</given-names> </name><name name-style="western"><surname>Maechler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bolker</surname><given-names>B</given-names> </name><name name-style="western"><surname>Walker</surname><given-names>S</given-names> </name></person-group><article-title>Fitting linear mixed-effects models using lme4</article-title><source>J Stat Softw</source><year>2015</year><volume>67</volume><issue>1</issue><fpage>2</fpage><lpage>48</lpage><pub-id pub-id-type="doi">10.18637/jss.v067.i01</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kuznetsova</surname><given-names>A</given-names> </name><name name-style="western"><surname>Brockhoff</surname><given-names>PB</given-names> </name><name name-style="western"><surname>Christensen</surname><given-names>RHB</given-names> </name></person-group><article-title>lmerTest package: tests in linear mixed effects models</article-title><source>J Stat Soft</source><year>2017</year><volume>82</volume><issue>13</issue><fpage>1</fpage><lpage>26</lpage><pub-id pub-id-type="doi">10.18637/jss.v082.i13</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Saxena</surname><given-names>K</given-names> </name><name name-style="western"><surname>Diamond</surname><given-names>R</given-names> </name><name name-style="western"><surname>Conant</surname><given-names>RF</given-names> </name><name name-style="western"><surname>Mitchell</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Gallopyn</surname><given-names>IG</given-names> </name><name name-style="western"><surname>Yakimow</surname><given-names>KE</given-names> </name></person-group><article-title>Provider adoption of speech recognition and its impact on satisfaction, documentation quality, efficiency, and cost in an inpatient EHR</article-title><source>AMIA Jt Summits Transl Sci Proc</source><year>2018</year><volume>2017</volume><fpage>186</fpage><lpage>195</lpage><pub-id pub-id-type="medline">29888069</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yen</surname><given-names>PY</given-names> </name><name name-style="western"><surname>Kelley</surname><given-names>M</given-names> </name><name name-style="western"><surname>Lopetegui</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Understanding and visualizing multitasking and task switching activities: a time motion study to capture nursing workflow</article-title><source>AMIA Annu Symp Proc</source><year>2016</year><volume>2016</volume><fpage>1264</fpage><lpage>1273</lpage><pub-id pub-id-type="medline">28269924</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Van Der Laan</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Heino</surname><given-names>A</given-names> </name><name name-style="western"><surname>De Waard</surname><given-names>D</given-names> </name></person-group><article-title>A simple procedure for the assessment of acceptance of advanced transport telematics</article-title><source>Transp Res Part C Emerg Technol</source><year>1997</year><month>02</month><volume>5</volume><issue>1</issue><fpage>1</fpage><lpage>10</lpage><pub-id pub-id-type="doi">10.1016/S0968-090X(96)00025-3</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Definition of task categories for observations at <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub>.</p><media xlink:href="jmir_v28i1e86078_app1.docx" xlink:title="DOCX File, 1766 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Self-developed questionnaire for assessment at <italic>t</italic><sub>0</sub> and <italic>t</italic><sub>1</sub>.</p><media xlink:href="jmir_v28i1e86078_app2.docx" xlink:title="DOCX File, 1768 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Statistical analysis rationale and detailed methods.</p><media xlink:href="jmir_v28i1e86078_app3.docx" xlink:title="DOCX File, 1771 KB"/></supplementary-material><supplementary-material id="app4"><label>Multimedia Appendix 4</label><p>Bayesian supplementary modeling and analysis.</p><media xlink:href="jmir_v28i1e86078_app4.docx" xlink:title="DOCX File, 2868 KB"/></supplementary-material><supplementary-material id="app5"><label>Multimedia Appendix 5</label><p>Results of exploratory sensitivity analyses.</p><media xlink:href="jmir_v28i1e86078_app5.docx" xlink:title="DOCX File, 2317 KB"/></supplementary-material></app-group></back></article>