<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v22i6e14827</article-id>
      <article-id pub-id-type="pmid">32442129</article-id>
      <article-id pub-id-type="doi">10.2196/14827</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Effect of Speech Recognition on Problem Solving and Recall in Consumer Digital Health Tasks: Controlled Laboratory Experiment</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Lee</surname>
            <given-names>Mikyoung</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zuccon</surname>
            <given-names>Guido</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kelly</surname>
            <given-names>Liadh</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Jessica</given-names>
          </name>
          <degrees>BEng, MBiomedEng</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Centre for Health Informatics</institution>
            <institution>Australian Institute of Health Innovation</institution>
            <institution>Macquarie University</institution>
            <addr-line>Level 6</addr-line>
            <addr-line>75 Talavera Rd</addr-line>
            <addr-line>North Ryde, NSW 2113</addr-line>
            <country>Australia</country>
            <phone>61 2 9850 2400</phone>
            <email>jessica.a.chen@mq.edu.au</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1798-905X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Lyell</surname>
            <given-names>David</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2695-0368</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Laranjo</surname>
            <given-names>Liliana</given-names>
          </name>
          <degrees>MD, MPH, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1020-3402</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Magrabi</surname>
            <given-names>Farah</given-names>
          </name>
          <degrees>BEng, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-8426-5588</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Centre for Health Informatics</institution>
        <institution>Australian Institute of Health Innovation</institution>
        <institution>Macquarie University</institution>
        <addr-line>North Ryde</addr-line>
        <country>Australia</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Jessica Chen <email>jessica.a.chen@mq.edu.au</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>6</month>
        <year>2020</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>1</day>
        <month>6</month>
        <year>2020</year>
      </pub-date>
      <volume>22</volume>
      <issue>6</issue>
      <elocation-id>e14827</elocation-id>
      <history>
        <date date-type="received">
          <day>28</day>
          <month>5</month>
          <year>2019</year>
        </date>
        <date date-type="rev-request">
          <day>3</day>
          <month>10</month>
          <year>2019</year>
        </date>
        <date date-type="rev-recd">
          <day>27</day>
          <month>11</month>
          <year>2019</year>
        </date>
        <date date-type="accepted">
          <day>29</day>
          <month>3</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Jessica Chen, David Lyell, Liliana Laranjo, Farah Magrabi. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 01.06.2020.</copyright-statement>
      <copyright-year>2020</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2020/6/e14827" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Recent advances in natural language processing and artificial intelligence have led to widespread adoption of speech recognition technologies. In consumer health applications, speech recognition is usually applied to support interactions with conversational agents for data collection, decision support, and patient monitoring. However, little is known about the use of speech recognition in consumer health applications and few studies have evaluated the efficacy of conversational agents in the hands of consumers. In other consumer-facing tools, cognitive load has been observed to be an important factor affecting the use of speech recognition technologies in tasks involving problem solving and recall. Users find it more difficult to think and speak at the same time when compared to typing, pointing, and clicking. However, the effects of speech recognition on cognitive load when performing health tasks has not yet been explored.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The aim of this study was to evaluate the use of speech recognition for documentation in consumer digital health tasks involving problem solving and recall.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>Fifty university staff and students were recruited to undertake four documentation tasks with a simulated conversational agent in a computer laboratory. The tasks varied in complexity determined by the amount of problem solving and recall required (simple and complex) and the input modality (speech recognition vs keyboard and mouse). Cognitive load, task completion time, error rate, and usability were measured.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Compared to using a keyboard and mouse, speech recognition significantly increased the cognitive load for complex tasks (<italic>Z</italic>=–4.08, <italic>P</italic>&#60;.001) and simple tasks (<italic>Z</italic>=–2.24, <italic>P</italic>=.03). Complex tasks took significantly longer to complete (<italic>Z</italic>=–2.52, <italic>P</italic>=.01) and speech recognition was found to be overall less usable than a keyboard and mouse (<italic>Z</italic>=–3.30, <italic>P</italic>=.001). However, there was no effect on errors.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Use of a keyboard and mouse was preferable to speech recognition for complex tasks involving problem solving and recall. Further studies using a broader variety of consumer digital health tasks of varying complexity are needed to investigate the contexts in which use of speech recognition is most appropriate. The effects of cognitive load on task performance and its significance also need to be investigated.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>speech recognition software</kwd>
        <kwd>consumer health informatics</kwd>
        <kwd>ergonomics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Recent advances in natural language processing and artificial intelligence have led to improvements in and widespread adoption of speech recognition technologies [<xref ref-type="bibr" rid="ref1">1</xref>]. Speech recognition is an input modality that translates human speech into computerized text [<xref ref-type="bibr" rid="ref2">2</xref>]. In consumer applications, speech recognition is usually applied as a way to interact with conversational agents, which are systems that mimic human conversation using text or spoken language [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. Consumer conversational agents (such as Amazon Alexa and Google Assistant) can improve patient workflow by allowing patients to call nurses [<xref ref-type="bibr" rid="ref5">5</xref>]. In health care, conversational agents have been utilized for a variety of purposes, including data collection, decision support, and patient monitoring [<xref ref-type="bibr" rid="ref5">5</xref>-<xref ref-type="bibr" rid="ref8">8</xref>]. </p>
      <p>Problems with the use of digital health technology represent a well-documented safety concern in the literature [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>]. However, little is known about the problems associated with conversational agents that pose actual or potential risks of harm to consumers [<xref ref-type="bibr" rid="ref9">9</xref>]. One study in which the participants were asked to interact with conversational agents identified significant safety concerns arising from the quality of information provided in response to health-related questions [<xref ref-type="bibr" rid="ref11">11</xref>]. For example, incorrect information provided by a conversational agent in response to a question about the amount of alcohol that could be consumed while taking oxycodone could lead to severe harm, including death. However, few studies have evaluated conversational agents in the hands of consumers [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>].</p>
      <p>In other consumer-facing tools, cognitive load has been observed to be an important factor affecting the use of speech recognition technologies. Users reported finding it more difficult to think and speak at the same time when compared to typing, pointing, and clicking [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. However, the effects of speech recognition use on cognitive load when performing health tasks has not yet been explored. Cognitive load is the amount of workload imposed on the brain’s working or short-term memory, which has limited capacity [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>] and a short duration [<xref ref-type="bibr" rid="ref16">16</xref>], particularly when performing tasks requiring problem solving and recall [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. An example of a problem-solving task is using basic arithmetic to calculate nutritional information, whereas recall involves memorizing and reporting exercise information. Due to the cognitive load, certain tasks may be more difficult to perform because speaking shares the same cognitive resources in the brain as those required for problem solving and recall (ie, working memory) [<xref ref-type="bibr" rid="ref12">12</xref>]. Therefore, a possible challenge with the use of speech recognition is that it can increase the cognitive load in tasks requiring more problem solving and recall.</p>
      <p>Although studies in other domains have investigated the effects of speech recognition use on cognitive load, to our knowledge, no study has measured its effects in digital health tasks [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. Thus, the aim of the present study was to evaluate the use of speech recognition for documentation in consumer digital health tasks such as recording diet and exercise information in comparison to using a conventional keyboard and mouse. The following hypotheses were tested: (1) cognitive load is higher for speech recognition compared to keyboard and mouse use in complex tasks requiring more problem solving and recall; (2) the percentage of errors is higher for speech recognition compared with keyboard and mouse use in complex tasks requiring more problem solving and recall; (3) task completion time is lower for speech recognition compared with keyboard and mouse use in simple tasks requiring less problem solving and recall; and (4) speech recognition is less usable than a keyboard and mouse for both simple and complex tasks requiring more problem solving and recall.</p>
      <p/>
      <p>These findings will shed light on the characteristics of consumer digital health tasks that make them most suitable for using speech recognition as an input modality.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Participants</title>
        <p>Fifty-two university students and staff participated in this study (see Results for a summary of the participant demographics). The participants were either students or staff who met the minimum English language proficiency for admission to a university program or workplace. These participants also had working knowledge of computer technology and systems as required for their degree or profession. Hence, there were no requirements regarding level of English language, health literacy, or technology proficiency for inclusion. The participants responded to advertisements sent by email or published in a university newsletter. Consenting adults aged 18 years or older were eligible to participate. Ethical approval was granted by Macquarie University’s Human Research Ethics Committee to recruit people within the Faculty of Medicine and Health Sciences. Participants were not offered any remuneration or gifts to incentivize participation.</p>
      </sec>
      <sec>
        <title>Experimental Design and Tasks</title>
        <p>The study included two within-subject factors: human-computer interaction modality (speech recognition vs keyboard and mouse) and task complexity (simple vs complex) providing four experimental conditions (<xref rid="figure1" ref-type="fig">Figure 1</xref>).</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Experimental design.</p>
          </caption>
          <graphic xlink:href="jmir_v22i6e14827_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>Each participant was asked to complete four consumer digital health tasks to document nutritional and exercise information: two simple and two complex tasks, using speech recognition and a keyboard and mouse to interact with a simulated conversational agent (<xref rid="figure1" ref-type="fig">Figure 1</xref>). For each documentation task, participants were asked to adopt a persona within a hypothetical scenario focusing on physical activity and diet, and to answer the conversational agent’s questions (see <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The tasks were designed to assess problem solving and recall using the information provided within the scenario.</p>
        <p>Complexity was measured by the number of information items that participants needed to manipulate in working memory. The relationship between the number of information items requiring manipulation in working memory and cognitive load has been well established in the literature [<xref ref-type="bibr" rid="ref20">20</xref>]; that is, human performance is affected when the cognitive load is too high or exceeds the limits of working memory [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. For simple tasks, the conversational agent displayed the hypothetical scenario on the same screen as the questions. Complex tasks were designed to impose a higher cognitive load by increasing the number of pieces of information needed to perform arithmetic and commit to memory (6 total items for simple tasks vs 17 total items for complex tasks; see <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). Simple tasks required participants to problem solve using 5 items of information, including 5 days, 20 minutes, line dancing, 10:00-10.50 am, whereas complex tasks required participants to recall the same 5 items and were provided with 12 items for problem solving. A higher cognitive load was also imposed by displaying the scenario in a pop-up modal window, requiring participants to commit key information in the scenario to working memory when transferring between the windows [<xref ref-type="bibr" rid="ref23">23</xref>]. Participants could not copy and paste their answers.</p>
        <p>The tasks were developed in consultation with a health informatics researcher (DL) and a primary care physician (LL). Pilot testing was performed by asking 7 individuals to complete the tasks using a prototype of the system and provide feedback. Any issues were iteratively fixed before the next pilot test. This pilot testing ensured that the prescribed tasks and the system were understandable and functional. Individuals who assisted with pilot testing were excluded from participating in the experiment. The correct answers to tasks were predetermined and validated by the health informatics researcher.</p>
      </sec>
      <sec>
        <title>Simulated Conversational Agent</title>
        <p>Participants attended a computer laboratory at the university where a workstation was set up with the tasks running on a web app connected to a keyboard, mouse, and microphone. Presented as a text message conversation, user responses could either be typed using the keyboard and mouse or transcribed using speech recognition depending on the experimental condition (<xref rid="figure2" ref-type="fig">Figure 2</xref>). Participants were able to view their responses before submitting. Web Speech API, a JavaScript-based general purpose speech recognition application programming interface, was used to implement a live audio transcription function for this system [<xref ref-type="bibr" rid="ref24">24</xref>]. To activate speech-to-text using speech recognition, the participants were instructed to press the “Record” button. The system was launched using Google Chrome on a local machine.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Graphical user interface of the simulated conversational agent on Google Chrome.</p>
          </caption>
          <graphic xlink:href="jmir_v22i6e14827_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Procedure</title>
        <p>After obtaining informed consent, the participants completed a short survey about use of speech recognition technology and provided demographic information. They were then briefed that the tasks consisted of problem solving and recall and given instructions about how to use the speech recognition interface before commencing the experiment. Training continued until the participant clearly understood all aspects required to perform the experimental tasks; no practice tasks were undertaken. Participants completed two simple and two complex tasks, half of which were randomly assigned to using speech recognition (<xref rid="figure3" ref-type="fig">Figure 3</xref>). At the end of each task, cognitive load was assessed. To avoid order effects, the assignment of simple and complex tasks for the two modalities and task sequence were randomized. All voice recordings for the speech recognition tasks were captured independently to determine errors. At the end of all four tasks, the System Usability Scale (SUS) questionnaire [<xref ref-type="bibr" rid="ref25">25</xref>] was completed for each modality followed by a feedback interview.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Experimental flow diagram.</p>
          </caption>
          <graphic xlink:href="jmir_v22i6e14827_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Outcome Measures and Analysis</title>
        <p>Participants’ responses to the lifestyle management tasks with speech recognition and keyboard and mouse were compared using data extracted from the computer log, audio recordings of participant responses, and paper surveys.</p>
        <p>Cognitive load was measured using a cognitive load inventory that was adapted from a validated instrument to reflect the nature of the tasks in the present study [<xref ref-type="bibr" rid="ref26">26</xref>] (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>). Self-ratings of using this inventory have shown it to be reliable, unobtrusive, and sensitive to small differences [<xref ref-type="bibr" rid="ref23">23</xref>]. This inventory has been widely used [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>], including in controlled studies of clinical decision making [<xref ref-type="bibr" rid="ref28">28</xref>]. The inventory was administered on paper at the end of each condition.</p>
        <p>Task completion time was measured in seconds and calculated from computer logs as the difference between the task start and end time.</p>
        <p>The error rate for a given task was defined as the number of failed recall and problem-solving responses, calculated as a percentage of the maximum score for the task. Participants were asked to answer each of the conversational agent’s questions correctly by recalling information or solving for numerical answers. Each task had predefined correct answers along with numbers associated with correct problem-solving and recall responses (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). For example, if the correct answers to a task were “line dancing” and “50 minutes,” then the maximum score was 2. If a participant incorrectly recalled “line dancing” but correctly answered “50 minutes,” then the error rate would be calculated as 50%. To ensure that transcription errors were not mistakenly attributed to participants, the voice recording on each response was used to validate the answers for tasks that were completed with speech recognition.</p>
        <p>Usability was assessed using the SUS, a validated 10-item questionnaire [<xref ref-type="bibr" rid="ref25">25</xref>], resulting in overall usability scores and scores of subelements (usability and learnability).</p>
      </sec>
      <sec>
        <title>Statistical Analyses</title>
        <p>The effects of speech recognition on cognitive load, task completion time, error rate, and SUS scores were tested using repeated-measures analysis to control for individual differences. The Wilcoxon signed-rank test was used because the results were not normally distributed. We compared speech recognition and keyboard and mouse across both levels of complexity. Cohen criteria were used to calculate and interpret effect sizes (<italic>r</italic>), where 0.1 indicates a small effect, 0.3 indicates a medium effect, and 0.5 indicates a large effect [<xref ref-type="bibr" rid="ref29">29</xref>]. We estimated that a sample size of 42 was required to detect a difference of 25% and a standard deviation of 20% for each outcome measure for all tasks with 90% power and <italic>P</italic>&#60;.05 [<xref ref-type="bibr" rid="ref30">30</xref>]. No baseline measures could be derived from the literature. Descriptive statistics were used to summarize demographic information. All statistical analyses were undertaken using SPSS v24.0.0.0 software (IBM Corp, Armonk, NY, USA).</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Participants</title>
        <p>Fifty-two university staff and students participated in the experiment. One participant did not complete the experiment and data from another was lost due to a technical error, leaving a total of 50 participants for inclusion in the analysis. The participants were aged 18-64 years, 30% (15/50) were 25-34 years and 54% (27/50) were women. More than half of the participants (56%, 28/50) reported never using speech recognition in their daily lives, 28% (14/50) reported using it once or twice a week, 14% (7/50) reported using it more than multiple times a week, and 4% (2/50) used speech recognition multiple times a day.</p>
      </sec>
      <sec>
        <title>Effects of Speech Recognition on Cognitive Load</title>
        <p>Participants experienced a significantly higher cognitive load when using speech recognition to perform the prescribed tasks (<xref ref-type="table" rid="table1">Table 1</xref>). These findings were consistent across both levels of task complexity, although the effect size increased from medium for simple tasks to high for complex tasks.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Comparison of cognitive load between keyboard and mouse and speech recognition use by task complexity.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="220"/>
            <col width="250"/>
            <col width="240"/>
            <col width="90"/>
            <col width="90"/>
            <col width="110"/>
            <thead>
              <tr valign="bottom">
                <td>Task complexity</td>
                <td>Keyboard and mouse, median (IQR)</td>
                <td>Speech recognition, median (IQR)</td>
                <td><italic>Z</italic> value<sup>a</sup></td>
                <td><italic>P</italic> value</td>
                <td>Effect size (<italic>r</italic>)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Complex</td>
                <td>3.2 (2.1-4.2)</td>
                <td>5.2 (3.6-6.3)</td>
                <td>–4.22</td>
                <td>&#60;.001</td>
                <td>0.60</td>
              </tr>
              <tr valign="top">
                <td>Simple</td>
                <td>2.2 (2.2-4.4)</td>
                <td>3.3 (2.2-5.6)</td>
                <td>–2.24</td>
                <td>.02</td>
                <td>0.32</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>Wicoxon signed-rank test.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Task Completion Time and Error Rate</title>
        <p>Participants took significantly longer to complete tasks using speech recognition than the keyboard and mouse for complex tasks; however, there was no difference observed for simple tasks (<xref ref-type="table" rid="table2">Table 2</xref>). For complex tasks, there was a statistically significant increase in task completion time with a medium effect size (<italic>r</italic>=0.36); however, there was no difference for simple tasks.</p>
        <p>There was no difference in error rates for both simple and complex tasks (<xref ref-type="table" rid="table2">Table 2</xref>). For complex tasks, we examined error rates by their type and found no difference for both problem solving (<italic>Z</italic>=–1.96, <italic>P</italic>=.05) and recall error rates (<italic>Z</italic>=–1.55, <italic>P</italic>=.12). </p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Comparison of task completion time and error rates between keyboard and mouse and speech recognition use by task complexity.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="340"/>
            <col width="210"/>
            <col width="200"/>
            <col width="110"/>
            <col width="110"/>
            <thead>
              <tr valign="bottom">
                <td colspan="2">Task complexity</td>
                <td>Keyboard and mouse</td>
                <td>Speech recognition</td>
                <td><italic>Z</italic> value<sup>a</sup></td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  <bold>Complex, median (IQR)</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Completion time (seconds)</td>
                <td>162 (124-192)</td>
                <td>173 (136-223)</td>
                <td>2.52</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Error rate (%)</td>
                <td>11 (0-25)</td>
                <td>11 (0-36)</td>
                <td>1.80</td>
                <td>.07</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>Simple, median (IQR)</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Completion time (seconds)</td>
                <td>90 (74-124)</td>
                <td>83 (68-11)</td>
                <td>7.90</td>
                <td>.43</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Error rate (%)</td>
                <td>0 (0-33)</td>
                <td>0 (0-33)</td>
                <td>–0.33</td>
                <td>.74</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>Wilcoxon signed-rank test.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Usability and Participant Perceptions About Speech Recognition</title>
        <p>Participants found speech recognition to be significantly less usable than the keyboard and mouse. This was consistent with the SUS factor analysis, which revealed that speech recognition was perceived to be significantly less usable and harder to learn with medium and large effect sizes, respectively (<xref ref-type="table" rid="table3">Table 3</xref>).</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Usability (SUSa scores) of speech recognition compared to keyboard and mouse.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="220"/>
            <col width="250"/>
            <col width="240"/>
            <col width="90"/>
            <col width="90"/>
            <col width="110"/>
            <thead>
              <tr valign="bottom">
                <td>SUS category</td>
                <td>Keyboard and mouse, median (IQR)</td>
                <td>Speech recognition, median (IQR)</td>
                <td><italic>Z</italic> value<sup>b</sup></td>
                <td><italic>P</italic> value</td>
                <td>Effect size (<italic>r</italic>)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Overall</td>
                <td>85 (72-90)</td>
                <td>75 (62-85)</td>
                <td>–3.30</td>
                <td>.001</td>
                <td>0.47</td>
              </tr>
              <tr valign="top">
                <td>Usability</td>
                <td>100 (97-100)</td>
                <td>100 (75-100)</td>
                <td>–2.98</td>
                <td>.003</td>
                <td>0.42</td>
              </tr>
              <tr valign="top">
                <td>Learnability</td>
                <td>81 (68-88)</td>
                <td>72 (59-84)</td>
                <td>–3.54</td>
                <td>&#60;.001</td>
                <td>0.50</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>SUS: System Usability Scale.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>Wilcoxon signed-rank test.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>Overall, participants commented that the simulated conversational agent was “very straightforward” to use. Some reported that not having to think about spelling and grammar was an advantage of speech recognition. Although the participants found speech recognition to be a possibly more convenient way to enter free-text information, many observed that the major pitfalls of speech recognition were any transcription errors generated by the software, an inability to retract and edit answers once a sentence was uttered, and unwanted filler utterances such as “um” and “er” appearing in their responses. Participants also commented about the extra time taken to check and correct the output of speech recognition for transcribing errors when such errors could be prevented by using a keyboard and mouse in the first place. Others raised privacy issues as they were self-conscious about strangers eavesdropping on their responses.</p>
        <p>When commenting on difficulties, some participants stated that they had struggled to formulate answers for scenarios and construct sentences to dictate to the conversational agent via speech recognition at the same time. They also commented that this action consumed more time to complete the task. By contrast, typing was perceived to be easier because it was “what we do every day.” This also allowed for using the textbox input to record answers as opposed to “keeping more in the brain” when storing information in cognitive, working memory. Speech was also perceived as a novelty because many were familiar with using a keyboard and mouse and saw it as a standard mode for human-computer interaction.</p>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Main Findings</title>
        <p>When using speech recognition, participants reported a higher cognitive load for both simple and complex tasks compared to using a keyboard and mouse. Some participants reported an inability to think while speaking, which is consistent with previous observations [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. A possible reason is due to an extra step introduced in processing the information and then having to form complete sentences before speaking. This extra processing step, which occurs once participants begin to formulate responses to the simulated conversational agent’s questions, may have involved working memory and contributed to participants experiencing a higher cognitive load while using speech recognition.</p>
        <p>The higher cognitive load also provides a potential explanation for the more time required to complete complex tasks when using speech recognition and for it to be perceived less usable. Participants also noted a time delay required for speech recognition to process responses. It is thus possible that the extra time was introduced by the system and the time participants spent in checking the live transcription provided by speech recognition. If this was true, a consistent increase in time across all conditions would be expected. However, there was no significant time difference found for simple tasks. This could mean that the difference more likely arose from the task characteristics themselves. Further experiments are required to examine the interaction effect of modality and task complexity on cognitive load.</p>
        <p>Despite previous reports about higher error rates with speech recognition for clinical documentation tasks [<xref ref-type="bibr" rid="ref31">31</xref>], we found no difference in error rates between the two modalities for both simple and complex tasks. One possible reason that the second hypothesis was not supported is that most participants were observed to formulate their responses with caution by double-checking their answers when using speech recognition. For simple tasks, this meant looking at the scenario section that was on the same screen. For complex tasks, participants repeatedly clicked the “review scenario” button until they were satisfied with their answer.</p>
        <p>Speech recognition was perceived to be less usable than a keyboard and mouse because it was harder to learn. A major factor contributing to this effect is that many participants reported being more comfortable with typing because it was an everyday human-computer interaction, especially in their profession. In addition, 56% of the participants did not use speech recognition in their daily lives, which possibly meant that they are less proficient in using speech recognition. Another possible factor is that the conversational agent was purposely designed to limit participants by not allowing users to correct transcription errors. This may be a common source of frustration that affects perceptions of usability for conversational agents. However, the error rate was unaffected because we used independent voice recordings to determine errors and the impact on usability was captured in the overall score.</p>
        <p>Some participants reported lack of privacy as another major factor for speech recognition being less usable than a keyboard and mouse. A major advantage with typing was that a third party could not eavesdrop on interactions with the conversational agent. Thus, less confidence was placed on using the conversational agent with speech recognition compared with a keyboard and mouse. Although there was no difference in error rates, the increased cognitive load and task completion time may also have affected user experience.</p>
      </sec>
      <sec>
        <title>Implications</title>
        <p>Our findings suggest that speech recognition may not be uniformly suitable for the different contexts of health care. An important implication is for system designers to consider task characteristics and the resulting impact on cognitive load when selecting modalities for humans to interact with computers. In general, lower cognitive load, fewer errors, less time, and better usability are desirable. Speech recognition may be more suitable for frequent tasks such as documentation of notes in electronic health records (EHRs), which may generally not involve problem solving and recall (eg, when a doctor has already established a diagnosis during the patient consultation and uses the EHR to record their notes). However, it may not be suitable for occasional tasks such as incident reporting, which involves problem solving and recall, requiring clinicians to recollect the sequence of events and identify problems that led to an incident [<xref ref-type="bibr" rid="ref32">32</xref>]. For such complex documentation tasks, the higher cognitive load and greater time required to use speech recognition, along with the lower usability, suggest that a keyboard and mouse may be a better input modality. By contrast, speech recognition may be necessary for contexts that require use of the hands and eyes.</p>
        <p>One practical strategy for designers to assess the suitability of speech recognition as an input modality is to test user interfaces in the prototype stage with the cognitive load inventory, which is readily applicable to different modalities and systems. Importantly, designers need to consider privacy requirements when using speech recognition for busy health care settings, especially when sensitive health information is being handled. Our findings also suggest that privacy requirements may present a barrier for the use of speech recognition in some contexts such as health apps that require users to document their personal health information. These considerations will need to be evaluated on a case-by-case basis.</p>
      </sec>
      <sec>
        <title>Comparison With the Literature</title>
        <p>To the best of our knowledge, no previous studies have measured the effects of speech recognition in consumer digital health tasks. In clinical applications, the use of speech recognition for clinical documentation was found to increase error rates and task completion time in a controlled laboratory setting [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. One possible reason for the disparity with our results may be the difference in the source of complexity. In the previous studies, complexity was distributed between the clinical scenario and the user interactions with an EHR, which required users to navigate to different sections of the record. By contrast, in our study, the type of user interaction remained the same for the different task types. The complexity instead arose from the scenario itself, which required users to problem solve and recall information from memory.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>There are several limitations to the design of the current study. We focused on the use of speech recognition as an input modality within a computer laboratory, which may not be representative of a real-world setting where environmental factors such as background noise and interruptions affect consumer interactions with digital health technologies. A general-purpose speech recognition engine that was not specifically optimized for the consumer health domain was tested on a desktop computer. This may have affected participant perceptions about usability and time required to use speech recognition. Participants were university students and staff, almost half of whom reported not using speech recognition technology in their daily lives. Therefore, our sample may not be representative of the general population of health consumers who might use conversational agents. However, because the participants were from a cohort that regularly used a keyboard and mouse, we were able to undertake a realistic assessment of the effort to learn and use speech recognition in the context of the conversational agent. For regular speech recognition users, expectations about the robustness and accuracy could have affected SUS scores. Although it is possible that individual differences such as health literacy, native language, pronunciation, fluency, and experience with speech recognition and a keyboard and mouse may have impacted the outcome variables, we attempted to control for these differences by using a within-subjects design. Further studies are needed to explore the influence of health and nutritional literacy. It is also possible that the quality of the speech recognition could have affected task completion time, but the effect would be consistent across experimental conditions. The error rate was unaffected because we used independent voice recordings to determine correct answers; therefore, the risk of the speech recognition mistranscribing speech by recording responses was controlled.</p>
        <p>Despite these limitations, this study has contributed evidence related to the use of speech recognition as an input modality in human-computer interaction, particularly in a consumer digital health context. These results provide baseline measures of cognitive load in using speech recognition. Further studies using a more representative population of conversational agent users are needed to investigate the effects of cognitive load on task performance when speech recognition is integrated with consumer digital health technologies in real-world settings, including mobile devices such as smartphones and tablet computers.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>This study found that using a keyboard and mouse was preferable to speech recognition for complex tasks involving problem solving and recall. This may be due to the higher cognitive load reported when using speech recognition and that the participants were more comfortable using a keyboard and mouse. Our results suggest that task characteristics need to be considered by designers when selecting the most appropriate input modality for human-computer interaction. Further studies using a broader variety of consumer digital health tasks of varying complexity are needed to investigate the contexts in which use of speech recognition is appropriate. The effects of cognitive load on task performance and its significance also need to be investigated.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Tasks.</p>
        <media xlink:href="jmir_v22i6e14827_app1.docx" xlink:title="DOCX File , 22 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Cognitive load inventory.</p>
        <media xlink:href="jmir_v22i6e14827_app2.docx" xlink:title="DOCX File , 16 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">EHR</term>
          <def>
            <p>electronic health records</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">SUS</term>
          <def>
            <p>System Usability Scale</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>We thank the participants who gave up their time to participate in this study. We thank Vitaliy Kim and Owen Sun for sharing their expertise to build the simulated conversational agent.</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>FM conceptualized the study design. JC conducted the research and primary analysis, and drafted the paper. LL helped to formulate the scenarios and DL helped to adapt the cognitive load inventory. JC is responsible for the integrity of the work as the guarantor of the study. All authors participated in writing and revising the paper. All aspects of the study (including design; collection, analysis, and interpretation of data; writing of the report; and decision to publish) were led by the authors.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Dahl</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Mohamed</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Jaitly</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Senior</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Vanhoucke</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sainath</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Kingsbury</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups</article-title>
          <source>IEEE Signal Process Mag</source>
          <year>2012</year>
          <month>11</month>
          <volume>29</volume>
          <issue>6</issue>
          <fpage>82</fpage>
          <lpage>97</lpage>
          <pub-id pub-id-type="doi">10.1109/MSP.2012.2205597</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hodgson</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the usability of speech recognition to create clinical documentation using a commercial electronic health record</article-title>
          <source>Int J Med Inform</source>
          <year>2018</year>
          <month>05</month>
          <volume>113</volume>
          <fpage>38</fpage>
          <lpage>42</lpage>
          <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2018.02.011</pub-id>
          <pub-id pub-id-type="medline">29602431</pub-id>
          <pub-id pub-id-type="pii">S1386-5056(18)30082-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kumah-Crystal</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Pirtle</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Whyte</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Goode</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Anders</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lehmann</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Electronic Health Record Interactions through Voice: A Review</article-title>
          <source>Appl Clin Inform</source>
          <year>2018</year>
          <month>07</month>
          <day>18</day>
          <volume>9</volume>
          <issue>3</issue>
          <fpage>541</fpage>
          <lpage>552</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/30040113"/>
          </comment>
          <pub-id pub-id-type="doi">10.1055/s-0038-1666844</pub-id>
          <pub-id pub-id-type="medline">30040113</pub-id>
          <pub-id pub-id-type="pmcid">PMC6051768</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Laranjo</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Dunn</surname>
              <given-names>AG</given-names>
            </name>
            <name name-style="western">
              <surname>Tong</surname>
              <given-names>HL</given-names>
            </name>
            <name name-style="western">
              <surname>Kocaballi</surname>
              <given-names>AB</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bashir</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Surian</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Gallego</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Lau</surname>
              <given-names>AYS</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Conversational agents in healthcare: a systematic review</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2018</year>
          <month>09</month>
          <day>01</day>
          <volume>25</volume>
          <issue>9</issue>
          <fpage>1248</fpage>
          <lpage>1258</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/30010941"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocy072</pub-id>
          <pub-id pub-id-type="medline">30010941</pub-id>
          <pub-id pub-id-type="pii">5052181</pub-id>
          <pub-id pub-id-type="pmcid">PMC6118869</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Black</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>McTear</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Black</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Harper</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Lemon</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Appraisal of a conversational artefact and its utility in remote patient monitoring</article-title>
          <year>2005</year>
          <conf-name>18th IEEE Symposium on Computer-Based Medical Systems</conf-name>
          <conf-date>June 23-24, 2005</conf-date>
          <conf-loc>Dublin</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cbms.2005.33</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Levin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Levin</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of spoken dialogue technology for real-time health data collection</article-title>
          <source>J Med Internet Res</source>
          <year>2006</year>
          <month>12</month>
          <day>11</day>
          <volume>8</volume>
          <issue>4</issue>
          <fpage>e30</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2006/4/e30/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/jmir.8.4.e30</pub-id>
          <pub-id pub-id-type="medline">17213048</pub-id>
          <pub-id pub-id-type="pii">v8i4e30</pub-id>
          <pub-id pub-id-type="pmcid">PMC1794008</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beveridge</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Fox</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Automatic generation of spoken dialogue from medical plans and ontologies</article-title>
          <source>J Biomed Inform</source>
          <year>2006</year>
          <month>10</month>
          <volume>39</volume>
          <issue>5</issue>
          <fpage>482</fpage>
          <lpage>499</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(06)00003-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2005.12.008</pub-id>
          <pub-id pub-id-type="medline">16495159</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(06)00003-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Miner</surname>
              <given-names>AS</given-names>
            </name>
            <name name-style="western">
              <surname>Milstein</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schueller</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hegde</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mangurian</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Linos</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Smartphone-Based Conversational Agents and Responses to Questions About Mental Health, Interpersonal Violence, and Physical Health</article-title>
          <source>JAMA Intern Med</source>
          <year>2016</year>
          <month>05</month>
          <day>01</day>
          <volume>176</volume>
          <issue>5</issue>
          <fpage>619</fpage>
          <lpage>625</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/26974260"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jamainternmed.2016.0400</pub-id>
          <pub-id pub-id-type="medline">26974260</pub-id>
          <pub-id pub-id-type="pii">2500043</pub-id>
          <pub-id pub-id-type="pmcid">PMC4996669</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>MO</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Problems with health information technology and their effects on care delivery and patient outcomes: a systematic review</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2017</year>
          <month>03</month>
          <day>01</day>
          <volume>24</volume>
          <issue>2</issue>
          <fpage>246</fpage>
          <lpage>250</lpage>
          <pub-id pub-id-type="doi">10.1093/jamia/ocw154</pub-id>
          <pub-id pub-id-type="medline">28011595</pub-id>
          <pub-id pub-id-type="pii">ocw154</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Akbar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Safety concerns with consumer-facing mobile health applications and their consequences: a scoping review</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2020</year>
          <month>02</month>
          <day>01</day>
          <volume>27</volume>
          <issue>2</issue>
          <fpage>330</fpage>
          <lpage>340</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/31599936"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocz175</pub-id>
          <pub-id pub-id-type="medline">31599936</pub-id>
          <pub-id pub-id-type="pii">5585394</pub-id>
          <pub-id pub-id-type="pmcid">PMC7025360</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bickmore</surname>
              <given-names>TW</given-names>
            </name>
            <name name-style="western">
              <surname>Trinh</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Olafsson</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>O'Leary</surname>
              <given-names>TK</given-names>
            </name>
            <name name-style="western">
              <surname>Asadi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rickles</surname>
              <given-names>NM</given-names>
            </name>
            <name name-style="western">
              <surname>Cruz</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Patient and Consumer Safety Risks When Using Conversational Assistants for Medical Information: An Observational Study of Siri, Alexa, and Google Assistant</article-title>
          <source>J Med Internet Res</source>
          <year>2018</year>
          <month>09</month>
          <day>04</day>
          <volume>20</volume>
          <issue>9</issue>
          <fpage>e11510</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmir.org/2018/9/e11510/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/11510</pub-id>
          <pub-id pub-id-type="medline">30181110</pub-id>
          <pub-id pub-id-type="pii">v20i9e11510</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shneiderman</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>The limits of speech recognition</article-title>
          <source>Commun ACM</source>
          <year>2000</year>
          <volume>43</volume>
          <issue>9</issue>
          <fpage>63</fpage>
          <lpage>65</lpage>
          <pub-id pub-id-type="doi">10.1145/348941.348990</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ruan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wobbrock</surname>
              <given-names>JO</given-names>
            </name>
            <name name-style="western">
              <surname>Liou</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ng</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Landay</surname>
              <given-names>JA</given-names>
            </name>
          </person-group>
          <article-title>Comparing Speech and Keyboard Text Entry for Short Messages in Two Languages on Touchscreen Phones</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2018</year>
          <month>01</month>
          <day>08</day>
          <volume>1</volume>
          <issue>4</issue>
          <fpage>1</fpage>
          <lpage>23</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/10.1145/3161187"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3161187</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cowan</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>The magical number 4 in short-term memory: a reconsideration of mental storage capacity</article-title>
          <source>Behav Brain Sci</source>
          <year>2001</year>
          <month>02</month>
          <volume>24</volume>
          <issue>1</issue>
          <fpage>87</fpage>
          <lpage>114</lpage>
          <pub-id pub-id-type="doi">10.1017/s0140525x01003922</pub-id>
          <pub-id pub-id-type="medline">11515286</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>GA</given-names>
            </name>
          </person-group>
          <article-title>The magical number seven, plus or minus two: some limits on our capacity for processing information</article-title>
          <source>Psychol Rev</source>
          <year>1956</year>
          <volume>63</volume>
          <issue>2</issue>
          <fpage>81</fpage>
          <lpage>97</lpage>
          <pub-id pub-id-type="doi">10.1037/h0043158</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Peterson</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Peterson</surname>
              <given-names>MJ</given-names>
            </name>
          </person-group>
          <article-title>Short-term retention of individual verbal items</article-title>
          <source>J Exp Psychol</source>
          <year>1959</year>
          <volume>58</volume>
          <issue>3</issue>
          <fpage>193</fpage>
          <lpage>198</lpage>
          <pub-id pub-id-type="doi">10.1037/h0049234</pub-id>
          <pub-id pub-id-type="medline">14432252</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sweller</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Cognitive load during problem solving: effects on learning</article-title>
          <source>Cognitive Sci</source>
          <year>1988</year>
          <volume>12</volume>
          <issue>2</issue>
          <fpage>257</fpage>
          <lpage>285</lpage>
          <pub-id pub-id-type="doi">10.1207/s15516709cog1202_4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Limerick</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Moore</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Coyle</surname>
              <given-names>D</given-names>
            </name>
            <collab>editors</collab>
          </person-group>
          <article-title>Empirical evidence for a diminished sense of agency in speech interfaces</article-title>
          <year>2015</year>
          <conf-name>Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems; : ACM</conf-name>
          <conf-date>2015</conf-date>
          <conf-loc>Seoul</conf-loc>
          <fpage>3967</fpage>
          <lpage>3970</lpage>
          <pub-id pub-id-type="doi">10.1145/2702123.2702379</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Leitner</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ahlström</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Hitz</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Usability of Mobile Computing in Emergency Response Systems – Lessons Learned and Future Directions</article-title>
          <year>2007</year>
          <month>11</month>
          <conf-name>Symposium of the Austrian HCI and Usability Engineering Group</conf-name>
          <conf-date>November 22, 2007</conf-date>
          <conf-loc>Graz</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-540-76805-0_20</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sweller</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Cognitive load theory, learning difficulty, and instructional design</article-title>
          <source>Learn Instruct</source>
          <year>1994</year>
          <month>1</month>
          <volume>4</volume>
          <issue>4</issue>
          <fpage>295</fpage>
          <lpage>312</lpage>
          <pub-id pub-id-type="doi">10.1016/0959-4752(94)90003-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ayres</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sweller</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Locus of Difficulty in Multistage Mathematics Problems</article-title>
          <source>Am J Psychol</source>
          <year>1990</year>
          <volume>103</volume>
          <issue>2</issue>
          <fpage>167</fpage>
          <lpage>193</lpage>
          <pub-id pub-id-type="doi">10.2307/1423141</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ayres</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Systematic Mathematical Errors and Cognitive Load</article-title>
          <source>Contemp Educ Psychol</source>
          <year>2001</year>
          <month>04</month>
          <volume>26</volume>
          <issue>2</issue>
          <fpage>227</fpage>
          <lpage>248</lpage>
          <pub-id pub-id-type="doi">10.1006/ceps.2000.1051</pub-id>
          <pub-id pub-id-type="medline">11273658</pub-id>
          <pub-id pub-id-type="pii">S0361476X00910518</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sweller</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ayres</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Kalyuga</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Intrinsic and Extraneous Cognitive Load</article-title>
          <source>Cognitive Load Theory. Explorations in the Learning Sciences, Instructional Systems and Performance Technologies Vol. 1</source>
          <year>2011</year>
          <publisher-loc>New York</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>57</fpage>
          <lpage>69</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beverloo</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Bringert</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Capiel</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Carter</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kharidi</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Mazzoni</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <source>W3C Community</source>
          <year>2012</year>
          <access-date>2018-03-03</access-date>
          <comment>Web Speech API Specification<ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.w3.org/community/speech-api/">https://www.w3.org/community/speech-api/</ext-link>
                                                </comment>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brooke</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>SUS: a 'quick and dirty' usability scale</article-title>
          <source>Usability Evaluation In Industry</source>
          <year>1996</year>
          <publisher-loc>London</publisher-loc>
          <publisher-name>Taylor &#38; Francis</publisher-name>
          <fpage>189</fpage>
          <lpage>194</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Paas</surname>
              <given-names>FG</given-names>
            </name>
          </person-group>
          <article-title>Training strategies for attaining transfer of problem-solving skill in statistics: A cognitive-load approach</article-title>
          <source>J Education Psychol</source>
          <year>1992</year>
          <volume>84</volume>
          <issue>4</issue>
          <fpage>429</fpage>
          <lpage>434</lpage>
          <pub-id pub-id-type="doi">10.1037/0022-0663.84.4.429</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Paas</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Tuovinen</surname>
              <given-names>JE</given-names>
            </name>
            <name name-style="western">
              <surname>Tabbers</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Van Gerven</surname>
              <given-names>PWM</given-names>
            </name>
          </person-group>
          <article-title>Cognitive Load Measurement as a Means to Advance Cognitive Load Theory</article-title>
          <source>Education Psychol</source>
          <year>2003</year>
          <month>03</month>
          <volume>38</volume>
          <issue>1</issue>
          <fpage>63</fpage>
          <lpage>71</lpage>
          <pub-id pub-id-type="doi">10.1207/S15326985EP3801_8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lyell</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>The Effect of Cognitive Load and Task Complexity on Automation Bias in Electronic Prescribing</article-title>
          <source>Hum Factors</source>
          <year>2018</year>
          <month>11</month>
          <day>25</day>
          <volume>60</volume>
          <issue>7</issue>
          <fpage>1008</fpage>
          <lpage>1021</lpage>
          <pub-id pub-id-type="doi">10.1177/0018720818781224</pub-id>
          <pub-id pub-id-type="medline">29939764</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pallant</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Non-parametric statistics</article-title>
          <source>SPSS Survival Manual, 3rd ed</source>
          <year>2010</year>
          <publisher-loc>Adelaide</publisher-loc>
          <publisher-name>McGraw-Hill Education</publisher-name>
          <fpage>210</fpage>
          <lpage>231</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Faul</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Erdfelder</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lang</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Buchner</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences</article-title>
          <source>Behav Res Methods</source>
          <year>2007</year>
          <month>05</month>
          <volume>39</volume>
          <issue>2</issue>
          <fpage>175</fpage>
          <lpage>191</lpage>
          <pub-id pub-id-type="doi">10.3758/bf03193146</pub-id>
          <pub-id pub-id-type="medline">17695343</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hodgson</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the Efficiency and Safety of Speech Recognition within a Commercial Electronic Health Record System: A Replication Study</article-title>
          <source>Appl Clin Inform</source>
          <year>2018</year>
          <month>04</month>
          <day>16</day>
          <volume>9</volume>
          <issue>2</issue>
          <fpage>326</fpage>
          <lpage>335</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29768633"/>
          </comment>
          <pub-id pub-id-type="doi">10.1055/s-0038-1649509</pub-id>
          <pub-id pub-id-type="medline">29768633</pub-id>
          <pub-id pub-id-type="pmcid">PMC5955718</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Using Voice-Activated Conversational Interfaces for Reporting Patient Safety Incidents: A Technical Feasibility and Pilot Usability Study</article-title>
          <source>Stud Health Technol Inform</source>
          <year>2018</year>
          <volume>252</volume>
          <fpage>139</fpage>
          <lpage>144</lpage>
          <pub-id pub-id-type="medline">30040696</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hodgson</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Risks and benefits of speech recognition for clinical documentation: a systematic review</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2016</year>
          <month>04</month>
          <day>17</day>
          <volume>23</volume>
          <issue>e1</issue>
          <fpage>e169</fpage>
          <lpage>e179</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/26578226"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocv152</pub-id>
          <pub-id pub-id-type="medline">26578226</pub-id>
          <pub-id pub-id-type="pii">ocv152</pub-id>
          <pub-id pub-id-type="pmcid">PMC4954615</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hodgson</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Efficiency and safety of speech recognition for documentation in the electronic health record</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2017</year>
          <month>11</month>
          <day>01</day>
          <volume>24</volume>
          <issue>6</issue>
          <fpage>1127</fpage>
          <lpage>1133</lpage>
          <pub-id pub-id-type="doi">10.1093/jamia/ocx073</pub-id>
          <pub-id pub-id-type="medline">29016971</pub-id>
          <pub-id pub-id-type="pii">4049461</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
