<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="review-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v24i10e38472</article-id>
      <article-id pub-id-type="pmid">36239999</article-id>
      <article-id pub-id-type="doi">10.2196/38472</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Review</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Review</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>The Effectiveness of Supervised Machine Learning in Screening and Diagnosing Voice Disorders: Systematic Review and Meta-analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Kukafka</surname>
            <given-names>Rita</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Ahmed</surname>
            <given-names>Arfan</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Alhuwail</surname>
            <given-names>Dari</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Al-Hussain</surname>
            <given-names>Ghada</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9120-2868</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Shuweihdi</surname>
            <given-names>Farag</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1199-2992</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Alali</surname>
            <given-names>Haitham</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-9367-0078</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Househ</surname>
            <given-names>Mowafa</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3648-6271</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Abd-alrazaq</surname>
            <given-names>Alaa</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff5" ref-type="aff">5</xref>
          <address>
            <institution>AI Center for Precision Health</institution>
            <institution>Weill Cornell Medicine</institution>
            <addr-line>Education City, Qatar Foundation</addr-line>
            <addr-line>PO Box 24144</addr-line>
            <addr-line>Doha</addr-line>
            <country>Qatar</country>
            <phone>974 55708549</phone>
            <email>alaa_alzoubi88@yahoo.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7695-4626</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Unified Health Record</institution>
        <institution>Lean for Business Services</institution>
        <addr-line>Riyadh</addr-line>
        <country>Saudi Arabia</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Leeds Institute of Health Sciences, School of Medicine, University of Leads</institution>
        <addr-line>Leeds</addr-line>
        <country>United Kingdom</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Health Management Department, Faculty of Medical and Health Sciences, Liwa College of Technology</institution>
        <addr-line>Abu Dhabi</addr-line>
        <country>United Arab Emirates</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Division of Information and Computing Technology, College of Science and Engineering, Hamad Bin Khalifa University, Qatar Foundation</institution>
        <addr-line>Doha</addr-line>
        <country>Qatar</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>AI Center for Precision Health</institution>
        <institution>Weill Cornell Medicine</institution>
        <addr-line>Doha</addr-line>
        <country>Qatar</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Alaa Abd-alrazaq <email>alaa_alzoubi88@yahoo.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>10</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>14</day>
        <month>10</month>
        <year>2022</year>
      </pub-date>
      <volume>24</volume>
      <issue>10</issue>
      <elocation-id>e38472</elocation-id>
      <history>
        <date date-type="received">
          <day>4</day>
          <month>4</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>22</day>
          <month>4</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>17</day>
          <month>6</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>28</day>
          <month>7</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Ghada Al-Hussain, Farag Shuweihdi, Haitham Alali, Mowafa Househ, Alaa Abd-alrazaq. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 14.10.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2022/10/e38472" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>When investigating voice disorders a series of processes are used when including voice screening and diagnosis. Both methods have limited standardized tests, which are affected by the clinician’s experience and subjective judgment. Machine learning (ML) algorithms have been used as an objective tool in screening or diagnosing voice disorders. However, the effectiveness of ML algorithms in assessing and diagnosing voice disorders has not received sufficient scholarly attention.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This systematic review aimed to assess the effectiveness of ML algorithms in screening and diagnosing voice disorders.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>An electronic search was conducted in 5 databases. Studies that examined the performance (accuracy, sensitivity, and specificity) of any ML algorithm in detecting pathological voice samples were included. Two reviewers independently selected the studies, extracted data from the included studies, and assessed the risk of bias. The methodological quality of each study was assessed using the Quality Assessment of Diagnostic Accuracy Studies 2 tool via RevMan 5 software (Cochrane Library). The characteristics of studies, population, and index tests were extracted, and meta-analyses were conducted to pool the accuracy, sensitivity, and specificity of ML techniques. The issue of heterogeneity was addressed by discussing possible sources and excluding studies when necessary.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Of the 1409 records retrieved, 13 studies and 4079 participants were included in this review. A total of 13 ML techniques were used in the included studies, with the most common technique being least squares support vector machine. The pooled accuracy, sensitivity, and specificity of ML techniques in screening voice disorders were 93%, 96%, and 93%, respectively. Least squares support vector machine had the highest accuracy (99%), while the K-nearest neighbor algorithm had the highest sensitivity (98%) and specificity (98%). Quadric discriminant analysis achieved the lowest accuracy (91%), sensitivity (89%), and specificity (89%).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>ML showed promising findings in the screening of voice disorders. However, the findings were not conclusive in diagnosing voice disorders owing to the limited number of studies that used ML for diagnostic purposes; thus, more investigations are needed. While it might not be possible to use ML alone as a substitute for current diagnostic tools, it may be used as a decision support tool for clinicians to assess their patients, which could improve the management process for assessment.</p>
        </sec>
        <sec sec-type="Trial Registration">
          <title>Trial Registration</title>
          <p>PROSPERO CRD42020214438; https://www.crd.york.ac.uk/prospero/display_record.php?RecordID=214438</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>machine learning</kwd>
        <kwd>voice disorders</kwd>
        <kwd>systematic review</kwd>
        <kwd>meta-analysis</kwd>
        <kwd>diagnose</kwd>
        <kwd>screening</kwd>
        <kwd>mobile phone</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Voice disorders are abnormalities in voice production that could be due to lesions or abnormal modifications in the structure of vocal folds [<xref ref-type="bibr" rid="ref1">1</xref>]. In 2019, it was estimated that 16.9% of the population in Sweden had voice disorders [<xref ref-type="bibr" rid="ref2">2</xref>], and in 2014, it was found that 1 in 13 adults in the United States develops voice disorders every year [<xref ref-type="bibr" rid="ref3">3</xref>]. This led to a loss of US $845 million in the United States owing to missed working days among employees with voice disorders [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. At the individual level, voice disorders can severely affect a patient’s social life and mental health compared with other chronic disorders such as back pain [<xref ref-type="bibr" rid="ref6">6</xref>]. Thus, 4.3% of the patients with voice disorders reported that they were unable to do certain job-related tasks due to the disorder [<xref ref-type="bibr" rid="ref7">7</xref>]; this especially affects professions that have a high demand on the voice, for instance, teachers [<xref ref-type="bibr" rid="ref8">8</xref>], singers, or telephone operators [<xref ref-type="bibr" rid="ref9">9</xref>]. Therefore, screening or diagnosing voice disorders is essential to detect other related health conditions such as laryngeal lesions that could be a symptom of cancer [<xref ref-type="bibr" rid="ref10">10</xref>]; thus, the diagnosis should be made as soon as possible [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>].</p>
        <p>Diagnosing and screening voice disorders involve auditory-perceptual and instrumental assessments. The auditory-perceptual assessment is carried out by a qualified speech and language therapist (SLT); in this assessment, the SLT determines the quality of patients’ voice by listening to their sustained vowel production; for example, the, aa, or sound or continuous speech [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. Furthermore, the instrumental assessment involves laryngeal imaging to examine the structure and function of vocal cords while the patient produces a vowel sound; other techniques are also used including video laryngeoendoscopy and video laryngostroboscopy examinations. In addition, acoustic instruments were used to analyze acoustic features (frequency, pitch, volume, and quality of sound) of voice samples of patients to assess voice disorders by using computer software [<xref ref-type="bibr" rid="ref13">13</xref>-<xref ref-type="bibr" rid="ref15">15</xref>]. Although the aforementioned assessments are recognized by the American Speech-Language-Hearing Association [<xref ref-type="bibr" rid="ref13">13</xref>] and American Academy of Otolaryngology-Head and Neck Surgery [<xref ref-type="bibr" rid="ref16">16</xref>], there is still a lack of standardized methods and guidelines to regulate these or other assessments [<xref ref-type="bibr" rid="ref17">17</xref>]. Therefore, several limitations may pose a risk to the current assessment [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. Although each case is evaluated objectively (via instrumental techniques, eg, stroboscopy), these objective tests include acoustic and visual imaging and videos; the acoustic techniques reveal the speech characteristics of the patient’s speech sample, specifically, the frequency, intensity, loudness, and pitch, to give the clinician insight into other indicators such as the patient’s rate of speech or voice; for example, the voice may be breathy or tremored [<xref ref-type="bibr" rid="ref18">18</xref>]. Although these instrumental methods enable clinicians to perform objective tests, the validity of the tests largely depends on the auditory-perceptual skills of the clinician [<xref ref-type="bibr" rid="ref18">18</xref>]. This is because the clinician first assesses the instrumental management or the patients’ pathway and then chooses the type of instrumental assessment to be used. Naturally, any mistake in the auditory-perceptual assessment would affect the instrumental management, and thus, the whole management of the case; such subjective judgment might not be reliable as it relies on the clinician’s skills and experience [<xref ref-type="bibr" rid="ref18">18</xref>]. As the condition of each diagnosis or screening and the level of experience differ in each case, severe cases might be easier to diagnose or screen than mild cases; therefore, the experience of the SLT and the reliability of their judgment on each patient’s condition differ, and low interrater correlations may occur (&#60;0.9) [<xref ref-type="bibr" rid="ref19">19</xref>]. Moreover, the agreement between experienced and inexperienced SLTs was found to be &#60;75%, making the experience an essential part of the diagnosis or screening [<xref ref-type="bibr" rid="ref20">20</xref>].</p>
        <p>Machine learning (ML) was introduced for speech sounds in the early 1980s [<xref ref-type="bibr" rid="ref21">21</xref>]. ML can be performed automatically by analyzing acoustic features either from voice recordings samples that are previously stored in a database such as the Massachusetts Eye and Ear Infirmary (MEEI), which are databases that stores a recordings of voice samples from patients in clinical environments, these recordings either recorded patients’ voices while pronouncing vowels such as in MEEI [<xref ref-type="bibr" rid="ref22">22</xref>] or continuous speech, or phrases such as in the Saarbruecken Voice Database [<xref ref-type="bibr" rid="ref23">23</xref>]. ML is also used to analyze patients in the clinic by recording their voices via a microphone [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref24">24</xref>]. ML was applied either as a differential diagnosis for s, which involves diagnosing the voice sample as 1 of 2 diseases (voice disorders a or voice disorders b), or for screening different voice samples as either healthy or pathological voice. This method has been used to improve the diagnosis and screening process to be more objective. ML involves 2 different models: classification (supervised learning) and clustering or categorization (unsupervised learning) [<xref ref-type="bibr" rid="ref25">25</xref>]. In the unsupervised model, the algorithm categorizes and identifies relationships within a data set [<xref ref-type="bibr" rid="ref26">26</xref>]. By contrast, classification is a prediction model that defines labels, for example, disease or not disease, in clinical diagnosis [<xref ref-type="bibr" rid="ref26">26</xref>], making it more common in diagnosing [<xref ref-type="bibr" rid="ref27">27</xref>].</p>
      </sec>
      <sec>
        <title>Research Problem and Aim</title>
        <p>Although several studies have investigated the effectiveness of ML algorithms in detecting and diagnosing voice disorders, to the best of our knowledge, only 1 review attempted to summarize the evidence resulting from these studies [<xref ref-type="bibr" rid="ref27">27</xref>]. However, there are several limitations in the review, including the following: it did not exclude studies that did not validate their ML outcomes by using validation techniques; it included studies that relied on scientific but not technical or objective solutions, and they relied on subjective assessment only; and it did not assess the included studies against any risk of bias assessment. Accordingly, this systematic review aimed to assess the effectiveness of supervised ML algorithms in screening and diagnosing voice disorders. Thus, only supervised ML techniques were considered because supervised ML algorithms are more commonly used for diagnosing and detecting disorders.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <p>This systematic review followed the Cochrane Library’s systematic reviews for diagnostic test accuracy (DTA) guidelines [<xref ref-type="bibr" rid="ref28">28</xref>] to meet the objectives of this review. The protocol for this review was registered with PROSPERO (CRD42020214438).</p>
      <sec>
        <title>Search Strategy</title>
        <sec>
          <title>Search Sources</title>
          <p>The following 5 databases were searched on June 24, 2021: MEDLINE (via Ovid), Embase, Scopus, Web of Science, and ACM Digital Library. No language limitations were applied, and non-English articles were translated to check their applicability to the review. The retrieved references were exported and managed using EndNote 9.</p>
        </sec>
        <sec>
          <title>Search Terms</title>
          <p>A total of 2 groups of keywords were used to search the databases: one group representing the target diagnosis (ie, voice disorders) and the other group representing the intervention of interest (ie, ML algorithms). The terms were derived from ML and speech therapy experts. Medical Subject Headings were also included to maximize the sensitivity of the search in MEDLINE and Embase. The detailed search strategy that was applied to MEDLINE and Web of Science is shown in Multimedia Appendixes 1 and 2, respectively.</p>
        </sec>
      </sec>
      <sec>
        <title>Eligibility Criteria</title>
        <sec>
          <title>Inclusion Criteria</title>
          <p>The population of interest in this review included patients diagnosed with a voice disorder. No restrictions were applied to the type of population characteristics (eg, age, gender, and ethnicity). With regard to index tests, we focused on supervised ML techniques (classification) that were used to screen or diagnose voice disorders in binary outcomes (eg, pathological voice vs healthy voice or voice disorder a vs b) by using voice samples collected in a controlled environment (eg, speech laboratories, hospitals, clinics, and databases). The reference standards of interest in this review are instrumental assessment and auditory-perceptual assessment, as both follow the recommendations of the American Speech-Language-Hearing Association [<xref ref-type="bibr" rid="ref17">17</xref>] and American Academy of Otolaryngology [<xref ref-type="bibr" rid="ref16">16</xref>]. To be included in this review, studies had to assess the diagnostic performance of ML algorithms by using at least one of the following outcomes: accuracy, sensitivity, and specificity. We included only peer-reviewed articles and empirical studies regardless of their study design. No restrictions were applied on the country of publication, year of publication, or language of publication.</p>
        </sec>
        <sec>
          <title>Exclusion Criteria</title>
          <p>We excluded studies that relied on clinicians’ judgments only without using any instrumental tools to ensure the validity and reliability of the review, as relying on subjective assessment may be affected by the clinician’s level of experience. Unsupervised ML methods were excluded. Conference papers, reviews, reports, editorials, ongoing studies, non–peer-reviewed articles, studies that assessed accuracy only, and those that did not assess sensitivity and specificity were also excluded.</p>
        </sec>
      </sec>
      <sec>
        <title>Study Selection</title>
        <p>Study selection was first conducted by screening the titles and abstracts of the retrieved studies. Although we excluded studies whose titles and abstracts did not meet any of the eligibility criteria, all studies that met the eligibility criteria or were unclear owing to a lack of information in their titles and abstracts were retained. We then read the full texts of the studies that remained after the title and abstract screening to assess their eligibility for this review. The study selection process was performed by 2 reviewers.</p>
      </sec>
      <sec>
        <title>Data Extraction</title>
        <p>The 2 reviewers created a data extraction form (<xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>) and extracted the data from each included study. If a study did not report a required piece of information, we contacted the corresponding authors to obtain any missing information. If the corresponding authors did not reply within 2 weeks, we sent 2 reminders. If we did not receive a reply after 2 weeks of the second reminder, the missing piece of information was referred to as <italic>n/a: not applicable</italic> data were extracted in an Excel spreadsheet.</p>
      </sec>
      <sec>
        <title>Evaluation of Methodological Quality</title>
        <p>The risk of bias in the included studies was assessed using a revised tool for the Quality Assessment of Diagnostic Accuracy Studies (QUADAS)-2 [<xref ref-type="bibr" rid="ref29">29</xref>], which is highly recommended by the Cochrane Collaboration [<xref ref-type="bibr" rid="ref30">30</xref>]. QUADAS-2 assessed the risk of bias in 4 domains in the included studies: patient selection, index test, reference standards, and flow and timing (<xref ref-type="supplementary-material" rid="app4">Multimedia Appendices 4</xref>-7). Furthermore, QUADAS-2 appraised the applicability of the included studies to this review in terms of 3 domains: patient selection, index test, and reference standards. QUADAS-2 was modified to fit this review (<xref ref-type="supplementary-material" rid="app8">Multimedia Appendix 8</xref>). The 2 reviewers assessed the methodological quality of all included studies by using Review Manager (RevMan version 5.4).</p>
      </sec>
      <sec>
        <title>Data Synthesis and Analysis</title>
        <p>Narrative and quantitative syntheses were conducted to analyze the outcome of each ML technique (accuracy, sensitivity, and specificity). If &#62;1 study used the same ML technique, and the difference between the outcomes was not significant (&#60;5%), the best outcome was considered in the meta-analysis. All outcomes are presented in the extraction table (<xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>). In addition, if a study used voice samples from 2 different databases, each sample was included to account for the sample size (referred to as sample A and sample B in the forest plot).</p>
        <p>The accuracy, sensitivity, and specificity of ML methods extracted from the eligible studies were analyzed using the random effect proportional meta-analysis to estimate a pooled proportion and 95% CI, which are based on the Wilson score [<xref ref-type="bibr" rid="ref31">31</xref>] procedures. To stabilize the variances, the pooled estimate was calculated using the Freeman-Tukey double arcsine transformation [<xref ref-type="bibr" rid="ref32">32</xref>], and heterogeneity was calculated using the <italic>I</italic><sup>2</sup> measure [<xref ref-type="bibr" rid="ref33">33</xref>]. A value of ≤50% is considered low, 51% to 75% moderate, and ≥76% high [<xref ref-type="bibr" rid="ref33">33</xref>]. All results were plotted and presented in a forest plot. Studies were included in the meta-analysis if their scope of using ML was for screening. Statistical software STATA 16 was used to perform random effects meta-analyses.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Search Results</title>
        <p>As presented in <xref rid="figure1" ref-type="fig">Figure 1</xref>, a total of 1409 hits were identified by searching the 5 databases. No additional records were obtained from different resources. After removing duplicates, 95.31% (1343/1409) of articles were left. After scanning the titles and abstracts, 93.89% (1261/1343) of records were excluded, leaving 82 (6.11%) records for full-text reading. We excluded further 84% (69/82) of articles after full-text reading; therefore, only 16% (13/82) of studies were included in this review [<xref ref-type="bibr" rid="ref34">34</xref>-<xref ref-type="bibr" rid="ref46">46</xref>].</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Flowchart of the study selection process.</p>
          </caption>
          <graphic xlink:href="jmir_v24i10e38472_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Study Characteristics</title>
        <sec>
          <title>Study Metadata</title>
          <p>As shown in <xref ref-type="table" rid="table1">Table 1</xref>, the 13 included studies were conducted between 2000 and 2020. However, most of the studies (11/13, 85%) were conducted between 2010 and 2020. The year that witnessed the largest number of studies (3/13, 23%) was 2016. The included studies were conducted in 12 different countries, and approximately 30% (4/13) of them were conducted in Iran. All the studies were observational studies, peer-reviewed articles, and written in English.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Metadata of the included studies.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="380"/>
              <col width="80"/>
              <col width="280"/>
              <col width="260"/>
              <thead>
                <tr valign="top">
                  <td>Study</td>
                  <td>Year</td>
                  <td>Country</td>
                  <td>Publication language</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Akbari and Arjmandi [<xref ref-type="bibr" rid="ref34">34</xref>]</td>
                  <td>2015</td>
                  <td>Iran</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td>
                  <td>2011</td>
                  <td>Greece</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi and Pooyan [<xref ref-type="bibr" rid="ref36">36</xref>]</td>
                  <td>2012</td>
                  <td>Iran</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td>
                  <td>2011</td>
                  <td>Iran</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Cordeiro et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td>
                  <td>2017</td>
                  <td>Portugal</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Ghasemzadeh et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td>
                  <td>2015</td>
                  <td>Iran</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Godino-Llorente and Gómez-Vilda [<xref ref-type="bibr" rid="ref40">40</xref>]</td>
                  <td>2004</td>
                  <td>Spain</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Hadjitodorov et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td>
                  <td>2000</td>
                  <td>Bulgaria and France</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Hariharan et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td>
                  <td>2014</td>
                  <td>Turkey</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                  <td>2017</td>
                  <td>Brazil</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Mohmmad et al [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                  <td>2020</td>
                  <td>Saudi Arabia and Malaysia</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Souissi and Cherif [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                  <td>2016</td>
                  <td>Tunis</td>
                  <td>English</td>
                </tr>
                <tr valign="top">
                  <td>Wang et al [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                  <td>2011</td>
                  <td>China</td>
                  <td>English</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
        </sec>
        <sec>
          <title>Participants or Sample Characteristics</title>
          <p>The number of participants or voice samples ranged from 40 to 960, with a total of 4019 and an average of 309 (<xref ref-type="table" rid="table2">Table 2</xref>). The included studies collected data from 6 different sources. The MEEI database was the most commonly used database among the included studies (9/13, 69%). Voice samples were collected from male and females and intersex participants in most included studies (12/13, 92%); however, 8% (1/13) of studies used voice samples from female participants only [<xref ref-type="bibr" rid="ref43">43</xref>]. Participants’ ages in the included studies ranged from 13 to 86 years, with an average age of 45 years (mean 46, SD 29.5 years).</p>
          <table-wrap position="float" id="table2">
            <label>Table 2</label>
            <caption>
              <p>Characteristics of participants or sample.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="250"/>
              <col width="120"/>
              <col width="130"/>
              <col width="90"/>
              <col width="240"/>
              <col width="170"/>
              <thead>
                <tr valign="top">
                  <td>Study</td>
                  <td>Voice sample size, n</td>
                  <td>Age (years), range</td>
                  <td>Male (%)</td>
                  <td>Setting or database</td>
                  <td>Database accessibility</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Akbari and Arjmandi [<xref ref-type="bibr" rid="ref34">34</xref>]</td>
                  <td>293</td>
                  <td>13-82</td>
                  <td>40</td>
                  <td>MEEI<sup>a</sup> database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td>
                  <td>628</td>
                  <td>19-70</td>
                  <td>—<sup>b</sup></td>
                  <td>MEEI and UPM<sup>c</sup> databases</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi and Pooyan [<xref ref-type="bibr" rid="ref36">36</xref>]</td>
                  <td>120</td>
                  <td>18-86</td>
                  <td>56</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td>
                  <td>100</td>
                  <td>16-85</td>
                  <td>67</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Cordeiro et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td>
                  <td>154</td>
                  <td>—</td>
                  <td>34</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Ghasemzadeh et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td>
                  <td>393</td>
                  <td>—</td>
                  <td>—</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Godino-Llorente and Gómez-Vilda [<xref ref-type="bibr" rid="ref40">40</xref>]</td>
                  <td>135</td>
                  <td>—</td>
                  <td>—</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Hadjitodorov et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td>
                  <td>400</td>
                  <td>—</td>
                  <td>—</td>
                  <td>Phoniatric Department of the University Hospital in Sofia</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Hariharan et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td>
                  <td>274</td>
                  <td>20-68</td>
                  <td>—</td>
                  <td>MEEI and MAPACI databases</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                  <td>279</td>
                  <td>18-65</td>
                  <td>0</td>
                  <td>Voice laboratory</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Mohmmad et al [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                  <td>960</td>
                  <td>—</td>
                  <td>—</td>
                  <td>SVD<sup>d</sup></td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Souissi and Cherif [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                  <td>120</td>
                  <td>—</td>
                  <td>—</td>
                  <td>SVD</td>
                  <td>Private</td>
                </tr>
                <tr valign="top">
                  <td>Wang et al [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                  <td>226</td>
                  <td>26-58</td>
                  <td>—</td>
                  <td>MEEI database</td>
                  <td>Private</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table2fn1">
                <p><sup>a</sup>MEEI: Massachusetts Eye and Ear Infirmary.</p>
              </fn>
              <fn id="table2fn2">
                <p><sup>b</sup>Not available.</p>
              </fn>
              <fn id="table2fn3">
                <p><sup>c</sup>UPM: Universidad Autónoma de Madrid.</p>
              </fn>
              <fn id="table2fn4">
                <p><sup>d</sup>SVD: Saarbruecken Voice Database.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Index Test Characteristics</title>
          <p>The included studies used 12 ML algorithms (<xref ref-type="table" rid="table3">Table 3</xref>). Least-squares support-vector machines (LS-SVMs) were the most used algorithms across studies (9/13, 69%), followed by quadratic discriminant analysis (QDA) (3/13, 23%) and K-nearest neighbor (K-NN) (4/13, 31%). The feature-extraction technique was reported in 85% (11/13) of studies. While 61% (8/13) of studies extracted short-term features (eg, mel frequency cepstral coefficients), 23% (3/13) extracted long-term features (eg, jitter and shimmer and fundamental frequency). A total of 3 feature reduction techniques were used in the included studies; linear discriminant analysis was the most used technique (4/13, 31%), and training-test split validation was the most prominent technique used in the included studies (10/13, 77%), followed by cross-validation technique (4/13, 31%).</p>
          <table-wrap position="float" id="table3">
            <label>Table 3</label>
            <caption>
              <p>Index test characteristics.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="190"/>
              <col width="140"/>
              <col width="330"/>
              <col width="140"/>
              <col width="200"/>
              <thead>
                <tr valign="top">
                  <td>Study</td>
                  <td>Machine learning method</td>
                  <td>Feature extraction</td>
                  <td>Feature reduction</td>
                  <td>Validation</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Akbari and Arjmandi [<xref ref-type="bibr" rid="ref34">34</xref>]</td>
                  <td>LS-SVM<sup>a</sup></td>
                  <td>Mean, variance, skewness, kurtosis of coefficient, wavelet subband coefficients</td>
                  <td>Linear prediction analysis and LDA<sup>b</sup></td>
                  <td>70% training and 30% 0% testing</td>
                </tr>
                <tr valign="top">
                  <td>Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td>
                  <td>LS-SVM</td>
                  <td>12 MFCC<sup>c</sup> and MSMR<sup>d</sup></td>
                  <td>MSMR and LS-SVM</td>
                  <td>75% training and 25% testing (cross-validation–test split validation)</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi and Pooyan [<xref ref-type="bibr" rid="ref36">36</xref>]</td>
                  <td>QDA<sup>e</sup>, NMC<sup>f</sup>, K-NN<sup>g</sup>, LS-SVM, ML-NN<sup>h</sup>, and PC<sup>i</sup></td>
                  <td>PCA<sup>j</sup> and LDA; feature selection: IFS<sup>k</sup>, FFS<sup>l</sup>, BFS<sup>m</sup>, and BBFS<sup>n</sup></td>
                  <td>PCA and LDA</td>
                  <td>70% training and 30% validation</td>
                </tr>
                <tr valign="top">
                  <td>Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td>
                  <td>QDA, NMC, PC, K-NN, LS-SVM, and ML-NN</td>
                  <td>Fundamental frequency (average, high, and low variation), STD<sup>o</sup>, PFR<sup>p</sup>, jitter, shimmer, RAP<sup>q</sup>, PPQ<sup>r</sup>, smoothed PPQ, vAm<sup>s</sup>, NHR<sup>t</sup>, VTI<sup>u</sup>, SPI<sup>v</sup>, FTRI<sup>w</sup>, ATRI<sup>x</sup>, Tsam<sup>y</sup>, T0<sup>z</sup>, shimmer in dB, DVB<sup>aa</sup>, DSH<sup>ab</sup>, DUV<sup>ac</sup>, NVB<sup>ad</sup>, NSH<sup>ae</sup>, and total number of segments pitch period during the period-to-period pitch extraction</td>
                  <td>PCA and LDA</td>
                  <td>70% training and 30% testing</td>
                </tr>
                <tr valign="top">
                  <td>Cordeiro et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td>
                  <td>SVM and DA<sup>af</sup></td>
                  <td>MFCCs, line spectral frequencies, and delta-MFCC</td>
                  <td>N/A<sup>ag</sup></td>
                  <td>75% training and 25% testing (k-fold cross-validation method, k=4; training-test split validation)</td>
                </tr>
                <tr valign="top">
                  <td>Ghasemzadeh et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td>
                  <td>ANN<sup>a</sup><sup>h</sup> and LS-SVM</td>
                  <td>False neighbor fraction and mutual information</td>
                  <td>LDA and LS-SVM</td>
                  <td>70% training and 30% testing using cross-validation</td>
                </tr>
                <tr valign="top">
                  <td>Godino-Llorente and Gómez-Vilda [<xref ref-type="bibr" rid="ref40">40</xref>]</td>
                  <td>LVQ<sup>a</sup><sup>i</sup></td>
                  <td>MFCC coefficient, energy, and first and second temporal derivatives</td>
                  <td>MFCC</td>
                  <td>70% training and 30% test split validation</td>
                </tr>
                <tr valign="top">
                  <td>Hadjitodorov et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td>
                  <td>K-NN</td>
                  <td>Pitch period (To), PPQ, APQ<sup>aj</sup>, STAB<sup>a</sup><sup>k</sup>, the degree of the dissimilarity of the shape [<xref ref-type="bibr" rid="ref47">47</xref>] of the pitch pulses, LHER<sup>a</sup><sup>l</sup>, NHR, HNR<sup>am</sup>, and energy in the pitch impulse-incepstra</td>
                  <td>LDA</td>
                  <td>Training-test split validation stage (200 phonation); testing (200 phonation)</td>
                </tr>
                <tr valign="top">
                  <td>Hariharan et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td>
                  <td>K-NN, LS-SVM, and GRNN<sup>an</sup></td>
                  <td>5 level WPT<sup>ao</sup> decomposition</td>
                  <td>N/A</td>
                  <td>70% training and 30% testing using conventional validation and cross-validation</td>
                </tr>
                <tr valign="top">
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                  <td>QDA</td>
                  <td>F0 measurements (mean and SD, jitter, shimmer, and GNE<sup>ap</sup>)</td>
                  <td>N/A</td>
                  <td>Cross-validation</td>
                </tr>
                <tr valign="top">
                  <td>Mohmmad et al [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                  <td>CNN<sup>aq</sup></td>
                  <td>Octaves and its first and second derivatives</td>
                  <td>N/A</td>
                  <td>10-fold cross-validation</td>
                </tr>
                <tr valign="top">
                  <td>Souissi and Cherif [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                  <td>LS-SVM and ANN</td>
                  <td>MFCC and first and second derivatives</td>
                  <td>MFCC, LDA, and delta</td>
                  <td>70% training; and 30% testing</td>
                </tr>
                <tr valign="top">
                  <td>Wang et al [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                  <td>LS-SVM and GMM<sup>a</sup><sup>r</sup></td>
                  <td>36 dimensional MFCC parameters with 1 derivative were calculated every frame of 18-mel-cepstral coefficient</td>
                  <td>8, 16, and 32 mixture</td>
                  <td>10-fold cross-validation</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table3fn1">
                <p><sup>a</sup>LS-SVM: least-squares support-vector machine.</p>
              </fn>
              <fn id="table3fn2">
                <p><sup>b</sup>LDA: linear discriminant analysis.</p>
              </fn>
              <fn id="table3fn3">
                <p><sup>c</sup>MFCC: mel frequency cepstral coefficient.</p>
              </fn>
              <fn id="table3fn4">
                <p><sup>d</sup>MSMR: modulation spectra minimum redundancy.</p>
              </fn>
              <fn id="table3fn5">
                <p><sup>e</sup>QDA: quadric discriminant analysis.</p>
              </fn>
              <fn id="table3fn6">
                <p><sup>f</sup>NMC: neuromorphic computing.</p>
              </fn>
              <fn id="table3fn7">
                <p><sup>g</sup>K-NN: K-nearest neighbor.</p>
              </fn>
              <fn id="table3fn8">
                <p><sup>h</sup>ML-NN: multilayer neural network.</p>
              </fn>
              <fn id="table3fn9">
                <p><sup>i</sup>PC: Parzen classifier.</p>
              </fn>
              <fn id="table3fn10">
                <p><sup>j</sup>PCA: principal component analysis.</p>
              </fn>
              <fn id="table3fn11">
                <p><sup>k</sup>IFS: individual feature selection.</p>
              </fn>
              <fn id="table3fn12">
                <p><sup>l</sup>FFS: forward feature selection.</p>
              </fn>
              <fn id="table3fn13">
                <p><sup>m</sup>BFS: backward feature selection.</p>
              </fn>
              <fn id="table3fn14">
                <p><sup>n</sup>BBFS: branch-and-bound feature selection.</p>
              </fn>
              <fn id="table3fn15">
                <p><sup>o</sup>STD: SD of fundamental frequency.</p>
              </fn>
              <fn id="table3fn16">
                <p><sup>p</sup>PFR: phonatory fundamental frequency.</p>
              </fn>
              <fn id="table3fn17">
                <p><sup>q</sup>RAP: relative average perturbation.</p>
              </fn>
              <fn id="table3fn18">
                <p><sup>r</sup>PPQ: pitch perturbation quotient.</p>
              </fn>
              <fn id="table3fn19">
                <p><sup>s</sup>vAm: peak amplitude variation.</p>
              </fn>
              <fn id="table3fn20">
                <p><sup>t</sup>NHR: noise-to-harmonic ratio.</p>
              </fn>
              <fn id="table3fn21">
                <p><sup>u</sup>VTI: voice turbulence index.</p>
              </fn>
              <fn id="table3fn22">
                <p><sup>v</sup>SPI: soft phonation index.</p>
              </fn>
              <fn id="table3fn23">
                <p><sup>w</sup>FTRI: Fo-tremor intensity index.</p>
              </fn>
              <fn id="table3fn24">
                <p><sup>x</sup>ATRI: amplitude tremor intensity index.</p>
              </fn>
              <fn id="table3fn25">
                <p><sup>y</sup>Tsam: length in seconds of analyzed voice data sample.</p>
              </fn>
              <fn id="table3fn26">
                <p><sup>z</sup>T0: period of the average glottal period.</p>
              </fn>
              <fn id="table3fn27">
                <p><sup>aa</sup>DVB: degree of voice breaks.</p>
              </fn>
              <fn id="table3fn28">
                <p><sup>ab</sup>DSH: degree of subharmonic.</p>
              </fn>
              <fn id="table3fn29">
                <p><sup>ac</sup>DUV: degree of voicelessness.</p>
              </fn>
              <fn id="table3fn30">
                <p><sup>ad</sup>NVB: number of voice breaks.</p>
              </fn>
              <fn id="table3fn31">
                <p><sup>ae</sup>NSH: number of subharmonic segments.</p>
              </fn>
              <fn id="table3fn32">
                <p><sup>af</sup>DA: Discriminant analysis.</p>
              </fn>
              <fn id="table3fn33">
                <p><sup>ag</sup>N/A: not applicable.</p>
              </fn>
              <fn id="table3fn34">
                <p><sup>ah</sup>ANN: artificial neural network.</p>
              </fn>
              <fn id="table3fn35">
                <p><sup>ai</sup>LVQ: learning vector quantization.</p>
              </fn>
              <fn id="table3fn36">
                <p><sup>aj</sup>APQ: amplitude of the pitch pules.</p>
              </fn>
              <fn id="table3fn37">
                <p><sup>ak</sup>STAB: stability of the t0 generation.</p>
              </fn>
              <fn id="table3fn38">
                <p><sup>al</sup>LHER: low-high energy ratio.</p>
              </fn>
              <fn id="table3fn39">
                <p><sup>am</sup>HNR: harmonics noise ratio.</p>
              </fn>
              <fn id="table3fn40">
                <p><sup>an</sup>GRNN: general regression neural network.</p>
              </fn>
              <fn id="table3fn41">
                <p><sup>ao</sup>WPT: wavelet packet transform.</p>
              </fn>
              <fn id="table3fn42">
                <p><sup>ap</sup>GNE: glottal to noise excitation.</p>
              </fn>
              <fn id="table3fn43">
                <p><sup>aq</sup>CNN: conventional neural network.</p>
              </fn>
              <fn id="table3fn44">
                <p><sup>ar</sup>GMM: Gaussian mixture model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
      <sec>
        <title>Quality Assessment Results</title>
        <sec>
          <title>Risk of Bias</title>
          <p>In the patient selection domain, only 38% (5/13) of studies were judged to have a low risk of bias in patient sampling, as they used an appropriate sampling process to select voice samples (<xref ref-type="supplementary-material" rid="app9">Multimedia Appendix 9</xref>). The risk of bias in index tests was rated as high in all included studies owing to the nature of the supervised ML tests, and their results were interpreted with prior knowledge of the results of the reference standard test. Owing to the subjective nature of voice assessment, it was not clear whether the reference standard correctly classified the patients. This led to an unclear risk of bias in the reference standard domain in all studies although the reference standard was used before the index test, and the findings were not affected by the findings of the index test. Patient flow and timing were poorly reported in almost all the studies (12/13, 92%). Thus, these studies were judged to pose an unclear risk of bias in terms of patient flow and timing. <xref ref-type="supplementary-material" rid="app9">Multimedia Appendix 9</xref> shows the QUADAS-2 tool risk of bias judgment in each included study across all 3 domains as well as applicability concerns for each study.</p>
        </sec>
        <sec>
          <title>Applicability Concerns</title>
          <p>There are no applicability concerns regarding how patients were selected in all included studies, as the patients’ characteristics and the condition and setting of each test match the review question and criteria (<xref ref-type="supplementary-material" rid="app9">Multimedia Appendix 9</xref>). Similarly, all included studies were judged to have low applicability concern in the index test as the ML algorithms method in the included studies matched the review definition of ML. However, the applicability concern in the reference standard was rated as unclear in 84% (11/13) of studies, as the voice samples in those studies were collected from databases, and the detailed diagnosis process of each voice sample was not described.</p>
        </sec>
      </sec>
      <sec>
        <title>Performance of ML Algorithms</title>
        <sec>
          <title>Diagnosing Voice Disorders</title>
          <p>Only 8% (1/13) of studies used the QDA algorithm to differentiate between 2 [<xref ref-type="bibr" rid="ref43">43</xref>]. As shown in <xref ref-type="table" rid="table4">Table 4</xref>, the accuracy, sensitivity, and specificity of the QDA ranged from 70% to 77%, 20% to 65%, and 74.76% to 95%, respectively. See the following section for a description of how QDA was used as a screening tool. For breakdown of the diagnostic findings, please refer to <xref ref-type="supplementary-material" rid="app10">Multimedia Appendix 10</xref>.</p>
          <table-wrap position="float" id="table4">
            <label>Table 4</label>
            <caption>
              <p>The performance of machine learning in diagnosing voice disorders.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="130"/>
              <col width="280"/>
              <col width="140"/>
              <col width="160"/>
              <col width="160"/>
              <col width="130"/>
              <thead>
                <tr valign="top">
                  <td>Algorithm</td>
                  <td>Tested diseases</td>
                  <td>Accuracy (%)</td>
                  <td>Sensitivity (%)</td>
                  <td>Specificity (%)</td>
                  <td>Study</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>QDA<sup>a</sup></td>
                  <td>Vocal polyps vs healthy</td>
                  <td>70.56</td>
                  <td>50</td>
                  <td>74.76</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal cyst vs healthy</td>
                  <td>72.67</td>
                  <td>60.83</td>
                  <td>78.1</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Unilateral VF<sup>b</sup> paralysis or healthy</td>
                  <td>79.82</td>
                  <td>20</td>
                  <td>92.38</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Middle-posterior triangular gap vs healthy</td>
                  <td>71.11</td>
                  <td>45</td>
                  <td>80.43</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Sulcus vocalis vs healthy</td>
                  <td>78.75</td>
                  <td>50</td>
                  <td>83.33</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>VDDGER<sup>c</sup> vs healthy</td>
                  <td>72.44</td>
                  <td>33.33</td>
                  <td>90.71</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal nodules vs unilateral VF paralysis</td>
                  <td>76.61</td>
                  <td>20</td>
                  <td>88.57</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal nodules vs sulcus vocalis</td>
                  <td>72.68</td>
                  <td>50</td>
                  <td>75.95</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal nodules vs VDDGER</td>
                  <td>71</td>
                  <td>33.33</td>
                  <td>89.05</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal nodules vs sulcus vocalis</td>
                  <td>70</td>
                  <td>30</td>
                  <td>95</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal polyp vs healthy</td>
                  <td>75.14</td>
                  <td>65</td>
                  <td>78.33</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
                <tr valign="top">
                  <td>QDA</td>
                  <td>Vocal cyst vs healthy</td>
                  <td>73.22</td>
                  <td>62.5</td>
                  <td>78.57</td>
                  <td>Lopes et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table4fn1">
                <p><sup>a</sup>QDA: quadratic discriminant analysis.</p>
              </fn>
              <fn id="table4fn2">
                <p><sup>b</sup>VF: vocal fold.</p>
              </fn>
              <fn id="table4fn3">
                <p><sup>c</sup>VDDGER: voice disorder due to gastroesophageal reflux.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Screening Voice Disorders</title>
          <p>Of the 13 included studies in the systematic review, 10 (77%) were included in the meta-analysis and 3 (23%) were excluded [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>,<xref ref-type="bibr" rid="ref44">44</xref>,<xref ref-type="bibr" rid="ref46">46</xref>]. Of the 10 studies, 2 (20%) examined ML techniques by using 2 different databases: Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>] (MEEI and Universidad Autónoma de Madrid [UPM] databases) and Hariharan et al [<xref ref-type="bibr" rid="ref42">42</xref>] (MEEIEMPACI) Accordingly, the performance of ML techniques in these databases was included in the meta-analysis. More information about the performance in screening can be found in (<xref ref-type="supplementary-material" rid="app11">Multimedia Appendix 11</xref>).</p>
        </sec>
        <sec>
          <title>Accuracy</title>
          <p>The accuracy of ML techniques in assessing voice disorders was reported in 77% (10/13) of studies. These studies examined the accuracy of 9 ML techniques. The pooled accuracy of the 9 ML techniques was 96% (95% CI 93%-98%; <xref rid="figure2" ref-type="fig">Figure 2</xref>). Significant heterogeneity was shown in the meta-analyzed studies (<italic>I</italic><sup>2</sup>=93.51%; <italic>P</italic>&#60;.001), and the possible causes of this heterogeneity are discussed below. Regarding voice disorders assessment, the ML technique that achieved the highest accuracy was LS-SVM (99%), whereas the one that had the lowest accuracy was QDA (91%).</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>The forest plot shows the accuracy of machine learning algorithms in voice disorder screening. ANN: artificial neural network; GRNN: general regression neural network; K-NN: K-nearest neighbor; LS-SVM: least-squares support-vector machine; ML-NN: multilayer neural network; NMC: neuromorphic computing; PC: parzan Classifier; QDA: quadratic discriminant analysis; SVM: support vector machine.</p>
            </caption>
            <graphic xlink:href="jmir_v24i10e38472_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Sensitivity</title>
          <p>The sensitivity of ML techniques in assessing voice disorders was reported in 77% (10/13) of studies. These studies examined the sensitivity of 3 ML techniques. The pooled sensitivity of the 3 ML techniques was 96% (95% CI 91%-100%; <xref rid="figure3" ref-type="fig">Figure 3</xref>). The meta-analyzed studies showed significant heterogeneity (<italic>I</italic><sup>2</sup>=95.49%; <italic>P</italic>&#60;.001), and the possible causes of such heterogeneity are discussed in further sections. K-NN had the highest sensitivity (98%) among the 3 ML techniques, while QDA achieved the lowest sensitivity (89%).</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>The forest plot shows the sensitivity of machine learning algorithms in voice disorder screening. K-NN: K-nearest neighbor; QDA: quadratic discriminant analysis; SVM: support vector machine.</p>
            </caption>
            <graphic xlink:href="jmir_v24i10e38472_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Specificity</title>
          <p>The specificity of ML techniques in assessing voice disorders was examined in 77% (10/13) of studies and included the specificity of 3 ML techniques. The pooled specificity of the 3 ML techniques was 93% (95% CI 88%-97%; <xref rid="figure4" ref-type="fig">Figure 4</xref>). The meta-analyzed evidence showed significant heterogeneity (<italic>I</italic><sup>2</sup>=84.3%; <italic>P</italic>&#60;.001); the possible causes of heterogeneity are discussed below. The ML technique that achieved the highest specificity was K-NN (98%), whereas the one that had the lowest specificity was QDA (89%).</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>The forest plot shows the specificity of machine learning algorithms in voice disorder screening. K-NN: K-nearest neighbor; QDA: quadratic discriminant analysis; SVM: support vector machine.</p>
            </caption>
            <graphic xlink:href="jmir_v24i10e38472_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Heterogeneity and Pooled Performance</title>
          <p>The possible source of heterogeneity in the pooled performance was explored, and the possibility that studies that used short-term parameters, such as the study by Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>], increased the heterogeneity in K-NN and LS-SVM was found. In the K-NN algorithm, the heterogeneity was reduced to 69.73% when the study by Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>] (which used long-term parameters) was excluded; in specificity, it was 84.92% in sensitivity and 91.55% in accuracy. This was also found in the study by Hadjitodorov et al [<xref ref-type="bibr" rid="ref41">41</xref>], which also used long-term parameters, and when it was excluded, the heterogeneity in all K-NN outcomes was reduced (<xref ref-type="supplementary-material" rid="app12">Multimedia Appendix 12</xref> presents further details on the heterogeneity values when each study was removed). Similarly, when the study by Arjmandi et al [<xref ref-type="bibr" rid="ref37">37</xref>] was removed from the LS-SVM forest plot for sensitivity, a reduction was found in <italic>I</italic><sup>2</sup> test values, which decreased to 91.89%. Therefore, long-term parameters could affect the sensitivity of LS-SVM and all 3 outcomes in K-NN. Furthermore, the database used by Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>] and Souissi and Cherif [<xref ref-type="bibr" rid="ref45">45</xref>] might increase the heterogeneity in LS-SVM performance. Arias-Londoño et al [<xref ref-type="bibr" rid="ref35">35</xref>] used the UPM database, which is a Spanish sounds database, thus excluding the study from the sensitivity and specificity forest plot of LS-SVM, which decreased the heterogeneity to 58% and 71%, respectively. On the other hand, Souissi and Cherif [<xref ref-type="bibr" rid="ref45">45</xref>] used voice samples from S (German speech samples database), whereas the remaining studies used s from the English speech samples database (MEEI; <xref ref-type="supplementary-material" rid="app12">Multimedia Appendix 12</xref>).</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>This study systematically reviewed the performance of ML in assessing voice disorders, similar to another study by Syed et al [<xref ref-type="bibr" rid="ref27">27</xref>] that examined the accuracy of ML algorithms at the voice database level and qualitatively analyzed the accuracy of each ML algorithm technique. It was concluded that LS-SVM is the most common algorithm used in studies included in this research, which aligns with our findings. Furthermore, the performance showed the accuracy of LS-SVM to be &#62;93%, which was similar to our findings. Generally, ML performance was found to be more promising when it was used as a screening tool rather than in diagnosis, achieving &#62;90% in all 3 outcomes (accuracy, sensitivity, and specificity). Second, the findings differ significantly between the algorithms or even within the same algorithm in different studies. For example, LS-SVM was almost 100% in all 3 outcomes; however, Parzen classifier showed sensitivity ranging from 74% to 100%. Because of the limited number of studies, the performance of ML in ≤2 studies remains unclear. This was also noticed in ML algorithms that were used in the diagnosis, as only 1 study implemented ML algorithms to differentiate between different disorders (diagnosis). For example, the performance of QDA in screening showed 83% accuracy, 91% sensitivity, and 68% specificity. By contrast, it was found to be &#60;76% in diagnosis, and the percentage fell sharply in sensitivity and specificity in the same study [<xref ref-type="bibr" rid="ref43">43</xref>]. However, this finding could not be conclusive because of the limited number of studies that used ML for diagnosis (1 study).</p>
        <p>The analysis implies that K-NN and LS-SVM showed the highest accuracy. K-NN demonstrated increased specificity; however, LS-SVM was found to be better at detecting true positive cases. Because ML in the included studies was used as a screening tool (pathological voice vs healthy voice), the ability of ML to be more sensitive might be more important than the ability to be specific. This may be due to the consequences of diagnosing healthy voiced patients as pathological voice which will only lead to further examination (stroboscopy). Moreover, it will not cause any distress to the patient, as the diagnosis is not final, and patients would only be referred for further examination. However, in less sensitive tests, misdiagnosis of patients can lead to harmful consequences.</p>
      </sec>
      <sec>
        <title>Research and Practical Implications</title>
        <sec>
          <title>Practical Implications</title>
          <p>When a person’s strength, agility, and structure of vocal folds result in pathological noise and reduced acoustic tone, their vocal pathology may be serious enough to qualify as a voice disorder. These disorders can be caused by tissue diseases and changes in tissue, mechanical stress, surface discomfort, systemic changes, changes in muscles and nerves, and many other factors [<xref ref-type="bibr" rid="ref48">48</xref>]. Research on has achieved a wide scope, partly because of its societal benefits. Standard databases have been developed to mitigate disorders and include new features and emphasis on specific voice disorders while using deep neural networks. Recently, subjective and objective evaluations of vocal issues have received considerable attention in the research field [<xref ref-type="bibr" rid="ref49">49</xref>].</p>
          <p>Subjective assessments may be conducted by clinicians, as they focus on the patient’s voice and use different instruments to discern various vocal disorder diagnoses. ML can be used as a decision-support tool for clinicians conducting auditory-perceptual assessments [<xref ref-type="bibr" rid="ref14">14</xref>]. A second assessment, known as “target evaluated assessments,” focuses on the automatic, computer-based processing of acoustic signals. These signals assess and recognize the underlying vocal pathology, which may not be screened or diagnosed by a clinician [<xref ref-type="bibr" rid="ref50">50</xref>]. Consequently, this type of evaluation is nonsubjective. Furthermore, when using this type of assessment, voices can be captured and stored at a global level via cloud technologies by using various intelligent devices. This has been beneficial for researchers across the globe, who can access the data through different academic institutions.</p>
          <p>Using ML as an assessment tool may reduce the learning gap between experienced and inexperienced clinicians. Bassich and Ludlow [<xref ref-type="bibr" rid="ref20">20</xref>] found that the intrajudge test-retest agreement was &#60;75% when evaluating voice quality in patients with polyps or vocal fold nodules; thus, the overall reliance on experienced clinicians in voice assessment might be eliminated. Furthermore, the practice of using instrumental assessments in practice could be eliminated, as ML may reduce the need to conduct instrumental assessments for more typical cases [<xref ref-type="bibr" rid="ref27">27</xref>]. However, eliminating instrumental assessments altogether may lead to misdiagnoses, for example, if a patient with laryngeal cancer was screened “as healthy,” the clinician may not have performed a stroboscopic examination. Therefore, we aim to further our study by establishing an ideal and automatic ML-based system. We anticipate that this system will be sensitive, accurate, efficient, and successful in detecting and diagnosing various voice disorders quickly and effortlessly for both patients and practitioners.</p>
          <p>The review showed that ML provided optimum performance in screening and diagnosing voice disorders to inform clinicians of anomalies. A comparison of the performance of ML algorithms, including accuracy, specificity, and sensitivity, across studies is recommended owing to the different characteristics of each study. The most commonly used ML methods for diagnosing voice disorders in this review were LS-SVM and artificial neural network algorithms. However, the preference of applying 1 ML method to another was not clearly explained in the studies. All studies used internal validation (training and test splits and cross-validation) to evaluate the ML quality. However, external validation is a necessary procedure to evaluate the real quality of ML predictions for new data. Therefore, external validation is essential to implement ML in routine clinical practice to diagnose voice disorders. Therefore, external validation must be performed before using ML for any clinical diagnosis. None of the ML methods investigated in this review used external validation.</p>
        </sec>
        <sec>
          <title>Implications for Research</title>
          <p>This paper analyzes the literature related to the effectiveness of using ML algorithms to screen and diagnose voice disorders. It not only provides insight into the type of research conducted over the last 2 decades but also highlights the areas of research needing further experimentation and analysis. Researchers and practitioners can use this research to improve their objective screening or diagnosis of speech pathology. For instances, voice disorders [<xref ref-type="bibr" rid="ref23">23</xref>], MEEI [<xref ref-type="bibr" rid="ref22">22</xref>], and UPM databases [<xref ref-type="bibr" rid="ref51">51</xref>] are all accessible to researchers interested in voice disorders case studies. However, these data repositories are not without their flaws. For instance, certain databases are uniformly classified into healthy and unhealthy classes. These voices are, in turn, generally categorized as “healthy” and “pathological” in most of the research published using these data. Some databases do not specify the severity of voice disorders or provide sufficient details on the pathological symptoms during phonation. As such, some samples may appear healthy normal despite being labeled as pathological, and vice versa. In addition, &#62;1 disorder may be used to label documents, which can be challenging to incorporate or exclude samples in different languages [<xref ref-type="bibr" rid="ref52">52</xref>]. The nature of supervised ML, that is, “labeled,” tests require prior knowledge of the reference standard finding to the corresponding test. This may lead to a higher risk of bias in some quality assessment tools, such as the QUADAS-2 tool, which shows a high risk of bias in the index test domain. Future researchers may wish to consider providing information on how a reference standard was applied when examining the performance of ML. Furthermore, these repositories may determine a more specific judgment on suitable demographic characteristics and how to appropriately classify these specifics. Finally, differential diagnostic abilities for ML may be better examined by dividing both the outcomes of each disorder as well as their severity. This would allow for more definitive and specific findings about the type of patients for whom ML may be more effectively used.</p>
          <p>Because ML in the included studies was used as a screening tool (pathological voice vs healthy voice), the ability of ML to be more sensitive might be more important than its ability to be more specific. This may be due to the consequences of diagnosing healthy patients as unhealthy (patients with pathological voice), which will lead to further examination (stroboscopy) and not cause patient distress, as the diagnosis, at this point, is not final and patients would be referred for further examinations. Misdiagnosing patients (less sensitive tests) could lead to harmful consequences and distress, for example, if life-threatening diseases such as laryngeal cancer are misdiagnosed.</p>
          <p>It should also be considered that ML can be used as a decision-support tool by clinicians while subjectively judging patients’ voices to determine whether they should undergo further examinations. Applying the ML algorithm as a screening tool could help in predetermining the patient’s voice condition. Consequently, this could support the clinicians’ whole management process in voice disorders assessment, especially in their decision on whether to apply an instrumental examination for the patient, a decision that is currently being made subjectively. Therefore, applying ML as a screening tool would reduce the gap between experienced and inexperienced clinicians (the agreement was found to be &#60;75%) [<xref ref-type="bibr" rid="ref20">20</xref>], and the overall reliance on experienced clinicians in voice assessment might be eliminated. Furthermore, the use of instrumental assessments in practice could be eliminated, as not all patients will have to undergo instrumental assessments (ML might reduce the need to use them for healthy cases). Therefore, the cost of assessing voice disorders might be reduced.</p>
          <p>Our findings also imply that ML can be used in web-based methods to detect voice disorders. This means that the algorithms can be used in smartphone apps or users’ phone calls to detect the presence of voice disorders or even track the progress of their therapy. This might eliminate the amount of time spent by the clinician to screen or diagnose or record the progress of each follow-up. This study also found that researchers may want to consider investigating the applicability of various ML algorithms to identify and diagnose voice disorders. moreover, adding to previously established databases is recommended, which includes adding different languages, such as the Arabic voice pathology database, to other mainstream repositories.</p>
        </sec>
      </sec>
      <sec>
        <title>Strengths</title>
        <p>The key strength of this review is that it follows the DTA systematic review and search strategy. First, this review was in accordance with the Cochrane Library DTA systematic reviews, and second, it used a variety of medical, computer, and engineering databases. This increased the sensitivity of the review and broadened the search, overcoming the limited number of related articles. Moreover, in the screening process, in cases where the relevance of the abstract was not clear, the study was included in the full-text scanning. This eliminated any chance of eliminating relevant articles from the review. In addition, in the reference standard test, the inclusion criteria were restricted to a controlled environment, which might have ensured a more accurate and reliable result.</p>
        <p>This is the first review to systematically assess the performance of different ML algorithms in the assessment and diagnosis of voice disorders. A total of 13 observational studies were included, which recruited patients from both genders and different age groups (13-85 years). In all, 14 ML techniques were tested, 9 of which were included in the meta-analysis, and their pooled accuracies, sensitivities, and specificities were estimated.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>The main weakness of this review is the limited reporting by primary studies; for example, the criteria for selecting voice samples from the databases or the patient recruitment process, the poor reporting of the demographic characteristics of the sample, and the severity of the voice disorders in each case. This hindered the ability to find sources of heterogeneity, as subgroup analysis based on gender, age group or type, or severity of each disease could not be investigated. Furthermore, the main outcomes of the review could not be more specific to a certain gender or age group or the type or severity of the disease. Mentioning these details could have allowed for further investigation of which factors—voice disorders, gender, or age group—would determine the accuracy of ML performance. In the patient selection domain, more than half (8/13, 60%) of the included studies demonstrated an unclear risk of bias. The poor reporting of how voice samples were chosen from the database led to the estimated accuracy being subject to bias. The bias increased when the voice samples were not chosen randomly, as they might have been chosen based on unreported severities. However, removing these studies from the meta-analysis was not possible owing to the limited number of included studies.</p>
        <p>All included studies (13/13, 100%) failed to report how the reference standard was used, thus leading to an “unclear” risk of bias assessment in the overall reference standard. This is mainly due to the use of voice samples from a database; therefore, the clinicians’ assessment was not performed by the authors of the primary studies. Moreover, the clinicians’ assessment, which was applied by the chosen database, was not reported in the studies. Not knowing how the assessment was performed increased the risk of bias, and the outcome of the review was found to be unclear. Although the authors were contacted to request further details about the choice of voice samples and reference standard assessment, no response was received. Poor reporting led to an unclear risk of bias in the flow and timing of patients in almost all included studies (12/13, 92%), especially the lack of reporting of the time intervals between clinicians’ assessment and the recording of patients’ voices. For example, if the recordings were made at intervals of a few months after the clinician’s assessment, the patients’ condition could have changed from when the first recording was made. Consequently, this increased the chance of misclassification or misdiagnosis, as the voice sample diagnosis could be different from the clinician’s diagnosis. Better reporting of patients’ diagnosis and recruitment process would lead to a clearer risk of bias assessment.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>ML showed promising findings in screening, as its accuracy, sensitivity, and specificity showed high performance. The findings also suggested that ML can be further used in new smartphone apps for screening purposes and that screening can be conducted on the web. In scholarly research, more research with specific patient demographics and disorders is recommended. However, definitive conclusions could not be drawn about the effectiveness of ML in diagnosing owing to the limited number of studies (only 1). Therefore, we recommend using ML as a decision-support tool for clinicians during screening. For more definitive conclusions regarding the use of ML in diagnosis, more studies are suggested to be conducted, and risk of bias assessment that suits the application of ML for medical purposes and supervised ML is encouraged.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Search strategy in Web of Science.</p>
        <media xlink:href="jmir_v24i10e38472_app1.png" xlink:title="PNG File , 196 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Search strategy in MEDLINE.</p>
        <media xlink:href="jmir_v24i10e38472_app2.png" xlink:title="PNG File , 117 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>Extraction table.</p>
        <media xlink:href="jmir_v24i10e38472_app3.docx" xlink:title="DOCX File , 14 KB"/>
      </supplementary-material>
      <supplementary-material id="app4">
        <label>Multimedia Appendix 4</label>
        <p>Quality Assessment of Diagnostic Accuracy Studies 2 tool; patient selection domain.</p>
        <media xlink:href="jmir_v24i10e38472_app4.png" xlink:title="PNG File , 97 KB"/>
      </supplementary-material>
      <supplementary-material id="app5">
        <label>Multimedia Appendix 5</label>
        <p>Quality Assessment of Diagnostic Accuracy Studies 2 tool; index test domain.</p>
        <media xlink:href="jmir_v24i10e38472_app5.png" xlink:title="PNG File , 99 KB"/>
      </supplementary-material>
      <supplementary-material id="app6">
        <label>Multimedia Appendix 6</label>
        <p>Quality Assessment of Diagnostic Accuracy Studies 2 tool; reference standard domain.</p>
        <media xlink:href="jmir_v24i10e38472_app6.png" xlink:title="PNG File , 101 KB"/>
      </supplementary-material>
      <supplementary-material id="app7">
        <label>Multimedia Appendix 7</label>
        <p>Quality Assessment of Diagnostic Accuracy Studies 2 tool; flow and timing domain.</p>
        <media xlink:href="jmir_v24i10e38472_app7.png" xlink:title="PNG File , 102 KB"/>
      </supplementary-material>
      <supplementary-material id="app8">
        <label>Multimedia Appendix 8</label>
        <p>Modified Quality Assessment of Diagnostic Accuracy Studies 2 tool for this systematic review.</p>
        <media xlink:href="jmir_v24i10e38472_app8.png" xlink:title="PNG File , 200 KB"/>
      </supplementary-material>
      <supplementary-material id="app9">
        <label>Multimedia Appendix 9</label>
        <p>Quality Assessment of Diagnostic Accuracy Studies 2 tool risk of bias judgment in each included study across all domains and applicability concerns.</p>
        <media xlink:href="jmir_v24i10e38472_app9.png" xlink:title="PNG File , 96 KB"/>
      </supplementary-material>
      <supplementary-material id="app10">
        <label>Multimedia Appendix 10</label>
        <p>Summary of the performance of machine learning algorithms that were used for diagnosis.</p>
        <media xlink:href="jmir_v24i10e38472_app10.png" xlink:title="PNG File , 40 KB"/>
      </supplementary-material>
      <supplementary-material id="app11">
        <label>Multimedia Appendix 11</label>
        <p>Summary of the performance of machine learning algorithms that were used in screening.</p>
        <media xlink:href="jmir_v24i10e38472_app11.png" xlink:title="PNG File , 39 KB"/>
      </supplementary-material>
      <supplementary-material id="app12">
        <label>Multimedia Appendix 12</label>
        <p>The heterogeneity values after removing heterogenous studies from each forest plot.</p>
        <media xlink:href="jmir_v24i10e38472_app12.png" xlink:title="PNG File , 62 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">DTA</term>
          <def>
            <p>diagnostic test accuracy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">K-NN</term>
          <def>
            <p>K-nearest neighbor</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">LS-SVM</term>
          <def>
            <p>least-squares support-vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">MEEI</term>
          <def>
            <p>Massachusetts Eye and Ear Infirmary</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">ML</term>
          <def>
            <p>machine learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">QDA</term>
          <def>
            <p>quadratic discriminant analysis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">QUADAS</term>
          <def>
            <p>Quality Assessment of Diagnostic Accuracy Studies</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">SLT</term>
          <def>
            <p>speech and language therapist</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">UPM</term>
          <def>
            <p>Universidad Autónoma de Madrid</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>The authors would like to express their sincere gratitude to Munerah Al-Abdulsalam, biomedical engineer, for her support and guidance.</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Baken</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Orlikoff</surname>
              <given-names>RF</given-names>
            </name>
          </person-group>
          <source>Clinical Measurement of Speech and Voice. 2nd edition</source>
          <year>2000</year>
          <publisher-loc>Baltimore, MD, USA</publisher-loc>
          <publisher-name>Singular Publishing</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lyberg-Åhlander</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Rydell</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Fredlund</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Magnusson</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wilén</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Prevalence of voice disorders in the general population, based on the Stockholm public health cohort</article-title>
          <source>J Voice</source>
          <year>2019</year>
          <month>11</month>
          <volume>33</volume>
          <issue>6</issue>
          <fpage>900</fpage>
          <lpage>5</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2018.07.007</pub-id>
          <pub-id pub-id-type="medline">30126692</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(18)30175-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bhattacharyya</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>The prevalence of voice problems among adults in the United States</article-title>
          <source>Laryngoscope</source>
          <year>2014</year>
          <month>10</month>
          <volume>124</volume>
          <issue>10</issue>
          <fpage>2359</fpage>
          <lpage>62</lpage>
          <pub-id pub-id-type="doi">10.1002/lary.24740</pub-id>
          <pub-id pub-id-type="medline">24782443</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zabret</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hočevar Boltežar</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Šereg Bahar</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>The importance of the occupational vocal load for the occurence and treatment of organic voice disorders</article-title>
          <source>Zdr Varst</source>
          <year>2018</year>
          <month>03</month>
          <volume>57</volume>
          <issue>1</issue>
          <fpage>17</fpage>
          <lpage>24</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/29651311"/>
          </comment>
          <pub-id pub-id-type="doi">10.2478/sjph-2018-0003</pub-id>
          <pub-id pub-id-type="medline">29651311</pub-id>
          <pub-id pub-id-type="pii">sjph-2018-0003</pub-id>
          <pub-id pub-id-type="pmcid">PMC5894365</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Roy</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Asche</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Courey</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>The impact of laryngeal disorders on work-related dysfunction</article-title>
          <source>Laryngoscope</source>
          <year>2012</year>
          <month>07</month>
          <volume>122</volume>
          <issue>7</issue>
          <fpage>1589</fpage>
          <lpage>94</lpage>
          <pub-id pub-id-type="doi">10.1002/lary.23197</pub-id>
          <pub-id pub-id-type="medline">22549455</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Benninger</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Ahuja</surname>
              <given-names>AS</given-names>
            </name>
            <name name-style="western">
              <surname>Gardner</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Grywalski</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Assessing outcomes for dysphonic patients</article-title>
          <source>J Voice</source>
          <year>1998</year>
          <month>12</month>
          <volume>12</volume>
          <issue>4</issue>
          <fpage>540</fpage>
          <lpage>50</lpage>
          <pub-id pub-id-type="doi">10.1016/s0892-1997(98)80063-5</pub-id>
          <pub-id pub-id-type="medline">9988041</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(98)80063-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Roy</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Merrill</surname>
              <given-names>RM</given-names>
            </name>
            <name name-style="western">
              <surname>Gray</surname>
              <given-names>SD</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>EM</given-names>
            </name>
          </person-group>
          <article-title>Voice disorders in the general population: prevalence, risk factors, and occupational impact</article-title>
          <source>Laryngoscope</source>
          <year>2005</year>
          <month>11</month>
          <volume>115</volume>
          <issue>11</issue>
          <fpage>1988</fpage>
          <lpage>95</lpage>
          <pub-id pub-id-type="doi">10.1097/01.mlg.0000179174.32345.41</pub-id>
          <pub-id pub-id-type="medline">16319611</pub-id>
          <pub-id pub-id-type="pii">00005537-200511000-00016</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vilkman</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Voice problems at work: a challenge for occupational safety and health arrangement</article-title>
          <source>Folia Phoniatr Logop</source>
          <year>2000</year>
          <volume>52</volume>
          <issue>1-3</issue>
          <fpage>120</fpage>
          <lpage>5</lpage>
          <pub-id pub-id-type="doi">10.1159/000021519</pub-id>
          <pub-id pub-id-type="medline">10474011</pub-id>
          <pub-id pub-id-type="pii">21519</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Koufman</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Isaacson</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>The spectrum of vocal dysfunction</article-title>
          <source>Otolaryngol Clin North Am</source>
          <year>1991</year>
          <month>10</month>
          <volume>24</volume>
          <issue>5</issue>
          <fpage>985</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="medline">1754226</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Resouly</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Hope</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thomas</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A rapid access husky voice clinic: useful in diagnosing laryngeal pathology</article-title>
          <source>J Laryngol Otol</source>
          <year>2001</year>
          <month>12</month>
          <volume>115</volume>
          <issue>12</issue>
          <fpage>978</fpage>
          <lpage>80</lpage>
          <pub-id pub-id-type="doi">10.1258/0022215011909800</pub-id>
          <pub-id pub-id-type="medline">11779327</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dankovičová</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Sovák</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Drotár</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Vokorokos</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Machine learning approach to dysphonia detection</article-title>
          <source>Appl Sci</source>
          <year>2018</year>
          <month>10</month>
          <day>15</day>
          <volume>8</volume>
          <issue>10</issue>
          <fpage>1927</fpage>
          <pub-id pub-id-type="doi">10.3390/app8101927</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tsanas</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Little</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>McSharry</surname>
              <given-names>PE</given-names>
            </name>
            <name name-style="western">
              <surname>Ramig</surname>
              <given-names>LO</given-names>
            </name>
          </person-group>
          <article-title>Nonlinear speech analysis algorithms mapped to a standard metric achieve clinically useful quantification of average Parkinson's disease symptom severity</article-title>
          <source>J R Soc Interface</source>
          <year>2011</year>
          <month>06</month>
          <day>06</day>
          <volume>8</volume>
          <issue>59</issue>
          <fpage>842</fpage>
          <lpage>55</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/21084338"/>
          </comment>
          <pub-id pub-id-type="doi">10.1098/rsif.2010.0456</pub-id>
          <pub-id pub-id-type="medline">21084338</pub-id>
          <pub-id pub-id-type="pii">rsif.2010.0456</pub-id>
          <pub-id pub-id-type="pmcid">PMC3104343</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="web">
          <article-title>Autism Spectrum Disorder</article-title>
          <source>American Speech-Language-Hearing Association</source>
          <year>2022</year>
          <access-date>2022-03-08</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.asha.org/practice-portal/clinical-topics/autism/">https://www.asha.org/practice-portal/clinical-topics/autism/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kempster</surname>
              <given-names>GB</given-names>
            </name>
            <name name-style="western">
              <surname>Gerratt</surname>
              <given-names>BR</given-names>
            </name>
            <name name-style="western">
              <surname>Verdolini Abbott</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Barkmeier-Kraemer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hillman</surname>
              <given-names>RE</given-names>
            </name>
          </person-group>
          <article-title>Consensus auditory-perceptual evaluation of voice: development of a standardized clinical protocol</article-title>
          <source>Am J Speech Lang Pathol</source>
          <year>2009</year>
          <month>05</month>
          <volume>18</volume>
          <issue>2</issue>
          <fpage>124</fpage>
          <lpage>32</lpage>
          <pub-id pub-id-type="doi">10.1044/1058-0360(2008/08-0017)</pub-id>
          <pub-id pub-id-type="medline">18930908</pub-id>
          <pub-id pub-id-type="pii">1058-0360_2008_08-0017</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fawcus</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Fawcus</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Hyperfunctional voice: the misuse and abuse syndrome</article-title>
          <source>Voice Disorders and their Management</source>
          <year>1991</year>
          <publisher-loc>New York, NY, USA</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>139</fpage>
          <lpage>75</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Stachler</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Francis</surname>
              <given-names>DO</given-names>
            </name>
            <name name-style="western">
              <surname>Schwartz</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Damask</surname>
              <given-names>CC</given-names>
            </name>
            <name name-style="western">
              <surname>Digoy</surname>
              <given-names>GP</given-names>
            </name>
            <name name-style="western">
              <surname>Krouse</surname>
              <given-names>HJ</given-names>
            </name>
            <name name-style="western">
              <surname>McCoy</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ouellette</surname>
              <given-names>DR</given-names>
            </name>
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>RR</given-names>
            </name>
            <name name-style="western">
              <surname>Reavis</surname>
              <given-names>CC</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>LJ</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Strode</surname>
              <given-names>SW</given-names>
            </name>
            <name name-style="western">
              <surname>Woo</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nnacheta</surname>
              <given-names>LC</given-names>
            </name>
          </person-group>
          <article-title>Clinical practice guideline: hoarseness (dysphonia) (update) executive summary</article-title>
          <source>Otolaryngol Head Neck Surg</source>
          <year>2018</year>
          <month>03</month>
          <volume>158</volume>
          <issue>3</issue>
          <fpage>409</fpage>
          <lpage>26</lpage>
          <pub-id pub-id-type="doi">10.1177/0194599817751031</pub-id>
          <pub-id pub-id-type="medline">29494316</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>RR</given-names>
            </name>
            <name name-style="western">
              <surname>Awan</surname>
              <given-names>SN</given-names>
            </name>
            <name name-style="western">
              <surname>Barkmeier-Kraemer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Courey</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Deliyski</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Eadie</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Paul</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Švec</surname>
              <given-names>JG</given-names>
            </name>
            <name name-style="western">
              <surname>Hillman</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Recommended protocols for instrumental assessment of voice: American speech-language-hearing association expert panel to develop a protocol for instrumental assessment of vocal function</article-title>
          <source>Am J Speech Lang Pathol</source>
          <year>2018</year>
          <month>08</month>
          <day>06</day>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>887</fpage>
          <lpage>905</lpage>
          <pub-id pub-id-type="doi">10.1044/2018_AJSLP-17-0009</pub-id>
          <pub-id pub-id-type="medline">29955816</pub-id>
          <pub-id pub-id-type="pii">2686671</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Duffy</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <source>Motor Speech Disorders: Substrates, Differential Diagnosis, and Management</source>
          <year>2019</year>
          <publisher-loc>Boston, MA, USA</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Maier</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Haderlein</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Eysholdt</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Rosanowski</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Batliner</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schuster</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Nöth</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>PEAKS – a system for the automatic evaluation of voice and speech disorders</article-title>
          <source>Speech Commun</source>
          <year>2009</year>
          <month>5</month>
          <volume>51</volume>
          <issue>5</issue>
          <fpage>425</fpage>
          <lpage>37</lpage>
          <pub-id pub-id-type="doi">10.1016/j.specom.2009.01.004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bassich</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ludlow</surname>
              <given-names>CL</given-names>
            </name>
          </person-group>
          <article-title>The use of perceptual methods by new clinicians for assessing voice quality</article-title>
          <source>J Speech Hear Disord</source>
          <year>1986</year>
          <month>05</month>
          <volume>51</volume>
          <issue>2</issue>
          <fpage>125</fpage>
          <lpage>33</lpage>
          <pub-id pub-id-type="doi">10.1044/jshd.5102.125</pub-id>
          <pub-id pub-id-type="medline">3702360</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hegde</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Shetty</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Rai</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Dodderi</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>A survey on machine learning approaches for automatic detection of voice disorders</article-title>
          <source>J Voice</source>
          <year>2019</year>
          <month>11</month>
          <volume>33</volume>
          <issue>6</issue>
          <fpage>947.e11</fpage>
          <lpage>33</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2018.07.014</pub-id>
          <pub-id pub-id-type="medline">30316551</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(18)30143-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="web">
          <article-title>Lab Database &#124; Laboratory on the Physiology, Acoustics, and Perception of Speech</article-title>
          <source>MIT OpenCourseWare</source>
          <year>2022</year>
          <access-date>2022-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ocw.mit.edu/courses/electrical-engineering-and-computer-science/6-542j-laboratory-on-the-physiology-acoustics-and-perception-of-speech-fall-2005/lab-database/">https://ocw.mit.edu/courses/electrical-engineering-and-computer-science/6-542j-laboratory-on-the-physiology-acoustics-​and-perception-of-speech-fall-2005/lab-database/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pützer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Barry</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Saarbrücken Voice Database-Handbook</article-title>
          <source>Deutsche Forschungsgemeinschaft</source>
          <year>2022</year>
          <access-date>2022-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.stimmdatenbank.coli.uni-saarland.de/help_en.php4">http://www.stimmdatenbank.coli.uni-saarland.de/help_en.php4</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jackson-Menaldi</surname>
              <given-names>MC</given-names>
            </name>
          </person-group>
          <source>La voz patológica</source>
          <year>2002</year>
          <publisher-loc>Madrid, Spain</publisher-loc>
          <publisher-name>Editorial Médica Panamericana</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Larrañaga</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Calvo</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Santana</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bielza</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Galdiano</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Inza</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Lozano</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Armañanzas</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Santafé</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Pérez</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Robles</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Machine learning in bioinformatics</article-title>
          <source>Brief Bioinform</source>
          <year>2006</year>
          <month>03</month>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>86</fpage>
          <lpage>112</lpage>
          <pub-id pub-id-type="doi">10.1093/bib/bbk007</pub-id>
          <pub-id pub-id-type="medline">16761367</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rauschert</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Raubenheimer</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Melton</surname>
              <given-names>PE</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>RC</given-names>
            </name>
          </person-group>
          <article-title>Machine learning and clinical epigenetics: a review of challenges for diagnosis and classification</article-title>
          <source>Clin Epigenetics</source>
          <year>2020</year>
          <month>04</month>
          <day>03</day>
          <volume>12</volume>
          <issue>1</issue>
          <fpage>51</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://clinicalepigeneticsjournal.biomedcentral.com/articles/10.1186/s13148-020-00842-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s13148-020-00842-4</pub-id>
          <pub-id pub-id-type="medline">32245523</pub-id>
          <pub-id pub-id-type="pii">10.1186/s13148-020-00842-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC7118917</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Syed</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Rashid</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hussain</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Meta-analysis of voice disorders databases and applied machine learning techniques</article-title>
          <source>Math Biosci Eng</source>
          <year>2020</year>
          <month>11</month>
          <day>11</day>
          <volume>17</volume>
          <issue>6</issue>
          <fpage>7958</fpage>
          <lpage>79</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aimspress.com/article/10.3934/mbe.2020404"/>
          </comment>
          <pub-id pub-id-type="doi">10.3934/mbe.2020404</pub-id>
          <pub-id pub-id-type="medline">33378928</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Macaskill</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Gatsonis</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Deeks</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Harbord</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Takwoingi</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Cochrane handbook for systematic reviews of diagnostic test accuracy</article-title>
          <source>Cochrane Methods</source>
          <year>2010</year>
          <access-date>2022-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://methods.cochrane.org/sdt/">https://methods.cochrane.org/sdt/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Whiting</surname>
              <given-names>PF</given-names>
            </name>
            <name name-style="western">
              <surname>Rutjes</surname>
              <given-names>AW</given-names>
            </name>
            <name name-style="western">
              <surname>Westwood</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Mallett</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Deeks</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Reitsma</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Leeflang</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Sterne</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Bossuyt</surname>
              <given-names>PM</given-names>
            </name>
            <collab>QUADAS-2 Group</collab>
          </person-group>
          <article-title>QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies</article-title>
          <source>Ann Intern Med</source>
          <year>2011</year>
          <month>10</month>
          <day>18</day>
          <volume>155</volume>
          <issue>8</issue>
          <fpage>529</fpage>
          <lpage>36</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.acpjournals.org/doi/abs/10.7326/0003-4819-155-8-201110180-00009?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub%3dpubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.7326/0003-4819-155-8-201110180-00009</pub-id>
          <pub-id pub-id-type="medline">22007046</pub-id>
          <pub-id pub-id-type="pii">155/8/529</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Deeks</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Bossuyt</surname>
              <given-names>PM</given-names>
            </name>
            <name name-style="western">
              <surname>Gatsonis</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Cochrane handbook for systematic reviews of diagnostic test accuracy</article-title>
          <source>The Cochrane Collaboration</source>
          <year>2009</year>
          <access-date>2022-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://methods.cochrane.org/sdt/">https://methods.cochrane.org/sdt/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Newcombe</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>Two-sided confidence intervals for the single proportion: comparison of seven methods</article-title>
          <source>Stat Med</source>
          <year>1998</year>
          <month>04</month>
          <day>30</day>
          <volume>17</volume>
          <issue>8</issue>
          <fpage>857</fpage>
          <lpage>72</lpage>
          <pub-id pub-id-type="doi">10.1002/(sici)1097-0258(19980430)17:8&#60;857::aid-sim777&#62;3.0.co;2-e</pub-id>
          <pub-id pub-id-type="medline">9595616</pub-id>
          <pub-id pub-id-type="pii">10.1002/(SICI)1097-0258(19980430)17:8&#60;857::AID-SIM777&#62;3.0.CO;2-E</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Freeman</surname>
              <given-names>MF</given-names>
            </name>
            <name name-style="western">
              <surname>Tukey</surname>
              <given-names>JW</given-names>
            </name>
          </person-group>
          <article-title>Transformations related to the angular and the square root</article-title>
          <source>Ann Math Statist</source>
          <year>1950</year>
          <month>12</month>
          <volume>21</volume>
          <issue>4</issue>
          <fpage>607</fpage>
          <lpage>11</lpage>
          <pub-id pub-id-type="doi">10.1214/aoms/1177729756</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Higgins</surname>
              <given-names>JP</given-names>
            </name>
            <name name-style="western">
              <surname>Thompson</surname>
              <given-names>SG</given-names>
            </name>
            <name name-style="western">
              <surname>Deeks</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Altman</surname>
              <given-names>DG</given-names>
            </name>
          </person-group>
          <article-title>Measuring inconsistency in meta-analyses</article-title>
          <source>BMJ</source>
          <year>2003</year>
          <month>09</month>
          <day>06</day>
          <volume>327</volume>
          <issue>7414</issue>
          <fpage>557</fpage>
          <lpage>60</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/12958120"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmj.327.7414.557</pub-id>
          <pub-id pub-id-type="medline">12958120</pub-id>
          <pub-id pub-id-type="pii">327/7414/557</pub-id>
          <pub-id pub-id-type="pmcid">PMC192859</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Akbari</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Arjmandi</surname>
              <given-names>MK</given-names>
            </name>
          </person-group>
          <article-title>Employing linear prediction residual signal of wavelet sub-bands in automatic detection of laryngeal pathology</article-title>
          <source>Biomed Signal Process Control</source>
          <year>2015</year>
          <month>04</month>
          <volume>18</volume>
          <fpage>293</fpage>
          <lpage>302</lpage>
          <pub-id pub-id-type="doi">10.1016/j.bspc.2015.02.008</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arias-Londoño</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Godino-Llorente</surname>
              <given-names>JI</given-names>
            </name>
            <name name-style="western">
              <surname>Markaki</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Stylianou</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>On combining information from modulation spectra and mel-frequency cepstral coefficients for automatic detection of pathological voices</article-title>
          <source>Logoped Phoniatr Vocol</source>
          <year>2011</year>
          <month>07</month>
          <volume>36</volume>
          <issue>2</issue>
          <fpage>60</fpage>
          <lpage>9</lpage>
          <pub-id pub-id-type="doi">10.3109/14015439.2010.528788</pub-id>
          <pub-id pub-id-type="medline">21073260</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arjmandi</surname>
              <given-names>MK</given-names>
            </name>
            <name name-style="western">
              <surname>Pooyan</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>An optimum algorithm in pathological voice quality assessment using wavelet-packet-based features, linear discriminant analysis and support vector machine</article-title>
          <source>Biomed Signal Process Control</source>
          <year>2012</year>
          <month>1</month>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>3</fpage>
          <lpage>19</lpage>
          <pub-id pub-id-type="doi">10.1016/j.bspc.2011.03.010</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arjmandi</surname>
              <given-names>MK</given-names>
            </name>
            <name name-style="western">
              <surname>Pooyan</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mikaili</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Vali</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Moqarehzadeh</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Identification of voice disorders using long-time features and support vector machine with different feature reduction methods</article-title>
          <source>J Voice</source>
          <year>2011</year>
          <month>11</month>
          <volume>25</volume>
          <issue>6</issue>
          <fpage>e275</fpage>
          <lpage>89</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2010.08.003</pub-id>
          <pub-id pub-id-type="medline">21186096</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(10)00157-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cordeiro</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Fonseca</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Guimarães</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Meneses</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Hierarchical classification and system combination for automatically identifying physiological and neuromuscular laryngeal pathologies</article-title>
          <source>J Voice</source>
          <year>2017</year>
          <month>05</month>
          <volume>31</volume>
          <issue>3</issue>
          <fpage>384.e9</fpage>
          <lpage>14</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2016.09.003</pub-id>
          <pub-id pub-id-type="medline">27743845</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(16)30265-X</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ghasemzadeh</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Tajik Khass</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Khalil Arjmandi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pooyan</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Detection of vocal disorders based on phase space parameters and Lyapunov spectrum</article-title>
          <source>Biomed Signal Process Control</source>
          <year>2015</year>
          <month>09</month>
          <volume>22</volume>
          <fpage>135</fpage>
          <lpage>45</lpage>
          <pub-id pub-id-type="doi">10.1016/j.bspc.2015.07.002</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Godino-Llorente</surname>
              <given-names>JI</given-names>
            </name>
            <name name-style="western">
              <surname>Gómez-Vilda</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Automatic detection of voice impairments by means of short-term cepstral parameters and neural network based detectors</article-title>
          <source>IEEE Trans Biomed Eng</source>
          <year>2004</year>
          <month>02</month>
          <volume>51</volume>
          <issue>2</issue>
          <fpage>380</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1109/TBME.2003.820386</pub-id>
          <pub-id pub-id-type="medline">14765711</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hadjitodorov</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Boyanov</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Teston</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Laryngeal pathology detection by means of class-specific neural maps</article-title>
          <source>IEEE Trans Inf Technol Biomed</source>
          <year>2000</year>
          <month>03</month>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>68</fpage>
          <lpage>73</lpage>
          <pub-id pub-id-type="doi">10.1109/4233.826861</pub-id>
          <pub-id pub-id-type="medline">10761776</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hariharan</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Polat</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Yaacob</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A new feature constituting approach to detection of vocal fold pathology</article-title>
          <source>Int J Syst Sci</source>
          <year>2014</year>
          <volume>45</volume>
          <issue>8</issue>
          <fpage>1622</fpage>
          <lpage>34</lpage>
          <pub-id pub-id-type="doi">10.1080/00207721.2013.794905</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lopes</surname>
              <given-names>LW</given-names>
            </name>
            <name name-style="western">
              <surname>Batista Simões</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Delfino da Silva</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>da Silva Evangelista</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>da Nóbrega E Ugulino</surname>
              <given-names>AC</given-names>
            </name>
            <name name-style="western">
              <surname>Oliveira Costa Silva</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Jefferson Dias Vieira</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Accuracy of acoustic analysis measurements in the evaluation of patients with different laryngeal diagnoses</article-title>
          <source>J Voice</source>
          <year>2017</year>
          <month>05</month>
          <volume>31</volume>
          <issue>3</issue>
          <fpage>382.e15</fpage>
          <lpage>26</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2016.08.015</pub-id>
          <pub-id pub-id-type="medline">27742492</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(16)30158-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mohammed</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Abdulkareem</surname>
              <given-names>KH</given-names>
            </name>
            <name name-style="western">
              <surname>Mostafa</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Khanapi Abd Ghani</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Maashi</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Garcia-Zapirain</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Oleagordia</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Alhakami</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>AL-Dhief</surname>
              <given-names>FT</given-names>
            </name>
          </person-group>
          <article-title>Voice pathology detection and classification using convolutional neural network model</article-title>
          <source>Appl Sci</source>
          <year>2020</year>
          <month>05</month>
          <day>27</day>
          <volume>10</volume>
          <issue>11</issue>
          <fpage>3723</fpage>
          <pub-id pub-id-type="doi">10.3390/app10113723</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Souissi</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Cherif</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Artificial neural networks and support vector machine for voice disorders identification</article-title>
          <source>Int J Adv Comput Sci Appl</source>
          <year>2016</year>
          <volume>7</volume>
          <issue>5</issue>
          <fpage>339</fpage>
          <lpage>44</lpage>
          <pub-id pub-id-type="doi">10.14569/ijacsa.2016.070546</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Discrimination between pathological and normal voices using GMM-SVM approach</article-title>
          <source>J Voice</source>
          <year>2011</year>
          <month>01</month>
          <volume>25</volume>
          <issue>1</issue>
          <fpage>38</fpage>
          <lpage>43</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2009.08.002</pub-id>
          <pub-id pub-id-type="medline">20137892</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(09)00119-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="web">
          <article-title>Systematic reviews: CRD's guidance for undertaking reviews in healthcare</article-title>
          <source>Centre for Reviews and Dissemination</source>
          <year>2009</year>
          <access-date>2022-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.york.ac.uk/media/crd/Systematic_Reviews.pdf">https://www.york.ac.uk/media/crd/Systematic_Reviews.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hariharan</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Polat</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Sindhu</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yaacob</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A hybrid expert system approach for telemonitoring of vocal fold pathology</article-title>
          <source>Appl Soft Comput</source>
          <year>2013</year>
          <month>10</month>
          <volume>13</volume>
          <issue>10</issue>
          <fpage>4148</fpage>
          <lpage>61</lpage>
          <pub-id pub-id-type="doi">10.1016/j.asoc.2013.06.004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mehta</surname>
              <given-names>DD</given-names>
            </name>
            <name name-style="western">
              <surname>Hillman</surname>
              <given-names>RE</given-names>
            </name>
          </person-group>
          <article-title>Voice assessment: updates on perceptual, acoustic, aerodynamic, and endoscopic imaging methods</article-title>
          <source>Curr Opin Otolaryngol Head Neck Surg</source>
          <year>2008</year>
          <month>06</month>
          <volume>16</volume>
          <issue>3</issue>
          <fpage>211</fpage>
          <lpage>5</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/18475073"/>
          </comment>
          <pub-id pub-id-type="doi">10.1097/MOO.0b013e3282fe96ce</pub-id>
          <pub-id pub-id-type="medline">18475073</pub-id>
          <pub-id pub-id-type="pii">00020840-200806000-00007</pub-id>
          <pub-id pub-id-type="pmcid">PMC3775647</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mesallam</surname>
              <given-names>TA</given-names>
            </name>
            <name name-style="western">
              <surname>Farahat</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Malki</surname>
              <given-names>KH</given-names>
            </name>
            <name name-style="western">
              <surname>Alsulaiman</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Nasheri</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Muhammad</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Development of the Arabic voice pathology database and its evaluation by using speech features and machine learning algorithms</article-title>
          <source>J Healthc Eng</source>
          <year>2017</year>
          <volume>2017</volume>
          <fpage>8783751</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1155/2017/8783751"/>
          </comment>
          <pub-id pub-id-type="doi">10.1155/2017/8783751</pub-id>
          <pub-id pub-id-type="medline">29201333</pub-id>
          <pub-id pub-id-type="pmcid">PMC5672151</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arias-Londoño</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Godino-Llorente</surname>
              <given-names>JI</given-names>
            </name>
            <name name-style="western">
              <surname>Sáenz-Lechón</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Osma-Ruiz</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Castellanos-Domínguez</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>An improved method for voice pathology detection by means of a HMM-based feature space transformation</article-title>
          <source>Pattern Recognit</source>
          <year>2010</year>
          <month>9</month>
          <volume>43</volume>
          <issue>9</issue>
          <fpage>3100</fpage>
          <lpage>12</lpage>
          <pub-id pub-id-type="doi">10.1016/j.patcog.2010.03.019</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Harar</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Galaz</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Alonso-Hernandez</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Mekyska</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Burget</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Smekal</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Towards robust voice pathology detection</article-title>
          <source>Neural Comput Applic</source>
          <year>2018</year>
          <month>04</month>
          <day>04</day>
          <volume>32</volume>
          <issue>20</issue>
          <fpage>15747</fpage>
          <lpage>57</lpage>
          <pub-id pub-id-type="doi">10.1007/s00521-018-3464-7</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
