<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v22i7e18697</article-id>
      <article-id pub-id-type="pmid">32673247</article-id>
      <article-id pub-id-type="doi">10.2196/18697</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Diagnosing Parkinson Disease Through Facial Expression Recognition: Video Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhang</surname>
            <given-names>Ping</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Benis</surname>
            <given-names>Arriel</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Jin</surname>
            <given-names>Bo</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4094-7499</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Qu</surname>
            <given-names>Yue</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2133-4208</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>Liang</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <address>
            <institution>Dongbei University of Finance and Economics</institution>
            <addr-line>217 Jianshan St, Shahekou District</addr-line>
            <addr-line>Dalian</addr-line>
            <country>China</country>
            <phone>86 15524709655</phone>
            <email>liang.zhang@dufe.edu.cn</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-5055-1527</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Gao</surname>
            <given-names>Zhan</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1751-5153</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Dalian University of Technology</institution>
        <addr-line>Dalian</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Dongbei University of Finance and Economics</institution>
        <addr-line>Dalian</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Beijing Haoyisheng Cloud Hospital Management Technology Ltd</institution>
        <addr-line>Beijing</addr-line>
        <country>China</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Liang Zhang <email>liang.zhang@dufe.edu.cn</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>7</month>
        <year>2020</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>10</day>
        <month>7</month>
        <year>2020</year>
      </pub-date>
      <volume>22</volume>
      <issue>7</issue>
      <elocation-id>e18697</elocation-id>
      <history>
        <date date-type="received">
          <day>12</day>
          <month>3</month>
          <year>2020</year>
        </date>
        <date date-type="rev-request">
          <day>3</day>
          <month>4</month>
          <year>2020</year>
        </date>
        <date date-type="rev-recd">
          <day>15</day>
          <month>4</month>
          <year>2020</year>
        </date>
        <date date-type="accepted">
          <day>19</day>
          <month>4</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Bo Jin, Yue Qu, Liang Zhang, Zhan Gao. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 10.07.2020.</copyright-statement>
      <copyright-year>2020</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2020/7/e18697" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>The number of patients with neurological diseases is currently increasing annually, which presents tremendous challenges for both patients and doctors. With the advent of advanced information technology, digital medical care is gradually changing the medical ecology. Numerous people are exploring new ways to receive a consultation, track their diseases, and receive rehabilitation training in more convenient and efficient ways. In this paper, we explore the use of facial expression recognition via artificial intelligence to diagnose a typical neurological system disease, Parkinson disease (PD).</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study proposes methods to diagnose PD through facial expression recognition.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We collected videos of facial expressions of people with PD and matched controls. We used relative coordinates and positional jitter to extract facial expression features (facial expression amplitude and shaking of small facial muscle groups) from the key points returned by Face++. Algorithms from traditional machine learning and advanced deep learning were utilized to diagnose PD.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The experimental results showed our models can achieve outstanding facial expression recognition ability for PD diagnosis. Applying a long short-term model neural network to the positions of the key features, precision and F1 values of 86% and 75%, respectively, can be reached. Further, utilizing a support vector machine algorithm for the facial expression amplitude features and shaking of the small facial muscle groups, an F1 value of 99% can be achieved.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This study contributes to the digital diagnosis of PD based on facial expression recognition. The disease diagnosis model was validated through our experiment. The results can help doctors understand the real-time dynamics of the disease and even conduct remote diagnosis.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>Parkinson disease</kwd>
        <kwd>face landmarks</kwd>
        <kwd>machine learning</kwd>
        <kwd>artificial intelligence</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>The population overall is currently aging. While an aging population represents the triumph of medical and social advances over disease, it also presents daunting challenges. Age is a crucial parameter for the occurrence, development, and diagnosis of diseases. As age increases, the central nervous system’s morphology, metabolism, and function undergo different degrees of decline, which results in certain neurological diseases [<xref ref-type="bibr" rid="ref1">1</xref>]. One typical example is Parkinson disease (PD), which is caused by a decrease in dopamine secretion. PD, also known as tremor palsy, is a common neurodegenerative disease; the manifestations are mainly bradykinesia, myotonia, resting tremors, and unstable posture [<xref ref-type="bibr" rid="ref2">2</xref>]. Clinical symptoms may also include nonmotor symptoms such as olfactory function decline, constipation, and depression. As a progressive disease, the various symptoms, both motor and nonmotor, become more serious as the disease course develops, and various complications, such as the “on-off” phenomenon, dyskinesia, and decline in drug efficacy, often occur in the later stages. Patients with serious illness may be plagued by balance disorders, frozen gait, falls, and speech disorders, resulting in an inability to take care of themselves and a decrease in quality of life [<xref ref-type="bibr" rid="ref3">3</xref>]. Famous people like Xiaoping Deng, Jingrun Chen, Jin Ba, and Muhammad Ali have all been deeply affected by PD. Hence, an accurate diagnosis and medication that works immediately are important.</p>
      <p>In recent years, with the development of computer vision technology, facial image recognition has been used for disease diagnosis. In 2017, the National Human Genome Research Institute developed facial diagnosis software to identify whether a child has DiGeorge syndrome [<xref ref-type="bibr" rid="ref4">4</xref>]. It is a very rare disease, with a pediatric incidence rate ranging from 1/6000 to 1/3000 worldwide. With such a wide range of morbidity, disease diagnosis can be difficult. However, all children with DiGeorge syndrome have clinically obvious facial features, which were utilized by the National Human Genome Research Institute for disease diagnosis. In 2016, FDNA Inc developed the Face2Gene system, which can help doctors diagnose genetic diseases via facial recognition [<xref ref-type="bibr" rid="ref5">5</xref>]. Some relatively rare genetic diseases can even be discovered by comparing photos of patients with healthy people. This helps patients detect their disease and obtain treatment in a relatively short time.</p>
      <p>For PD, the disease-specific facial expressions of patients have attracted researchers’ attention. In the 1860s, Charcot first described the characteristics of “masked face” in patients with PD [<xref ref-type="bibr" rid="ref6">6</xref>]. Based on years of research, “mask face” is considered one of the common symptoms of PD. Its symptoms involve a faceless and binocular gaze [<xref ref-type="bibr" rid="ref7">7</xref>]. Facial expressions in humans are expressed by the superficial muscles of the face. These muscles are called the “facial muscles” or “expression muscles.” Smiles and other facial expressions of patients with PD often appear to be unresponsive and have significantly smaller amplitude. The time it takes to form an expression is also extended. To many patients’ families, the “mask face” symptom causes issues because it is an expression of sullenness. Therefore, exploring a diagnostic method using facial feature point recognition is promising. Note that detection of a speech disorder in patients with PD, which is also a noninvasive diagnostic method, has been shown to be effective [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. Hence, facial expression recognition can be combined with speech disorder recognition to obtain a more comprehensive, multidirectional, noninvasive, remote diagnosis.</p>
      <p>Several researchers have utilized machine learning and computer vision technology to explore the inner relationship in the “mask face.” Bandini et al [<xref ref-type="bibr" rid="ref10">10</xref>] proposed an automatic method in 2017 to analyze videos of facial expressions in patients with PD. They extracted the average distances between facial key points using face tracing. Then, they found that patients with PD have much smaller distances of facial movement. In contrast to video-based research, Rajnoha et al [<xref ref-type="bibr" rid="ref11">11</xref>] designed an automatic detection method based on static facial images using convolutional neural network models. However, the best achieved accuracy was only 67.33%. In 2019, Langevin et al [<xref ref-type="bibr" rid="ref12">12</xref>] designed the Parkinson's Analysis with Remote Kinetic-tasks framework to analyze PD characteristics, finding that facial features and motion features in the Movement Disorder Society Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) could be extracted from videos.</p>
      <p>In order to explore facial expressions in humans and test whether they are accompanied by tremors, we need to extract the key points of the face. In this paper, we investigated the ability to diagnose PD by recognizing changes in key points of the face in a short video. At present, to the best of our knowledge, we are the first to utilize sequential changes in key points of the face to diagnose PD.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Data Collection</title>
        <p>People with PD usually suffer from loss of facial expression on both sides of the face. Moreover, their eye movement is reduced, like wearing a mask. This symptom is usually called the “mask face” [<xref ref-type="bibr" rid="ref13">13</xref>]. Motion retardation is one of the main symptoms of PD. When facial symptoms appear, movement in facial expressions is significantly reduced, accompanied by local tremor symptoms of the small tendon group. When patients with PD try to smile or form other expressions, the facial muscles move slowly and often show excessive expressions. Based on the main facial expression symptoms of patients with PD, we recorded the smiles of patients with PD. Each video was 5 seconds long, and we recorded 2-3 videos per patient. The expression of a smile can be used to distinguish a patient with PD from someone without PD by comparing the magnitude of the expression and trembling of the small muscles of the face.</p>
        <p>Patients with PD at the First Affiliated Hospital of Dalian Medical University were recruited and provided video data. Patients provided written informed consent before collecting data. Data for the control group, or people without PD, were randomly collected from senior people who liked to exercise frequently. Finally, we collected data from 64 senior citizens, including 33 people with PD. Each person recorded a smile expression 3 times. The length of every video was 5 seconds. After filtering out the videos that could not be used, we finally collected 176 records. The label is the subject status (ie, whether he or she had PD), as shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Smiles of a patient with Parkinson disease (left) and person without Parkinson disease (right).</p>
          </caption>
          <graphic xlink:href="jmir_v22i7e18697_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <sec>
          <title>Acquisition Equipment</title>
          <p>In order to make the captured video clear enough and not affect the experimental results due to human jitter, the videos were captured using a Canon 700D camera placed on a standard tripod.</p>
        </sec>
        <sec>
          <title>Collection Plan</title>
          <p>In the process of recording facial expressions, we showed emoticons, which were printed on photos, to the patients with PD and asked them to imitate the emoticons. This data collection process aimed to explore the vibration of the small muscle groups of the face to distinguish patients with PD from people without PD. The underlying reason for asking the participants to imitate an emoticon in a photo was to prevent the subjective emotions of the patients with PD from affecting the data collection. At the same time, it guaranteed consistency of data collection and ensured that each patient’s understanding of a smile or anger was identical. The data collection process was as follows. First, we recorded the participant’s gender, age, and history of other neurological diseases. Then, the subjects filled out the video data collection registration form, with full awareness of the PD detection research project. <xref ref-type="table" rid="table1">Table 1</xref> displays an example of the information collected via the registration form. Third, the emoticon photos were given to the subjects, and they were asked to imitate them. We collected 3 videos of smiling facial expressions, each of which lasted for 5 seconds. After all the recordings were completed, the videos were classified into 2 classes: patients with PD and people without PD. Different folders were sorted according to the patient ID.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Example data collected using the registration form to collect data via video of patients with Parkinson disease.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="130"/>
              <col width="100"/>
              <col width="100"/>
              <col width="120"/>
              <col width="190"/>
              <col width="200"/>
              <col width="160"/>
              <thead>
                <tr valign="top">
                  <td>Patient Number</td>
                  <td>Age (years)</td>
                  <td>Gender</td>
                  <td>Confirmed</td>
                  <td>Other neurological disease</td>
                  <td>Length of disease (month)</td>
                  <td>Date of collection</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>1</td>
                  <td>60</td>
                  <td>Male</td>
                  <td>Yes</td>
                  <td>No</td>
                  <td>10</td>
                  <td>11/13/2017</td>
                </tr>
                <tr valign="top">
                  <td>2</td>
                  <td>55</td>
                  <td>Male</td>
                  <td>Yes</td>
                  <td>No</td>
                  <td>24</td>
                  <td>11/13/2017</td>
                </tr>
                <tr valign="top">
                  <td>3</td>
                  <td>60</td>
                  <td>Male</td>
                  <td>Yes</td>
                  <td>No</td>
                  <td>10</td>
                  <td>11/13/2017</td>
                </tr>
                <tr valign="top">
                  <td>4</td>
                  <td>63</td>
                  <td>Female</td>
                  <td>Yes</td>
                  <td>No</td>
                  <td>14</td>
                  <td>11/13/2017</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
        </sec>
        <sec>
          <title>Converting Video to Images</title>
          <p>To obtain the participant’s facial information, we split the recorded video into individual frames that were extracted directly from the video every 0.1 seconds. In this experiment, we used ffmpeg to implement this function. ffmpeg is a set of open source computer programs that convert digital audio and video into streams [<xref ref-type="bibr" rid="ref14">14</xref>]. In practice, processing multiple videos in multiple folders involves the files’ operation commands. We leveraged a subprocess module in the python programming environment that encapsulates the running terminal commands, which perfectly met our needs. The subprocess package mainly executes external commands and programs and uses the function <italic>subprocess.call</italic>() to call external commands. The video was converted into several images according to a preset time interval and stored in the specified folder for analysis.</p>
        </sec>
      </sec>
      <sec>
        <title>Data Preprocessing</title>
        <p>Face++ is a well-known service for face recognition and facial landmark detection. There are many interesting applications for using Face++, such as inferring the demographics of social media users from profile pictures [<xref ref-type="bibr" rid="ref15">15</xref>]. We chose Face++ because of its outstanding ability to localize facial landmarks. Based on previous literature and market research on key points of faces, we found that the Face++ interface can provide 106 coordinate points to create very accurate faces at this stage.</p>
        <p>Users simply log in to the Face++ official website and apply for an API key. By calling the “Detect API” interface while using the “POST” method and setting the <italic>return_landmark</italic> parameter value to 2, the system will return 106 key points of the human face, as shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>. These 106 points cover most of the key points of a human’s face, including the facial contour, eyebrows, eyes, nose, and mouth that can describe almost all facial expressions. The returned coordinates are numbers using pixels as the unit. The top left corner of the image is used as the origin point.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Face key points (n=106) returned by the Face++ interface.</p>
          </caption>
          <graphic xlink:href="jmir_v22i7e18697_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>The facial muscles of patients with PD tend to be more rigid than those of people without PD, which causes a smaller facial expression range. We obtained the magnitude of the patient’s expression by calculating the coordinate positions’ range. However, during the video recording process, we found that although the video equipment is stable, it is difficult to ensure the participant, especially a patient with PD, does not move. To capture a relatively accurate magnitude of the facial expression, we converted the absolute coordinates into relative coordinates. Through the variation of the relative coordinates, the magnitude of the change in the facial expression of the patient can be reflected, and the error caused by postural changes of the body can be avoided.</p>
        <p>To transform absolute coordinates into relative coordinates, we used the midpoint between the inside corners of the eyes as the origin of the coordinate (0,0). Then, we set the line connecting the inside corners of the eye as the x axis and the line connecting the nose and the origin as the y axis. A non-Cartesian coordinate system could then be created. As for the units, the coordinates of the inside corners of both eyes were quantized as (–1,0) and (1,0), and the coordinates of the nose were quantized as (0,–1), as shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>. The black coordinate system was used to record the position of the pixels in the image, so we called this the absolute coordinate system. The blue coordinate system represents the relative coordinate system.</p>
        <p>Assuming that the unit vector along the x axis in the relative coordinate system is vector <italic>a</italic> (<italic>a</italic><sub>1</sub>, <italic>a</italic><sub>2</sub>) and the unit vector along the y axis direction is vector <italic>b</italic> (<italic>b</italic><sub>1</sub>, <italic>b</italic><sub>2</sub>), we only need to compute (<italic>x</italic>,<italic>y</italic>), as shown in <xref rid="figure4" ref-type="fig">Figure 4</xref>. In <xref rid="figure4" ref-type="fig">Figure 4</xref>, the absolute coordinates (<italic>m</italic><sub>1</sub>, <italic>n</italic><sub>1</sub>), (<italic>m</italic><sub>2</sub>, <italic>n</italic><sub>2</sub>), (<italic>a</italic><sub>1</sub>, <italic>a</italic><sub>2</sub>), (<italic>b</italic><sub>1</sub>, <italic>b</italic><sub>2</sub>) can be converted to the relative coordinates (0, 0), (<italic>x</italic>, <italic>y</italic>), (1, 0), (0, 1), respectively.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Relative face coordinate system, where the Cartesian, or absolute, coordinate system is represented by the black coordinate system, which was used to record the position of pixels in the image, and the non-Cartesian, or relative, coordinate system is represented by the blue coordinate system, which was used to record the relative position of key points on the face.</p>
          </caption>
          <graphic xlink:href="jmir_v22i7e18697_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Coordinate system conversion, where the absolute coordinates (m<sub>1</sub>,n<sub>1</sub>), (m<sub>2</sub>, n<sub>2</sub>), (a<sub>1</sub>, a<sub>2</sub>), and (b<sub>1</sub>, b<sub>2</sub>) can be converted to the relative coordinates (0,0), (x,y), (1,0), and (0,1), respectively.</p>
          </caption>
          <graphic xlink:href="jmir_v22i7e18697_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>The following relationships were satisfied between vectors:</p>
        <graphic xlink:href="jmir_v22i7e18697_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>After calculation, the converted relative coordinates (<italic>x</italic>, <italic>y</italic>) were:</p>
        <graphic xlink:href="jmir_v22i7e18697_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>At this time, the value of (<italic>x</italic>, <italic>y</italic>) was the coordinate position in the relative coordinate system, which was converted by (<italic>m</italic><sub>2</sub>, <italic>n</italic><sub>2</sub>) in the absolute coordinate system.</p>
      </sec>
      <sec>
        <title>Key Facial Feature Extraction</title>
        <p>Based on the facial expressions and tremors in patients with PD, we evaluated and extracted features from two main dimensions. One was the amplitude magnitude of the patient’s facial expression, which can be used to detect whether there is “mask face.” The other was judging the patient’s facial tremor through the face’s key points. When the tremor effects of some patients are obvious, head and elbow vibration will drive the head to perform regular tremors. For the 106 key points on the face, they were divided into two main parts, which were amplitude features of the facial expression and shaking features of the facial small muscle group.</p>
        <sec>
          <title>Extraction of the Features of Facial Expression Amplitude</title>
          <p>We defined a key point <italic>z</italic>’s position at the <italic>i</italic>th frame with a vector: <bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>,</sub><italic><sub>z</sub></italic>=(<italic>x<sub>i</sub></italic><sub>,</sub><italic><sub>z</sub></italic>,<italic>y<sub>i</sub></italic><sub>,</sub><italic><sub>z</sub></italic>), in which <italic>x<sub>i</sub></italic><sub>,</sub><italic><sub>z</sub></italic> and <italic>y<sub>i</sub></italic><sub>,</sub><italic><sub>z</sub></italic> are the relative position’s horizontal coordinate and vertical coordinate, respectively. It is called the <bold><italic>p</italic></bold>’s vector position.</p>
          <p>For range <italic>R</italic><sub>max</sub>, we defined the key point <italic>z</italic>’s range in the x-axis direction as <italic>R</italic><sub>x</sub>, then:</p>
          <p><italic>R</italic><sub>x_max</sub><italic>= x</italic><sub>z_max</sub><italic>– x</italic><sub>z_min</sub> (<bold>3</bold>)</p>
          <p>Similarly, the key point’s range in the y-axis direction was:</p>
          <p><italic>R</italic><sub>y_max</sub><italic>= x</italic><sub>z_max</sub><italic>– x</italic><sub>z_min</sub> (<bold>4</bold>)</p>
          <p>Intuitively, covariance, Cov(X,Y), represents the expectation of the overall error of two variables, which can reflect the correlation of the patient’s expression amplitude changes in two directions, and it is calculated as follows:</p>
          <p>Cov(X,Y) = <italic>E</italic>(XY) – <italic>E</italic>(X)<italic>E</italic>(Y) (<bold>5</bold>)</p>
          <p>where <italic>E</italic>(X) and <italic>E</italic>(Y) are the expected values of <italic>x</italic> and <italic>y</italic>, respectively.</p>
          <p>For absolute covariance, Cov(X_abs, Y_abs), we also calculated the covariance between the absolute coordinate X_abs, Y_abs. To some extent, it reflects the shaking of the human head.</p>
        </sec>
        <sec>
          <title>Feature Extraction of the Tremor at Facial Key Points</title>
          <p>Because patients with Parkinson disease patients cannot control their muscles like people without PD, the key points will tremble when they are trying to maintain facial expressions. To obtain the extent of tremor, we adopted absolute coordinates to measure the patient’s vibration. More specifically, we used <bold><italic>p</italic></bold><sub>i,x</sub> = (x<sub>iz</sub>,y<sub>iz</sub>) to denote the position of the key point <italic>z</italic> at the <italic>i</italic>th frame. The Euclidean distance between the two locations was defined as:</p>
          <graphic xlink:href="jmir_v22i7e18697_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>Moreover, when there were many position vectors, the set <bold>P</bold> was defined as:</p>
          <p><bold>P</bold> = {<bold><italic>p</italic></bold><sub>a(1,1), a(1,2)</sub>, <bold><italic>p</italic></bold><sub>a(2,1)a(2,2)</sub>,…, <bold><italic>p</italic></bold><sub>a(N,1)a(N,2)</sub>} (<bold>7</bold>)</p>
          <p>in which there exists <italic>N</italic> position vectors. Then, the average position of the set <bold>P</bold> was defined as:</p>
          <graphic xlink:href="jmir_v22i7e18697_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>Jitter is a good measure of tremor [<xref ref-type="bibr" rid="ref16">16</xref>]. However, it was created to calculate the tremor of a single signal. If we want to evaluate the positional Jitter of organ <italic>z</italic>, we take <italic>N</italic> frames of the organ, and the average position of the <italic>N</italic> frames is <bold><italic>p</italic></bold>_ave(<bold>P</bold><italic><sub>z</sub></italic>). Then, positional jitter can be defined.</p>
          <p>Jitter_abs is the relative Jitter of the key points (ie, the average absolute value of the positional distance between two adjacent frames).</p>
          <graphic xlink:href="jmir_v22i7e18697_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>Jitter_PPQ5 represents the adjacent 5 points of Jitter at the key point (ie, the average absolute value of the difference between the position of a certain frame and the average position of the adjacent 5 frames).</p>
          <graphic xlink:href="jmir_v22i7e18697_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <bold>P</bold><italic><sub>z,i,</sub></italic><sub>5</sub>={<bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>-2</sub><italic><sub>,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>-1</sub><italic><sub>,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>+1</sub><italic><sub>,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>+2</sub><italic><sub>,z</sub></italic>} denotes the average position of 5 adjacent frames (including the <italic>i</italic>-th frame) before and after the <italic>i</italic>-th frame of the key point <italic>z</italic>.</p>
          <p>Jitter_rap represents the adjacent 3 points of jitter at the key point (ie, the average absolute value of the difference between the position of a certain frame and the average position of the adjacent 3 frames).</p>
          <graphic xlink:href="jmir_v22i7e18697_fig13.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <bold>P</bold><italic><sub>z,i,</sub></italic><sub>5</sub>={<bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>-1</sub><italic><sub>,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i,z</sub></italic>, <bold><italic>p</italic></bold><italic><sub>i</sub></italic><sub>+1</sub><italic><sub>,z</sub></italic>} is the average position of 3 frames (including the <italic>i</italic>th frame) before and after the <italic>i</italic>th frame of the key point <italic>z</italic>.</p>
          <p>Jitter_ddp represents the difference between the adjacent 3 points of the key points’ jitter (ie, the difference between the distances of each adjacent 2 frames in the adjacent 3 frames). Then, the average absolute value was calculated as:</p>
          <graphic xlink:href="jmir_v22i7e18697_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>When we analyzed each patient’s key points, statistical analysis was performed on 106 key points. There were 8 features extracted for each key point, including 4 facial expression change amplitude features (<italic>R<sub>x</sub></italic><sub>_max</sub>, <italic>R<sub>y</sub></italic><sub>_max</sub>, Cov(X,Y), Cov(X_abs, Y_abs)) and 4 facial tremor features (Jitter, Jitter_PPQ5, Jitter_rap, Jitter_ddp). Since the collected data were not evaluated by a professional doctor using the UPDRS and we only knew whether the person had PD, we performed a diagnostic (classification) experiment. The final data statistics used in our experiment are shown in <xref ref-type="table" rid="table2">Table 2</xref>.</p>
          <table-wrap position="float" id="table2">
            <label>Table 2</label>
            <caption>
              <p>Video data statistics.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="500"/>
              <col width="500"/>
              <thead>
                <tr valign="top">
                  <td>Data statistics</td>
                  <td>Video data</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Creation date</td>
                  <td>3/15/2018</td>
                </tr>
                <tr valign="top">
                  <td>Number of patients with Parkinson disease</td>
                  <td>33<break/>  <break/>  </td>
                </tr>
                <tr valign="top">
                  <td>Number of people without Parkinson disease</td>
                  <td>31</td>
                </tr>
                <tr valign="top">
                  <td>Number of records</td>
                  <td>176</td>
                </tr>
                <tr valign="top">
                  <td>Number of features</td>
                  <td>848</td>
                </tr>
                <tr valign="top">
                  <td>Task</td>
                  <td>Classification</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <p>In this paper, we performed a categorization task for facial expressions using widely used machine learning algorithms such as logistic regression (LR) [<xref ref-type="bibr" rid="ref17">17</xref>], support vector machine (SVM) [<xref ref-type="bibr" rid="ref18">18</xref>], decision tree [<xref ref-type="bibr" rid="ref19">19</xref>], and random forest (RF) [<xref ref-type="bibr" rid="ref20">20</xref>]. <xref ref-type="table" rid="table3">Table 3</xref> shows that the methods based on facial key points can distinguish patients with PD from people without PD. The diagnostic result is relatively good.</p>
      <table-wrap position="float" id="table3">
        <label>Table 3</label>
        <caption>
          <p>Experimental results of common machine learning algorithms.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <thead>
            <tr valign="top">
              <td>Algorithm</td>
              <td>Precision</td>
              <td>Recall</td>
              <td>F1 value</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td>LR<sup>a</sup></td>
              <td>0.98</td>
              <td>0.98</td>
              <td>0.98</td>
            </tr>
            <tr valign="top">
              <td>SVM<sup>b</sup></td>
              <td>0.99</td>
              <td>0.99</td>
              <td>0.99</td>
            </tr>
            <tr valign="top">
              <td>DT<sup>c</sup></td>
              <td>0.93</td>
              <td>0.93</td>
              <td>0.93</td>
            </tr>
            <tr valign="top">
              <td>RF<sup>d</sup></td>
              <td>0.98</td>
              <td>0.98</td>
              <td>0.98</td>
            </tr>
          </tbody>
        </table>
        <table-wrap-foot>
          <fn id="table3fn1">
            <p><sup>a</sup>LR: logistic regression.</p>
          </fn>
          <fn id="table3fn2">
            <p><sup>b</sup>SVM: support vector machine.</p>
          </fn>
          <fn id="table3fn3">
            <p><sup>c</sup>DT: decision tree.</p>
          </fn>
          <fn id="table3fn4">
            <p><sup>d</sup>RF: random forest.</p>
          </fn>
        </table-wrap-foot>
      </table-wrap>
      <p>In order to verify the significance of original features, we conducted hypothesis testing. In this experiment, the threshold value α, which is referred to as the level of significance, was set to .05 and .005 separately.</p>
      <p><xref ref-type="table" rid="table4">Table 4</xref> shows the number of key points that reached significance for each of the 8 features types. At a <italic>P</italic> value &#60;.05, all 106 key points were significant for all 4 tremor features at facial key points. At a <italic>P</italic> value &#60;.005, all key points were significant for only 3 tremor features at facial key points. Therefore, we found that tremor features at facial key points may be more important than facial expression amplitude features.</p>
      <table-wrap position="float" id="table4">
        <label>Table 4</label>
        <caption>
          <p>Number of points that reached significance for each feature type.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="500"/>
          <col width="200"/>
          <col width="300"/>
          <thead>
            <tr valign="top">
              <td>Feature name</td>
              <td colspan="2">Number of key points that reached significance</td>
            </tr>
            <tr valign="top">
              <td>
                <break/>
              </td>
              <td><italic>P</italic> value &#60;.05</td>
              <td><italic>P</italic> value &#60;.005</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td>
                <italic>R</italic>
                <sub>x_max</sub>
              </td>
              <td>83</td>
              <td>69</td>
            </tr>
            <tr valign="top">
              <td>
                <italic>R</italic>
                <sub>y_max</sub>
              </td>
              <td>56</td>
              <td>45</td>
            </tr>
            <tr valign="top">
              <td>Cov(X, Y)</td>
              <td>97</td>
              <td>87</td>
            </tr>
            <tr valign="top">
              <td>Cov(X_abs, Y_abs)</td>
              <td>13</td>
              <td>12</td>
            </tr>
            <tr valign="top">
              <td>Jitter</td>
              <td>106</td>
              <td>106</td>
            </tr>
            <tr valign="top">
              <td>Jitter_PPQ5</td>
              <td>106</td>
              <td>106</td>
            </tr>
            <tr valign="top">
              <td>Jitter_rap</td>
              <td>106</td>
              <td>90</td>
            </tr>
            <tr valign="top">
              <td>Jitter_ddp</td>
              <td>106</td>
              <td>106</td>
            </tr>
            <tr valign="top">
              <td>Total</td>
              <td>673</td>
              <td>621</td>
            </tr>
          </tbody>
        </table>
      </table-wrap>
      <p>The least absolute shrinkage and selection operator (LASSO) is a linear regression method using L1 regularization [<xref ref-type="bibr" rid="ref21">21</xref>]. LASSO can make part of the learned feature weights be 0, so it achieves the function of sparseness and feature selection simultaneously. In this experiment, we used LASSO to compress the features.</p>
      <p>If a feature weight is nonzero in LASSO, this feature will be left; otherwise, this feature will be abandoned. The red line in <xref rid="figure5" ref-type="fig">Figure 5</xref> shows that, as the coefficient α in LASSO increases, the number of features after compression decreases gradually. Then, we used the compressed features to perform the final classification task. As the number of features used in the training model decreases, the accuracy of the training set of the LR and SVM models changes (as shown by the gray and yellow lines, respectively). At an α value of .004, the two models (LR and SVM) had the highest prediction accuracy. However, as the original features were gradually reduced, the accuracy did not decrease linearly, which can prove the remaining features are more important than the discarded features.</p>
      <fig id="figure5" position="float">
        <label>Figure 5</label>
        <caption>
          <p>The effects of least absolute shrinkage and selection operator (LASSO) feature compression on logistic regression (LR) and support vector machine (SVM) models.</p>
        </caption>
        <graphic xlink:href="jmir_v22i7e18697_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>Then, we used LASSO with the best hyperparameters to obtain the most relevant features to the target (PD or not). At the same time, RF was used to sort out the importance of the features. These points are near the corner of the right eye and the lips on a human. In <xref rid="figure6" ref-type="fig">Figure 6</xref>, the features chosen by LASSO are marked with red circles, and the features chosen by RF are marked with blue ovals.</p>
      <fig id="figure6" position="float">
        <label>Figure 6</label>
        <caption>
          <p>The key points that have a large influence on the classification result.</p>
        </caption>
        <graphic xlink:href="jmir_v22i7e18697_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>In this paper, we also utilized the advanced deep learning technique long short-term memory (LSTM) [<xref ref-type="bibr" rid="ref22">22</xref>] to perform the classification. We converted the position changes of the key points in the x axis and y axis with time into time-series data. Then, we applied an order difference to the time series, that is, ∆<italic>x<sub>i</sub></italic> = <italic>x<sub>i</sub></italic><sub>-1</sub> - <italic>x<sub>i</sub></italic>, ∆<italic>y<sub>i</sub></italic> = <italic>y<sub>i</sub></italic><sub>-1</sub> - <italic>y<sub>i</sub></italic>. LSTM utilized these new time series in the model training. The results are shown in <xref ref-type="table" rid="table5">Table 5</xref>. The experimental results are relatively good too. Compared with the pure recurrent neural network technique [<xref ref-type="bibr" rid="ref23">23</xref>], which has no effect on the classification, LSTM demonstrated it is more practical when dealing with sequential video data. Note that the facial video data used in this paper are limited. LSTM should be able to achieve better results if more data can be acquired.</p>
      <table-wrap position="float" id="table5">
        <label>Table 5</label>
        <caption>
          <p>Experimental results of neural network models.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <thead>
            <tr valign="top">
              <td>Algorithm</td>
              <td>Precision</td>
              <td>Recall</td>
              <td>F1 value</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td>LSTM<sup>a</sup></td>
              <td>0.86</td>
              <td>0.66</td>
              <td>0.75</td>
            </tr>
            <tr valign="top">
              <td>RNN<sup>b</sup></td>
              <td>0.48</td>
              <td>0.46</td>
              <td>0.47</td>
            </tr>
          </tbody>
        </table>
        <table-wrap-foot>
          <fn id="table5fn1">
            <p><sup>a</sup>LSTM: long short-term memory.</p>
          </fn>
          <fn id="table5fn2">
            <p><sup>b</sup>RNN: recurrent neural network.</p>
          </fn>
        </table-wrap-foot>
      </table-wrap>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Limitations</title>
        <p>The amount of data that we collected was not sufficient. We hope to collect more data not only from patients with PD and people without PD but also from patients with other neurological diseases.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>In the case of PD, to the best of our knowledge, there have only been a few software options, similar to ours, providing patients a convincing diagnosis using the facial recognition technique. These are compared to our work in <xref ref-type="table" rid="table6">Table 6</xref>.</p>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Comparison with a selection of prior work.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="150"/>
            <col width="270"/>
            <col width="150"/>
            <col width="240"/>
            <col width="190"/>
            <thead>
              <tr valign="top">
                <td>Work</td>
                <td>Target and result</td>
                <td>Data</td>
                <td>Feature</td>
                <td>Technology</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Bandini et al [<xref ref-type="bibr" rid="ref10">10</xref>]</td>
                <td>Found PD<sup>a</sup> patients have lower average facial expression movement distance; facial expression recognition for PD</td>
                <td>17 PD patients,<break/>17 healthy control subjects</td>
                <td>Average distance of 49 facial key points in the facial expression movement</td>
                <td>Face tracing, SVM<sup>b</sup></td>
              </tr>
              <tr valign="top">
                <td>Rajnoha et al [<xref ref-type="bibr" rid="ref11">11</xref>]</td>
                <td>Identified PD hypomimia by analyzing static facial images; less accurate compared with video-recording processing method.</td>
                <td>50 PD patients,<break/>50 healthy control subjects</td>
                <td>128 facial measures (embedding) by CNN<sup>c</sup></td>
                <td>Face detector-based (HOG<sup>d</sup>), CNN, traditional classiﬁers (eg, random forests, XGBoost)</td>
              </tr>
              <tr valign="top">
                <td>PARK<sup>e</sup> framework by Langevin et al [<xref ref-type="bibr" rid="ref12">12</xref>]</td>
                <td>PARK instructs and guides users through 6 motor tasks and 1 audio task selected from MDS-UPDRS<sup>f</sup> and records their performance by videos</td>
                <td>127 PD patients,<break/>127 healthy control subjects</td>
                <td>Facial features: facial action units (AUs);<break/>motion features: motion magnitude metric of fingers and hands based on FFT<sup>g</sup></td>
                <td>OpenFace tool version 2, FFT</td>
              </tr>
              <tr valign="top">
                <td>Our method</td>
                <td>Proposed facial landmark features from videos to diagnose PD using facial expressions and achieved outstanding performance</td>
                <td>33 PD patients,<break/>31 healthy control subjects,<break/>176 records</td>
                <td>848 facial expression amplitude features and tremor features of facial key points;<break/>65 features were left after feature compression</td>
                <td>Face ++, traditional classiﬁers (LR<sup>h</sup>, SVM, DT<sup>i</sup>, RF<sup>j</sup>), LSTM<sup>k</sup>, LASSO<sup>l</sup></td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>PD: Parkinson disease.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>SVM: support vector machine.</p>
            </fn>
            <fn id="table6fn3">
              <p><sup>c</sup>CNN: convolutional neural network.</p>
            </fn>
            <fn id="table6fn4">
              <p><sup>d</sup>HOG: histogram of oriented gradients.</p>
            </fn>
            <fn id="table6fn5">
              <p><sup>e</sup>PARK: Parkinson's Analysis with Remote Kinetic-tasks.</p>
            </fn>
            <fn id="table6fn6">
              <p><sup>f</sup>MDS-UPDRS: Movement Disorder Society Unified Parkinson Disease Rating Scale.</p>
            </fn>
            <fn id="table6fn7">
              <p><sup>g</sup>FFT: fast fourier transform.</p>
            </fn>
            <fn id="table6fn8">
              <p><sup>h</sup>LR: logistic regression.</p>
            </fn>
            <fn id="table6fn9">
              <p><sup>i</sup>DT: decision tree.</p>
            </fn>
            <fn id="table6fn10">
              <p><sup>j</sup>RF: random forest.</p>
            </fn>
            <fn id="table6fn11">
              <p><sup>k</sup>LSTM: long short-term memory.</p>
            </fn>
            <fn id="table6fn12">
              <p><sup>l</sup>LASSO: least absolute shrinkage and selection operator.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this paper, we established a diagnostic model for PD based on facial expressions. In the model, we formulated the diagnostic task into a classification problem. Then, we solved it by using algorithms from the area of traditional machine learning and the LSTM model from the field of deep learning research. When constructing video features, we conducted feature extraction according to the expression amplitude and degree of tremor. Using a fixed time interval and conversion of the coordinate system, the image was intercepted. This method converted the video into frame data. Further, the LSTM model was applied to the diagnosis of PD based on the generated time series data. Experimental results proved our model is effective and can be used as an efficient tool in PD diagnosis.</p>
        <p>This article is a preliminary exploration of neurological diseases in the context of machine learning. The proposed method is designed to help patients get more comprehensive treatment and help doctors to understand the real-time dynamics of the disease. At the same time, it also aimed to relieve the problem of registering patients who have difficulty moving and to relieve the pressure on repeated patient diagnoses by doctors. With the development of science and technology, the introduction and application of artificial intelligence will bring more convenient and rapid diagnostic and treatment technologies.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">DT</term>
          <def>
            <p>decision tree</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">FFT</term>
          <def>
            <p>fast fourier transform</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">HOG</term>
          <def>
            <p>histogram of oriented gradients</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">LASSO</term>
          <def>
            <p>least absolute shrinkage and selection operator</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">LR</term>
          <def>
            <p>logistic regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">LSTM</term>
          <def>
            <p>long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">MDS-UPDRS</term>
          <def>
            <p>Movement Disorder Society Unified Parkinson Disease Rating Scale</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">PARK</term>
          <def>
            <p>Parkinson's Analysis with Remote Kinetic-tasks</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">PD</term>
          <def>
            <p>Parkinson disease</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">RF</term>
          <def>
            <p>random forest</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">RNN</term>
          <def>
            <p>recurrent neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>We are deeply indebted to people from the Department of Neurology, the First Affiliated Hospital of Dalian Medical University, who supported us in data collection and provided considerable professional advice. In addition, we owe our thanks to all the volunteers, who allowed us to collect their video data. This study was supported by the Program of National Key R&#38;D Program of China (2018YFC0116800), National Natural Science Foundation of China (No. 61772110), CERNET Innovation Project (NGII20170711), and Introducing Talents of Discipline to Universities (Plan 111) (No. B20070).</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>YQ developed the conceptual framework and research protocol for the study. BJ and LZ conducted the publications review, data collection, and analysis. ZG interpreted the data. BJ drafted the manuscript, and LZ made major revisions. All authors approved the final version of the manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lexell</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Evidence for nervous system degeneration with advancing age</article-title>
          <source>J Nutr</source>
          <year>1997</year>
          <month>05</month>
          <volume>127</volume>
          <issue>5 Suppl</issue>
          <fpage>1011S</fpage>
          <lpage>1013S</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1093/jn/127.5.1011S"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jn/127.5.1011S</pub-id>
          <pub-id pub-id-type="medline">9164286</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thomson</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Disease briefing: Parkinson's disease</article-title>
          <source>Journal of International Pharmaceutical Research</source>
          <year>2015</year>
          <issue>3</issue>
          <fpage>338</fpage>
          <lpage>345</lpage>
          <pub-id pub-id-type="doi">10.13220/j.cnki.jipr.2015.03.009</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Martinez-Martín</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez-Blazquez</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Paz</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Forjaz</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Frades-Payo</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Cubo</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>de Pedro-Cuesta</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lizán</surname>
              <given-names>L</given-names>
            </name>
            <collab>ELEP Group</collab>
          </person-group>
          <article-title>Parkinson Symptoms and Health Related Quality of Life as Predictors of Costs: A Longitudinal Observational Study with Linear Mixed Model Analysis</article-title>
          <source>PLoS One</source>
          <year>2015</year>
          <month>09</month>
          <volume>10</volume>
          <issue>12</issue>
          <fpage>e0145310</fpage>
          <lpage>33</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://dx.plos.org/10.1371/journal.pone.0145310"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pone.0145310</pub-id>
          <pub-id pub-id-type="medline">26698860</pub-id>
          <pub-id pub-id-type="pii">PONE-D-15-30698</pub-id>
          <pub-id pub-id-type="pmcid">PMC4689528</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kruszka</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Addissie</surname>
              <given-names>YA</given-names>
            </name>
            <name name-style="western">
              <surname>McGinn</surname>
              <given-names>DE</given-names>
            </name>
            <name name-style="western">
              <surname>Porras</surname>
              <given-names>AR</given-names>
            </name>
            <name name-style="western">
              <surname>Biggs</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Share</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Crowley</surname>
              <given-names>TB</given-names>
            </name>
            <name name-style="western">
              <surname>Chung</surname>
              <given-names>BHY</given-names>
            </name>
            <name name-style="western">
              <surname>Mok</surname>
              <given-names>GTK</given-names>
            </name>
            <name name-style="western">
              <surname>Mak</surname>
              <given-names>CCY</given-names>
            </name>
            <name name-style="western">
              <surname>Muthukumarasamy</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Thong</surname>
              <given-names>MK</given-names>
            </name>
            <name name-style="western">
              <surname>Sirisena</surname>
              <given-names>ND</given-names>
            </name>
            <name name-style="western">
              <surname>Dissanayake</surname>
              <given-names>VHW</given-names>
            </name>
            <name name-style="western">
              <surname>Paththinige</surname>
              <given-names>CS</given-names>
            </name>
            <name name-style="western">
              <surname>Prabodha</surname>
              <given-names>LBL</given-names>
            </name>
            <name name-style="western">
              <surname>Mishra</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Shotelersuk</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ekure</surname>
              <given-names>EN</given-names>
            </name>
            <name name-style="western">
              <surname>Sokunbi</surname>
              <given-names>OJ</given-names>
            </name>
            <name name-style="western">
              <surname>Kalu</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Ferreira</surname>
              <given-names>CR</given-names>
            </name>
            <name name-style="western">
              <surname>Duncan</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Patil</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>KL</given-names>
            </name>
            <name name-style="western">
              <surname>Kaplan</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Abdul-Rahman</surname>
              <given-names>OA</given-names>
            </name>
            <name name-style="western">
              <surname>Uwineza</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mutesa</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Moresco</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Obregon</surname>
              <given-names>MG</given-names>
            </name>
            <name name-style="western">
              <surname>Richieri-Costa</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gil-da-Silva-Lopes</surname>
              <given-names>VL</given-names>
            </name>
            <name name-style="western">
              <surname>Adeyemo</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Summar</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zackai</surname>
              <given-names>EH</given-names>
            </name>
            <name name-style="western">
              <surname>McDonald-McGinn</surname>
              <given-names>DM</given-names>
            </name>
            <name name-style="western">
              <surname>Linguraru</surname>
              <given-names>MG</given-names>
            </name>
            <name name-style="western">
              <surname>Muenke</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>22q11.2 deletion syndrome in diverse populations</article-title>
          <source>Am J Med Genet A</source>
          <year>2017</year>
          <month>04</month>
          <volume>173</volume>
          <issue>4</issue>
          <fpage>879</fpage>
          <lpage>888</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/28328118"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/ajmg.a.38199</pub-id>
          <pub-id pub-id-type="medline">28328118</pub-id>
          <pub-id pub-id-type="pmcid">PMC5363275</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Basel-Vanagaite</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wolf</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Orin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Larizza</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gervasini</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Krantz</surname>
              <given-names>ID</given-names>
            </name>
            <name name-style="western">
              <surname>Deardoff</surname>
              <given-names>MA</given-names>
            </name>
          </person-group>
          <article-title>Recognition of the Cornelia de Lange syndrome phenotype with facial dysmorphology novel analysis</article-title>
          <source>Clin Genet</source>
          <year>2016</year>
          <month>05</month>
          <volume>89</volume>
          <issue>5</issue>
          <fpage>557</fpage>
          <lpage>63</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/cge.12716"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/cge.12716</pub-id>
          <pub-id pub-id-type="medline">26663098</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goetz</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>The history of Parkinson's disease: early clinical descriptions and neurological therapies</article-title>
          <source>Cold Spring Harb Perspect Med</source>
          <year>2011</year>
          <month>09</month>
          <volume>1</volume>
          <issue>1</issue>
          <fpage>a008862</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/22229124"/>
          </comment>
          <pub-id pub-id-type="doi">10.1101/cshperspect.a008862</pub-id>
          <pub-id pub-id-type="medline">22229124</pub-id>
          <pub-id pub-id-type="pii">a008862</pub-id>
          <pub-id pub-id-type="pmcid">PMC3234454</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tickle-Degnen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Doyle Lyons</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Practitioners’ impressions of patients with Parkinson's disease: the social ecology of the expressive mask</article-title>
          <source>Social Science &#38; Medicine</source>
          <year>2004</year>
          <month>2</month>
          <volume>58</volume>
          <issue>3</issue>
          <fpage>603</fpage>
          <lpage>614</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/S0277-9536(03)00213-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/s0277-9536(03)00213-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tsanas</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Little</surname>
              <given-names>Ma</given-names>
            </name>
            <name name-style="western">
              <surname>McSharry</surname>
              <given-names>Pe</given-names>
            </name>
            <name name-style="western">
              <surname>Spielman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ramig</surname>
              <given-names>Lo</given-names>
            </name>
          </person-group>
          <article-title>Novel Speech Signal Processing Algorithms for High-Accuracy Classification of Parkinson's Disease</article-title>
          <source>IEEE Trans. Biomed. Eng</source>
          <year>2012</year>
          <month>05</month>
          <volume>59</volume>
          <issue>5</issue>
          <fpage>1264</fpage>
          <lpage>1271</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/TBME.2012.2183367"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/tbme.2012.2183367</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Little</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>McSharry</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Hunter</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Spielman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ramig</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Suitability of dysphonia measurements for telemonitoring of Parkinson’s disease</article-title>
          <source>Nat Prec</source>
          <year>2008</year>
          <month>9</month>
          <day>12</day>
          <fpage>1</fpage>
          <lpage>1</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/npre.2008.2298.1"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/npre.2008.2298.1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bandini</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Orlandi</surname>
              <given-names>Silvia</given-names>
            </name>
            <name name-style="western">
              <surname>Escalante</surname>
              <given-names>Hugo Jair</given-names>
            </name>
            <name name-style="western">
              <surname>Giovannelli</surname>
              <given-names>Fabio</given-names>
            </name>
            <name name-style="western">
              <surname>Cincotta</surname>
              <given-names>Massimo</given-names>
            </name>
            <name name-style="western">
              <surname>Reyes-Garcia</surname>
              <given-names>Carlos A</given-names>
            </name>
            <name name-style="western">
              <surname>Vanni</surname>
              <given-names>Paola</given-names>
            </name>
            <name name-style="western">
              <surname>Zaccara</surname>
              <given-names>Gaetano</given-names>
            </name>
            <name name-style="western">
              <surname>Manfredi</surname>
              <given-names>Claudia</given-names>
            </name>
          </person-group>
          <article-title>Analysis of facial expressions in Parkinson's disease through video-based automatic methods</article-title>
          <source>J Neurosci Methods</source>
          <year>2017</year>
          <month>04</month>
          <day>01</day>
          <volume>281</volume>
          <fpage>7</fpage>
          <lpage>20</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.jneumeth.2017.02.006"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jneumeth.2017.02.006</pub-id>
          <pub-id pub-id-type="medline">28223023</pub-id>
          <pub-id pub-id-type="pii">S0165-0270(17)30048-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rajnoha</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mekyska</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Burget</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Eliasova</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Kostalova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rektorova</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Towards Identification of Hypomimia in Parkinson's Disease Based on Face Recognition Methods</article-title>
          <year>2018</year>
          <conf-name>10th International Congress on Ultra Modern Telecommunications and Control Systems and Workshops (ICUMT)</conf-name>
          <conf-date>November 5-9, 2018</conf-date>
          <conf-loc>Moskva · Russian Federation</conf-loc>
          <fpage>1</fpage>
          <lpage>4</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/ICUMT.2018.8631249"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Langevin</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>MR</given-names>
            </name>
            <name name-style="western">
              <surname>Sen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Snyder</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Myers</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Dorsey</surname>
              <given-names>ER</given-names>
            </name>
            <name name-style="western">
              <surname>Hoque</surname>
              <given-names>ME</given-names>
            </name>
          </person-group>
          <article-title>The PARK Framework for Automated Analysis of Parkinson's Disease Characteristics</article-title>
          <source>Proc. ACM Interact. Mob. Wearable Ubiquitous Technol</source>
          <year>2019</year>
          <conf-name>Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies</conf-name>
          <conf-date>September 11-13, 2019</conf-date>
          <conf-loc>London, UK</conf-loc>
          <fpage>1</fpage>
          <lpage>22</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3328925"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Madeley</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Ellis</surname>
              <given-names>AW</given-names>
            </name>
            <name name-style="western">
              <surname>Mindham</surname>
              <given-names>RHS</given-names>
            </name>
          </person-group>
          <article-title>Facial Expressions and Parkinson’s Disease</article-title>
          <source>Behavioural Neurology</source>
          <year>1995</year>
          <volume>8</volume>
          <issue>2</issue>
          <fpage>115</fpage>
          <lpage>119</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3233/BEN-1995-8207"/>
          </comment>
          <pub-id pub-id-type="doi">10.1155/1995/750425</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Newmarch</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Ffmpeg/libav</article-title>
          <source>Linux Sound Programming</source>
          <year>2017</year>
          <publisher-loc>Berkeley</publisher-loc>
          <publisher-name>Apress</publisher-name>
          <fpage>227</fpage>
          <lpage>234</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jung</surname>
              <given-names>SG</given-names>
            </name>
            <name name-style="western">
              <surname>An</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kwak</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Salminen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jansen</surname>
              <given-names>BJ</given-names>
            </name>
          </person-group>
          <article-title>Inferring social media users’ demographics from profile pictures: A Face++ analysis on Twitter users</article-title>
          <year>2017</year>
          <conf-name>Proceedings of the 17th International Conference on Electronic Business</conf-name>
          <conf-date>August, 2015</conf-date>
          <conf-loc>Seoul, Republic of Korea</conf-loc>
          <publisher-loc>New York, NY, United States</publisher-loc>
          <publisher-name>Association for Computing Machinery</publisher-name>
          <fpage>140</fpage>
          <lpage>145</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://aisel.aisnet.org/iceb2017/22/"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boersma</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Weenink</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Praat v. 4.0. 8</article-title>
          <source>A system for doing phonetics by computer. Institute of Phonetic Sciences of the University of Amsterdam</source>
          <year>2002</year>
          <volume>218</volume>
          <fpage>1</fpage>
          <lpage>2</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kleinbaum</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Dietz</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Gail</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Klein</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Logistic regression</article-title>
          <source>Logistic regression</source>
          <year>2002</year>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hearst</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dumais</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Osuna</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Platt</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Scholkopf</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Support vector machines</article-title>
          <source>IEEE Intell. Syst. Their Appl</source>
          <year>1998</year>
          <month>7</month>
          <volume>13</volume>
          <issue>4</issue>
          <fpage>18</fpage>
          <lpage>28</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/5254.708428"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/5254.708428</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Quinlan</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Simplifying decision trees</article-title>
          <source>International Journal of Man-Machine Studies</source>
          <year>1987</year>
          <month>9</month>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>221</fpage>
          <lpage>234</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/S0020-7373(87)80053-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/s0020-7373(87)80053-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liaw</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wiener</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Classification and regression by random Forest</article-title>
          <source>R news</source>
          <year>2002</year>
          <volume>2</volume>
          <fpage>18</fpage>
          <lpage>22</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.researchgate.net/publication/228451484"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tibshirani</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Regression Shrinkage and Selection Via the Lasso</article-title>
          <source>Journal of the Royal Statistical Society: Series B (Methodological)</source>
          <year>2018</year>
          <month>12</month>
          <day>05</day>
          <volume>58</volume>
          <issue>1</issue>
          <fpage>267</fpage>
          <lpage>288</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/j.2517-6161.1996.tb02080.x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/j.2517-6161.1996.tb02080.x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sundermeyer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schlüter</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ney</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>LSTM neural networks for language modeling</article-title>
          <source>Thirteenth annual conference of the international speech communication association</source>
          <year>2012</year>
          <conf-name>Thirteenth annual conference of the international speech communication association</conf-name>
          <conf-date>September 9-13, 2012</conf-date>
          <conf-loc>Portland, Oregon</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.isca-speech.org/archive/interspeech_2012/i12_0194"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mikolov</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Karafiát</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Burget</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Černocký</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Khudanpur</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Recurrent neural network based language model</article-title>
          <year>2010</year>
          <conf-name>Eleventh annual conference of the international speech communication association</conf-name>
          <conf-date>September 26-30, 2010</conf-date>
          <conf-loc>Makuhari, Chiba, Japan</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.isca-speech.org/archive/interspeech_2010/i10_1045"/>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
