<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
    <article-id pub-id-type="publisher-id">v20i7e240</article-id>
    <article-id pub-id-type="pmid">30021708</article-id>
    <article-id pub-id-type="doi">10.2196/jmir.9300</article-id>
    <article-categories>
      <subj-group subj-group-type="heading">
        <subject>Original Paper</subject>
      </subj-group>
      <subj-group subj-group-type="article-type">
        <subject>Original Paper</subject>
      </subj-group>
    </article-categories>
    <title-group>
      <article-title>Evaluating Doctor Performance: Ordinal Regression-Based Approach</article-title>
    </title-group>
    <contrib-group>
      <contrib contrib-type="editor">
        <name>
          <surname>Eysenbach</surname>
          <given-names>Gunther</given-names>
        </name>
      </contrib>
    </contrib-group>
    <contrib-group>
      <contrib contrib-type="reviewer">
        <name>
          <surname>Allin</surname>
          <given-names>Sonya</given-names>
        </name>
      </contrib>
      <contrib contrib-type="reviewer">
        <name>
          <surname>Zhang</surname>
          <given-names>Hongbo</given-names>
        </name>
      </contrib>
      <contrib contrib-type="reviewer">
        <name>
          <surname>Liu</surname>
          <given-names>Yupeng</given-names>
        </name>
      </contrib>
    </contrib-group>
    <contrib-group>
      <contrib contrib-type="author" id="contrib1">
        <name name-style="western">
          <surname>Shi</surname>
          <given-names>Yong</given-names>
        </name>
        <degrees>PhD</degrees>
        <xref rid="aff1" ref-type="aff">1</xref>
        <xref rid="aff2" ref-type="aff">2</xref>
        <xref rid="aff3" ref-type="aff">3</xref>
        <xref rid="aff4" ref-type="aff">4</xref>
        <ext-link ext-link-type="orcid">http://orcid.org/0000-0003-0895-9868</ext-link>
      </contrib>
      <contrib contrib-type="author" id="contrib2">
        <name name-style="western">
          <surname>Li</surname>
          <given-names>Peijia</given-names>
        </name>
        <degrees>PhD</degrees>
        <xref rid="aff5" ref-type="aff">5</xref>
        <xref rid="aff6" ref-type="aff">6</xref>
        <ext-link ext-link-type="orcid">http://orcid.org/0000-0003-2989-8777</ext-link>
      </contrib>
      <contrib contrib-type="author" id="contrib3">
        <name name-style="western">
          <surname>Yu</surname>
          <given-names>Xiaodan</given-names>
        </name>
        <degrees>PhD</degrees>
        <xref rid="aff7" ref-type="aff">7</xref>
        <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-1465-4190</ext-link>
      </contrib>
      <contrib contrib-type="author" id="contrib4">
        <name name-style="western">
          <surname>Wang</surname>
          <given-names>Huadong</given-names>
        </name>
        <degrees>PhD</degrees>
        <xref rid="aff2" ref-type="aff">2</xref>
        <xref rid="aff3" ref-type="aff">3</xref>
        <ext-link ext-link-type="orcid">http://orcid.org/0000-0003-3260-4119</ext-link>
      </contrib>
      <contrib contrib-type="author" id="contrib5" corresp="yes">
      <name name-style="western">
        <surname>Niu</surname>
        <given-names>Lingfeng</given-names>
      </name>
      <degrees>PhD</degrees>
      <xref rid="aff1" ref-type="aff">1</xref>
      <xref rid="aff2" ref-type="aff">2</xref>
      <xref rid="aff3" ref-type="aff">3</xref>
      <address>
        <institution>Research Center on Fictitious Economy and Data Science</institution>
        <institution>Chinese Academy of Sciences</institution>
        <addr-line>80 Zhongguancun East Road</addr-line>
        <addr-line>Haidian District</addr-line>
        <addr-line>Beijing, 100190</addr-line>
        <country>China</country>
        <phone>86 15600616246</phone>
        <email>niulf@ucas.ac.cn</email>
      </address>  
      <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-5827-8449</ext-link></contrib>
    </contrib-group>
    <aff id="aff1">
    <sup>1</sup>
    <institution>School of Economics and Management</institution>
    <institution>University of Chinese Academy of Sciences</institution>  
    <addr-line>Beijing</addr-line>
    <country>China</country></aff>
    <aff id="aff2">
    <sup>2</sup>
    <institution>Key Laboratory of Big Data Mining and Knowledge Management</institution>
    <institution>Chinese Academy of Sciences</institution>  
    <addr-line>Beijing</addr-line>
    <country>China</country></aff>
    <aff id="aff3">
    <sup>3</sup>
    <institution>Research Center on Fictitious Economy and Data Science</institution>
    <institution>Chinese Academy of Sciences</institution>  
    <addr-line>Beijing</addr-line>
    <country>China</country></aff>
    <aff id="aff4">
    <sup>4</sup>
    <institution>College of Information Science and Technology</institution>
    <institution>University of Nebraska at Omaha</institution>  
    <addr-line>Omaha, NE</addr-line>
    <country>United States</country></aff>
    <aff id="aff5">
    <sup>5</sup>
    <institution>School of Computer and Control Engineering</institution>
    <institution>University of Chinese Academy of Sciences</institution>  
    <addr-line>Beijing</addr-line>
    <country>China</country></aff>
    <aff id="aff6">
      <sup>6</sup>
      <institution>Development and Planning Research Institute</institution>
      <addr-line>Ningbo</addr-line>
      <country>China</country>
    </aff>
    <aff id="aff7">
    <sup>7</sup>
    <institution>School of Information Technology and Management</institution>
    <institution>University of International Business and Economics</institution>  
    <addr-line>Beijing</addr-line>
    <country>China</country></aff>
    <author-notes>
      <corresp>Corresponding Author: Lingfeng Niu 
      <email>niulf@ucas.ac.cn</email></corresp>
    </author-notes>
    <pub-date pub-type="collection"><month>07</month><year>2018</year></pub-date>
    <pub-date pub-type="epub">
      <day>18</day>
      <month>07</month>
      <year>2018</year>
    </pub-date>
    <volume>20</volume>
    <issue>7</issue>
    <elocation-id>e240</elocation-id>
    <!--history from ojs - api-xml-->
    <history>
      <date date-type="received">
        <day>30</day>
        <month>10</month>
        <year>2017</year>
      </date>
      <date date-type="rev-request">
        <day>15</day>
        <month>3</month>
        <year>2018</year>
      </date>
      <date date-type="rev-recd">
        <day>11</day>
        <month>4</month>
        <year>2018</year>
      </date>
      <date date-type="accepted">
        <day>8</day>
        <month>5</month>
        <year>2018</year>
      </date>
    </history>
    <!--(c) the authors - correct author names and publication date here if necessary. Date in form ', dd.mm.yyyy' after jmir.org-->
    <copyright-statement>©Yong Shi, Peijia Li, Xiaodan Yu, Huadong Wang, Lingfeng Niu. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 18.07.2018.</copyright-statement>
    <copyright-year>2018</copyright-year>
    <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
      <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.</p>
    </license>  
    <self-uri xlink:href="http://www.jmir.org/2018/7/e240/" xlink:type="simple"/>
    <abstract>
      <sec sec-type="background">
        <title>Background</title>
        <p>Doctor’s performance evaluation is an important task in mobile health (mHealth), which aims to evaluate the overall quality of online diagnosis and patient outcomes so that customer satisfaction and loyalty can be attained. However, most patients tend not to rate doctors’ performance, therefore, it is imperative to develop a model to make doctor’s performance evaluation automatic. When evaluating doctors’ performance, we rate it into a score label that is as close as possible to the true one.</p>
      </sec>
      <sec sec-type="objective">
        <title>Objective</title>
        <p>This study aims to perform automatic doctor’s performance evaluation from online textual consultations between doctors and patients by way of a novel machine learning method.</p>
      </sec>
      <sec sec-type="methods">
        <title>Methods</title>
        <p>We propose a solution that models doctor’s performance evaluation as an ordinal regression problem. In doing so, a support vector machine combined with an ordinal partitioning model (SVMOP), along with an innovative predictive function will be developed to capture the hidden preferences of the ordering labels over doctor’s performance evaluation. When engineering the basic text features, eight customized features (extracted from over 70,000 medical entries) were added and further boosted by the Gradient Boosting Decision Tree algorithm.</p>
      </sec>
      <sec sec-type="results">
        <title>Results</title>
        <p>Real data sets from one of the largest mobile doctor/patient communication platforms in China are used in our study. Statistically, 64% of data on mHealth platforms lack the evaluation labels from patients. Experimental results reveal that our approach can support an automatic doctor performance evaluation. Compared with other auto-evaluation models, SVMOP improves mean absolute error (MAE) by 0.1, mean square error (MSE) by 0.5, pairwise accuracy (PAcc) by 5%; the suggested customized features improve MAE by 0.1, MSE by 0.2, PAcc by 3%. After boosting, performance is further improved. Based on SVMOP, predictive features like politeness and sentiment words can be mined, which can be further applied to guide the development of mHealth platforms.</p>
      </sec>
      <sec sec-type="conclusions">
        <title>Conclusions</title>
        <p>The initial modelling of doctor performance evaluation is an ordinal regression problem. Experiments show that the performance of our proposed model with revised prediction function is better than many other machine learning methods on MAE, MSE, as well as PAcc. With this model, the mHealth platform could not only make an online auto-evaluation of physician performance, but also obtain the most effective features, thereby guiding physician performance and the development of mHealth platforms.</p>
      </sec>
    </abstract>
    <kwd-group>
      <kwd>performance evaluation</kwd>
      <kwd>ordinal regression</kwd>
      <kwd>mHealth</kwd>
      <kwd>support vector machines</kwd>
      <kwd>ordinal partitioning</kwd>
    </kwd-group></article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>With the advancement of the internet and electronic devices, mobile heath (mHealth), is defined by the World Health Organization as “medical and public health practice supported by mobile devices,” is becoming increasingly popular. mHealth has strong links with electronic health [<xref ref-type="bibr" rid="ref1">1</xref>] with some differences [<xref ref-type="bibr" rid="ref2">2</xref>]. According to an mHealth survey [<xref ref-type="bibr" rid="ref3">3</xref>], 80% of physicians use smartphones and medical apps and 61% of people have downloaded a medical app. Meanwhile, 93% of physicians believe that mHealth apps can help to improve patients’ health. The doctor/patient communication platform is one of the most common areas in mHealth, for example, “Dermatologist-on-Call” in America and, in China, “Chunyu-Doctor-online” and “Good-Doctor-online.” These platforms digitally connect doctors and patients and offer a convenient channel for doctor/patient communication and help doctors use time more efficiently. Additionally, the mHealth platforms are more beneficial to under-developed countries, especially when medical resources are scarce, and quality medical care is difficult to access.</p>
      <p>Many doctor/patient communication platforms face the challenge of how to evaluate the performance of doctors online. Doctor performance evaluation serves to increase the probability for patients to have a positive experience and improve patient satisfaction [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. Meanwhile, doctor performance evaluation also helps doctors to improve medical practice [<xref ref-type="bibr" rid="ref7">7</xref>]. In this paper, we address the issue of doctor performance evaluation (DPE).</p>
      <p>Various methods have been attempted that address the issue of DPEs. Ratings by patients is the most common method, which averages patient ratings when evaluating physicians. Physician ratings are usually based on the following labels: (1) very unsatisfied, (2) unsatisfied, (3) neutral, (4) satisfied, and (5) very satisfied. Statistics show that only a small proportion of patients rate their doctors on mHealth platforms, and in China, only 36% of patients rated their doctor at the end of the consultation.</p>
      <p>A physician expert assesses the doctors’ professional skills and services. In this combined method, experts re-rate the patient’s unsatisfied consultations and judge whether the doctors are qualified. It is an advanced evaluation method, which not only considers patient ratings but also imbues prior professional knowledge. Therefore, this method is recommended but heavily depends on the patient’s ratings.</p>
      <p>Considering the amount of data generated from doctor/patient communication platforms every day, machine learning techniques are recommended. In the machine learning area, some scholars have rated patient satisfaction into standard classification algorithms [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>] but ignore the ordering information between labels. The ordering information between labels as mentioned above are, (1) very unsatisfied, (2) unsatisfied, (3) neutral, (4) satisfied, and (5) very satisfied. The label “unsatisfied” is adjacent to the label “very unsatisfied” and the label “neutral,” while the label “very unsatisfied” is not next to the label “neutral,” therefore the ordering of information is extremely important. Rating a “very unsatisfied” doctor consultation as an “unsatisfied” one is less of an error than rating it as “very satisfied.” Therefore, it is important that the predicted labels are not only accurate but as close to the true labels as possible. This method of classifying the instances into the nearest ordinal labels, is called ordinal regression (OR) [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>], and the overall evaluation model is ordinal regression for doctor performance evaluation (OR-DPE).</p>
      <p>In supervised learning OR resides between multi-classification and metric regression. The difference between the two is that the labels of the latter are in a limited but unordered set. The difference between OR and metric regression is that the OR labels do not represent numerical values. Although standard multi-classification and regression algorithms can be used to solve OR problems, they ignore the ordering information between labels. Some researchers [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>] have proved that ordering information benefits modelling greatly. There exist many models especially designed for OR. The “Proportional Odds” model, designed in 1980 [<xref ref-type="bibr" rid="ref13">13</xref>], is one of the earliest such models. Since then a wide range of OR models have been proposed including support vector machine (SVM)-based models [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref17">17</xref>], Neural Network-based models [<xref ref-type="bibr" rid="ref18">18</xref>], Gaussian Process models [<xref ref-type="bibr" rid="ref19">19</xref>], and more. An excellent survey [<xref ref-type="bibr" rid="ref11">11</xref>], provides a comprehensive literature review about OR. The SVM-based model is one of the most popular models used in the field.</p>
      <p>In the OR-DPE model, the consultation text is the input, and one label from the set of (1) very unsatisfied, (2) unsatisfied, (3) neutral, (4) satisfied, and (5) very satisfied is the output. The workflow of the OR-DPE model is shown in <xref ref-type="fig" rid="figure1">Figure 1</xref>. OR-DPE comprises of text preprocessing, representation, model training, and predictability. Because the communication between doctor and patient is through a text message, the DPE task is like text mining. The consultation texts are preprocessed and displayed as high dimensional vectors. Because the SVM-based model with linear kernel [<xref ref-type="bibr" rid="ref14">14</xref>] performs excellently on large-scale data and is well suited for text mining fields, this model is preferred to address the DPE. In this paper, a new SVM-based Ordinal Partitioning model (SVMOP) is proposed as the OR model for DPE. With the SVMOP model, OR-DPE can, not only make sure that the predicted labels are as correct as possible, but also ensure that the incorrect labels are as close to true as possible. To our knowledge, this is the first time that the issue of DPE has been conceptualized as an ordinal regression task. Empirical studies on real data sets from one of the largest mobile doctor/patient communication platforms in China show that the model can achieve state-of-the-art performance from multiple metrics.</p>
      <fig id="figure1" position="float">
        <label>Figure 1</label>
        <caption>
          <p>The general workflow of the ordinal regression for doctor performance evaluation (OR-DPE) model.</p>
        </caption>
        <graphic xlink:href="jmir_v20i7e240_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Preprocessing and Text Representation</title>
        <p>The original corpus should be preprocessed, and each sample should be represented as an input vector. In the preprocessing step, punctuation and stop words will be removed. If the experimental data is written in Chinese, the words must be segmented as in Chinese text. Sentences are represented as character strings without natural delimiters. Chinese Word Segmentation (CWS) is used to identify word sequences in a sentence and mark boundaries in appropriate places. For example, CWS can put the character sequence “天花” together as a Chinese word for “smallpox” rather than the individual Chinese character “天” (sky) and “花” (flower) respectively. Word segmentation is a preliminary and important step for preprocessing. Most methods take the CWS as a sequence labeling problem [<xref ref-type="bibr" rid="ref20">20</xref>], which can be formalized as supervised learning methods with customized features. Additionally, domain dictionaries with technical terms as ancillary resources, are beneficial for CWS and medical feature extraction. Here, 3 medical dictionaries are employed; one for Illness, one for Symptoms and one for Medicine. Most terms in the dictionaries are customized by medical experts and extended with new word detection techniques. We have collected 49,758 illness and symptom terms and 24,975 medical terms. Information about the dictionaries are shown in <xref ref-type="table" rid="table1">Table 1</xref>. For this purpose, we combined the dictionaries with Jieba tool, an open sourced Chinese segmentation software, for word segmentation.</p>
        <p>For text representation, each sample is represented as an input vector where each dimension of the vector represents a feature. The element is the corresponding feature value. Feature engineering plays an important role in text mining. Apart from the basic text features such as Bag of Words (BOW) [<xref ref-type="bibr" rid="ref21">21</xref>], unigrams, and bigrams, the custom medical features that can mirror some characteristics of the platform are utilized. These are specifically designed for the doctor/patient communication platform by domain experts and most are based on medical dictionaries. Typical text and medical features used in OR-DPE are presented in <xref ref-type="table" rid="table2">Table 2</xref>. Customized features (F1-F8) can capture domain knowledge: the count of medicine and symptom names in doctors’ answers reflects the doctors’ professional level; the number of Chinese characters in doctors’ answers mirrors the service attitudes, and more. Likewise, the text features (F9 and F10) cover most consultation information. The feature value is the numerical value of the feature while the feature value of text features is the term frequency inverted document frequency (TF-IDF) [<xref ref-type="bibr" rid="ref22">22</xref>]. TF-IDF reflects how important a word is to a document. If a word occurs rarely but appears frequently in a sample, it is most likely to reflect the characteristics of this sample. Specifically, TF-IDF is the product of two statistics: term frequency and inverse document frequency, where the former represents the frequency and the latter represents the inverse frequency of occurrence in all samples.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>The details about the medical dictionaries. “1≤terms≤3” means the number of terms having a character length less than 3 but greater than 1.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="220"/>
            <col width="430"/>
            <col width="350"/>
            <thead>
              <tr valign="top">
                <td>Number of phrases</td>
                <td colspan="2">Dictionary Name</td>
              </tr>
              <tr valign="top">
                <td><break/></td>
                <td>Illness and Symptom Dictionary (N=49,758)</td>
                <td>Medicine Dictionary (N=24,975)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>1≤terms≤3, n (%)</td>
                <td>32840 (66.00)</td>
                <td>3746 (15.00)</td>
              </tr>
              <tr valign="top">
                <td>4≤terms≤6, n (%)</td>
                <td>16918 (34.00)</td>
                <td>14486 (58.00)</td>
              </tr>
              <tr valign="top">
                <td>terms≥7, n (%)</td>
                <td>0 (0)</td>
                <td>6743 (27.00)</td>
              </tr>
              <tr valign="top">
                <td>Representative examples</td>
                <td>神经衰弱症 (Neurosis), 高血压 (HTN), 天花 (smallpox)</td>
                <td>帕罗西汀 (Paroxetine), 盐酸环苯扎林 (Flexeril)</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>F1-F8 represent the customized medical features, while F9 and F10 are the text features.</p>
          </caption>
          <table width="1000" cellpadding="1" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Feature</td>
                <td>Description</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>F1</td>
                <td>The number of symptom names in doctors’ answers</td>
              </tr>
              <tr valign="top">
                <td>F2</td>
                <td>The number of illness names in doctors’ answers</td>
              </tr>
              <tr valign="top">
                <td>F3</td>
                <td>The number of medicine names in doctors’ answers</td>
              </tr>
              <tr valign="top">
                <td>F4</td>
                <td>The number of patients’ questions</td>
              </tr>
              <tr valign="top">
                <td>F5</td>
                <td>The number of doctors’ answers</td>
              </tr>
              <tr valign="top">
                <td>F6</td>
                <td>The response time for the patient’s first question</td>
              </tr>
              <tr valign="top">
                <td>F7</td>
                <td>The number of Chinese characters in patients’ questions</td>
              </tr>
              <tr valign="top">
                <td>F8</td>
                <td>The number of Chinese characters in doctors’ answers</td>
              </tr>
              <tr valign="top">
                <td>F9</td>
                <td>Unigrams</td>
              </tr>
              <tr valign="top">
                <td>F10</td>
                <td>Bigrams</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>The quantity of text features is so large that the customized features (see <xref ref-type="table" rid="table2">Table 2</xref>) can easily be overshadowed. To highlight the importance of customized features, they are boosted by the Gradient Boosting Decision Tree (GBDT) [<xref ref-type="bibr" rid="ref23">23</xref>]. GBDT is a powerful tool in many industrial communities [<xref ref-type="bibr" rid="ref24">24</xref>]. GBDT mines the most effective features and feature combinations by a decision tree to boost the performance of regression and classification tasks. This technique is applied to increase the number of custom medical feature combinations. The main idea of GBDT is to combine weak learners into a single, strong learner like other boosting methods. GBDT is an iteration algorithm, which is composed of multiple decision trees. In the <italic>m</italic>-th iteration of GBDT, assumes that there are some imperfect models, <italic>F</italic><sub>m</sub>. The GBDT would construct a better model <italic>F</italic><sub>m+1</sub> to approach the best model by adding an estimator <italic>h</italic>, namely <italic>F</italic><sub>m+1</sub> = <italic>F</italic><sub>m</sub><italic>(x)</italic> + <italic>h(x).</italic> Then the problem is transformed by the question of how to find <italic>h(x)</italic>. As the above equations imply, a perfect <italic>h</italic> should satisfy the equation:</p>
        <disp-formula>
        <italic>h(x)</italic> = 
        <italic>F</italic>
        <sub>m+1</sub> – 
        <italic>F</italic>
        <sub>m</sub>
        <italic>(x)</italic> ≈ 
        <italic>y – F</italic>
        <sub>m</sub>
        <italic>(x)</italic></disp-formula>
        <p>where <italic>y</italic> is the true label, <italic>y – F</italic><sub>m</sub><italic>(x)</italic> is called a loss function. In practice, a general way is to apply square loss function is: ½(<italic>y</italic> –<italic>F</italic><sub>m</sub><italic>(x))</italic><sup>2</sup>. Because the residual is exactly the negative gradients of the squared loss function. The problem on the left can then be solved directly by gradient descent algorithms. In our work, we apply GBDT to boost the 8 customized features shown in <xref ref-type="table" rid="table2">Table 2</xref> to generate several effective feature combinations. According to the statistics, the number of features is 363,336 with text features, and 363,344 if adding the 8 customized features. After boosting the customized features, the number becomes 370,858. Another 7514 combined customized feature combinations have been added. The performances of various features are shown in Section Results.</p>
      </sec>
      <sec>
        <title>Model Training</title>
        <sec>
          <title>How the Ordinal Regression Method for the Ordinal Regression for Doctor Performance Evaluation Model Was Chosen</title>
          <p>There are many different models of OR. Referring to an OR survey [<xref ref-type="bibr" rid="ref11">11</xref>], the models are grouped into three categories, namely the (1) naive approach, (2) threshold approach, and (3) ordinal partitioning approach. These models have corresponding strengths and weakness. The naive approach considers OR naively, as a standard classification task or a regression task [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. At the same time, the ordering information between labels has been ignored. The threshold approach is based on the idea of approximating a real value predictor and then dividing the real line into intervals [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. Assuming <italic>P</italic> is the number of categories, the objective of threshold-based OR models is to seek <italic>P</italic> –1 parallel hyperplanes further dividing the data into ordered classes. The ordinal partitioning approach uses the ordering information to decompose the ordinal regression into several binary classification tasks. For binary classification, there are many models to choose from. For example, Frank and Hall [<xref ref-type="bibr" rid="ref16">16</xref>], applied decision trees as submodels while Waegeman and Boullart [<xref ref-type="bibr" rid="ref17">17</xref>] used weighted SVMs as binary classifiers.</p>
          <p>Since the ordering of information is conducive to model building [<xref ref-type="bibr" rid="ref11">11</xref>], we chose the OR model from the latter two methods. As the number of samples is large and the dimension of the representative vectors is high, a model was chosen that can handle large-scale and high dimensional data. So, the ordinal partitioning approach is used instead of the threshold approach for OR problems depending on paralleled hyperplanes. There are many binary classifiers that can be chosen from the submodels. Hsieh et al [<xref ref-type="bibr" rid="ref14">14</xref>] showed that the linear SVM is a robust tool that can deal with large-scale and high dimensional data. Inspired by these, we want to combine SVM with Ordinal Partitioning (SVMOP) as the OR model for the OR-DPE.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>The demo that shows how a combined support vector machine and ordinal partitioning scheme model (SVMOP) works on ordinal data.</p>
            </caption>
            <graphic xlink:href="jmir_v20i7e240_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>SVMOP Model and Training Algorithm</title>
          <p>The OR problem can be described as follows: given a training set <inline-graphic xlink:href="jmir_v20i7e240_fig7.png" mimetype="image" xlink:type="simple"/>where <italic>x</italic> ϵ <italic>R</italic><sup>l</sup> is the <italic>i</italic>-th input vector (<italic>i</italic>=1,2,…,n), where <italic>n</italic> is the number of instances, <italic>l</italic> is the number of features, and <italic>y<sub>i</sub></italic> ϵ <italic>Y</italic><sub>i</sub> is the label of <italic>x</italic><sub>i</sub>. Assuming there are <italic>P</italic> categories and without loss of generality, we take the label set <italic>Y</italic>={1,2,…P}. The goal of OR is to find a function ƒ: <italic>X</italic> → <italic>Y</italic> to predict the label of a new instance <italic>x</italic>. As mentioned earlier, SVMOP will be embedded into the OR-DPE model. <xref ref-type="fig" rid="figure2">Figure 2</xref> illustrates the SVMOP procedure. In this figure, five ordinal categories of data are represented by different colors and shapes. The idea of SVMOP is to partition the overall model into <italic>P</italic> –1 binary classifications. Then the associated question: “Is the rank of the input greater than p?” can be asked. Here <italic>p</italic>=1,2,…,<italic>P</italic> –1. Therefore, the rank of <italic>x</italic> can be determined by a sequence of these binary classification problems. Specifically, when training the <italic>p</italic>-th binary classifier, the label <italic>y</italic><sub>i</sub> is retransformed to a new class label depending on whether the label ŷ<sub>pi</sub> is greater than <italic>p</italic> or not, namely:</p>
          <graphic xlink:href="jmir_v20i7e240_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <italic>i</italic>=1,2,…,<italic>n</italic>. Therefore, the problem can be reformulated: given a training set <inline-graphic xlink:href="jmir_v20i7e240_fig9.png" mimetype="image" xlink:type="simple"/>, where <italic>x</italic> ϵ <italic>R</italic><sup>l</sup> is the <italic>i</italic>-th input sample, ŷ<sub>pi</sub> ϵ{-1,1} is defined by equation 1. The model aims to find a function to predict the ordered labels of new instances.</p>
          <p>Linear SVM is one of the best candidates among the binary classifiers dealing with high dimensional data. Then linear SVM is taken as the <italic>p</italic>-th sub-model:</p>
          <graphic xlink:href="jmir_v20i7e240_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>Where <italic>w</italic><sub>p</sub> represents the parameter of the <italic>p</italic>-th submodel, ξ<sub>pi</sub> is the slack variable of the <italic>p</italic>-th submodel. As for the optimization solver, we chose the Dual Coordinate Descent algorithm (DCD) as the training algorithm of SVM [<xref ref-type="bibr" rid="ref14">14</xref>]. DCD is one of the most effective training algorithms for linear SVMs. It solves the model in equation 2 by the Lagrange dual form. The dual form of the <italic>p</italic>-th sub-model in equation 2 is given as equation 3. Without loss of generality, we ignore the subscript p in the dual form:</p>
          <graphic xlink:href="jmir_v20i7e240_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <inline-graphic xlink:href="jmir_v20i7e240_fig12.png" mimetype="image" xlink:type="simple"/>is to employ a classic divide-and-conquer method for optimizing high dimensional problems. It starts from an initial zero vector <bold>α</bold><sup>0</sup>=<bold>0</bold> and generates a sequence of vectors <inline-graphic xlink:href="jmir_v20i7e240_fig13.png" mimetype="image" xlink:type="simple"/>. For each iteration step, the algorithm sequentially selects one dimension associated with α to optimize by fixing other dimensions. Suppose <bold>α</bold>* is the solution of equation 3 then the optimal value of <italic>w</italic><sub>p</sub> for equation 2 can be computed as follows:</p>
          <graphic xlink:href="jmir_v20i7e240_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </sec>
      </sec>
      <sec>
        <title>Model Prediction</title>
        <p>For model prediction, the research [<xref ref-type="bibr" rid="ref11">11</xref>] shows that it is important to construct an effective rule for predicting new instances in the ordinal partitioning-based OR models. Many existing ways are based on the probability manipulation or outcomes by submodels to predict the label of a new instance. In the work by Frank et al [<xref ref-type="bibr" rid="ref16">16</xref>], when estimating the probabilities for the first and the last class, the authors were dependent on a corresponding classifier. However, it needs to rely on two adjacent classifiers when computing the middle classes. This prediction method is simple and easy to implement, but may lead to a negative probability [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref28">28</xref>]. Another example in the work [<xref ref-type="bibr" rid="ref17">17</xref>], the authors combined the outcomes of all the submodels to predict the label of a new instance <italic>x</italic>. However, their prediction function may cause ambiguities for some test samples.</p>
        <p>To alleviate the problem with the above prediction functions, we propose a new prediction function as shown in equation 5:</p>
        <graphic xlink:href="jmir_v20i7e240_fig15.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>where <italic>r(x</italic> <italic>)</italic>=1 if none of <italic>w</italic><sup>T</sup><sub>p</sub><italic>x</italic> is greater than 0. This prediction function relies on the discriminant planes and joins all binary classifiers to obtain a single classification. The <italic>p</italic>-th binary classifier provides the answer to the associated question: “Is the rank of the input <italic>x</italic> greater than <italic>p</italic>?”, where <italic>p</italic>=1,2,…,<italic>P</italic> –1. That is, for prediction, the new sample <italic>x</italic> would be asked by a sequence of the questions above. And last, the predicted label equals <italic>r(x</italic> <italic>)</italic> which represents the satisfaction degree. The greater <italic>r(x</italic> <italic>)</italic>, the more satisfied.</p>
      </sec>
      <sec>
        <title>Statistical Methods and Evaluation Metrics</title>
        <p>To better highlight the characteristics of ordinal regression models, we evaluated the performance with the following three common evaluation measures: (1) mean absolute error (MAE) [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref29">29</xref>], (2) mean square error (MSE) [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref32">32</xref>], and (3) pairwise accuracy (PAcc) [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. MAE and MSE can directly measure the degree of deviation between the true label (<italic>gold</italic><sub>i</sub>) and predicted label (<italic>predicted</italic>). They can be defined by the following equations:</p>
        <graphic xlink:href="jmir_v20i7e240_fig16.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <graphic xlink:href="jmir_v20i7e240_fig17.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>Since they are metrics measuring the error, the lower they are, the better their performance. PAcc is widely applied in the medical data analysis, ranking and statistics fields with the name of concordance index or Kendall τ [<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]. PAcc could reflect the correct ratio of ranking between pairwise instances. Specifically, the set of preference evaluation pairs is represented as <italic>S,S</italic>={(<italic>i,j</italic>) &#124; <italic>gold<sub>i</sub> &#62; gold<sub>j</sub></italic> }.</p>
        <p>The PAcc is given by</p>
        <graphic xlink:href="jmir_v20i7e240_fig18.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>where “&#124; <italic>S</italic> &#124;” represents the number of the set <italic>S</italic>. It accords with the rule: the greater, the better.</p>
      </sec>
      <sec>
        <title>Mining Predictive Features</title>
        <p>Apart from rating doctors’ performance, we continue to explore the most predictive features among text features and customized features in DPE. In general, predictive features always play significant and instructive roles on the platform construction. In this case, the most important features were extracted by analyzing the weight matrix <italic>W ϵ R<sup>l x (P-1))</sup></italic>, where <italic>l</italic> and <italic>P</italic> –1 are the dimensions of the matrix. As mentioned, <italic>l</italic> is the total number of all the features (that is, <italic>l</italic>=363<italic>,</italic> 344) and <italic>P</italic> is the number of categories, where <italic>P</italic> –1 is the number of the submodels. In equation 2, <italic>W</italic> is composed of the weight parameters, with <italic>w</italic> in each submodel, namely <italic>W</italic>=(<italic>w</italic><sub>1</sub><italic>,w</italic><sub>2</sub><italic>,…,w</italic><sub>P–1</sub>). We denote <italic>W(j,:)</italic> as the <italic>j</italic>-th row vector and the absolute value of the elements in the row vector represents the contributions to each submodel for the <italic>j</italic>-th feature. The larger the value is, the more predictive property the feature has. For every feature in each kind of text feature or customized feature, described in <xref ref-type="table" rid="table2">Table 2</xref>, it owns its corresponding weight vector <italic>W(j,:)</italic>, where 1 ≤ <italic>j</italic> ≤ <italic>l</italic>. We compute the total contribution Con<sub>j</sub> of the <italic>j</italic>-th feature to the model decision by equation 9:</p>
        <graphic xlink:href="jmir_v20i7e240_fig19.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>where “∥ ∥<sub>2</sub>” represents the L2-norm of a vector. When the contributions of all the features have been computed, they would be ranked and hence obtain the top-most predictive features.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Preparation of Datasets</title>
        <p>To validate the proposed model on real data, the data from one of the popular doctor/patient consultation platforms (Platform X) in China was chosen as the experimental data sets. In practice, the platform maintains long-term cooperation with us. However, in order to comply with the confidentiality agreement with the platform, we are not able to use the real name of the platform in the paper but instead we use the name Platform X throughout. On Platform X, the consultation mainly consists of patients’ questions (eg, “医生您好，如何补钙” [Hi doctor, how can I add more calcium?]) and the response (eg, “很高兴为您服务，您可以通过牛奶，豆制品，鱼等食物，也可以口服碳酸钙和葡萄酸钙” [“Glad to help you,” “you can eat foods such as milk, bean products, and fish” and “You can also take calcium gluconate and calcium carbonate directly”]). To introduce experimental data, an actual consultation letter was used (<xref ref-type="fig" rid="figure3">Figure 3</xref>). Based on analysis of patients’ questions, multiple question types are proposed. Most questions are about ailments that are not serious or related to personal privacy, like chronic pharyngitis and dermatosis. And, because they are flexible and convenient, most consultations are done through mobile software applications. Platform X faces the same problem when evaluating doctor’s online performance. Platform X did not receive direct customer ratings or feedback since most patients tend to rate the very good or bad and at times feedback was not received because, for example, a customer may have been offline.</p>
        <p>Of a sampling of 2,337,828 clinical data collected over the last twenty days, only 841,618 (36%) of the data was labeled by patients, which proves that most patients do not like to provide feedback. From the labeled data, only 720 instances, 1712 instances, and 8737 instances were labeled very unsatisfied, unsatisfied, and neutral respectively. The unbalanced data challenges the model. To alleviate the issue of unbalanced data and collect more instances of “very unsatisfied,” we chose sample training data from Platform X’s database. As previously mentioned, we have a long-term association with Platform X. It takes approximately two hours to access the entire database. Data collected is valid for about 18 months. After that, the same number of instances from each category are sampled. After filtering the data, (removing consultations with a length of less than 10 words), we have approximately 112,485 instances to use as experimental datasets. Each category contains approximately 22,497 samples which are randomly split into five sections with four sections serving as the training sets and the remaining one as the test set.</p>
      </sec>
      <sec>
        <title>Baseline Methods</title>
        <p>To better reflect the effectiveness of the proposed model with the above metrics, the following baseline methods will be compared with our model. These methods are popular and representative in OR fields. To solve the high dimensional data efficiently, the following models all use the linear kernel. The DCD algorithm is adopted to solve the following models. These are implemented by modifying the open source package LIBLINEAR [<xref ref-type="bibr" rid="ref36">36</xref>] directly and all the codes related to the experiment are uploaded to a Github website. The following methods were used to compare with our model.</p>
        <list list-type="order">
          <list-item>
            <p>SVC [<xref ref-type="bibr" rid="ref14">14</xref>]: Support vector classification with one versus the rest. This model belongs to the naive approach.</p>
          </list-item>
          <list-item>
            <p>SVR [<xref ref-type="bibr" rid="ref37">37</xref>]: Support vector regression. The ordinal labels are treated as continuous values. When predicting new instances, the predictions for test instances are rounded to the nearest ordinal label and the model belongs to the naive approach.</p>
          </list-item>
          <list-item>
            <p>LR [<xref ref-type="bibr" rid="ref38">38</xref>]: Logistic regression one versus the rest. This model belongs to the naive approach.</p>
          </list-item>
          <list-item>
            <p>SVOR [<xref ref-type="bibr" rid="ref10">10</xref>]: Support vector ordinal regression. This model aims to optimize multiple thresholds to define parallel discriminant hyperplanes. The SVOR model is used with implicit constraints and belongs to the threshold approach.</p>
          </list-item>
          <list-item>
            <p>RedSVM [<xref ref-type="bibr" rid="ref39">39</xref>]: Reduction support vector machine. A threshold approach and it is a reduction framework from ordinal ranking to weighted binary classification by extending examples.</p>
          </list-item>
        </list>
      </sec>
      <sec>
        <title>Evaluation Performance</title>
        <p>First, we compare the performance of five different baselines with our SVMOP model using different sets of features, including (1) text features (T), (2) text and customized features (T+C), and (3) text, customized, and boosted features (T+C+B). Three metrics, (ie, MAE, MSE, and PAcc) are used to evaluate model performance. <xref ref-type="table" rid="table3">Table 3</xref> shows the results of the experiment. The best performance for each metric is represented by the footnote “k” while the best “one of” feature sets is represented by the footnote “e.” In <xref ref-type="table" rid="table3">Table 3</xref>, the SVMOP model outperforms other baselines on MAE, MSE and PAcc with each type of feature sets, demonstrating the effectiveness of our model. On different set of features, all models achieved better performance with feature set T+C and feature set T+C+B. Furthermore, compared with feature set T+C+B, feature set T+C attained more improvement. In other words, using customized features are important for performance improvement</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>An example of consultation letters on Platform X. The left subfigure is the real consultation on Platform X by mobile software applications but without sensitive information such as doctors’ photos. The right one is the version in English.</p>
          </caption>
          <graphic xlink:href="jmir_v20i7e240_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Performances of various models having multiple feature sets (T, T+C, T+C+B) are shown in this table.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="0"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <thead>
              <tr valign="top">
                <td>Method</td>
                <td colspan="3">Text (T)</td>
                <td colspan="3">Text and Customized (T+C)</td>
                <td colspan="4">Text, Customized, and Booster (T+C+B)</td>
              </tr>
              <tr valign="bottom">
                <td><break/></td>
                <td>MAE<sup>a</sup></td>
                <td>MSE<sup>b</sup></td>
                <td>PAcc<sup>c</sup> (%)</td>
                <td>MAE</td>
                <td>MSE</td>
                <td colspan="2">PAcc (%)</td>
                <td>MAE</td>
                <td>MSE</td>
                <td>PAcc (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="bottom">
                <td>SVC<sup>d</sup></td>
                <td>0.7925</td>
                <td>1.7613</td>
                <td>53.32</td>
                <td>0.6726</td>
                <td>1.3759</td>
                <td colspan="2">57.32</td>
                <td>0.6212<sup>e</sup></td>
                <td>1.1981<sup>e</sup></td>
                <td>59.05<sup>e</sup></td>
              </tr>
              <tr valign="bottom">
                <td>SVR<sup>f</sup></td>
                <td>0.8023</td>
                <td>1.3302</td>
                <td>49.74</td>
                <td>0.7050</td>
                <td>1.1106</td>
                <td colspan="2">54.24</td>
                <td>0.6906<sup>e</sup></td>
                <td>1.0332<sup>e</sup></td>
                <td>56.37<sup>e</sup></td>
              </tr>
              <tr valign="bottom">
                <td>LR<sup>g</sup></td>
                <td>0.7716</td>
                <td>1.6883</td>
                <td>53.86</td>
                <td>0.6359</td>
                <td>1.2606</td>
                <td colspan="2">57.77</td>
                <td>0.5978<sup>e</sup></td>
                <td>1.1310<sup>e</sup></td>
                <td>59.50<sup>e</sup></td>
              </tr>
              <tr valign="bottom">
                <td>SVOR<sup>h</sup></td>
                <td>0.8086</td>
                <td>1.3742</td>
                <td>49.58</td>
                <td>0.7170</td>
                <td>1.1167</td>
                <td colspan="2">54.09</td>
                <td>0.6665<sup>e</sup></td>
                <td>1.0143<sup>e</sup></td>
                <td>57.20<sup>e</sup></td>
              </tr>
              <tr valign="bottom">
                <td>RedSVM<sup>i</sup></td>
                <td>0.8046</td>
                <td>1.3715</td>
                <td>50.11</td>
                <td>0.7168</td>
                <td>1.1127</td>
                <td colspan="2">54.00</td>
                <td>0.6718<sup>e</sup></td>
                <td>1.0236<sup>e</sup></td>
                <td>57.21<sup>e</sup></td>
              </tr>
              <tr valign="bottom">
                <td>SVMOP<sup>j</sup></td>
                <td>0.7054<sup>k</sup></td>
                <td>1.2706<sup>k</sup></td>
                <td>54.11<sup>k</sup></td>
                <td>0.6130<sup>k</sup></td>
                <td>1.0108<sup>k</sup></td>
                <td colspan="2">57.92<sup>k</sup></td>
                <td>0.5864<sup>e,k</sup></td>
                <td>0.9605<sup>e,k</sup></td>
                <td>59.65<sup>e,k</sup></td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>MAE: mean absolute error.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>MSE: mean standard error.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>PAcc: pairwise accuracy.</p>
            </fn>
            <fn id="table3fn4">
              <p><sup>d</sup>SVC: support vector classification.</p>
            </fn>
            <fn id="table3fn5">
              <p><sup>e</sup>Best “one of” feature sets.</p>
            </fn>
            <fn id="table3fn6">
              <p><sup>f</sup>SVR: support vector regression.</p>
            </fn>
            <fn id="table3fn7">
              <p><sup>g</sup>LR: logistic regression.</p>
            </fn>
            <fn id="table3fn8">
              <p><sup>h</sup>SVOR: support vector ordinal regression.</p>
            </fn>
            <fn id="table3fn9">
              <p><sup>i</sup>RedSVM: reduction support vector machine.</p>
            </fn>
            <fn id="table3fn10">
              <p><sup>j</sup>SVMOP: a combined support vector machine and ordinal partitioning scheme model.</p>
            </fn>
            <fn id="table3fn11">
              <p><sup>k</sup>Best performance for each metric.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Mean absolute error (MAE), mean square error (MSE), and pairwise accuracy (PAcc) varying from different models and different feature sets. LR: logistic regression; RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; T: text features; T+C: text and customized features; T+C+B: text, customized, and boosted features.</p>
          </caption>
          <graphic xlink:href="jmir_v20i7e240_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p><xref ref-type="fig" rid="figure4">Figure 4</xref> displays the performances of 6 models on 3 measures, namely MAS, MSE, and PAcc. As we can see, SVMOP greatly outperforms the other models on MAE, MSE, and PAcc. Additionally, the models that consider ordering information, namely, SVOR, RedSVM, and SVMOP, perform better than the rest on MSE; and SVC and LR achieve comparable performances with SVMOP on PAcc. To investigate the influence of the parameter, we show the various performance of each model as we change the parameter log<sub>2</sub><italic>C</italic> in a range [-5,5]. In <xref ref-type="fig" rid="figure5">Figure 5</xref>, we find that the performances vary as the parameters change and the model can achieve the best performance in this range.</p>
        <p>Additionally, the confusion matrices were used to further discuss the differences among the performance of different models. Each confusion matrix is generated by the corresponding model on feature set T+C+B. As shown in <xref ref-type="fig" rid="figure6">Figure 6</xref>, models that consider ordering information, such as SVOR, RedSVM, and SVMOP, misclassify the incorrectly labeled samples into the closest categories. For example, in the confusion matrix of SVMOP, the third cell of the third row shows that 63% of most (17%) misclassified instances fell into Category 2. In contrast to the confusion matrix of SVMOP, when looking at the third row of the confusion matrix of SVC, we find that most (19%) of the misclassified instances fall into Category 1. For this study, this example illustrates that the nonordering information methods, such as SVC and LR, can misclassify doctors having <italic>neutral</italic> performance levels into the <italic>very unsatisfied</italic> category. However, methods that consider ordering information of doctors’ performances, such as SVOR, RedSVM and SVMOP, are more likely to place misclassified neutral doctors into the <italic>unsatisfied</italic> category.</p>
      </sec>
      <sec>
      <title>Predictive Features Analysis</title>
      <p>As for mining the most predictive features, (see equation 9), and after computing and sorting each feature, we find that 4 customized features are in the top 5 most predictive features, including F3, F4, F5, and F7. Feature F6 ranks 8th and feature F8 ranks 23rd. In other words, most of the customized features play the most predictive roles in DPE, which is consistent with our numerical results. In view of the most predictive text features, we find that the features that contribute most to the model decision are the polite expressions like: “不客气” (“You’re welcome”), “谢谢您” (“Thank you”), “很乐意帮助您” (“Glad to help you”) and sensitive words such as “好评” (“good rating”), “态度”(“attitude”), “五星”(“five-star”). Some words like “禁忌辛辣食物” (“avoid spicy foods”), “对身体有害” (“bad for health”) are helpful, and indicates that the doctor is explaining some issues in more detail. These features cannot guide in questionnaire design but are beneficial for platform building and optimization.</p>
      <fig id="figure5" position="float">
        <label>Figure 5</label>
        <caption>
          <p>The different performances with different parameters in training process with the text, customized, and boosted feature set (T+C+B). LR: logistic regression; MAE: mean absolute error; MSE: mean square error; PAcc: pairwise accuracy; RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; SVR: support vector regression.</p>
        </caption>
        <graphic xlink:href="jmir_v20i7e240_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <fig id="figure6" position="float">
        <label>Figure 6</label>
        <caption>
          <p>The confusion matrices of different models with the text, customized, and boosted feature set (T+C+B). RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; SVR: support vector regression.</p>
        </caption>
        <graphic xlink:href="jmir_v20i7e240_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig> </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Results</title>
        <p>Statistically, 36% of data on doctor/patient communication platforms has been labeled by patients, showing that 64% of the clinical data lack evaluation. Considering that doctors’ performance could affect patient satisfaction, we take the DPE task as an ordinal regression problem, ensuring the automatically predicted labels are as close as possible to the true ones. The OR-DPE, SVMOP model with revised prediction is applied as the core model, and the metrics of MAE, MSE, PAcc, and SVMOP models with feature set T+C+B could achieve state-of-the-art performance. Compared with other auto-evaluation models, SVMOP improves MAE by 0.1, MSE by 0.5, and PAcc by 5%. The customized features improve MAE by 0.1, MSE by 0.2, and PAcc by 3%. Additionally, with the boosting technique, the performance of SVMOP can be further improved. Furthermore, based on OR-DPE model, predictive features like polite expressions and sentiment words can also be mined, which can be used to guide the development of mHealth platforms.</p>
      </sec>
      <sec>
      <title>Comparison with Prior Work</title>
      <p>The experiments conducted on real data have validated the effectiveness of SVMOP. Because of the noise in the real data, we continue to experiment on benchmark OR datasets [<xref ref-type="bibr" rid="ref19">19</xref>] in a precisely controlled environment. The datasets can be downloaded from the public website. In this experiment, we compare our model with all the baselines mentioned in the paper. The details about the benchmark datasets and the results are shown in <xref ref-type="table" rid="table4">Tables 4</xref> and <xref ref-type="table" rid="table5">5</xref>.</p>
      <p>We find that SVMOP always performs better than other baselines on MAE, MSE, and PAcc. The results verify the effectiveness of SVMOP on clean data. Therefore, the good results benefit from the SVMOP model but not the experimental data about DPE, which further demonstrates the correctness of choosing SVMOP as the core model of OR-DPE.</p>  
      </sec>
      <sec>
      <title>Limitations</title>
      <p>Although this study has solved the problem of doctors’ auto-evaluation on doctor/patient communication platforms by the ordinal regression approach, there are limitations. Firstly, the definition of a good consult here is related to user satisfaction, not to medical accuracy or clinical utility. A good doctor seems to be a likable one, but a likable one may make incorrect medical decisions. Secondly, Farmer et al [<xref ref-type="bibr" rid="ref40">40</xref>] point out that doctors’ work should be evaluated by multiple complex professional factors. In other words, a good consult is not only related to patients but also to many other factors. One way to handle this issue is to multisource feedback [<xref ref-type="bibr" rid="ref41">41</xref>], which is called 360-degree evaluation in which key performance behaviors are simultaneously rated by peers, patients, and coworkers. Considering the characteristics of doctor/patient communication platforms, peer evaluation can be achieved by questionnaires, and the predictive features generated by the OR-DPE model may, in turn guide the questionnaire design.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>The authors are the first to conceptualize the problem of DPE as an ordinal regression task and develop an OR-DPE model to address it. Apart from the basic text features, we use eight customized features suggested by domain experts as important features to improve model performance. Furthermore, we applied GBDT to boost the 8 customized features. Additionally, we proposed a new model called SVMOP which has a reasonable and effective prediction function. Experiments show that the performance of SVMOP is better than many other machine learning methods on MAE, MSE, and PAcc. In summary, with the OR-DPE model, the mHealth platform could not only make an auto-evaluation of online doctors’ performance but also mine the most effective features which can then be further applied to guide the promotion of doctors and platforms. In the future, we hope our model can also be explored and applied to other medical service-oriented issues in medical education.</p>
      <table-wrap position="float" id="table4">
        <label>Table 4</label>
        <caption>
          <p>Benchmark datasets. “#ins” is the number of instances. “#fea” is the number of features. “#class” is the number of classes.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <col width="250"/>
          <thead>
            <tr valign="top">
              <td>Datasets</td>
              <td>#ins</td>
              <td>#fea</td>
              <td>#class</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td>housing-5</td>
              <td>10120</td>
              <td>14</td>
              <td>5</td>
            </tr>
            <tr valign="top">
              <td>machine-5</td>
              <td>4180</td>
              <td>7</td>
              <td>5</td>
            </tr>
            <tr valign="top">
              <td>abalone-5</td>
              <td>83540</td>
              <td>11</td>
              <td>5</td>
            </tr>
            <tr valign="top">
              <td>housing-10</td>
              <td>10120</td>
              <td>14</td>
              <td>10</td>
            </tr>
            <tr valign="top">
              <td>machine-10</td>
              <td>4180</td>
              <td>7</td>
              <td>10</td>
            </tr>
            <tr valign="top">
              <td>abalone-10</td>
              <td>83540</td>
              <td>11</td>
              <td>10</td>
            </tr>
          </tbody>
        </table>
      </table-wrap>
      <table-wrap position="float" id="table5">
        <label>Table 5</label>
        <caption>
          <p>The mean absolute error (MAE), mean standard error (MSE), and pairwise accuracy (PAcc) performances of different models on benchmark datasets. The best result is indicated by a footnote.</p>
        </caption>
        <table width="1000" cellpadding="7" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="30"/>
          <col width="250"/>
          <col width="120"/>
          <col width="120"/>
          <col width="120"/>
          <col width="120"/>
          <col width="120"/>
          <col width="120"/>
          <thead>
            <tr valign="bottom">
              <td colspan="2">Datasets</td>
              <td>SVC<sup>a</sup></td>
              <td>SVR<sup>b</sup></td>
              <td>LR<sup>c</sup></td>
              <td>SVOR<sup>d</sup></td>
              <td>RedSVM<sup>e</sup></td>
              <td>SVMOP<sup>f</sup></td>
            </tr>
          </thead>
          <tbody>
            <tr valign="bottom">
              <td colspan="2"><bold>Mean absolute error (MAE)</bold></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>housing-5</td>
              <td>0.517</td>
              <td>0.454</td>
              <td>0.435</td>
              <td>0.398</td>
              <td>0.403</td>
              <td>0.366<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-5</td>
              <td>0.606</td>
              <td>0.550</td>
              <td>0.451</td>
              <td>0.390</td>
              <td>0.424</td>
              <td>0.369<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-5</td>
              <td>0.798</td>
              <td>0.712</td>
              <td>0.700</td>
              <td>0.683</td>
              <td>0.675</td>
              <td>0.648<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>housing-10</td>
              <td>1.513</td>
              <td>0.962</td>
              <td>0.999</td>
              <td>0.859</td>
              <td>0.848</td>
              <td>0.757<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-10</td>
              <td>1.425</td>
              <td>1.151</td>
              <td>0.986</td>
              <td>0.935</td>
              <td>0.927</td>
              <td>0.841<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-10</td>
              <td>1.959</td>
              <td>1.451</td>
              <td>1.557</td>
              <td>1.435</td>
              <td>1.434</td>
              <td>1.391<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td colspan="2"><bold>Mean standard error (MSE)</bold></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>housing-5</td>
              <td>0.665</td>
              <td>0.545</td>
              <td>0.612</td>
              <td>0.494</td>
              <td>0.524</td>
              <td>0.446<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-5</td>
              <td>0.994</td>
              <td>0.634</td>
              <td>0.648</td>
              <td>0.469</td>
              <td>0.505</td>
              <td>0.429<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-5</td>
              <td>1.450</td>
              <td>0.992</td>
              <td>1.244</td>
              <td>1.042</td>
              <td>0.991</td>
              <td>0.962<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>Housing-10</td>
              <td>4.564</td>
              <td>1.858</td>
              <td>2.560</td>
              <td>1.694</td>
              <td>1.642</td>
              <td>1.453<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-10</td>
              <td>3.998</td>
              <td>2.487</td>
              <td>2.277</td>
              <td>1.786</td>
              <td>1.720</td>
              <td>1.547<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-10</td>
              <td>7.222</td>
              <td>3.703</td>
              <td>5.091</td>
              <td>3.586</td>
              <td>3.783</td>
              <td>3.635<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td colspan="2"><bold>Pairwise accuracy (PAcc)</bold></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
              <td><break/></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>housing-5</td>
              <td>0.614</td>
              <td>0.638</td>
              <td>0.658</td>
              <td>0.663</td>
              <td>0.659</td>
              <td>0.676<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-5</td>
              <td>0.602</td>
              <td>0.604</td>
              <td>0.652</td>
              <td>0.666</td>
              <td>0.655</td>
              <td>0.680<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-5</td>
              <td>0.547</td>
              <td>0.553</td>
              <td>0.584</td>
              <td>0.584</td>
              <td>0.577</td>
              <td>0.589<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>Housing-10</td>
              <td>0.552</td>
              <td>0.623</td>
              <td>0.609</td>
              <td>0.635</td>
              <td>0.637</td>
              <td>0.642<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>machine-10</td>
              <td>0.488</td>
              <td>0.562</td>
              <td>0.597</td>
              <td>0.601</td>
              <td>0.599</td>
              <td>0.612<sup>g</sup></td>
            </tr>
            <tr valign="bottom">
              <td><break/></td>
              <td>abalone-10</td>
              <td>0.514</td>
              <td>0.568</td>
              <td>0.566</td>
              <td>0.565</td>
              <td>0.568</td>
              <td>0.569<sup>g</sup></td>
            </tr>
          </tbody>
        </table>
        <table-wrap-foot>
          <fn id="table5fn1">
            <p><sup>a</sup>SVC: support vector classification.</p>
          </fn>
          <fn id="table5fn2">
            <p><sup>b</sup>SVR: support vector regression.</p>
          </fn>
          <fn id="table5fn3">
            <p><sup>c</sup>LR: logistic regression.</p>
          </fn>
          <fn id="table5fn4">
            <p><sup>d</sup>SVOR: support vector ordinal regression.</p>
          </fn>
          <fn id="table5fn5">
            <p><sup>e</sup>RedSVM: reduction support vector machine.</p>
          </fn>
          <fn id="table5fn6">
            <p><sup>f</sup>SVMOP: a combined support vector machine and ordinal partitioning scheme model.</p>
          </fn>
          <fn id="table5fn7">
            <p><sup>g</sup>Best result.</p>
          </fn>
        </table-wrap-foot>
      </table-wrap>
      </sec>
    </sec>
  </body>
  <back>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">BOW</term>
          <def>
            <p>bag of words</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">CWS</term>
          <def>
            <p>Chinese word segmentation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">DCD</term>
          <def>
            <p>Dual Coordinate Descent algorithm</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">DPE</term>
          <def>
            <p>doctor performance evaluation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">GBDT</term>
          <def>
            <p>Gradient Boosting Decision Tree</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">LR</term>
          <def>
            <p>logistic regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">MAE</term>
          <def>
            <p>mean absolute error</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">mHealth</term>
          <def>
            <p>mobile health</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">MSE</term>
          <def>
            <p>mean square error</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">OR</term>
          <def>
            <p>ordinal regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">OR-DPE</term>
          <def>
            <p>ordinal regression for doctor performance evaluation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">PAcc</term>
          <def>
            <p>pairwise accuracy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">RedSVM</term>
          <def>
            <p>reduction support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">SVC</term>
          <def>
            <p>support vector classification</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">SVMOP</term>
          <def>
            <p>a combined support vector machine and ordinal partitioning scheme model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb17">SVOR</term>
          <def>
            <p>support vector ordinal regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb18">SVR</term>
          <def>
            <p>support vector regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb19">T</term>
          <def>
            <p>text features</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb20">T+C</term>
          <def>
            <p>text and customizable features</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb21">T+C+B</term>
          <def>
            <p>text, customizable, and boosted features</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb22">TF-IDF</term>
          <def>
            <p>term frequency inverted document frequency</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was supported by the National Natural Science Foundation of China (Grant No. 91546201, No. 71331005, No. 71110107026, No. 11671379, No. 11331012, No. 71501044), the Fundamental Research Funds for the Central Universities in UIBE (No. 16YQ07), UCAS Grant (No. Y55202LY00).</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Eysenbach</surname>
            <given-names>G</given-names>
          </name>
        </person-group>
        <article-title>What is e-health?</article-title>
        <source>Journal of Medical Internet Research</source>  
        <year>2001</year>  
        <month>6</month>  
        <volume>3</volume>  
        <issue>2</issue>  
        <fpage>E20</fpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmir.org/2001/2/e20/"/>
        </comment>  
        <pub-id pub-id-type="doi">10.2196/jmir.3.2.e20</pub-id>
        <pub-id pub-id-type="medline">11720962</pub-id>
        <pub-id pub-id-type="pmcid">PMC1761894</pub-id></nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Borrelli</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Ritterband</surname>
            <given-names>LM</given-names>
          </name>
        </person-group>
        <article-title>Special issue on eHealth and mHealth: Challenges and future directions for assessment, treatment, and dissemination</article-title>
        <source>Health Psychology</source>  
        <year>2015</year>  
        <month>12</month>  
        <volume>34S</volume>  
        <fpage>1205</fpage>  
        <pub-id pub-id-type="doi">10.1037/hea0000323</pub-id>
        <pub-id pub-id-type="medline">26651461</pub-id>
        <pub-id pub-id-type="pii">2015-56045-001</pub-id></nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="web">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Jonathan</surname>
            <given-names>Govette</given-names>
          </name>
        </person-group>
        <source>30 Amazing Mobile Health Technology Statistics for Today's Physician</source>  
        <access-date>2017-10-01</access-date>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://getreferralmd.com/2015/08/mobile-healthcare-technology-statistics/">https://getreferralmd.com/2015/08/mobile-healthcare-technology-statistics/</ext-link>
          <ext-link ext-link-type="webcite" xlink:href="70SvHWknK"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Caruana</surname>
            <given-names>A</given-names>
          </name>
        </person-group>
        <article-title>Service loyalty: The effects of service quality and the mediating role of customer satisfaction</article-title>
        <source>European journal of marketing</source>  
        <year>2002</year>  
        <volume>36</volume>  
        <issue>7/8</issue>  
        <fpage>811</fpage>  
        <lpage>828</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1108/03090560210430818"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1108/03090560210430818</pub-id>
        <pub-id pub-id-type="medline">10786546</pub-id></nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Gremler</surname>
            <given-names>DD</given-names>
          </name>
          <name name-style="western">
            <surname>Brown</surname>
            <given-names>SW</given-names>
          </name>
        </person-group>
        <article-title>Service loyalty: its nature, importance, and implications</article-title>
        <source>Advancing service quality: A global perspective</source>  
        <year>1996</year>  
        <volume>5</volume>  
        <fpage>171</fpage>  
        <lpage>181</lpage>  
        <pub-id pub-id-type="doi">10.1017/CBO9781107415324.004</pub-id></nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Mosahab</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Mahamad</surname>
            <given-names>O</given-names>
          </name>
          <name name-style="western">
            <surname>Ramayah</surname>
            <given-names>T</given-names>
          </name>
        </person-group>
        <article-title>Service quality, customer satisfaction and loyalty: A test of mediation</article-title>
        <source>International business research</source>  
        <year>2010</year>  
        <volume>3</volume>  
        <issue>4</issue>  
        <fpage>72</fpage>  
        <pub-id pub-id-type="doi">10.5539/ibr.v3n4p72</pub-id></nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Overeem</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Faber</surname>
            <given-names>M J</given-names>
          </name>
          <name name-style="western">
            <surname>Arah</surname>
            <given-names>O A</given-names>
          </name>
        </person-group>
        <article-title>Doctor performance assessment in daily practise: does it help doctors or not? A systematic review</article-title>
        <source>Medical education</source>  
        <year>2007</year>  
        <month>12</month>  
        <volume>41</volume>  
        <issue>11</issue>  
        <fpage>1039</fpage>  
        <lpage>1049</lpage>  
        <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2006.09.001</pub-id>
        <pub-id pub-id-type="medline">17049917</pub-id></nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Li</surname>
            <given-names>P</given-names>
          </name>
          <name name-style="western">
            <surname>Yan</surname>
            <given-names>Y</given-names>
          </name>
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Customer voice sensor: A comprehensive opinion mining system for call center conversation</article-title>
        <year>2016</year>  
        <conf-name>IEEE International Conference on Cloud Computing and Big Data Analysis (ICCCBDA)</conf-name>
        <conf-date>2016</conf-date>
        <conf-loc>Chengdu, China</conf-loc>
        <fpage>324</fpage>  
        <lpage>329</lpage>  
        <pub-id pub-id-type="doi">10.1109/ICCCBDA.2016.7529578</pub-id></nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Paprzycki</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Abraham</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Guo</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Data mining approach for analyzing call center performance</article-title>
        <source>Innovations in Applied Artificial Intelligence</source>  
        <year>2004</year>  
        <fpage>1092</fpage>  
        <lpage>1101</lpage>  
        <pub-id pub-id-type="doi">10.1007/978-3-540-24677-0_112</pub-id></nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Chu</surname>
            <given-names>W</given-names>
          </name>
          <name name-style="western">
            <surname>Keerthi</surname>
            <given-names>SS</given-names>
          </name>
        </person-group>
        <article-title>Support vector ordinal regression</article-title>
        <source>Neural computation</source>  
        <year>2007</year>  
        <volume>19</volume>  
        <issue>3</issue>  
        <fpage>792</fpage>  
        <lpage>815</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1162/neco.2007.19.3.792"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Gutiérrez</surname>
            <given-names>PA</given-names>
          </name>
          <name name-style="western">
            <surname>Pérez-Ortiz</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Sanchez-Monedero</surname>
            <given-names>J</given-names>
          </name>
        </person-group>
        <article-title>Ordinal regression method survey and experimental study</article-title>
        <source>IEEE Transactions on Knowledge and Data Engineering</source>  
        <year>2016</year>  
        <volume>28</volume>  
        <issue>1</issue>  
        <fpage>127</fpage>  
        <lpage>146</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/TKDE.2015.2457911"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Huhn</surname>
            <given-names>JC</given-names>
          </name>
          <name name-style="western">
            <surname>Hullermeier</surname>
            <given-names>E</given-names>
          </name>
        </person-group>
        <article-title>Is an ordinal class structure useful in classifier learning?</article-title>
        <source>International Journal of Data Mining, Modelling and Management</source>  
        <year>2008</year>  
        <volume>1</volume>  
        <issue>1</issue>  
        <fpage>45</fpage>  
        <lpage>67</lpage>  
        <pub-id pub-id-type="doi">10.1504/IJDMMM.2008.022537</pub-id></nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>McCullagh</surname>
            <given-names>P</given-names>
          </name>
        </person-group>
        <article-title>Regression models for ordinal data</article-title>
        <source>Journal of the royal statistical society</source>  
        <year>1980</year>  
        <volume>42</volume>  
        <issue>2</issue>  
        <fpage>109</fpage>  
        <lpage>142</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jstor.org/stable/2984952?seq=1#page_scan_tab_contents"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Hsieh</surname>
            <given-names>CJ</given-names>
          </name>
          <name name-style="western">
            <surname>Chang</surname>
            <given-names>KW</given-names>
          </name>
          <name name-style="western">
            <surname>Lin</surname>
            <given-names>CJ</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>A dual coordinate descent method for large-scale linear SVM</article-title>
        <year>2008</year>  
        <conf-name>Proceedings of the 25th international conference on Machine learning. ACM</conf-name>
        <conf-date>2008</conf-date>
        <conf-loc>New York, USA</conf-loc>
        <fpage>408</fpage>  
        <lpage>415</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/1390156.1390208"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Chu</surname>
            <given-names>W</given-names>
          </name>
          <name name-style="western">
            <surname>Keerthi</surname>
            <given-names>SS</given-names>
          </name>
        </person-group>
        <article-title>New approaches to support vector ordinal regression</article-title>
        <year>2005</year>  
        <conf-name>The 22nd international conference on Machine learning</conf-name>
        <conf-date>2005</conf-date>
        <conf-loc>New York, USA</conf-loc>
        <fpage>145</fpage>  
        <lpage>152</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/1102351.1102370"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Frank</surname>
            <given-names>E</given-names>
          </name>
          <name name-style="western">
            <surname>Hall</surname>
            <given-names>M</given-names>
          </name>
        </person-group>
        <article-title>A simple approach to ordinal classification</article-title>
        <year>2001</year>  
        <conf-name>The European Conference on Machine Learning</conf-name>
        <conf-date>2001</conf-date>
        <conf-loc>United Kingdom</conf-loc>
        <fpage>145</fpage>  
        <lpage>156</lpage>  
        <pub-id pub-id-type="doi">10.1007/3-540-44795-4_13</pub-id></nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Waegeman</surname>
            <given-names>W</given-names>
          </name>
          <name name-style="western">
            <surname>Boullart</surname>
            <given-names>L</given-names>
          </name>
        </person-group>
        <article-title>An ensemble of weighted support vector machines for ordinal regression</article-title>
        <source>International Journal of Computer Systems Science and Engineering</source>  
        <year>2009</year>  
        <volume>3</volume>  
        <issue>1</issue>  
        <fpage>47</fpage>  
        <lpage>51</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/1102351.1102370"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Deng</surname>
            <given-names>WY</given-names>
          </name>
          <name name-style="western">
            <surname>Zheng</surname>
            <given-names>QH</given-names>
          </name>
          <name name-style="western">
            <surname>Lian</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Ordinal extreme learning machine</article-title>
        <source>Neurocomputing</source>  
        <year>2010</year>  
        <volume>74</volume>  
        <issue>1</issue>  
        <fpage>447</fpage>  
        <lpage>456</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.neucom.2010.08.022"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Chu</surname>
            <given-names>W</given-names>
          </name>
          <name name-style="western">
            <surname>Ghahramani</surname>
            <given-names>Z</given-names>
          </name>
        </person-group>
        <article-title>Gaussian processes for ordinal regression</article-title>
        <source>Journal of machine learning research</source>  
        <year>2005</year>  
        <volume>6</volume>  
        <issue>3</issue>  
        <fpage>1019</fpage>  
        <lpage>1041</lpage> </nlm-citation>
      </ref>
      
      
      
      
      
      
    
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Xue</surname>
            <given-names>N</given-names>
          </name>
        </person-group>
        <article-title>Chinese word segmentation as character tagging</article-title>
        <source>International Journal of Computational Linguistics &#38; Chinese Language Processing, Special Issue on Word Formation and Chinese Language Processing</source>  
        <year>2003</year>  
        <volume>8</volume>  
        <issue>1</issue>  
        <fpage>29</fpage>  
        <lpage>48</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/fxsjy/jieba"/>
        </comment> </nlm-citation>
      </ref>
      
      
      
      
      
        <ref id="ref21">
        <label>21</label>
       <nlm-citation citation-type="journal"> <person-group person-group-type="author"> <name name-style="western"> <surname>Zhang</surname> <given-names>Y</given-names> </name> <name name-style="western"> <surname>Jin</surname> <given-names>R</given-names> </name> <name name-style="western"> <surname>Zhou</surname> <given-names>Z</given-names> </name> </person-group> <article-title>Understanding bag-of-words model: a statistical framework</article-title> <source>International Journal of Machine Learning and Cybernetics</source> <year>2010</year> <volume>1</volume> <issue>1-4</issue> <fpage>43</fpage> <lpage>52</lpage> <pub-id pub-id-type="doi">10.1007/s13042-010-0001-0</pub-id> </nlm-citation>
      </ref>    
      
      
      
      
      
      
      
      
      
      
      
      
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Witten</surname>
            <given-names>IH</given-names>
          </name>
          <name name-style="western">
            <surname>Frank</surname>
            <given-names>E</given-names>
          </name>
          <name name-style="western">
            <surname>Hall</surname>
            <given-names>MA</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Data Mining: Practical machine learning tools and techniques</article-title>
        <source>Morgan Kaufmann, ELSEVIER</source>  
        <year>2016</year>  
        <publisher-loc>Cambridge</publisher-loc>
        <publisher-name>ELSEVIER</publisher-name></nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Friedman</surname>
            <given-names>JH</given-names>
          </name>
        </person-group>
        <article-title>Greedy function approximation: a gradient boosting machine</article-title>
        <source>Annals of statistics</source>  
        <year>2001</year>  
        <volume>29</volume>  
        <issue>5</issue>  
        <fpage>1189</fpage>  
        <lpage>1232</lpage>  
        <pub-id pub-id-type="doi">10.1214/aos/1013203451</pub-id></nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Chapelle</surname>
            <given-names>O</given-names>
          </name>
          <name name-style="western">
            <surname>Shivaswamy</surname>
            <given-names>P</given-names>
          </name>
          <name name-style="western">
            <surname>Vadrevu</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Multi-task learning for boosting with application to web search ranking</article-title>
        <year>2010</year>  
        <conf-name>Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM</conf-name>
        <conf-date>2010</conf-date>
        <conf-loc>Washington, DC, USA</conf-loc>
        <fpage>1189</fpage>  
        <lpage>1198</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/tqchen/xgboost"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1145/1835804.1835953</pub-id></nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Kramer</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Widmer</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Pfahringer</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Prediction of ordinal classes using regression trees</article-title>
        <source>Fundamenta Informaticae</source>  
        <year>2001</year>  
        <volume>47</volume>  
        <issue>1-2</issue>  
        <fpage>1</fpage>  
        <lpage>13</lpage>  
        <pub-id pub-id-type="doi">10.1007/3-540-39963-1_45</pub-id></nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Herbrich</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Graepel</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Obermayer</surname>
            <given-names>K</given-names>
          </name>
        </person-group>
        <article-title>Support vector learning for ordinal regression</article-title>
        <year>1999</year>  
        <conf-name>The Ninth International Conference on Artificial Neural Networks</conf-name>
        <conf-date>1999</conf-date>
        <conf-loc>Edinburgh, UK</conf-loc>
        <fpage>97</fpage>  
        <lpage>102</lpage> </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Gu</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Sheng</surname>
            <given-names>VS</given-names>
          </name>
          <name name-style="western">
            <surname>Tay</surname>
            <given-names>KY</given-names>
          </name>
          <name name-style="western">
            <surname>Romano</surname>
            <given-names>W</given-names>
          </name>
          <name name-style="western">
            <surname>Li</surname>
            <given-names>S</given-names>
          </name>
        </person-group>
        <article-title>Incremental Support Vector Learning for Ordinal Regression</article-title>
        <source>IEEE Transactions on Neural networks and learning systems</source>  
        <year>2015</year>  
        <volume>26</volume>  
        <issue>7</issue>  
        <fpage>1403</fpage>  
        <lpage>1416</lpage>  
        <pub-id pub-id-type="doi">10.1109/TNNLS.2014.2342533</pub-id>
        <pub-id pub-id-type="medline">25134094</pub-id></nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Cardoso</surname>
            <given-names>JS</given-names>
          </name>
          <name name-style="western">
            <surname>Costa</surname>
            <given-names>JF</given-names>
          </name>
        </person-group>
        <article-title>Learning to classify ordinal data: The data replication method</article-title>
        <source>Journal of Machine Learning Research</source>  
        <year>2007</year>  
        <volume>8</volume>  
        <issue>7</issue>  
        <fpage>1393</fpage>  
        <lpage>1429</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmlr.org/papers/v8/cardoso07a.html?dabb7088"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Gutiérrez</surname>
            <given-names>PA</given-names>
          </name>
          <name name-style="western">
            <surname>Pérez-Ortiz</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Fernández-Navarro</surname>
            <given-names>F</given-names>
          </name>
        </person-group>
        <article-title>An experimental study of different ordinal regression methods and measures</article-title>
        <year>2012</year>  
        <conf-name>International Conference on Hybrid Artificial Intelligence Systems</conf-name>
        <conf-date>2012</conf-date>
        <conf-loc>Berlin, Heidelberg</conf-loc>
        <fpage>296</fpage>  
        <lpage>307</lpage>  
        <pub-id pub-id-type="doi">10.1007/978-3-642-28931-6_29</pub-id></nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Diao</surname>
            <given-names>Q</given-names>
          </name>
          <name name-style="western">
            <surname>Qiu</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Wu</surname>
            <given-names>CY</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>Jointly modeling aspects, ratings and sentiments for movie recommendation (jmars)</article-title>
        <year>2014</year>  
        <conf-name>The 20th ACM SIGKDD international conference on Knowledge discovery and data mining</conf-name>
        <conf-date>2014</conf-date>
        <conf-loc>New York</conf-loc>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/2623330.2623758"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Shimada</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Endo</surname>
            <given-names>T</given-names>
          </name>
        </person-group>
        <article-title>Seeing several stars: A rating inference task for a document containing several evaluation criteria</article-title>
        <source>Advances in Knowledge Discovery and Data Mining</source>  
        <year>2008</year>  
        <fpage>1006</fpage>  
        <lpage>1014</lpage>  
        <pub-id pub-id-type="doi">10.1007/978-3-540-68125-0_106</pub-id></nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Tang</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Qin</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Liu</surname>
            <given-names>T</given-names>
          </name>
        </person-group>
        <article-title>Document Modeling with Gated Recurrent Neural Network for Sentiment Classification</article-title>
        <year>2015</year>  
        <conf-name>International Conference on Empirical Methods in Natural Language Processing</conf-name>
        <conf-date>2015</conf-date>
        <conf-loc>Lisbon, Portugal</conf-loc>
        <fpage>1422</fpage>  
        <lpage>1432</lpage>  
        <pub-id pub-id-type="doi">10.18653/v1/D15-1167</pub-id></nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Kuo</surname>
            <given-names>TM</given-names>
          </name>
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>CP</given-names>
          </name>
          <name name-style="western">
            <surname>Lin</surname>
            <given-names>CJ</given-names>
          </name>
        </person-group>
        <article-title>Large-scale kernel rankSVM</article-title>
        <year>2014</year>  
        <conf-name>Proceedings of the 2014 SIAM international conference on data mining</conf-name>
        <conf-date>2014</conf-date>
        <conf-loc>Philadelphia, Pennsylvania, USA</conf-loc>
        <fpage>812</fpage>  
        <lpage>820</lpage>  
        <pub-id pub-id-type="doi">10.1137/1.9781611973440.93</pub-id></nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Lin</surname>
            <given-names>C</given-names>
          </name>
        </person-group>
        <article-title>Large-scale linear rankSVM</article-title>
        <source>Neural computation</source>  
        <year>2014</year>  
        <volume>26</volume>  
        <issue>4</issue>  
        <fpage>781</fpage>  
        <lpage>817</lpage>  
        <pub-id pub-id-type="doi">10.1162/NECO_a_00571</pub-id>
        <pub-id pub-id-type="medline">24479776</pub-id></nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>KENDALL</surname>
            <given-names>MG</given-names>
          </name>
        </person-group>
        <article-title>A new measure of rank correlation</article-title>
        <source>Biometrika</source>  
        <year>1938</year>  
        <volume>30</volume>  
        <issue>1/2</issue>  
        <fpage>81</fpage>  
        <lpage>93</lpage>  
        <pub-id pub-id-type="doi">10.1093/biomet/30.1-2.81</pub-id></nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Fan</surname>
            <given-names>RE</given-names>
          </name>
          <name name-style="western">
            <surname>Chang</surname>
            <given-names>KW</given-names>
          </name>
          <name name-style="western">
            <surname>Hsieh</surname>
            <given-names>CJ</given-names>
          </name>
          <name name-style="western">
            <surname>et</surname>
            <given-names>al</given-names>
          </name>
        </person-group>
        <article-title>LIBLINEAR: A library for large linear classification</article-title>
        <source>Journal of machine learning research</source>  
        <year>2008</year>  
        <volume>9</volume>  
        <fpage>1871</fpage>  
        <lpage>1874</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/UcasResearch/OR-DPE"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1145/1390681.1442794</pub-id></nlm-citation>
      </ref>
      
      
      
      
      
      
      
      
      
      
          <ref id="ref37">
        <label>37</label>
<nlm-citation citation-type="journal"> <person-group person-group-type="author"> <name name-style="western"> <surname>Ho</surname> <given-names>CH</given-names> </name> <name name-style="western"> <surname>Lin</surname> <given-names>CJ</given-names> </name> </person-group> <article-title>Large-scale linear support vector regression</article-title> <source>Journal of Machine Learning Research</source> <year>2012</year> <volume>13</volume> <fpage>3323</fpage> <lpage>3348</lpage> <comment><ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmlr.org/papers/v13/ho12a.html"/></comment> </nlm-citation>
      </ref>  
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Yu</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Huang</surname>
            <given-names>F</given-names>
          </name>
          <name name-style="western">
            <surname>Lin</surname>
            <given-names>C</given-names>
          </name>
        </person-group>
        <article-title>Dual coordinate descent methods for logistic regression and maximum entropy models</article-title>
        <source>Machine Learning</source>  
        <year>2011</year>  
        <volume>85</volume>  
        <issue>1-2</issue>  
        <fpage>41</fpage>  
        <lpage>75</lpage>  
        <pub-id pub-id-type="doi">10.1007/s10994-010-5221-8</pub-id></nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Lin</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Li</surname>
            <given-names>L</given-names>
          </name>
        </person-group>
        <article-title>Reduction from cost-sensitive ordinal ranking to weighted binary classification</article-title>
        <source>Neural Computation</source>  
        <year>2012</year>  
        <volume>24</volume>  
        <issue>5</issue>  
        <fpage>1329</fpage>  
        <lpage>1367</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.gatsby.ucl.ac.uk/~chuwei/ordinalregression"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1162/NECO_a_00265</pub-id>
        <pub-id pub-id-type="medline">22295981</pub-id></nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Farmer</surname>
            <given-names>EA</given-names>
          </name>
          <name name-style="western">
            <surname>Beard</surname>
            <given-names>JD</given-names>
          </name>
          <name name-style="western">
            <surname>Dauphinee</surname>
            <given-names>WD</given-names>
          </name>
          <name name-style="western">
            <surname>LaDuca</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Mann</surname>
            <given-names>KV</given-names>
          </name>
        </person-group>
        <article-title>Assessing the performance of doctors in teams and systems</article-title>
        <source>Medical education</source>  
        <year>2002</year>  
        <volume>36</volume>  
        <issue>10</issue>  
        <fpage>942</fpage>  
        <lpage>948</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://onlinelibrary.wiley.com/doi/10.1046/j.1365-2923.2002.01311.x/full"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1046/j.1365-2923.2002.01311.x</pub-id></nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Lockyer</surname>
            <given-names>J</given-names>
          </name>
        </person-group>
        <article-title>Multisource feedback in the assessment of physician competencies</article-title>
        <source>Journal of Continuing Education in the Health Professions</source>  
        <year>2003</year>  
        <volume>23</volume>  
        <issue>1</issue>  
        <fpage>4</fpage>  
        <lpage>12</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://onlinelibrary.wiley.com/doi/10.1002/chp.1340230103/full"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1002/chp.1340230103</pub-id>
        <pub-id pub-id-type="medline">12739254</pub-id></nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
