<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v23i12e27008</article-id>
      <article-id pub-id-type="pmid">34958305</article-id>
      <article-id pub-id-type="doi">10.2196/27008</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>A Novel Deep Learning–Based System for Triage in the Emergency Department Using Electronic Medical Records: Retrospective Cohort Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Kukafka</surname>
            <given-names>Rita</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Teixeira</surname>
            <given-names>A</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Fernandes</surname>
            <given-names>Marta</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Yao</surname>
            <given-names>Li-Hung</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5194-339X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Leung</surname>
            <given-names>Ka-Chun</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1917-7554</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Tsai</surname>
            <given-names>Chu-Lin</given-names>
          </name>
          <degrees>MD, ScD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4639-1513</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Huang</surname>
            <given-names>Chien-Hua</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2981-4537</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Fu</surname>
            <given-names>Li-Chen</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Department of Computer Science and Information Engineering</institution>
            <institution>National Taiwan University</institution>
            <addr-line>CSIE Der Tian Hall</addr-line>
            <addr-line>No. 1, Sec. 4, Roosevelt Road</addr-line>
            <addr-line>Taipei, 10617</addr-line>
            <country>Taiwan</country>
            <phone>886 0935545846</phone>
            <email>lichen@ntu.edu.tw</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6947-7646</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Computer Science and Information Engineering</institution>
        <institution>National Taiwan University</institution>
        <addr-line>Taipei</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Emergency Medicine</institution>
        <institution>National Taiwan University Hospital and National Taiwan University College of Medicine</institution>
        <addr-line>Taipei</addr-line>
        <country>Taiwan</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Li-Chen Fu <email>lichen@ntu.edu.tw</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>12</month>
        <year>2021</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>27</day>
        <month>12</month>
        <year>2021</year>
      </pub-date>
      <volume>23</volume>
      <issue>12</issue>
      <elocation-id>e27008</elocation-id>
      <history>
        <date date-type="received">
          <day>7</day>
          <month>1</month>
          <year>2021</year>
        </date>
        <date date-type="rev-request">
          <day>12</day>
          <month>2</month>
          <year>2021</year>
        </date>
        <date date-type="rev-recd">
          <day>21</day>
          <month>3</month>
          <year>2021</year>
        </date>
        <date date-type="accepted">
          <day>26</day>
          <month>10</month>
          <year>2021</year>
        </date>
      </history>
      <copyright-statement>©Li-Hung Yao, Ka-Chun Leung, Chu-Lin Tsai, Chien-Hua Huang, Li-Chen Fu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 27.12.2021.</copyright-statement>
      <copyright-year>2021</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2021/12/e27008" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Emergency department (ED) crowding has resulted in delayed patient treatment and has become a universal health care problem. Although a triage system, such as the 5-level emergency severity index, somewhat improves the process of ED treatment, it still heavily relies on the nurse’s subjective judgment and triages too many patients to emergency severity index level 3 in current practice. Hence, a system that can help clinicians accurately triage a patient’s condition is imperative.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aims to develop a deep learning–based triage system using patients’ ED electronic medical records to predict clinical outcomes after ED treatments.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We conducted a retrospective study using data from an open data set from the National Hospital Ambulatory Medical Care Survey from 2012 to 2016 and data from a local data set from the National Taiwan University Hospital from 2009 to 2015. In this study, we transformed structured data into text form and used convolutional neural networks combined with recurrent neural networks and attention mechanisms to accomplish the classification task. We evaluated our performance using area under the receiver operating characteristic curve (AUROC).</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>A total of 118,602 patients from the National Hospital Ambulatory Medical Care Survey were included in this study for predicting hospitalization, and the accuracy and AUROC were 0.83 and 0.87, respectively. On the other hand, an external experiment was to use our own data set from the National Taiwan University Hospital that included 745,441 patients, where the accuracy and AUROC were similar, that is, 0.83 and 0.88, respectively. Moreover, to effectively evaluate the prediction quality of our proposed system, we also applied the model to other clinical outcomes, including mortality and admission to the intensive care unit, and the results showed that our proposed method was approximately 3% to 5% higher in accuracy than other conventional methods.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Our proposed method achieved better performance than the traditional method, and its implementation is relatively easy, it includes commonly used variables, and it is better suited for real-world clinical settings. It is our future work to validate our novel deep learning–based triage algorithm with prospective clinical trials, and we hope to use it to guide resource allocation in a busy ED once the validation succeeds.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>emergency department</kwd>
        <kwd>triage system</kwd>
        <kwd>deep learning</kwd>
        <kwd>hospital admission</kwd>
        <kwd>data to text</kwd>
        <kwd>electronic health record</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Overcrowding in the emergency department (ED) is already a global public health issue and is clearly an important patient safety issue [<xref ref-type="bibr" rid="ref1">1</xref>]. Many countries, such as Ireland, the United States, Canada, Germany, and Australia, have shown a continuous and significant increase in the number of ED visits [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref7">7</xref>]. In the United States, ED visits were estimated to increase from 136.9 million in 2015 to 145.6 million in 2016, an increase of 6.4%. The 10-year volume change was 24.7% and has increased by a total of 61.2% over the past 20 years (ED visits in 1996 were estimated at 90.3 million) [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. In Taiwan, ED visits were estimated to increase from 7.18 million in 2017 to 7.64 million in 2019, an increase of 6.4%. In retrospect, the number of ED visits has increased by a total of 23.6% over the past 19 years [<xref ref-type="bibr" rid="ref10">10</xref>].</p>
        <p>The increasing number of ED visits has also caused a periodic imbalance in the supply and demand of ED and hospital resources, which leads to longer waiting times and delays in critical medical treatments. ED crowding is related to several adverse clinical outcomes, including higher mortality and morbidity [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>]. Therefore, it is most important to design a method to properly identify urgent patients’ priorities in the ED [<xref ref-type="bibr" rid="ref13">13</xref>].</p>
      </sec>
      <sec>
        <title>Related Work</title>
        <p>Several research studies have focused on developing a system for predicting hospital admissions based on the patient’s ED electronic medical record (EMR) [<xref ref-type="bibr" rid="ref14">14</xref>]. Among these studies, the National Hospital Ambulatory Medical Care Survey (NHAMCS) data set [<xref ref-type="bibr" rid="ref9">9</xref>] is the most common data set to be analyzed. Despite using the NHAMCS data set, those studies might end up with different outcomes being achieved. Here, we briefly introduce some existing methods and implementation results, followed by a description of the concepts and methods that our system uses.</p>
        <p>Gligorijevic et al [<xref ref-type="bibr" rid="ref15">15</xref>] developed a system for predicting the number of resources that the patients would need. They built a bidirectional Long Short-Term Memory (biLSTM) model to extract continuous data features and medical text data features, which resulted in a binary model with prediction accuracy and area under the receiver operating characteristic curve (AUROC) of 0.792 and 0.879, respectively. Moreover, they showed that using nurses’ notes can provide a significant improvement in the prediction accuracy in comparison with using only standard continuous and categorical data.</p>
        <p>Zhang et al [<xref ref-type="bibr" rid="ref16">16</xref>] constructed a method for analyzing the patients’ reasons for a visit to predict hospital admission using principal component analysis and traditional natural language processing (NLP) combined with multilayer neural network models and logistic regression (LR) model. In their study, they tested the model using a 10-fold cross-validation method, and the AUROC was 0.84. Sun et al [<xref ref-type="bibr" rid="ref17">17</xref>] used the chi-square test to select the association between hospital admission and various possible risk factors and inputted the extracted association features into LR model for training to develop a prediction model, which was used to predict whether a need for hospital admission exists for ED patients. The involved variables included demographics (age, sex, and ethnic group), ED visit or hospital admission in the preceding 3 months, arrival mode, patient acuity category of the ED visit, and coexisting chronic diseases (diabetes, hypertension, and dyslipidemia). The AUROC for their study was 0.85.</p>
        <p>Graham et al [<xref ref-type="bibr" rid="ref18">18</xref>] used 3 machine learning algorithms to create the following models: LR, gradient boosted machines, and decision trees; these models were validated using a 10-fold cross-validation method repeated 5 times, whereby the accuracy of the best result in the gradient boosted machines model was 0.8. It turns out that their study can help clinicians plan the allocation of resources in advance and avoid the bottleneck of patient congestion.</p>
        <p>Wang et al [<xref ref-type="bibr" rid="ref19">19</xref>] developed a data-driven and evidence-based triage method to quickly identify acute and severe patients and prevent the waste of limited resources because of overdiagnosis. They proposed an attention-based biLSTM called <italic>DeepTriager</italic>, which processes both structured data and textual data from a clinical record to predict an ED patient’s acuity level. The method can not only predict the acuity and severity of the outpatient but can also provide visualizable and interpretable evidence on the clinical context to support decision-making. The AUROC for binary classification (acuity 1 and 2) can achieve 0.93, which is 0.03 higher than that of traditional machine learning methods.</p>
      </sec>
      <sec>
        <title>Study Aim</title>
        <p>The aim of this study is to establish an effective and efficient system for predicting whether patients will eventually require hospital admission to provide a reference to physicians to rank the priority of treatment of patients in advance. In this proposed system, our goal is to use both conventional structured data and unstructured data to design a binary classification model to help identify the hospitalization needs of the ED patient visits.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>System Overview</title>
        <p>This study focuses on establishing an effective and rapid system to predict whether patients will eventually need to be hospitalized to provide a reference to physicians to determine the priority of treatment of patients in advance. Moreover, to evaluate the effectiveness of our model for other clinical predictions, we also applied the model to other clinical outcomes and compared the obtained results with those from other algorithms.</p>
        <p>The system overview in <xref rid="figure1" ref-type="fig">Figure 1</xref> shows that our system is separated into two parts: the training part and the prediction part. The EMR values of each ED visit patient were used as input, and the patient’s hospital admission decision from the physician was used as the ground truth. There were 3 steps in model training. The first step was to preprocess the input data, such as feature selection and filtering of the unusable or missing data, and the detailed method will be explained in the <italic>Data Preparation</italic> section. Next, the processed data were transformed into the corresponding text type. Finally, the transformed transcript and ground truth were used to train the binary classification model. Once the training of the model was completed, it was tested against the unseen data. The unseen data were transformed to the text type, which was then fed into the model, and the output of the model was the probability of hospital admission for the ED patient visits.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>System overview. EMR: electronic medical record; ED: emergency department.</p>
          </caption>
          <graphic xlink:href="jmir_v23i12e27008_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Data Transformation</title>
        <p>To allow the mentioned data set to be more effectively handled by the proposed methods in our work, we first transformed the data into another text type. The method of transforming the original data into the text type is shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
        <p>The table in <xref rid="figure2" ref-type="fig">Figure 2</xref> shows the characteristics sampled from the original data set, including vital signs and other information, such as age, gender, blood pressure, oxygen, and pain index. The lower part of <xref rid="figure2" ref-type="fig">Figure 2</xref> illustrates the format after data transformation in English or Chinese. Note that the original format of the data sample shown in <xref rid="figure2" ref-type="fig">Figure 2</xref> reveals the features and their corresponding values. However, after transformation into text, all the feature names remained as words, but their corresponding values also appeared as words, so that the new format of the data sample now became a complete sentence. Then, we inputted the complete sentence into the model for training and analyzed the correlation between the features.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>The method of data transformation.</p>
          </caption>
          <graphic xlink:href="jmir_v23i12e27008_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Triage Engine</title>
        <sec>
          <title>Overview</title>
          <p>In this work, our triage engine comprised 2 different parts, where the first part was used for analyzing short sentences (only convolutional neural network [CNN] type) and the second part was used for analyzing long sentences (only recurrent neural network [RNN] type). Experiments were conducted to verify the effectiveness of the triage engine after we first examined the performances of its 2 parts. Our classification engine was composed of 2 different text processing modules (ie, 2 parts). In the following section, we have introduced the characteristics of the 2 modules step by step and elaborated on the logic behind their design. The network architecture of our proposed triage engine, comprising the RNN-type module and the CNN-type module, is shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Network architecture of the triage engine. BiGRU: bidirectional gated recurrent unit; CNN: convolutional neural network; GRU: gated recurrent unit; RNN: recurrent neural network.</p>
            </caption>
            <graphic xlink:href="jmir_v23i12e27008_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>In our system, we used the data that had been converted into text format as input. As the text data contained many features, there was a certain relationship between the different features, such as the relationship between systolic blood pressure and diastolic blood pressure, pain index and pain location, and triage level and pain index. However, this relationship might have been lost after the features were converted into static word embeddings in high-dimensional space. Therefore, to be able to analyze the entire sentence for text information, we adopted the RNN structure, which has been shown to form a very useful algorithm. In fact, the RNN architecture is focused on the relationship among all the words from a sentence, and it is more appropriate to analyze the meaning of long sentences.</p>
          <p>In addition, as the features and the corresponding values were converted into text sequentially in terms of a sentence, the name and value of a feature were in a neighboring relationship. To be able to accurately and effectively analyze the relationship among each feature and the corresponding value, the semantics of short sentences (among 3 words, 4 words, or 5 words) was very important, for example, <italic>body temperature: 36.5 degrees</italic>, <italic>pain index: 5</italic>, and <italic>respiratory rate: 15 times</italic>. In this short sentence analysis, CNN served as a very useful algorithm. This architecture focuses on the relationship between each word and its neighboring subjects. We will introduce it in detail later.</p>
          <p>To effectively use the characteristics of the 2 learning algorithms of the 2 parts of the triage engine, we paid attention to the way in which the outputs of the 2 neural network models were effectively fused, which was also the focus of this system. In other words, we formed an overall model by merging the 2 parts to accomplish the single task, that is, prediction of the need for hospital admission. Our strategy was as follows: first, the 2 models were individually trained based on different data sets, and second, their parameters were optimized according to their respective losses and the corresponding parameter settings. After training the 2 models, we deleted the output layers of both the models and concatenated the last 2 fully connected layers, which were located before the output layers of the 2 models and the new output layer. Then, we fine-tuned the overall triage model on the 2 data sets, and the dropout was executed before the output layer. Finally, the output of the triage model predicted the probability of hospital admission for each ED visit.</p>
        </sec>
        <sec>
          <title>Module for a Long Sentence (RNN Type)</title>
          <p>The focus of this part of the study was on how to analyze the integrity of the text data, where the strength of feature extraction was extensive as the correlation among the patterns of the different samples was searched. Such a correlation not only involves time but also involves space. By learning from the sequence of sentences, the resulting model was able to effectively process each of the complete textual data and thus possessed a memory attribute.</p>
          <p>The input of this model was the textual data transformed from the structural and unstructured data (EMR) of the ED visits, and the output was the vector that included the probability of hospital admission for the ED visits. <xref rid="figure4" ref-type="fig">Figure 4</xref> shows the network architecture of the RNN part of the triage engine.</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>Recurrent neural network-type part of the triage engine. BiGRU: Bidirectional gated recurrent unit.</p>
            </caption>
            <graphic xlink:href="jmir_v23i12e27008_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>As depicted in <xref rid="figure4" ref-type="fig">Figure 4</xref>, first, all the words transformed from the EMR were input into the word vector layer, which was used to convert each word into the corresponding word vector. In our work, the pretrained word vector library was FastText [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>], which is the most popular and useful library for learning word embeddings and text classification. First, as our work processed 2 data sets, one in English from NHAMCS and another in Chinese from the National Taiwan University Hospital (NTUH), we decided to develop our prediction model in both English and Chinese versions. Second, the word vectors were sequentially passed to the bidirectional gated recurrent unit (biGRU), in which the hidden sequences in 2 directions were concatenated at each time stamp to form a new hidden sequence. In general, the biGRU can obtain the features of a text more effectively. Thus, the hidden states from the biGRU were fed to the attention layer to evaluate the weights of each hidden state, and the dot product between the evaluated weight value and each hidden state was calculated. The attention layer was composed of 2 fully connected feed-forward neural networks, using exponential linear units [<xref ref-type="bibr" rid="ref22">22</xref>] as the activation function. For the output of the attention layer, Softmax, developed by Goodfellow et al [<xref ref-type="bibr" rid="ref23">23</xref>], was chosen as the activation function.</p>
          <p>The attention layer in this work was used to find the key information content units in different sentences from each ED visit’s record and assess whether the patients will eventually need hospitalization. Moreover, the 2-layer attention network was used in this work as it is more effective than one with only a single layer in all sentences. Thus, the proposed system with multiple attention layers was more effective for the subsequent evaluation of the prediction performance throughout the experiment. Finally, the output from the attention layer was fed into the fully connected layers with 64 neurons.</p>
        </sec>
        <sec>
          <title>Module for a Short Sentence (CNN Type)</title>
          <p>This part of the study focused on extracting the local features of the text. By extracting the keywords of the document or sentence as features and training the classifier based on these features, it was possible to effectively analyze the more important and critical contexts of the sentences.</p>
          <p>Similarly, <xref rid="figure5" ref-type="fig">Figure 5</xref> shows the network architecture of the CNN part module of the triage engine. An image-like vector, whose format is <italic>I</italic> <bold>∈ ℝ<sup>ℎ×𝑤</sup></bold>, was obtained by stacking the word vectors that are converted from the original text. More specifically, <italic>h</italic> and <italic>w</italic> denoted the height (number of words) and width (dimension of the word vector) of the image, respectively. In particular, Chinese words were processed by text segmentation first and then passed to the embedding layer. Then, the word vectors were used to perform convolution operations with 3 kernels of different sizes, which were 3, 4, and 5. Different kernels were used to find the relationships among short words, that is, various correlations between words.</p>
          <p>For the convolutional operation, we adopted the concept from the deep pyramid CNNs [<xref ref-type="bibr" rid="ref24">24</xref>], which is a low-complexity CNN architecture for text categorization that can efficiently represent long-range associations in a text. Instead of using the original CNN for text processing [<xref ref-type="bibr" rid="ref25">25</xref>], we applied a simple network architecture to obtain better accuracy by increasing the network depth without significantly increasing the computational cost. <xref rid="figure6" ref-type="fig">Figure 6</xref> shows the network architecture of the aforementioned pyramid CNN for text. Owing to the problem of degradation, the shortcut connections were expected to facilitate every few stacked layers to more easily fit a desired underlying mapping, and such thoughts of shortcut connections were the key concept for the pyramid CNN architecture. Therefore, according to the idea, the deep pyramid CNN under different kernel sizes was used in our work.</p>
          <fig id="figure5" position="float">
            <label>Figure 5</label>
            <caption>
              <p>Convolutional neural network–type part of the triage engine. CNN: convolutional neural network.</p>
            </caption>
            <graphic xlink:href="jmir_v23i12e27008_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure6" position="float">
            <label>Figure 6</label>
            <caption>
              <p>Architecture of pyramid convolutional neural network for text.</p>
            </caption>
            <graphic xlink:href="jmir_v23i12e27008_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Model Synthesis by Integration of the 2 Modules</title>
          <p>As shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>, our final triage engine model was to integrate the 2 <italic>pseudoengines</italic> shown in <xref rid="figure4" ref-type="fig">Figures 4</xref> and <xref rid="figure5" ref-type="fig">5</xref>. Technically speaking, after the convolution operation, a feature map with the corresponding size was obtained, and each convolutional layer was followed by a max pooling layer of the corresponding size. The CNN-type part of the integrated engine was implemented to further distinguish the keywords from the transformed sentence. Then, the output of the max pooling layers was input to the self-attention layer, which was used to evaluate the weights of keyword vectors and calculate the dot product between the evaluated weight value and each keyword vector. Finally, the outputs from the 3 attention layers were concatenated and fed into fully connected layers with 64 neurons. The vector of the probability of hospital admission was calculated by applying the Softmax activation function in the fully connected layer, and we formulated the probability as follows:</p>
          <p>
            <disp-formula>
              <italic>Probability = CNN (k<sup>3</sup>(X') + k<sup>4</sup>(X') + k<sup>5</sup>(X'))</italic>
              <bold>(1)</bold>
            </disp-formula>
          </p>
        </sec>
        <sec>
          <title>Model Training</title>
          <p>The RNN-type module was trained with a learning rate of 0.00001 using an optimizer called Adam, developed by Kingma et al [<xref ref-type="bibr" rid="ref26">26</xref>], which is a gradient descent method widely used in deep learning applications for computer vision and NLP. The batch size was set as 64, the number of iterations was set as 60, and the hidden states of the biGRU were set as 128. For the CNN-type module, the learning hyperparameters were the same as that of the RNN type. The size of the kernel was set to 3, 4, and 5, and the strides were set as 1. The loss function used for the integrated model was the cross-entropy sum between the predicted output and ground truth as follows:</p>
          <disp-formula><italic>l<sub>total</sub></italic> = <italic>l<sub>cnn</sub></italic> + <italic>l<sub>rnn</sub></italic> (<bold>2</bold>)</disp-formula>
          <disp-formula>
            <graphic xlink:href="jmir_v23i12e27008_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>where <italic>y<sub>i</sub></italic> is the ground truth of class <italic>I</italic>, and <italic>ŷ<sub>i</sub></italic> is the prediction of the model.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Overview</title>
        <p>A series of experiments were conducted to validate our design. To evaluate the effectiveness of our model, all the experiments were carefully conducted using stratified random sampling. The following procedure was performed separately on the NHAMCS and NTUH data sets. For internal comparison, 72% of the data were used as the training set, 8% of the data were used as the validation set, and 20% of the data were used as the hold-out testing set. The training and validation process was repeated 20 times, with 20 models generated, and the best-performing model in the hold-out testing set was selected as the final model. For external comparison, the same training and validation procedure was performed; however, only the best-performing model in the training and validation procedure was tested on the hold-out testing set to ensure a fair comparison.</p>
      </sec>
      <sec>
        <title>Experiment Platform</title>
        <p>We adopted Keras (Tensorflow-Graphics Processing Unit) to execute all the algorithms on computers with Nvidia GeForce GTX 1080Ti Graphics Processing Unit (with 11 GB RAM) and Intel Core i5 Central Processing Unit (with 64 GB RAM). In the processing of the loss function, we used the Adam optimizer with a learning rate of 0.00001, batch size of 64, and epochs of 60.</p>
      </sec>
      <sec>
        <title>Data Preparation</title>
        <p>In this study, 2 different data sets were used to evaluate the performance of the proposed system: the NHAMCS data set and the NTUH data set.</p>
        <sec>
          <title>NHAMCS Data Set</title>
          <p>In our study, the data from 118,602 ED patient visits collected between 2012 and 2016 were used. We selected 37 features, including <italic>month</italic>, <italic>week</italic>, <italic>arrival time</italic>, <italic>age</italic>, <italic>residence</italic>, <italic>sex</italic>, <italic>race</italic>, <italic>did he or she come by ambulance</italic>, <italic>pay by insurance</italic>, <italic>pay by Medicare</italic>, <italic>pay by Medicaid</italic>, <italic>pay by work compensation</italic>, <italic>pay by self</italic>, <italic>no charge to pay</italic>, <italic>temperature</italic>, <italic>heart rate</italic>, <italic>respiratory rate</italic>, <italic>systolic blood pressure</italic>, <italic>diastolic blood pressure</italic>, <italic>pulse oximetry</italic>, <italic>pain scale</italic>, <italic>triage level</italic>, <italic>been ED during last 72 hours</italic>, <italic>dementia</italic>, <italic>cancer</italic>, <italic>cerebrovascular</italic>, <italic>COPD</italic>, <italic>heart failure</italic>, <italic>HIV</italic>, <italic>ECG</italic>, <italic>X-ray</italic>, <italic>CT-scan</italic>, <italic>MRI</italic>, <italic>Ultrasound</italic>, <italic>CPR</italic>, <italic>admitted to ICU</italic>, and <italic>hospital admission</italic>.</p>
        </sec>
        <sec>
          <title>NTUH Data Set</title>
          <p>In our study, the data from 745,441 ED patient visits collected between 2013 and 2017 were used. We selected 31 features, including <italic>age</italic>, <italic>sex</italic>, <italic>day zone</italic>, <italic>weekend</italic>, <italic>month</italic>, <italic>is he or she getting fever?</italic>, <italic>clinics by</italic>, <italic>clinics for</italic>, <italic>is job-related?</italic>, <italic>on the job way</italic>, <italic>pain character</italic>, <italic>pain period</italic>, <italic>CPR</italic>, <italic>ICU</italic>, <italic>acute change</italic>, <italic>account sequence number</italic>, <italic>systolic blood pressure</italic>, <italic>diastolic blood pressure</italic>, <italic>pulse</italic>, <italic>oxygen</italic>, <italic>respiration rate</italic>, <italic>body temperature</italic>, <italic>pain index</italic>, <italic>gcse</italic>, <italic>gcsv</italic>, <italic>gcsm</italic>, <italic>triage level</italic>, <italic>pain body part</italic>, <italic>pain period description</italic>, <italic>judgement description</italic>, and <italic>hospital admission</italic>. All the features were recommended by Nottingham Trent University physicians.</p>
        </sec>
      </sec>
      <sec>
        <title>Performance on the NHAMCS Data Set and Baseline</title>
        <p>We verified our proposed fusion model using the NHAMCS data set and compared the results with the 2 parts of the model (RNNs and CNNs). As a result, the AUROC can achieve 0.872 using the proposed model. The other metrics of the performance of the proposed network are shown in <xref ref-type="table" rid="table1">Table 1</xref>. For our fusion model, the highest accuracy and specificity can reach 0.828 and 0.843, respectively.</p>
        <p>So far, most existing studies have used different data sets. Here, to effectively evaluate the prediction quality of our model, we chose the traditional machine learning algorithms commonly used in other studies as our baselines for comparison, including LR, extreme gradient boosting (XGBoost), and random forest. Furthermore, we compared our model with the Bidirectional Encoder Representations From Transformers (BERT) [<xref ref-type="bibr" rid="ref27">27</xref>] model, which is considered to be a milestone of NLP. Then, we compared the different results obtained from different methods under various metrics.</p>
        <p><xref ref-type="table" rid="table2">Table 2</xref> shows 6 metrics of each algorithm. It can be seen that our proposed model scored the highest in 4 out of 6 metrics, including specificity, precision, accuracy, and AUROC, while comparing with other models. These results suggest that our proposed deep learning algorithm seems to be more promising than the traditional machine learning algorithms.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Performance on the National Hospital Ambulatory Medical Care Survey data set using different methods.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="470"/>
            <col width="120"/>
            <col width="120"/>
            <col width="110"/>
            <col width="180"/>
            <thead>
              <tr valign="bottom">
                <td>Model</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Accuracy</td>
                <td>AUROC<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>BiLSTM<sup>b</sup> only</td>
                <td>0.756</td>
                <td>0.768</td>
                <td>0.767</td>
                <td>0.850</td>
              </tr>
              <tr valign="top">
                <td>BiLSTM+Att<sup>c</sup></td>
                <td>0.711</td>
                <td>0.822</td>
                <td>0.809</td>
                <td>0.854</td>
              </tr>
              <tr valign="top">
                <td>BiLSTM+2×Att</td>
                <td>0.745</td>
                <td>0.802</td>
                <td>0.796</td>
                <td>0.856</td>
              </tr>
              <tr valign="top">
                <td>BiGRU<sup>d</sup> only</td>
                <td>0.744</td>
                <td>0.78</td>
                <td>0.776</td>
                <td>0.854</td>
              </tr>
              <tr valign="top">
                <td>BiGRU+Att</td>
                <td>0.757</td>
                <td>0.804</td>
                <td>0.798</td>
                <td>0.863</td>
              </tr>
              <tr valign="top">
                <td>BiGRU+2×Att</td>
                <td>0.764</td>
                <td>0.809</td>
                <td>0.801</td>
                <td>0.866</td>
              </tr>
              <tr valign="top">
                <td>CNNs<sup>e</sup> (with 3 kernels)</td>
                <td>0.756</td>
                <td>0.768</td>
                <td>0.767</td>
                <td>0.85</td>
              </tr>
              <tr valign="top">
                <td>Pyramid CNN (3 kernels)</td>
                <td>0.727</td>
                <td>0.813</td>
                <td>0.804</td>
                <td>0.855</td>
              </tr>
              <tr valign="top">
                <td>Pyramid CNN (3 kernels) with attention layer</td>
                <td>0.731</td>
                <td>0.825</td>
                <td>0.819</td>
                <td>0.862</td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>0.755</td>
                <td>
                  <italic>0.843</italic>
                  <sup>f</sup>
                </td>
                <td>
                  <italic>0.828</italic>
                </td>
                <td>
                  <italic>0.872</italic>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>BiLSTM: bidirectional Long Short-Term Memory.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>Att: attention layer.</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup>BiGRU: bidirectional gated recurrent unit.</p>
            </fn>
            <fn id="table1fn5">
              <p><sup>e</sup>CNN: convolutional neural network.</p>
            </fn>
            <fn id="table1fn6">
              <p><sup>f</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Comparison with baseline algorithms in the National Hospital Ambulatory Medical Care Survey data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="210"/>
            <col width="120"/>
            <col width="130"/>
            <col width="130"/>
            <col width="110"/>
            <col width="110"/>
            <col width="190"/>
            <thead>
              <tr valign="bottom">
                <td>Model</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td>F1 score</td>
                <td>Accuracy</td>
                <td>AUROC<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Logistic regression</td>
                <td>0.747</td>
                <td>0.741</td>
                <td>0.745</td>
                <td>0.745</td>
                <td>0.744</td>
                <td>0.825</td>
              </tr>
              <tr valign="top">
                <td>XGBoost<sup>b</sup></td>
                <td>0.761</td>
                <td>0.736</td>
                <td>0.749</td>
                <td>0.748</td>
                <td>0.748</td>
                <td>0.834</td>
              </tr>
              <tr valign="top">
                <td>Random forest</td>
                <td>0.781</td>
                <td>0.715</td>
                <td>0.748</td>
                <td>0.747</td>
                <td>0.747</td>
                <td>0.828</td>
              </tr>
              <tr valign="top">
                <td>BERT<sup>c</sup></td>
                <td>0.789</td>
                <td>0.768</td>
                <td>0.773</td>
                <td>0.781</td>
                <td>0.779</td>
                <td>0.852</td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>0.755</td>
                <td>
                  <italic>0.843</italic>
                  <sup>d</sup>
                </td>
                <td>
                  <italic>0.818</italic>
                  <sup>d</sup>
                </td>
                <td>0.759</td>
                <td>
                  <italic>0.828</italic>
                </td>
                <td>
                  <italic>0.872</italic>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>XGBoost: extreme gradient boosting.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>BERT: Bidirectional Encoder Representations From Transformers.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models<bold>.</bold></p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Performance on the NTUH Data Set and Baseline</title>
        <p>We also verified our proposed fusion model using the NTUH data set and compared the results with the 2 parts of the model (RNNs and CNNs), which included 10 experiments. In the RNNs part, we experimented with 6 different combinations using biLSTM and biGRU with different layers of attention mechanisms to observe the changes in the 4 metrics (sensitivity, specificity, accuracy, and AUROC) under different combinations. In the CNN part, we experimented with 3 different combinations using a traditional CNN and pyramid CNN with an attention mechanism to observe the changes in the 4 metrics under different combinations. As a result, the AUROC can achieve 0.876 using the proposed model in our fusion model (<xref ref-type="table" rid="table3">Table 3</xref>).</p>
        <p>Similarly, to effectively evaluate the prediction quality of our model, we chose 3 kinds of traditional machine learning algorithms commonly used as our baselines—LR, XGBoost, and random forest. Furthermore, we also compared our model with 2 common methods, deep neural network for structural data with biGRU for textual data and the BERT [<xref ref-type="bibr" rid="ref27">27</xref>] model. Then, we compared the differences in the results between the different methods.</p>
        <p><xref ref-type="table" rid="table4">Table 4</xref> shows 6 metrics of each algorithm. It can be seen that our proposed model outperforms all the other algorithms. The result of our proposed model suggested a great improvement in predicting hospitalization when compared with other traditional methods.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Performance on the National Taiwan University Hospital data set using different methods.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="540"/>
            <col width="130"/>
            <col width="120"/>
            <col width="110"/>
            <col width="100"/>
            <thead>
              <tr valign="bottom">
                <td>Method</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Accuracy</td>
                <td>AUROC<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>BiLSTM<sup>b</sup> only</td>
                <td>0.748</td>
                <td>0.792</td>
                <td>0.77</td>
                <td>0.848</td>
              </tr>
              <tr valign="top">
                <td>BiLSTM+Att<sup>c</sup></td>
                <td>0.74</td>
                <td>0.822</td>
                <td>0.781</td>
                <td>0.862</td>
              </tr>
              <tr valign="top">
                <td>BiLSTM+2×Att</td>
                <td>0.774</td>
                <td>0.8</td>
                <td>0.785</td>
                <td>0.867</td>
              </tr>
              <tr valign="top">
                <td>BiGRU<sup>d</sup> only</td>
                <td>0.768</td>
                <td>0.78</td>
                <td>0.774</td>
                <td>0.855</td>
              </tr>
              <tr valign="top">
                <td>BiGRU+Att</td>
                <td>0.805</td>
                <td>0.767</td>
                <td>0.786</td>
                <td>0.866</td>
              </tr>
              <tr valign="top">
                <td>BiGRU+2×Att</td>
                <td>0.8</td>
                <td>0.785</td>
                <td>0.808</td>
                <td>0.872</td>
              </tr>
              <tr valign="top">
                <td>CNNs<sup>e</sup> (with 3 kernels)</td>
                <td>0.78</td>
                <td>0.803</td>
                <td>0.791</td>
                <td>0.868</td>
              </tr>
              <tr valign="top">
                <td>Pyramid CNN (3 kernels)</td>
                <td>0.784</td>
                <td>0.793</td>
                <td>0.798</td>
                <td>0.868</td>
              </tr>
              <tr valign="top">
                <td>Pyramid CNN (3 kernels) with attention layer</td>
                <td>0.754</td>
                <td>0.823</td>
                <td>0.788</td>
                <td>0.871</td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>0.768</td>
                <td>0.819</td>
                <td>
                  <italic>0.825</italic>
                  <sup>f</sup>
                </td>
                <td>
                  <italic>0.876</italic>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>BiLSTM: bidirectional Long Short-Term Memory.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>ATT: attention layer.</p>
            </fn>
            <fn id="table3fn4">
              <p><sup>d</sup>BiGRU: bidirectional gated recurrent unit.</p>
            </fn>
            <fn id="table3fn5">
              <p><sup>e</sup>CNN: convolutional neural network.</p>
            </fn>
            <fn id="table3fn6">
              <p><sup>f</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Comparison with baseline algorithms in the National Taiwan University Hospital data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="280"/>
            <col width="120"/>
            <col width="120"/>
            <col width="110"/>
            <col width="90"/>
            <col width="110"/>
            <col width="170"/>
            <thead>
              <tr valign="bottom">
                <td>Model</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td>F1 score</td>
                <td>Accuracy</td>
                <td>AUROC<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Logistic regression</td>
                <td>0.705</td>
                <td>0.805</td>
                <td>0.758</td>
                <td>0.755</td>
                <td>0.756</td>
                <td>0.83</td>
              </tr>
              <tr valign="top">
                <td>XGBoost<sup>b</sup></td>
                <td>0.745</td>
                <td>0.785</td>
                <td>0.766</td>
                <td>0.765</td>
                <td>0.765</td>
                <td>0.84</td>
              </tr>
              <tr valign="top">
                <td>Random forest</td>
                <td>0.739</td>
                <td>0.784</td>
                <td>0.762</td>
                <td>0.761</td>
                <td>0.762</td>
                <td>0.84</td>
              </tr>
              <tr valign="top">
                <td>DNN<sup>c</sup>+BiGRU<sup>d</sup></td>
                <td>0.744</td>
                <td>0.775</td>
                <td>0.771</td>
                <td>0.766</td>
                <td>0.771</td>
                <td>0.858</td>
              </tr>
              <tr valign="top">
                <td>BERT<sup>e</sup></td>
                <td>0.736</td>
                <td>0.789</td>
                <td>0.777</td>
                <td>0.756</td>
                <td>0.763</td>
                <td>0.844</td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>
                  <italic>0.768</italic>
                  <sup>f</sup>
                </td>
                <td>
                  <italic>0.819</italic>
                </td>
                <td>
                  <italic>0.81</italic>
                </td>
                <td>
                  <italic>0.788</italic>
                </td>
                <td>
                  <italic>0.825</italic>
                </td>
                <td>
                  <italic>0.876</italic>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>XGBoost: extreme gradient boosting.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>DNN: deep neural network.</p>
            </fn>
            <fn id="table4fn4">
              <p><sup>d</sup>BiGRU: bidirectional gated recurrent unit.</p>
            </fn>
            <fn id="table4fn5">
              <p><sup>e</sup>BERT: Bidirectional Encoder Representations From Transformers.</p>
            </fn>
            <fn id="table4fn6">
              <p><sup>f</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Comparison With Other Related Studies</title>
        <p>According to existing research on the prediction of hospitalization, most studies used specific feature selection methods combined with traditional machine learning algorithms. As shown in <xref ref-type="table" rid="table5">Table 5</xref>, their results show a variable performance on different metrics. In this section, to compare with other studies in a fair manner, only the best-performing model in the training and validation procedure was tested on the hold-out test set.</p>
        <p>According to <xref ref-type="table" rid="table5">Table 5</xref>, our model achieved the highest performance in AUROC while being compared in the same open data set, that is, the NHAMCS data set. In addition, our work also achieved an excellent score in accuracy in comparison with private data sets.</p>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Performance of different research studies.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="190"/>
            <col width="230"/>
            <col width="120"/>
            <col width="130"/>
            <col width="120"/>
            <col width="110"/>
            <col width="100"/>
            <thead>
              <tr valign="top">
                <td>Study</td>
                <td>Methods</td>
                <td>Data set</td>
                <td colspan="4">Performance</td>
              </tr>
              <tr valign="bottom">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Accuracy</td>
                <td>AUROC<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Raita et al [<xref ref-type="bibr" rid="ref28">28</xref>]</td>
                <td>DNN<sup>b</sup></td>
                <td>NHAMCS<sup>c</sup></td>
                <td>0.79</td>
                <td>0.71</td>
                <td>—<sup>d</sup></td>
                <td>0.82</td>
              </tr>
              <tr valign="top">
                <td>Zhang et al [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                <td>NLP<sup>e</sup>+PCA<sup>f</sup>+LR<sup>g</sup></td>
                <td>NHAMCS</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.846</td>
              </tr>
              <tr valign="top">
                <td>Yan Sun et al [<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                <td>LR</td>
                <td>Private</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.849</td>
              </tr>
              <tr valign="top">
                <td>Graham et al [<xref ref-type="bibr" rid="ref18">18</xref>]</td>
                <td>GBM<sup>h</sup></td>
                <td>Private</td>
                <td>0.535</td>
                <td>0.899</td>
                <td>0.8</td>
                <td>0.859</td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>BiGRU<sup>i</sup>+ Att<sup>j</sup>+ PyCNN<sup>k</sup></td>
                <td>NHAMCS</td>
                <td>0.654</td>
                <td>
                  <italic>0.856</italic>
                  <sup>l</sup>
                </td>
                <td>
                  <italic>0.834</italic>
                </td>
                <td>
                  <italic>0.856</italic>
                </td>
              </tr>
              <tr valign="top">
                <td>Our model</td>
                <td>BiGRU+ Att+ PyCNN</td>
                <td>NTUH<sup>m</sup></td>
                <td>0.606</td>
                <td>0.852</td>
                <td>0.806</td>
                <td>
                  <italic>0.821</italic>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table5fn2">
              <p><sup>b</sup>DNN: deep neural network.</p>
            </fn>
            <fn id="table5fn3">
              <p><sup>c</sup>NHAMCS: National Hospital Ambulatory Medical Care Survey.</p>
            </fn>
            <fn id="table5fn4">
              <p><sup>d</sup>Not available.</p>
            </fn>
            <fn id="table5fn5">
              <p><sup>e</sup>NLP: natural language processing.</p>
            </fn>
            <fn id="table5fn6">
              <p><sup>f</sup>PCA: principal component analysis.</p>
            </fn>
            <fn id="table5fn7">
              <p><sup>g</sup>LR: logistic regression.</p>
            </fn>
            <fn id="table5fn8">
              <p><sup>h</sup>GBM: gradient boosted machines.</p>
            </fn>
            <fn id="table5fn9">
              <p><sup>i</sup>BiGRU: bidirectional gated recurrent unit.</p>
            </fn>
            <fn id="table5fn10">
              <p><sup>j</sup>Att: attention layer.</p>
            </fn>
            <fn id="table5fn11">
              <p><sup>k</sup>PyCNN: pyramid convolutional neural network.</p>
            </fn>
            <fn id="table5fn12">
              <p><sup>l</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
            </fn>
            <fn id="table5fn13">
              <p><sup>m</sup>NTUH: National Taiwan University Hospital.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Applying on Other Clinical Outcomes</title>
        <p>To effectively evaluate the prediction quality of our model for other clinical results, we selected other common outcomes to test and set the results of traditional machine learning algorithms as our baselines and then compared the differences in the results between different methods in various metrics.</p>
        <sec>
          <title>For Mortality Rate Prediction</title>
          <p>As mortality rate has a high correlation with the emergency severity index (ESI) 5-level triage, we applied our model to predict the mortality rate on the NTUH data set, and the results are shown in <xref ref-type="table" rid="table6">Table 6</xref>. Owing to the small number of deceased patients, we chose not to test this data set because of convergence issues.</p>
          <p>Compared with other algorithms, including the 3 traditional machine learning algorithms, our proposed model outperforms all other methods except in <italic>sensitivity</italic>.</p>
          <table-wrap position="float" id="table6">
            <label>Table 6</label>
            <caption>
              <p>Performance of mortality rate prediction on the National Taiwan University Hospital data set.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="210"/>
              <col width="140"/>
              <col width="140"/>
              <col width="130"/>
              <col width="120"/>
              <col width="120"/>
              <col width="140"/>
              <thead>
                <tr valign="bottom">
                  <td>Model</td>
                  <td>Sensitivity</td>
                  <td>Specificity</td>
                  <td>Precision</td>
                  <td>F1 score</td>
                  <td>Accuracy</td>
                  <td>AUROC<sup>a</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Logistic regression</td>
                  <td>0.903</td>
                  <td>0.887</td>
                  <td>0.895</td>
                  <td>0.895</td>
                  <td>0.896</td>
                  <td>0.954</td>
                </tr>
                <tr valign="top">
                  <td>XGBoost<sup>b</sup></td>
                  <td>0.926</td>
                  <td>0.913</td>
                  <td>0.909</td>
                  <td>0.919</td>
                  <td>0.919</td>
                  <td>0.962</td>
                </tr>
                <tr valign="top">
                  <td>Random forest</td>
                  <td>0.933</td>
                  <td>0.898</td>
                  <td>0.915</td>
                  <td>0.915</td>
                  <td>0.916</td>
                  <td>0.958</td>
                </tr>
                <tr valign="top">
                  <td>Our model</td>
                  <td>0.917</td>
                  <td>
                    <italic>0.941</italic>
                    <sup>c</sup>
                  </td>
                  <td>
                    <italic>0.939</italic>
                  </td>
                  <td>
                    <italic>0.928</italic>
                  </td>
                  <td>
                    <italic>0.941</italic>
                  </td>
                  <td>
                    <italic>0.983</italic>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table6fn1">
                <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
              </fn>
              <fn id="table6fn2">
                <p><sup>b</sup>XGBoost: extreme gradient boosting.</p>
              </fn>
              <fn id="table6fn3">
                <p><sup>c</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>For Prediction of Intensive Care Unit Admission</title>
          <p>In the probability of intensive care unit admission, we tested our model on 2 data sets, and the results are shown in <xref ref-type="table" rid="table7">Tables 7</xref> and <xref ref-type="table" rid="table8">8</xref>. <xref ref-type="table" rid="table7">Table 7</xref> shows the comparison of the 4 algorithms on the NHAMCS data set, and <xref ref-type="table" rid="table8">Table 8</xref> shows the results of the 4 algorithms on the NTUH data set.</p>
          <p>Similarly, compared with other algorithms, including the 3 traditional machine learning algorithms, our proposed model outperformed all other methods except in <italic>sensitivity</italic>.</p>
          <table-wrap position="float" id="table7">
            <label>Table 7</label>
            <caption>
              <p>Performance of prediction of intensive care unit admission on the National Hospital Ambulatory Medical Care Survey data set.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="210"/>
              <col width="140"/>
              <col width="140"/>
              <col width="130"/>
              <col width="120"/>
              <col width="120"/>
              <col width="140"/>
              <thead>
                <tr valign="bottom">
                  <td>Model</td>
                  <td>Sensitivity</td>
                  <td>Specificity</td>
                  <td>Precision</td>
                  <td>F1 score</td>
                  <td>Accuracy</td>
                  <td>AUROC<sup>a</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Logistic regression</td>
                  <td>0.787</td>
                  <td>0.734</td>
                  <td>0.761</td>
                  <td>0.760</td>
                  <td>0.761</td>
                  <td>0.845</td>
                </tr>
                <tr valign="top">
                  <td>XGBoost<sup>b</sup></td>
                  <td>0.823</td>
                  <td>0.708</td>
                  <td>0.769</td>
                  <td>0.764</td>
                  <td>0.765</td>
                  <td>0.849</td>
                </tr>
                <tr valign="top">
                  <td>Random forest</td>
                  <td>0.876</td>
                  <td>0.707</td>
                  <td>0.800</td>
                  <td>0.790</td>
                  <td>0.792</td>
                  <td>0.861</td>
                </tr>
                <tr valign="top">
                  <td>Our model</td>
                  <td>0.805</td>
                  <td>
                    <italic>0.807</italic>
                    <sup>c</sup>
                  </td>
                  <td>
                    <italic>0.807</italic>
                  </td>
                  <td>
                    <italic>0.806</italic>
                  </td>
                  <td>
                    <italic>0.824</italic>
                  </td>
                  <td>
                    <italic>0.884</italic>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table7fn1">
                <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
              </fn>
              <fn id="table7fn2">
                <p><sup>b</sup>XGBoost: extreme gradient boosting.</p>
              </fn>
              <fn id="table7fn3">
                <p><sup>c</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <table-wrap position="float" id="table8">
            <label>Table 8</label>
            <caption>
              <p>Performance of prediction of intensive care unit admission on the National Taiwan University Hospital data set.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="210"/>
              <col width="140"/>
              <col width="140"/>
              <col width="130"/>
              <col width="120"/>
              <col width="120"/>
              <col width="140"/>
              <thead>
                <tr valign="bottom">
                  <td>Model</td>
                  <td>Sensitivity</td>
                  <td>Specificity</td>
                  <td>Precision</td>
                  <td>F1 score</td>
                  <td>Accuracy</td>
                  <td>AUROC<sup>a</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Logistic regression</td>
                  <td>0.811</td>
                  <td>0.846</td>
                  <td>0.829</td>
                  <td>0.828</td>
                  <td>0.828</td>
                  <td>0.905</td>
                </tr>
                <tr valign="top">
                  <td>XGBoost<sup>b</sup></td>
                  <td>0.831</td>
                  <td>0.831</td>
                  <td>0.829</td>
                  <td>0.832</td>
                  <td>0.830</td>
                  <td>0.917</td>
                </tr>
                <tr valign="top">
                  <td>Random forest</td>
                  <td>0.833</td>
                  <td>0.828</td>
                  <td>0.831</td>
                  <td>0.830</td>
                  <td>0.831</td>
                  <td>0.911</td>
                </tr>
                <tr valign="top">
                  <td>Our model</td>
                  <td>0.823</td>
                  <td>
                    <italic>0.872</italic>
                    <sup>c</sup>
                  </td>
                  <td>
                    <italic>0.865</italic>
                  </td>
                  <td>
                    <italic>0.843</italic>
                  </td>
                  <td>
                    <italic>0.870</italic>
                  </td>
                  <td>
                    <italic>0.920</italic>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table8fn1">
                <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
              </fn>
              <fn id="table8fn2">
                <p><sup>b</sup>XGBoost: extreme gradient boosting.</p>
              </fn>
              <fn id="table8fn3">
                <p><sup>c</sup>Italicization indicates that the best performance was shown by our model in the metric among the different models.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>There are several limitations to this study. First, the NTUH database used in our experiment belongs to the NTUH and is not publicly available. Hence, it is hard to fairly compare it with the models developed in other studies. However, the NHAMCS data set is publicly available and may be used to evaluate the performance of the models across studies. Second, all the evaluations are based on retrospective data, and future prospective evaluation is needed.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>ED crowding has become one of the biggest issues in health care services. Many countries have shown a steady but significant increase in the number of ED patient visits. Although the ESI system somewhat improves the process of treatment, it still relies on the nurse’s judgment and is prone to the problem where most patients are triaged to ESI level 3. Moreover, the main purpose of the ESI is to classify patients and reserve the more limited resources for those belonging to the high-acuity classes who may need them more urgently. Therefore, a system that can help physicians accurately triage a patient’s condition is imperative. In this work, we proposed a system based on the patients’ ED EMR to predict the need for hospitalizations after the assigned procedures in the ED are completed. This system uses CNNs combined with RNNs, together with an attention mechanism for classification.</p>
        <p>We validated the proposed triage engine based on the developed fusion model on 2 data sets, one of which is from an open data set (NHAMCS) that contains 118,602 ED patient visits in the United States, in which the accuracy and AUROC were 0.83 and 0.87, respectively. On the other hand, we also externally validated our work on the local NTUH data set that includes 745,441 ED patient visits in Taiwan, in which the accuracy and AUROC were 0.83 and 0.88, respectively. Moreover, to effectively evaluate the prediction ability of our proposed system, we also applied the model to other clinical outcomes, including mortality and admission to the intensive care unit. The results showed that our method is approximately 3% to 5% higher in accuracy than other common methods, including 3 traditional machine learning algorithms. Furthermore, the implementation of the proposed system is relatively easy, includes commonly used variables, and is better fitting for real-world clinical settings. It is our future work to validate our novel deep learning–based triage algorithm with prospective clinical trials, and we hope to use it to guide resource allocation in a busy ED once the validation succeeds.</p>
        <p>The unstructured data used in this work were recorded manually by a nurse. However, the text information should be directly described by the ED visits during the ED clinical examination. Therefore, future work may focus on using automatic speech recognition to directly convert and use the speech data of the ED visits. Moreover, although our work includes an analysis of short and long sentences, it does not deal with the relevance of global words. Thus, our future works may focus on combining different types of deep learning algorithms in this system to provide a more comprehensive system, such as a graph convolutional network or transformer.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AUROC</term>
          <def>
            <p>area under the receiver operating characteristic curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BERT</term>
          <def>
            <p>Bidirectional Encoder Representations From Transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">ED</term>
          <def>
            <p>emergency department</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">EMR</term>
          <def>
            <p>electronic medical record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">ESI</term>
          <def>
            <p>emergency severity index</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">LR</term>
          <def>
            <p>logistic regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">NHAMCS</term>
          <def>
            <p>National Hospital Ambulatory Medical Care Survey</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">NTUH</term>
          <def>
            <p>National Taiwan University Hospital</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">RNN</term>
          <def>
            <p>recurrent neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">XGBoost</term>
          <def>
            <p>extreme gradient boosting</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This research was supported by the Joint Research Center for AI Technology and All Vista Healthcare under the Ministry of Science and Technology of Taiwan under the grants 110-2634-F-002-042-, 110-2634-F-002-016-, 110-2634-F-002-046-, and 110-2634-F-002-049-, as well as by the Center for Artificial Intelligence and Advanced Robotics, National Taiwan University.</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pines</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Griffey</surname>
              <given-names>RT</given-names>
            </name>
          </person-group>
          <article-title>What we have learned from a decade of ED crowding research</article-title>
          <source>Acad Emerg Med</source>
          <year>2015</year>
          <month>08</month>
          <volume>22</volume>
          <issue>8</issue>
          <fpage>985</fpage>
          <lpage>7</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/acem.12716"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/acem.12716</pub-id>
          <pub-id pub-id-type="medline">26194441</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>BC</given-names>
            </name>
            <name name-style="western">
              <surname>Hsia</surname>
              <given-names>RY</given-names>
            </name>
            <name name-style="western">
              <surname>Weiss</surname>
              <given-names>RE</given-names>
            </name>
            <name name-style="western">
              <surname>Zingmond</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>McCreath</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Asch</surname>
              <given-names>SM</given-names>
            </name>
          </person-group>
          <article-title>Effect of emergency department crowding on outcomes of admitted patients</article-title>
          <source>Ann Emerg Med</source>
          <year>2013</year>
          <month>06</month>
          <volume>61</volume>
          <issue>6</issue>
          <fpage>605</fpage>
          <lpage>11</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/23218508"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.annemergmed.2012.10.026</pub-id>
          <pub-id pub-id-type="medline">23218508</pub-id>
          <pub-id pub-id-type="pii">S0196-0644(12)01699-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC3690784</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Morley</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Unwin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Peterson</surname>
              <given-names>GM</given-names>
            </name>
            <name name-style="western">
              <surname>Stankovich</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kinsman</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Emergency department crowding: a systematic review of causes, consequences and solutions</article-title>
          <source>PLoS One</source>
          <year>2018</year>
          <volume>13</volume>
          <issue>8</issue>
          <fpage>e0203316</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dx.plos.org/10.1371/journal.pone.0203316"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pone.0203316</pub-id>
          <pub-id pub-id-type="medline">30161242</pub-id>
          <pub-id pub-id-type="pii">PONE-D-18-06823</pub-id>
          <pub-id pub-id-type="pmcid">PMC6117060</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Morley</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Stankovich</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Peterson</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Kinsman</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Planning for the future: Emergency department presentation patterns in Tasmania, Australia</article-title>
          <source>Int Emerg Nurs</source>
          <year>2018</year>
          <month>05</month>
          <volume>38</volume>
          <fpage>34</fpage>
          <lpage>40</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1755-599X(17)30049-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.ienj.2017.09.001</pub-id>
          <pub-id pub-id-type="medline">28958418</pub-id>
          <pub-id pub-id-type="pii">S1755-599X(17)30049-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guttmann</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schull</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Vermeulen</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Stukel</surname>
              <given-names>TA</given-names>
            </name>
          </person-group>
          <article-title>Association between waiting times and short term mortality and hospital admission after departure from emergency department: population based cohort study from Ontario, Canada</article-title>
          <source>Br Med J</source>
          <year>2011</year>
          <month>06</month>
          <day>01</day>
          <volume>342</volume>
          <fpage>d2983</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/21632665"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmj.d2983</pub-id>
          <pub-id pub-id-type="medline">21632665</pub-id>
          <pub-id pub-id-type="pmcid">PMC3106148</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dinh</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Bein</surname>
              <given-names>KJ</given-names>
            </name>
            <name name-style="western">
              <surname>Latt</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chalkley</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Muscatello</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Age before acuity: the drivers of demand for emergency department services in the Greater Sydney Area</article-title>
          <source>Emerg Med J</source>
          <year>2015</year>
          <month>09</month>
          <volume>32</volume>
          <issue>9</issue>
          <fpage>708</fpage>
          <lpage>11</lpage>
          <pub-id pub-id-type="doi">10.1136/emermed-2014-204174</pub-id>
          <pub-id pub-id-type="medline">25532104</pub-id>
          <pub-id pub-id-type="pii">emermed-2014-204174</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lowthian</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Curtis</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jolley</surname>
              <given-names>DJ</given-names>
            </name>
            <name name-style="western">
              <surname>Stoelwinder</surname>
              <given-names>JU</given-names>
            </name>
            <name name-style="western">
              <surname>McNeil</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Cameron</surname>
              <given-names>PA</given-names>
            </name>
          </person-group>
          <article-title>Demand at the emergency department front door: 10-year trends in presentations</article-title>
          <source>Med J Aust</source>
          <year>2012</year>
          <month>02</month>
          <day>06</day>
          <volume>196</volume>
          <fpage>128</fpage>
          <lpage>32</lpage>
          <pub-id pub-id-type="doi">10.5694/mja11.10955</pub-id>
          <pub-id pub-id-type="medline">22304608</pub-id>
          <pub-id pub-id-type="pii">low10955_fm</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Augustine</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>The latest emergency department utilization numbers are in</article-title>
          <source>ACEPNow</source>
          <year>2019</year>
          <access-date>2021-11-26</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.acepnow.com/article/the-latest-emergency-department-utilization-numbers-are-in/">https://www.acepnow.com/article/the-latest-emergency-department-utilization-numbers-are-in/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rui</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ashman</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>National hospital ambulatory medical care survey: 2016 emergency department summary tables</article-title>
          <source>Centers for Disease Control and Prevention</source>
          <year>2016</year>
          <access-date>2021-11-27</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.cdc.gov/nchs/data/nhamcs/web_tables/2016_ed_web_tables.pdf">https://www.cdc.gov/nchs/data/nhamcs/web_tables/2016_ed_web_tables.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="web">
          <article-title>Statistics of Medical Care Institution's Status and Hospital Utilization 2019</article-title>
          <source>Ministry of Health and Welfare, Taiwan</source>
          <access-date>2021-11-27</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mohw.gov.tw/lp-4932-2.html">https://www.mohw.gov.tw/lp-4932-2.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pines</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Hilton</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Weber</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Alkemade</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Al Shabanah</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Anderson</surname>
              <given-names>PD</given-names>
            </name>
            <name name-style="western">
              <surname>Bernhard</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Bertini</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gries</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ferrandiz</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>VA</given-names>
            </name>
            <name name-style="western">
              <surname>Harjola</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Hogan</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Madsen</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Mason</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ohlén</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Rainer</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Rathlev</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Revue</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Richardson</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sattarian</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schull</surname>
              <given-names>MJ</given-names>
            </name>
          </person-group>
          <article-title>International perspectives on emergency department crowding</article-title>
          <source>Acad Emerg Med</source>
          <year>2011</year>
          <month>12</month>
          <volume>18</volume>
          <issue>12</issue>
          <fpage>1358</fpage>
          <lpage>70</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/j.1553-2712.2011.01235.x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/j.1553-2712.2011.01235.x</pub-id>
          <pub-id pub-id-type="medline">22168200</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chalfin</surname>
              <given-names>DB</given-names>
            </name>
            <name name-style="western">
              <surname>Trzeciak</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Likourezos</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Baumann</surname>
              <given-names>BM</given-names>
            </name>
            <name name-style="western">
              <surname>Dellinger</surname>
              <given-names>RP</given-names>
            </name>
            <collab>DELAY-ED Study Group</collab>
          </person-group>
          <article-title>Impact of delayed transfer of critically ill patients from the emergency department to the intensive care unit</article-title>
          <source>Crit Care Med</source>
          <year>2007</year>
          <month>06</month>
          <volume>35</volume>
          <issue>6</issue>
          <fpage>1477</fpage>
          <lpage>83</lpage>
          <pub-id pub-id-type="doi">10.1097/01.CCM.0000266585.74905.5A</pub-id>
          <pub-id pub-id-type="medline">17440421</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ahsan</surname>
              <given-names>KB</given-names>
            </name>
            <name name-style="western">
              <surname>Alam</surname>
              <given-names>MR</given-names>
            </name>
            <name name-style="western">
              <surname>Morel</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Karim</surname>
              <given-names>MA</given-names>
            </name>
          </person-group>
          <article-title>Emergency department resource optimisation for improved performance: a review</article-title>
          <source>J Ind Eng Int</source>
          <year>2019</year>
          <month>11</month>
          <day>27</day>
          <volume>15</volume>
          <issue>S1</issue>
          <fpage>253</fpage>
          <lpage>66</lpage>
          <pub-id pub-id-type="doi">10.1007/s40092-019-00335-x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fernandes</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Vieira</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Leite</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Palos</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Finkelstein</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sousa</surname>
              <given-names>JM</given-names>
            </name>
          </person-group>
          <article-title>Clinical decision support systems for triage in the emergency department using intelligent systems: a review</article-title>
          <source>Artif Intell Med</source>
          <year>2020</year>
          <month>01</month>
          <volume>102</volume>
          <fpage>101762</fpage>
          <pub-id pub-id-type="doi">10.1016/j.artmed.2019.101762</pub-id>
          <pub-id pub-id-type="medline">31980099</pub-id>
          <pub-id pub-id-type="pii">S0933-3657(19)30126-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gligorijevic</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Stojanovic</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Satz</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Stojkovic</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Schreyer</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Del Portal</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Obradovic</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Deep attention model for triage of emergency department patients</article-title>
          <source>Proceedings of the 2018 SIAM International Conference on Data Mining (SDM)</source>
          <year>2018</year>
          <conf-name>SIAM International Conference on Data Mining (SDM)</conf-name>
          <conf-date>May 3-5, 2018</conf-date>
          <conf-loc>San deigo CA, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1137/1.9781611975321.34</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Patzer</surname>
              <given-names>RE</given-names>
            </name>
            <name name-style="western">
              <surname>Pitts</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Patzer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schrager</surname>
              <given-names>JD</given-names>
            </name>
          </person-group>
          <article-title>Prediction of emergency department hospital admission based on natural language processing and neural networks</article-title>
          <source>Methods Inf Med</source>
          <year>2017</year>
          <month>10</month>
          <day>26</day>
          <volume>56</volume>
          <issue>5</issue>
          <fpage>377</fpage>
          <lpage>89</lpage>
          <pub-id pub-id-type="doi">10.3414/ME17-01-0024</pub-id>
          <pub-id pub-id-type="medline">28816338</pub-id>
          <pub-id pub-id-type="pii">17-01-0024</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Heng</surname>
              <given-names>BH</given-names>
            </name>
            <name name-style="western">
              <surname>Tay</surname>
              <given-names>SY</given-names>
            </name>
            <name name-style="western">
              <surname>Seow</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Predicting hospital admissions at emergency department triage using routine administrative data</article-title>
          <source>Acad Emerg Med</source>
          <year>2011</year>
          <month>08</month>
          <volume>18</volume>
          <issue>8</issue>
          <fpage>844</fpage>
          <lpage>50</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/j.1553-2712.2011.01125.x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/j.1553-2712.2011.01125.x</pub-id>
          <pub-id pub-id-type="medline">21843220</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Graham</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Bond</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Quinn</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mulvenna</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Using data mining to predict hospital admissions from the emergency department</article-title>
          <source>IEEE Access</source>
          <year>2018</year>
          <volume>6</volume>
          <fpage>10458</fpage>
          <lpage>69</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2018.2808843</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>DeepTriager: a neural attention model for emergency triage with electronic health records</article-title>
          <source>Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM)</source>
          <year>2019</year>
          <conf-name>IEEE International Conference on Bioinformatics and Biomedicine (BIBM)</conf-name>
          <conf-date>Nov. 18-21, 2019</conf-date>
          <conf-loc>San Diego, CA, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bibm47256.2019.8983093</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Joulin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Grave</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Bojanowski</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Mikolov</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Bag of tricks for efficient text classification</article-title>
          <source>Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics</source>
          <year>2017</year>
          <conf-name>15th Conference of the European Chapter of the Association for Computational Linguistics</conf-name>
          <conf-date>April 2017</conf-date>
          <conf-loc>Valencia, Spain</conf-loc>
          <fpage>427</fpage>
          <lpage>31</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/e17-2068</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bojanowski</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Grave</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Joulin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mikolov</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Enriching word vectors with subword information</article-title>
          <source>Transact Assoc Computat Linguist</source>
          <year>2017</year>
          <month>12</month>
          <volume>5</volume>
          <fpage>135</fpage>
          <lpage>46</lpage>
          <pub-id pub-id-type="doi">10.1162/tacl_a_00051</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Clevert</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Unterthiner</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Hochreiter</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Fast and accurate deep network learning by exponential linear units (ELUs)</article-title>
          <source>Proceedings of the 4th International Conference on Learning Representations, ICLR</source>
          <year>2016</year>
          <conf-name>4th International Conference on Learning Representations, ICLR</conf-name>
          <conf-date>May 2-4, 2016</conf-date>
          <conf-loc>San Juan, Puerto Rico</conf-loc>
          <fpage>2</fpage>
          <lpage>4</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1511.07289.pdf%5cnhttp://arxiv.org/abs/1511.07289%5cnhttp://arxiv.org/abs/1511.07289.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goodfellow</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Bengio</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Courville</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <source>Deep Learning</source>
          <year>2016</year>
          <publisher-loc>Cambridge, USA</publisher-loc>
          <publisher-name>The MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Deep pyramid convolutional neural networks for text categorization</article-title>
          <source>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</source>
          <year>2017</year>
          <conf-name>55th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>July, 2017</conf-date>
          <conf-loc>Vancouver, Canada</conf-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <pub-id pub-id-type="doi">10.18653/v1/p17-1052</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Convolutional neural networks for sentence classification</article-title>
          <source>Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          <year>2014</year>
          <conf-name>Conference on Empirical Methods in Natural Language Processing (EMNLP)</conf-name>
          <conf-date>October, 2014</conf-date>
          <conf-loc>Doha, Qatar</conf-loc>
          <pub-id pub-id-type="doi">10.3115/v1/d14-1181</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kingma</surname>
              <given-names>DP</given-names>
            </name>
            <name name-style="western">
              <surname>Ba</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Adam: a method for stochastic optimization</article-title>
          <source>Proceedings of the 3rd International Conference on Learning Representations</source>
          <year>2015</year>
          <conf-name>3rd International Conference on Learning Representations</conf-name>
          <conf-date>May 7-9, 2015</conf-date>
          <conf-loc>San Diego, CA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1412.6980"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>MW</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          <year>2019</year>
          <conf-name>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</conf-name>
          <conf-date>June, 2019</conf-date>
          <conf-loc>Minneapolis, Minnesota</conf-loc>
          <fpage>4171</fpage>
          <lpage>86</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/n18-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raita</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Goto</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Faridi</surname>
              <given-names>MK</given-names>
            </name>
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>DFM</given-names>
            </name>
            <name name-style="western">
              <surname>Camargo</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Hasegawa</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Emergency department triage prediction of clinical outcomes using machine learning models</article-title>
          <source>Crit Care</source>
          <year>2019</year>
          <month>02</month>
          <day>22</day>
          <volume>23</volume>
          <issue>1</issue>
          <fpage>64</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ccforum.biomedcentral.com/articles/10.1186/s13054-019-2351-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s13054-019-2351-7</pub-id>
          <pub-id pub-id-type="medline">30795786</pub-id>
          <pub-id pub-id-type="pii">10.1186/s13054-019-2351-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC6387562</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
