<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v24i3e27210</article-id>
      <article-id pub-id-type="pmid">35319481</article-id>
      <article-id pub-id-type="doi">10.2196/27210</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Kukafka</surname>
            <given-names>Rita</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Roberts</surname>
            <given-names>Kirk</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>De Carvalho</surname>
            <given-names>Rogerio</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Louren</surname>
            <given-names>Anália</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Mitchell</surname>
            <given-names>Joseph Ross</given-names>
          </name>
          <degrees>MSc, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2340-4708</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Szepietowski</surname>
            <given-names>Phillip</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5525-6132</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Howard</surname>
            <given-names>Rachel</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-7934-5081</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Reisman</surname>
            <given-names>Phillip</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2591-6544</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Jones</surname>
            <given-names>Jennie D</given-names>
          </name>
          <degrees>MSHI-HA</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3102-9469</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Lewis</surname>
            <given-names>Patricia</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-3188-370X</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Fridley</surname>
            <given-names>Brooke L</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7739-7956</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Rollison</surname>
            <given-names>Dana E</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <address>
            <institution>Department of Health Data Services</institution>
            <institution>H Lee Moffitt Cancer Center and Research Institute</institution>
            <addr-line>12902 Magnolia Drive</addr-line>
            <addr-line>Tampa, FL, 33612</addr-line>
            <country>United States</country>
            <phone>1 813 745 6530</phone>
            <email>Dana.Rollison@moffitt.org</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3171-8045</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Machine Learning</institution>
        <institution>H Lee Moffitt Cancer Center and Research Institute</institution>
        <addr-line>Tampa, FL</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Medicine, Faculty of Medicine &#38; Dentistry, and the Alberta Machine Intelligence Institute</institution>
        <institution>University of Alberta</institution>
        <addr-line>Edmonton, AB</addr-line>
        <country>Canada</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Alberta Health Services</institution>
        <addr-line>Edmonton, AB</addr-line>
        <country>Canada</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Department of Health Data Services</institution>
        <institution>H Lee Moffitt Cancer Center and Research Institute</institution>
        <addr-line>Tampa, FL</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>Department of Biostatistics and Bioinformatics</institution>
        <institution>H Lee Moffitt Cancer Center and Research Institute</institution>
        <addr-line>Tampa, FL</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Dana E Rollison <email>Dana.Rollison@moffitt.org</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>3</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>23</day>
        <month>3</month>
        <year>2022</year>
      </pub-date>
      <volume>24</volume>
      <issue>3</issue>
      <elocation-id>e27210</elocation-id>
      <history>
        <date date-type="received">
          <day>16</day>
          <month>1</month>
          <year>2021</year>
        </date>
        <date date-type="rev-request">
          <day>30</day>
          <month>7</month>
          <year>2021</year>
        </date>
        <date date-type="rev-recd">
          <day>22</day>
          <month>10</month>
          <year>2021</year>
        </date>
        <date date-type="accepted">
          <day>10</day>
          <month>11</month>
          <year>2021</year>
        </date>
      </history>
      <copyright-statement>©Joseph Ross Mitchell, Phillip Szepietowski, Rachel Howard, Phillip Reisman, Jennie D Jones, Patricia Lewis, Brooke L Fridley, Dana E Rollison. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 23.03.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2022/3/e27210" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Information in pathology reports is critical for cancer care. Natural language processing (NLP) systems used to extract information from pathology reports are often narrow in scope or require extensive tuning. Consequently, there is growing interest in automated deep learning approaches. A powerful new NLP algorithm, bidirectional encoder representations from transformers (BERT), was published in late 2018. BERT set new performance standards on tasks as diverse as question answering, named entity recognition, speech recognition, and more.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The aim of this study is to develop a BERT-based system to automatically extract detailed tumor site and histology information from free-text oncological pathology reports.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We pursued three specific aims: extract accurate tumor site and histology descriptions from free-text pathology reports, accommodate the diverse terminology used to indicate the same pathology, and provide accurate standardized tumor site and histology codes for use by downstream applications. We first trained a base language model to comprehend the technical language in pathology reports. This involved unsupervised learning on a training corpus of 275,605 electronic pathology reports from 164,531 unique patients that included 121 million words. Next, we trained a question-and-answer (Q&#38;A) model that connects a Q&#38;A layer to the base pathology language model to answer pathology questions. Our Q&#38;A system was designed to search for the answers to two predefined questions in each pathology report: <italic>What organ contains the tumor?</italic> and <italic>What is the kind of tumor or carcinoma?</italic> This involved supervised training on 8197 pathology reports, each with ground truth answers to these 2 questions determined by certified tumor registrars. The data set included 214 tumor sites and 193 histologies. The tumor site and histology phrases extracted by the Q&#38;A model were used to predict International Classification of Diseases for Oncology, Third Edition (ICD-O-3), site and histology codes. This involved fine-tuning two additional BERT models: one to predict site codes and another to predict histology codes. Our final system includes a network of 3 BERT-based models. We call this CancerBERT network (caBERTnet). We evaluated caBERTnet using a sequestered test data set of 2050 pathology reports with ground truth answers determined by certified tumor registrars.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>caBERTnet’s accuracies for predicting group-level site and histology codes were 93.53% (1895/2026) and 97.6% (1993/2042), respectively. The top 5 accuracies for predicting fine-grained ICD-O-3 site and histology codes with 5 or more samples each in the training data set were 92.95% (1794/1930) and 96.01% (1853/1930), respectively.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>We have developed an NLP system that outperforms existing algorithms at predicting ICD-O-3 codes across an extensive range of tumor sites and histologies. Our new system could help reduce treatment delays, increase enrollment in clinical trials of new therapies, and improve patient outcomes.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>natural language processing</kwd>
        <kwd>NLP</kwd>
        <kwd>BERT</kwd>
        <kwd>transformer</kwd>
        <kwd>pathology</kwd>
        <kwd>ICD-O-3</kwd>
        <kwd>deep learning</kwd>
        <kwd>cancer</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Much of the information in electronic medical records (EMRs) required for the practice of clinical oncology and cancer research is contained in unstructured text. As much as 80% of EMR data can be found in narrative text notes and scanned documents [<xref ref-type="bibr" rid="ref1">1</xref>], ranging from clinic or surgical notes, including pathology, radiology, and ambulatory care, to past medical or family history. The extraction of discrete data elements from this unstructured text, particularly those relating to disease diagnosis and commonly captured in routine pathology reports, is critical for the selection of treatment options, identification of patients eligible for clinical trials, and monitoring of adherence to established clinical treatment pathways.</p>
        <p>Natural language processing (NLP) has been used to extract information from medical text for several decades [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>], and a thorough review of NLP-based information extraction for cancer-related EMR notes can be found in the study by Datta et al [<xref ref-type="bibr" rid="ref7">7</xref>]. Application of NLP to pathology reports has also been prevalent in the literature during the course of the last decade [<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref19">19</xref>]. That said, many early studies used regular expression– and rule-based systems [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>] that require considerable up-front development and can be difficult to adapt and maintain.</p>
        <p>More recently, there has been growing interest in more highly automated deep learning approaches for clinical NLP [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. In late 2018, a powerful new deep learning NLP algorithm was released: bidirectional encoder representations from transformers (BERT) [<xref ref-type="bibr" rid="ref23">23</xref>]. BERT established new, state-of-the-art performance levels on common nonclinical NLP benchmarks [<xref ref-type="bibr" rid="ref24">24</xref>]. This success spawned rapid research and development of multiple BERT-inspired and transformer-based neural architectures [<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref33">33</xref>]. Several of these have, for the first time, achieved or surpassed human-level performance on tasks as diverse as question answering, named entity recognition, speech recognition, and more [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref33">33</xref>-<xref ref-type="bibr" rid="ref35">35</xref>]. BERT and related architectures have facilitated significant improvements in multiple medical applications, including processing of electronic health records [<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref37">37</xref>], outcome prediction [<xref ref-type="bibr" rid="ref38">38</xref>-<xref ref-type="bibr" rid="ref40">40</xref>], identification of medical terms and concepts [<xref ref-type="bibr" rid="ref41">41</xref>], medical chatbots [<xref ref-type="bibr" rid="ref42">42</xref>], sentiment analysis [<xref ref-type="bibr" rid="ref43">43</xref>], recommender systems [<xref ref-type="bibr" rid="ref44">44</xref>], and others. Despite BERT’s success, we are aware of only a single application of BERT to the already promising area of free-text pathology reports [<xref ref-type="bibr" rid="ref45">45</xref>]. The study focused on classification of text into only a few cancer-related categories, including afflicted organ (15 organ groups), disease type (noncancer, premalignant, or cancer), cancer reason (6 histology groups), and presence of metastatic disease (no, yes: in lymph nodes and yes: in non–lymph node tissue). Our goal is to develop and evaluate a BERT-based system to extract detailed tumor site and histology information from free-text pathology reports. The availability of manually curated data within the H Lee Moffitt Cancer Center and Research Institute (Moffitt) Cancer Registry (MCR) represented a unique opportunity to train a BERT-based system using a gold standard data set classified using a standard ontology.</p>
        <p>BERT’s proficiency at question answering prompted us to construct a question-and-answer (Q&#38;A) system to extract clinical data from pathology reports. This concept has long been compelling—Q&#38;A systems for medical data extraction have been pursued for &#62;40 years [<xref ref-type="bibr" rid="ref46">46</xref>]. Such a system would have several desirable properties: an intuitive user interface, the ability to extract additional data fields by searching for answers to additional questions, and the ability to generalize to other medical documents. Furthermore, it would allow us to make data available for clinical and research use close to real time, thus reducing treatment delays, increasing enrollment in clinical trials of new therapies, and improving patient outcomes. To train such a general-purpose Q&#38;A system on pathology reports, one would need a diverse set of questions on which to train it. Our task in this paper is more modest (and is in essence a classification task of site and histology); however, we view the Q&#38;A portion of our system as a small step toward this broader goal.</p>
      </sec>
      <sec>
        <title>Primary Contributions</title>
        <p>This work describes three primary contributions:</p>
        <list list-type="order">
          <list-item>
            <p>A new BERT language model for comprehension of pathology reports in oncology. We call this <italic>CancerBERT</italic>, or <italic>caBERT</italic> for short.</p>
          </list-item>
          <list-item>
            <p>A new Q&#38;A caBERT-based system, tolerant to varied terminologies, word orders, and spelling mistakes, to extract tumor site and histology descriptions from free-text pathology reports.</p>
          </list-item>
          <list-item>
            <p>A new caBERT network (caBERTnet) to predict International Classification of Diseases for Oncology, version 3.2 (ICD-O-3.2), codes from the extracted descriptions. This system can handle up to 332 organ sites and 1143 tumor histologies. On an unseen test data set with 214 sites and 193 histologies it achieved overall accuracies that are equal to, or above, those of existing systems, while also expanding on the number of site and histology classes captured by these systems. Although the results in practice still require human validation, they provide a means of early abstraction from unstructured pathology text over a very broad set of sites and histologies and in addition can provide an initial signal to assist expert cancer registrars in their case diagnosis–abstraction workflow.</p>
          </list-item>
        </list>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <p>Information on our software development tools is provided in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> [<xref ref-type="bibr" rid="ref47">47</xref>-<xref ref-type="bibr" rid="ref53">53</xref>]. To construct our system, we had to achieve three specific aims: (1) extract accurate tumor site and histology descriptions from complex free-text pathology reports, (2) accommodate the diverse terminology used to indicate the same pathology, and (3) provide accurate standardized tumor site and histology codes for use by downstream applications.</p>
      <sec>
        <title>Extract Tumor Site and Histology Descriptions</title>
        <sec>
          <title>Overview</title>
          <p>Constructing this system required us to first train a base language model to comprehend the technical language in pathology reports (<xref ref-type="boxed-text" rid="box1">Textbox 1</xref>). This involved <italic>unsupervised</italic> training on a large corpus of pathology text. For this we constructed a <italic>pathology language–model training data set</italic>, described in more detail in the next section (<italic>Pathology Language Model</italic>).</p>
          <boxed-text id="box1" position="float">
            <title>A fragment of text from a pathology report generated at H Lee Moffitt Cancer Center and Research Institute (Moffitt).</title>
            <p>
              <bold>Fragment of text from a pathology report generated at Moffitt</bold>
            </p>
            <list list-type="bullet">
              <list-item>
                <p>clinical history: not given. preoperative diagnosis: right lower lobe, squamous cell carcinoma. specimen types: a: right station 7 fs b: station # 10 c: right lung fs d: station 4r e: additional station 10 f: additional station 7 final diagnosis: a. lymph node right station 7 biopsy: anthracotic lymph node 1 with lymphoid hyperplasia negative for malignancy. b. lymph node station 10 biopsy: anthracotic lymph node 1 negative for malignancy. c. lung right pneumonectomy: moderate to poorly differentiated squamous cell carcinoma with extensive necrosis see key pathological findings. bronchial and vascular resection margins negative for malignancy. three of 10 hilar lymph nodes with metastatic squamous cell carcinoma. d. lymph nodes station 4r biopsies: anthracotic lymph nodes 5 negative for malignancy. e. lymph node additional station 10 biopsy: minute lymph node 1 negative for malignancy. f. lymph node station 7 biopsy: anthracotic lymph nodes 4 negative for malignancy. key pathological findings tumor type: squamous cell carcinoma with extensive necrosis. histological grade: moderate to poorly differentiated. tumor location: right lung involving right lower lobe. right upper and middle lobes free of tumor.</p>
              </list-item>
            </list>
          </boxed-text>
          <p>Next, it required us to train a Q&#38;A model that appends a Q&#38;A layer onto the pathology language model to answer pathology questions. Our Q&#38;A system was designed to search for the answers to two predefined questions in each pathology report:</p>
          <list list-type="order">
            <list-item>
              <p>What organ contains the tumor?</p>
            </list-item>
            <list-item>
              <p>What is the kind of tumor or carcinoma?</p>
            </list-item>
          </list>
          <p>For example, when presented with the report shown in <xref ref-type="boxed-text" rid="box1">Textbox 1</xref>, the system should respond to the question <italic>What organ contains the tumor?</italic> with <italic>C343: lower lobe, lung</italic>, and would respond to the question <italic>What is the kind of tumor or carcinoma?</italic> with <italic>8070/3: squamous cell carcinoma, nos</italic> (not otherwise specified).</p>
          <p>This involved <italic>supervised</italic> training on a set of pathology reports, each with ground truth answers to these 2 questions determined by human experts. To do this, we constructed a second <italic>fine-tuning training data set</italic>, described in more detail in the <italic>Pathology Q&#38;A Model</italic> section. Finally, we evaluated our system using a sequestered <italic>fine-tuning testing data set</italic>, described in more detail in the <italic>Pathology Q&#38;A Model</italic> section.</p>
        </sec>
        <sec>
          <title>Pathology Language Model</title>
          <p>Training a base language model to comprehend pathology reports leveraged prior work by several groups (<xref rid="figure1" ref-type="fig">Figure 1</xref>). Lee et al [<xref ref-type="bibr" rid="ref54">54</xref>] performed transfer learning on BERT using nearly 18 billion words extracted from PubMed abstracts. The result, BioBERT, is tuned for biomedical language comprehension tasks and is publicly available. Next, Alsentzer et al [<xref ref-type="bibr" rid="ref55">55</xref>] performed transfer learning on BioBERT to tune it for clinical language comprehension. They used EMR notes in the Medical Information Mart for Intensive Care, version 3 (MIMIC-III) data set [<xref ref-type="bibr" rid="ref56">56</xref>], which includes data from approximately 60,000 intensive care unit stays by patients at Beth-Israel Hospital in Boston, Massachusetts. The model created by Alsentzer et al [<xref ref-type="bibr" rid="ref55">55</xref>], ClinicalBERT, was also made publicly available.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>Sequence of transfer-learning steps used in training the CancerBERT base language model. BERT: bidirectional encoder representations from transformers; ICU: intensive care unit; MIMIC-III: Medical Information Mart for Intensive Care, version 3.</p>
            </caption>
            <graphic xlink:href="jmir_v24i3e27210_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Alsentzer et al [<xref ref-type="bibr" rid="ref55">55</xref>] created two models built upon BioBERT: one trained on all MIMIC-III notes and one trained on just the MIMIC-III discharge summaries. Initial pretraining experimentation revealed that the latter provided higher accuracies on a separate sample of our pathology reports. We noted that Moffitt pathology reports have a language structure closer to discharge summaries than to general clinical notes. Consequently, our model was initialized with weights from the latter of the two ClinicalBERT models: ClinicalBERT–Bio+Discharge Summary BERT Model.</p>
          <p>Transfer learning was accomplished by performing masked-language modeling [<xref ref-type="bibr" rid="ref23">23</xref>]. Briefly, 15% of the words in the corpus are selected at random and replaced with a <italic>mask</italic> token. The language model is then trained to predict the masked words. The word-masking process is performed automatically at the beginning of each training run.</p>
          <p>Our language-model training corpus included electronic pathology reports of solid tumors produced by pathologists at Moffitt between 1986 and 2020. The year 1986 was the earliest date on pathology reports cataloged in our enterprise data warehouse. The data set was restricted to solid tumors for two reasons: first, to focus the problem domain for this proof-of-concept study, and, second, Moffitt hematologic pathology reports follow a quasi-structured format, reducing the need for extraction of data from unstructured text.</p>
          <p>This data set contained both Health Level Seven International messages and plain-text pathology reports. These were minimally processed to extract and clean the text relevant to pathology (more details can be found in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The final language model training corpus included 275,605 electronic pathology reports from 164,531 unique patients and included 121 million words.</p>
        </sec>
        <sec>
          <title>Pathology Q&#38;A Model</title>
          <p>The pathology Q&#38;A lesson plan involved 3 stages, each intended to improve our system’s comprehension of pathology reports and thereby increase the accuracy of question answering (<xref rid="figure2" ref-type="fig">Figure 2</xref>). The three stages involved training the Q&#38;A model to (1) answer general English language questions, (2) answer technical biomedical science questions, and (3) answer questions from Moffitt pathology reports.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Lesson plan for the caBERT network consisting of one question and answering model A and two classification models, one for primary site (model B) and another for histology (model C). BioASQ: Biomedical Semantic Indexing and Question Answering; Q&#38;A: question and answer; SQuAD: Stanford Question Answering Dataset.</p>
            </caption>
            <graphic xlink:href="jmir_v24i3e27210_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Each training stage used supervised learning. This required a training data set that included passages of text, ≥1 questions related to each passage, and ground truth answers to those questions that appeared as contiguous phrases within the related passage. At the end of each stage we evaluated our system using the same sequestered test data set constructed from Moffitt pathology reports, as described in more detail later in this section. The experimental parameters used to train the Q&#38;A model were held constant over all stages and are listed in Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
          <p>For the first stage of training we used the Stanford Question Answering Dataset (SQuAD), version 1.1 [<xref ref-type="bibr" rid="ref57">57</xref>]. SQuAD consists of more than 100,000 questions and answers created by crowdworkers on Wikipedia articles. The SQuAD data format is widely used in NLP research. Therefore, we designed our system to read and process data sets in this format.</p>
          <p>For the second stage of training we used the large-scale Biomedical Semantic Indexing and Question Answering (BioASQ) data set [<xref ref-type="bibr" rid="ref58">58</xref>]. In particular, we used data from BioASQ Challenge 7b: Biomedical Semantic Question Answering. This data set contains 2747 training questions along with their ground truth answers. According to the BioASQ Challenge 7b description: “All the questions are constructed by biomedical experts from around Europe.” This data set was converted to SQuAD format by Yoon et al [<xref ref-type="bibr" rid="ref59">59</xref>] and made available for public use.</p>
          <p>Next, we constructed a Q&#38;A fine-tuning data set in SQuAD format based on Moffitt pathology reports. We obtained ground truth answers to our 2 questions from data abstracted by Moffitt certified tumor registrars (CTRs). CTRs undergo an extensive training and internship program to become proficient at extracting quantitative and categorical data from unstructured pathology reports. They are widely employed by cancer centers and other organizations to extract data for clinical and research applications and for reporting to state and national agencies. The MCR deploys state-of-the art quality assurance procedures: its benchmark for quality is 90% and its target accuracy is 95% [<xref ref-type="bibr" rid="ref60">60</xref>].</p>
          <p>Moffitt’s enterprise data warehouse was searched to find solid tumor pathology reports generated after 2006 with matched MCR data (<xref rid="figure3" ref-type="fig">Figure 3</xref>). These reports were screened to ensure that they contained a description of a positive diagnosis of a single primary tumor. Next, each report was processed to ensure that it contained an answer to at least one of the questions in the Q&#38;A model. This was accomplished programmatically by searching each report for a phrase contained in a table of acceptable answer phrases (Figure S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>), as described in more detail in the next section (<italic>Accommodate Diverse Terminology</italic>). Our search produced 16782 reports that met these inclusion criteria (<xref rid="figure3" ref-type="fig">Figure 3</xref>).</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Flowchart depicting the data curation process for creating the Moffitt fine-tuning data sets (used in the site and histology question-and-answer and classification tasks). MCR: Moffitt Cancer Registry; MRN: medical record number.</p>
            </caption>
            <graphic xlink:href="jmir_v24i3e27210_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Next, we curated these reports to ensure that (1) the relative frequencies of the 10 most common tumor sites and histologies in this collection matched the relative frequencies in Moffitt’s patient population as a whole and (2) all MCR-assigned tumor sites and histologies reported for Moffitt patients after 2006 were represented in the data set. The final curated collection contained 10,247 reports (<xref rid="figure3" ref-type="fig">Figure 3</xref>). We will refer to this data set as the <italic>fine-tuning</italic> data set to contrast it with the caBERT language model data set used to train the base language model as described in the <italic>Pathology Language Model</italic> section.</p>
          <p>The curated collection of reports was randomly divided to create two data sets: 79.99% (8197/10,247) of the reports were used to create the Moffitt fine-tuning training data set and 20.01% (2050/10,247) were used to create the Moffitt fine-tuning testing data set. Each data set was saved in SQuAD format using a custom-written Python program. The training data set was used for the final stage of Q&#38;A training and also for ICD-O-3 code predictions, described in more detail in the <italic>Accurate ICD-O-3 Codes</italic> section. The testing subset was used to evaluate the impact of Q&#38;A training at the end of each training stage (Figure S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>) and also to evaluate the performance of the final pipeline (<xref rid="figure4" ref-type="fig">Figure 4</xref>).</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>The final caBERT network (caBERTnet) connects caBERT instances A, B, and C, used for site and histology question and answering, International Classification of Diseases for Oncology, Third Edition (ICD-O-3) primary site and ICD-O-3 histology code classification, respectively.</p>
            </caption>
            <graphic xlink:href="jmir_v24i3e27210_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Question-answering accuracy was evaluated using two metrics: exact match and F1 score. Exact match is true if the caBERT-extracted phrase is an identical word-for-word match with the MCR phrase and false otherwise. We calculated the average number of true results across all test samples. The F1 score is a measure of the degree of overlap among the words in the caBERT-extracted phrase and the MCR phrase. This varies from 0 (no words in common) to 1 (all words in common but not necessarily in the same order). Each exact match corresponded to an F1 score of 1.0. We calculated the average F1 score across all test samples and expressed it as a percentage.</p>
          <p>After all training stages were complete, they were repeated using the initial ClinicalBERT model (<italic>ClinicalBERT–Bio+Discharge Summary BERT</italic>) with a new randomly initialized Q&#38;A layer. This allowed us to determine the impact of developing a pathology-tuned BERT model on extraction accuracy over the baseline accuracy of using ClinicalBERT alone (Figure S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The training parameters were set to those optimized for ClinicalBERT and reported by Alsentzer et al [<xref ref-type="bibr" rid="ref55">55</xref>].</p>
        </sec>
      </sec>
      <sec>
        <title>Accommodate Diverse Terminology</title>
        <p>Our pathology reports came supplied with ground truth labels for primary site and histology in the form of ICD-O-3 codes [<xref ref-type="bibr" rid="ref61">61</xref>], which were abstracted by Moffitt CTRs. To train the Q&#38;A model we needed a method to determine the precise location within each pathology report of the actual text corresponding to these labels. This proved nontrivial owing to the rather diverse terminology within each pathology report used to refer to each primary site and histology.</p>
        <p>To address this issue, we used data from several canonical sources. Our primary source was the ICD-O-3 standards [<xref ref-type="bibr" rid="ref62">62</xref>], which we used to define the primary <italic>preferred</italic> terminology for each code. Within the ICD-O-3 standards there are 332 unique site codes and 1143 unique histology codes, each with accompanying preferred terms. Along with the preferred term, many codes also have an additional set of synonyms, which we stored together with the preferred term in a table of acceptable phrases for each code. In addition to the ICD-O-3 tables, we also used terminology from the National Cancer Institute’s Surveillance, Epidemiology, and End Results (SEER) program Site/Histology Validation List [<xref ref-type="bibr" rid="ref62">62</xref>], along with the SEER Site-specific training module website [<xref ref-type="bibr" rid="ref63">63</xref>].</p>
        <p>To provide a little more detail, the specific sources we used to construct our acceptable phrase tables were as follows. For histology, we used the ICD-O-3.2 morphology table (version 15112019) [<xref ref-type="bibr" rid="ref64">64</xref>] and supplemented this with terms from the SEER Site/Histology Validation List (version 20150918), current versions of which are both available in Microsoft Excel format from their respective websites. For the site terms, we used the ICD-O-3 mapping table maintained by the National Cancer Institute [<xref ref-type="bibr" rid="ref65">65</xref>], supplemented again by the SEER Site/Histology Validation List. In addition, for the site terms we also scraped the tables contained in the SEER Site-specific learning module website for any new terms.</p>
        <p>The aforementioned sources have the benefit of being subject to an international standard and are useful in designating preferred terms for each histology and site code. However, we should note that there do exist slight discrepancies between the World Health Organization–maintained ICD-O-3 coding standards and the North American Association of Central Cancer Registries coding guidelines, which are followed in the SEER materials. For simplicity, we chose to base our model on the ICD-O-3 standards, but this caveat may prove relevant for any future cancer registry applications of the model.</p>
        <p>Although these sources provided us with preferred and alternative terminologies, they did not encompass the full range of language used for every label in our pathology reports, which often included things such as permutations of word orderings as well as acronyms and other typographical differences with the canonical terms. Of note, Moffitt CTRs routinely record a short description of the histology and site for every labeled pathology report in a text-based field. For each histology and site code, we appended these additional phrases to the list of synonyms of the preferred canonical terminology.</p>
        <p>Using the sources described earlier, we created two hierarchical tree structures as illustrated in Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>: one to hold histology terms and one to hold site terms. To construct these trees, the histology and site codes were first grouped into broad morphology and site groups as specified in the ICD-O-3 tables. Within each group are a collection of specific codes, where each code has an associated preferred term, and a list of synonyms. For efficient searching, these trees were stored as JSON objects that were imported into Python as nested dictionaries and lists. See Figure S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for an example of an entry in our acceptable phrase table.</p>
        <p>To search each pathology report for appropriate spans of text, we used the trees to construct a dictionary with keys provided by the specific site and histology codes and values provided by the associated acceptable phrase table. Using this dictionary, for each pathology report we implemented a simple search for an exact match from the list of preferred terms and synonyms for the labeled ground truth histology and site code, giving preference for the preferred term, followed by each synonym ordered by length (with the longest matching synonym given preference over the others).</p>
        <p>Even with the diverse terminology within the acceptable phrase table for each code, not every pathology report contained an exact match within the list of allowed terms. For pathology reports that did not contain an exact match, we further refined the search by allowing for matches that only overlapped with a subset of the word tokens within each phrase, again giving preference to the longest synonyms and also using a set of stop terms to avoid overly general terminology. To capture potential word-ordering differences, we allowed these word token subsets to be constructed in an arbitrarily permuted order, which was made efficient by using the <italic>itertools</italic> module available as part of the Python Standard Library.</p>
        <p>Using the aforementioned procedure, of the 10,247 pathology reports in the Moffitt fine-tuning training and testing data sets, we were able to find appropriate textual answers within 10,096 (98.53%) reports for primary site (with n=8070, 79.93%, in the training set and n=2026, 20.07%, in the testing set) and within 10,218 (99.72%) reports for histology (with n=8176, 80.02%, in the training set and n=2042, 19.98%, in the testing set).</p>
      </sec>
      <sec>
        <title>Provide Accurate ICD-O-3 Codes</title>
        <sec>
          <title>Overview</title>
          <p>The tumor site and histology phrases extracted by the Q&#38;A model (<xref rid="figure2" ref-type="fig">Figure 2</xref>) were used to predict ICD-O-3 site and histology codes. This involved fine-tuning two additional copies of caBERT: one to predict site codes and the second to predict histology codes. Our final system (<xref rid="figure4" ref-type="fig">Figure 4</xref>) includes a network of 3 caBERT-based models. We refer to this system as caBERTnet.</p>
        </sec>
        <sec>
          <title>Training the ICD-O-3 Site and Histology Code Classifiers</title>
          <p>Classifier training parameters are described in more detail in Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. Briefly, each caBERT instance was trained to perform a classification task: given an input phrase, predict the corresponding ICD-O-3 code. Classification tasks were trained using the Moffitt fine-tuning training data set (<italic>Extract Tumor Site and Histology Descriptions</italic> section). Training samples were screened to ensure that each contained ground truth site and histology codes and at least one site or histology phrase provided by the MCR. Missing site and histology phrases were filled using SEER preferred terms. These were identified by performing a lookup into the ICD-O-3 table using the site or histology code in the training sample.</p>
          <p>After screening, the ground truth phrases were labeled and concatenated to form a single combined phrase. For example, if the MCR phrases were <italic>lung lower lobe</italic> and <italic>squamous cell carcinoma,</italic> then the combined phrase would be <italic>site: lung lower lobe. histology: squamous cell carcinoma</italic>. The combined phrase was used to train both the caBERT site classification model and the caBERT histology classification model. The use of a combined phrase allowed caBERTnet to leverage any correlation between site and histology to improve its performance. For example, astrocytomas are brain tumors. When caBERTnet encountered a previously unseen pathology report during the test phase with the combined phrase <italic>site: frontal. histology: anaplastic astrocytoma</italic>, it correctly predicted a brain site of <italic>C711, frontal lobe</italic> and a histology of <italic>9401/3 astrocytoma anaplastic NOS</italic> (not otherwise specified).</p>
        </sec>
        <sec>
          <title>Testing the ICD-O-3 Site and Histology Code Classifiers</title>
          <p>After training of the site and histology ICD-O-3 code classification models was complete, caBERTnet performance was evaluated using the sequestered Moffitt fine-tuning test data set described in the <italic>Extract Tumor Site and Histology Descriptions</italic> section. For each test sample, the MCR-generated site and histology phrases were used to create a <italic>ground truth</italic> combined phrase. Next, the site and histology phrases extracted by the Q&#38;A stage of caBERTnet were used to create a <italic>predicted</italic> combined phrase. The predicted phrase was tokenized to prepare it for input into each classification model. Ground truth site and histology codes from the MCR were enumerated, as described earlier, and stored as true labels. Subsequently, the trained site and histology classification models were used to classify the tokenized <italic>predicted</italic> combined phrases for each test sample. The outputs from this classification, logits, were converted into probabilities, sorted, and converted back into ICD-O-3 codes as described earlier, labeled as <italic>predicted</italic> codes, and saved for further performance analysis.</p>
          <p>caBERTnet performance was evaluated in 3 different ways. First, the top 5 accuracies were determined. This metric (or its inverse, the top 5 error rate) is commonly used to evaluate classification algorithms [<xref ref-type="bibr" rid="ref66">66</xref>]. Briefly, it calculates the average probability that the correct site or histology code occurs within the top N predicted codes because N is varied from 1 through 5. Top 1 accuracy, the accuracy of the code scored most highly by the classification algorithm, is equivalent to precision, recall, and F1 score for this classification task.</p>
          <p>Second, we examined the effect of culling or removing infrequently occurring codes. Our hypothesis was that the caBERT site and histology code classifiers suffer when they do not have enough training data to learn from. Therefore, to examine the effect of training sample size, we iteratively eliminated site and histology codes from the full 2050-sample test data set when the number of examples with a particular code in the training set (alone) fell below a specified threshold. We varied that threshold from 0 samples (no culling) to 35 samples in increments of 5 samples. At each culling threshold we recalculated the top 5 performance of the site and histology classifiers.</p>
          <p>Third, we calculated the overall accuracy of predicting the correct <italic>group</italic> code for each site and histology code. <italic>Group</italic> codes occur higher up in the ontological tree and, as the name implies, encompass a group or range of related tumor sites or histologies. For example, the site codes <italic>C341 upper lobe, lung</italic> and <italic>C349 lung, NOS</italic> have the same group code: <italic>C34 bronchus and lung</italic>. The histology codes <italic>8070/3 squamous cell carcinoma, NOS</italic> and <italic>8051/0 verrucous carcinoma</italic> both have the same group code: <italic>805-808 squamous cell neoplasms</italic>. The ICD-O-3 ontology includes 82 group-level site codes covering the 332 fine-grained site codes. It includes 49 group-level histology codes covering the 1143 fine-grained histology codes.</p>
          <p>Group codes are useful for search and summary applications. The group codes for both the predicted and ground truth fine-grained codes were determined by searching in the tree data structures described in the <italic>Accommodate Diverse Terminology</italic> section. For each fine-grained code, the search started at that code’s location in the tree and proceeded upward. Finally, we calculated the overall accuracy of prediction within each group code for both site and histology predictions.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Model Accuracy</title>
        <p>The accuracy of the both the ClinicalBERT and caBERT Q&#38;A models when tested on the Moffitt fine-tuning testing data set improved at each Q&#38;A training stage (SQuAD, BioASQ, and Moffitt training data; Figure S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). ClinicalBERT had higher accuracy than caBERT on the Moffitt test set after each of the first 2 training stages. This suggests that the specialized pathology-language tuning reduced caBERT’s ability to learn from the SQuAD and BioASQ training data sets. However, caBERT outperformed ClinicalBERT after training on Moffitt pathology reports. This was true both for exact match (3254/4068, 79.99%, for caBERT vs 3069/4068, 75.44%, for ClinicalBERT) and F1 score (87.76% for caBERT vs 84.85% for ClinicalBERT).</p>
        <p>The top N accuracy of predicting fine-grained site codes ranged from 71.58% (1456/2034; top 1) to 91.05% (1852/2034; top 5), without culling (<xref rid="figure5" ref-type="fig">Figure 5</xref>). The accuracy for predicting histology codes ranged from 83.87% (1706/2034; top 1) to 94.79% (1928/2034; top 5). Culling 6.39% (130/2034) of the test samples—those site and histology codes with &#60;5 samples in the training data set—improved accuracy for site code prediction to 73.84% (1406/1904; +2.26%; top 1) and 93.28% (1776/1904; +2.23%; top 5). The same culling improved the accuracy of histology code prediction to 85.29% (1624/1904; +1.42%; top 1) and 96.27% (1833/1904; +1.48%; top 5).</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>The effect of culling rare tumor sites and histologies on the top N accuracy of predicting fine-grained International Classification of Diseases for Oncology, Third Edition codes.</p>
          </caption>
          <graphic xlink:href="jmir_v24i3e27210_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Group-Level Code Predictions</title>
        <p>We also computed the accuracy of the more coarse–grained group-level code predictions by mapping each top 1 code prediction to its corresponding group in the ICD-O-3 ontology. The accuracy of predicting group-level site codes was 93.53% (1895/2026) overall (<xref rid="figure6" ref-type="fig">Figure 6</xref>). The ten most commonly represented sites—(1) breast, (2) skin, (3) lung and bronchus, (4) prostate gland, (5) corpus uteri, (6) thyroid gland, (7) kidney, (8) large intestine, (9) rectum, and (10) ovary—included 79.42% (1609/2026) of the test samples and had an accuracy of 97.7% (1572/1609). Accuracies &#60;80% were observed for connective and soft tissues (15/29, 52%, with 29/2026, 1.43%, of the samples), stomach (14/27, 52%, with 27/2026, 1.33%, of the samples), gallbladder and extrahepatic bile ducts (14/18, 78%, with 18/2026, 0.89%, of the samples), small intestine (10/13, 77%, with 13/2026, 0.64%, of the samples), retroperitoneum and peritoneum (7/11, 64%, with 11/2026, 0.54%, of the samples), and other (52/90, 58%, a collection of 27 sites totaling 90/2026, 4.44%, of the samples).</p>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Accuracy of predicting tumor site group codes from unstructured and previously unseen pathology reports on solid tumors, broken down to show performance within each site group. The overall accuracy over all site groups was 93.53% (1895/2026).</p>
          </caption>
          <graphic xlink:href="jmir_v24i3e27210_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>The accuracy of predicting group-level histology codes was 97.6% (1993/2026) overall (<xref rid="figure7" ref-type="fig">Figure 7</xref>). The ten most commonly represented histologies—(1) adenomas and adenocarcinomas, (2) ductal and lobular neoplasms, (3) nevi and melanomas, (4) squamous cell carcinomas, (5) cystic mucinous and serous neoplasms, (6) transitional cell papillomas and carcinomas, (7) gliomas, (8) epithelial neoplasms, (9) complex mixed and stromal neoplasms, and (10) lipomatous neoplasms—included 95.84% (1957/2042) of the test samples and had an accuracy of 98.06% (1919/1957). An accuracy of &#60;80% was observed for epithelial neoplasms only (14/18, 78%, with 18/2042, 0.88%, of the samples).</p>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>Accuracy of predicting tumor histology group codes from unstructured and previously unseen pathology reports on solid tumors, broken down to show performance within each histology group. The overall accuracy over all histology groups was 97.6% (1993/2042).</p>
          </caption>
          <graphic xlink:href="jmir_v24i3e27210_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Primary Contributions</title>
        <p>This work describes 3 primary contributions. First, we created caBERT: a BERT-based language model for comprehension of cancer pathology reports. We are aware of only 1 other attempt to create a pathology- and oncology-specific BERT language model [<xref ref-type="bibr" rid="ref45">45</xref>]. That study included 290,438 pathology reports created between 2005 and 2015 from a tertiary teaching hospital in the United States. However, only 8870 of these reports were cases involving cancer. Our study included 275,605 pathology reports from patients with cancer diagnosed or treated at Moffitt. The larger corpus of cancer-specific reports should help our system achieve higher performance levels with cancer-related NLP tasks. However, because we did not have access to the system described in the study by Ma et al [<xref ref-type="bibr" rid="ref45">45</xref>], a direct comparison with ours was not possible.</p>
        <p>Second, we created a new Q&#38;A system to extract tumor site and histology descriptions from free-text pathology reports. This is the first functional Q&#38;A system for extracting information from pathology reports that we are aware of. The Q&#38;A format has 2 important benefits. First, it provides a user-friendly interface to the information extraction system. Second, incorporation of additional questions into our system is straightforward. With appropriate ground truth–labeled training data, this should allow us to extract additional data fields from free-text pathology reports.</p>
        <p>Third, we created a new caBERT network, caBERTnet, to predict fine-grained ICD-O-3 site and histology codes using the answers extracted through the initial Q&#38;A component. There has been considerable prior work using NLP methods to predict ICD-O codes from pathology reports [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref67">67</xref>]. Here, we compare our results to 5 of the most highly cited recent publications in this area.</p>
      </sec>
      <sec>
        <title>Comparisons With Prior Work</title>
        <p>Much of the prior work has focused on a single anatomical site or histology. For example, Coden et al [<xref ref-type="bibr" rid="ref16">16</xref>] described a system to extract information on tumor site, histology, grade, lymph nodes, tumor size, and reporting date from free-text pathology reports of colon cancer. They achieved precision and recall values of between 0.95 and 0.98 for both site and histology ICD-O codes. Their system used a rule-based NLP pipeline, with a large number of controlling parameters that required extensive manual tuning to obtain optimal results. In contrast, BERT-based NLP systems can both discover and tune the steps of a traditional NLP pipeline automatically [<xref ref-type="bibr" rid="ref68">68</xref>]. This has significant advantages in terms of reduced effort; in addition, it allows these systems to be quickly retuned for data sets from other institutions or different applications through transfer learning [<xref ref-type="bibr" rid="ref69">69</xref>].</p>
        <p>The BERT system from Ma et al [<xref ref-type="bibr" rid="ref45">45</xref>], mentioned earlier, was used to extract information on 15 <italic>primary cancer sites</italic>, 6 <italic>cancer reasons</italic>, and 3 <italic>metastatic disease</italic> states. In all, 11 of their cancer sites corresponded to ICD-O-3 group-level site classifications (eg, breast, lung, or bronchus). The others were broader groupings (eg, colorectal, upper gastrointestinal, and head and neck). Of their cancer reasons, 4 corresponded to ICD-O-3 group-level histology classifications (eg, melanoma and soft tissue sarcoma). The remaining two cancer reasons were very broad groupings: carcinoma and blastoma. They achieved accuracies on the full test set of 96.7% and 98.5% for cancer site and cancer reason, respectively. However, they did not predict ICD-O-3 group or fine-grained codes.</p>
        <p>Nguyen et al [<xref ref-type="bibr" rid="ref70">70</xref>] developed a system to monitor Health Level Seven International electronic pathology reports from across the state of Queensland in Australia. Their system relied on business rules and symbolic reasoning using Systematized Nomenclature of Medicine codes. They tuned their system using 201 pathology reports and tested it on 220 unseen reports. They extracted 8 different cancer characteristics from these reports. These characteristics included ICD-O-3 site codes (both fine-grained, Cxxx, and group, Cxx) and histological type. Their data set included 66 sites and 94 histologies. They achieved F1 scores of 61.1%, 73.2% and 63.7% on fine-grained site codes, group site codes, and histology codes, respectively.</p>
        <p>Alawad et al [<xref ref-type="bibr" rid="ref67">67</xref>] developed a multistage system of deep convolutional neural networks to extract the primary site, histological grade, and laterality from pathology reports. They achieved an F1 score of 77.5% over 12 ICD-O-3 site codes.</p>
        <p>Qiu et al [<xref ref-type="bibr" rid="ref11">11</xref>] also developed a deep convolutional neural network to extract ICD-O-3 codes from breast and lung cancer pathology reports. Training was based on 942 pathology reports annotated by cancer registry experts. The data set included 7 breast sites and 5 lung sites. Of the 12 sites, 6 had at least 50 samples per code. The remaining 6 sites had 10 to 50 samples each. They evaluated their system using a 10-fold cross-validation. Their overall F1 score for predicting tumor sites across all 12 ICD-O-3 codes was 72.2%.</p>
        <p>Our study included a greater diversity of cancer cases than previous studies (214 site codes and 193 histology codes), while obtaining similar or better accuracy scores. Many of the site and histology codes in our training data set included ≤5 samples, whereas prior studies reported ≥10 training samples per code. Culling codes from the test set with ≤5 samples in the training set reduced the size of our test data set by 6.39% (130/2034; 130 pathology reports). However, this increased our top-1 accuracy on the test data to 73.84% (1406/1904; +2.26%) and 85.29% (1624/1904; +1.42%) for site and histology, respectively. Our system also ranks and reports the top 5 predictions for ICD-O-3 site and histology codes. This has useful clinical applications: often there is a degree of uncertainty or <italic>hedging</italic> in pathology reports [<xref ref-type="bibr" rid="ref16">16</xref>]. Listing the top 5 predicted codes could help to reduce this uncertainty. For example, an artificial intelligence–assisted abstraction system that provides the top 5 predicted ICD-O-3 codes for a particular pathology report (in a pull-down menu, for example) could aid the process of abstraction and enhance the workflow in cancer registries. Our top 5 accuracies for fine-grained codes with ≥5 training samples were 92.95% (1794/1930) and 96.01% (1853/1930) for site and histology, respectively.</p>
      </sec>
      <sec>
        <title>Additional Insights From the Results</title>
        <p><xref rid="figure5" ref-type="fig">Figure 5</xref> shows the top 5 results at various levels of rare-code elimination from the test data set, and it provides 3 additional insights. First, as N increased from 1 to 5, the improvement in accuracy for sites was larger than that for histologies. This suggests that there is more uncertainty predicting site codes than in predicting histology codes. Second, eliminating rare codes, for example, going from E=0 (green lines) to E=5 (blue lines; <xref rid="figure5" ref-type="fig">Figure 5</xref>), improved site accuracy more than it improved histology accuracy. This suggests that site prediction was more dependent on sample size. Third, site accuracy failed to improve for E&#62;20. This suggests that 20 samples per code were required to maximize site code prediction accuracy.</p>
        <p>The overall accuracy for predicting site group codes was 93.53% (1895/2026) (<xref rid="figure6" ref-type="fig">Figure 6</xref>). Nevertheless, several site group codes had accuracies &#60;80%. Here, we will discuss the <italic>Other</italic> group (52/90, 58%, accuracy), along with two of the site group codes with the lowest accuracies: <italic>C49 Connective, Subcutaneous, and Other Soft Tissues</italic> (15/29, 52%, accuracy) and <italic>C16 Stomach</italic> (14/27, 52%, accuracy).</p>
        <p>The <italic>Other</italic> site category included 27 group codes. Together, these group codes contained 61 fine-grained codes with at least one sample pathology report each in the training data set, as determined by the MCR. The mean and median number of reports in the training data set for each fine-grained code in the <italic>Other</italic> category were 6.3 (SD 6.3) and 4 (IQR 5), respectively. Consequently, caBERTnet accuracy on these rare sites was likely limited by the availability of training data.</p>
        <p>caBERTnet failed to predict the MCR site code for 14 test cases in the group <italic>C49 Connective, Subcutaneous, and Other Soft Tissues</italic>. We manually reviewed 50% (7/14) of these cases, all of which were labeled by the MCR as soft tissue of the limb, shoulder, and hip or pelvis (codes <italic>C491</italic>, <italic>C492</italic>, and <italic>C495</italic>). In 14% (1/7) of these cases, the information required to determine the correct site was not present in the pathology report text. In these situations, CTRs would use additional information in the patient record. However, this information was not available to caBERTnet. In the remaining 86% (6/7) of the cases, the pathology report described characteristics of a lesion that had metastasized from the limb, shoulder, hip, or pelvis to another location. The MCR recorded the originating organ as the tumor site, whereas caBERTnet predicted the metastasis site.</p>
        <p>caBERTnet failed to predict the MCR site code for 13 test cases in the group <italic>C16 Stomach</italic>. All these cases were labeled by the MCR as <italic>C160 Cardia, NOS</italic>, and by caBERTnet as lesions of the lower third of the esophagus (<italic>C155</italic>; 12/13, 92%) or as overlapping lesions of the esophagus (<italic>C158</italic>; 1/13, 8%). The MCR labels are due to a rule in the American Joint Committee on Cancer Staging Manual, Eighth Edition [<xref ref-type="bibr" rid="ref71">71</xref>]. On page 189 in that manual it states as follows:</p>
        <disp-quote>
          <p>Cancers involving the Esophagogastric Junction (EGJ) that have their epicenter within the proximal 2 cm of the cardia (Siewert types I/II) are to be staged as esophageal cancers. Cancers whose epicenter is more than 2 cm distal from the EGJ, even if the EGJ is involved, will be staged using the stomach cancer TNM (primary tumor, lymph nodes, and distant metastases) and stage groupings (see Chapter 17).</p>
        </disp-quote>
        <p>The pathology reports on these cases did not mention the spatial location of the tumor sample in relation to the EGJ. Consequently, measurement of the tumor location in pretreatment imaging was required to determine the correct tumor site code.</p>
        <p>The overall accuracy for predicting histology group codes was 97.7% (<xref rid="figure7" ref-type="fig">Figure 7</xref>). Only one group code had an accuracy &#60;80%: <italic>801-804 Epithelial Neoplasms, NOS</italic> (77.8%). caBERTnet failed to predict the MCR histology code for 5 of these cases. In 80% (4/5) of these cases, the pathology report was based on histology at the metastatic site of disease. The MCR coded these as the histology of the originating tumor, whereas caBERTnet predicted the histology at the metastatic site. In 20% (1/5) of the cases, the information required to determine the correct histology code was not present in the pathology report and required the CTR to conduct a review of the patient medical record.</p>
        <p>The last case was quite interesting because the pathology report included an initial intraoperative diagnosis that disagreed with the final diagnosis. The former indicated a histological type of <italic>spindle cell carcinoma</italic>. The latter included the following statements: “the differential diagnosis includes sarcomatoid carcinoma and inflammatory myofibroblastic tumor...the histomorphologic and immunoprofile support the diagnosis of sarcomatoid carcinoma.” The MCR coded the histology as <italic>8032/3 spindle cell carcinoma, NOS</italic>, based on the intraoperative statements, whereas caBERTnet predicted <italic>8033/3 pseudosarcomatous carcinoma</italic>. The phrase <italic>sarcomatoid carcinoma</italic> is an alternative form of the ICD-O-3 preferred phrase <italic>pseudosarcomatous carcinoma</italic>. Although caBERTnet’s prediction did not agree with that of the MCR, downstream applications may still value automatic prediction and codification of the final diagnosis.</p>
      </sec>
      <sec>
        <title>Potential Applications</title>
        <p>There are multiple potential applications of caBERTnet at Moffitt. For example, there is a delay of several months between initial pathology report dictation and CTR abstraction because the CTRs typically wait for enough time to have elapsed for the first course treatment to have been administered to minimize the number of times they have to review the medical record. caBERTnet can be used to extract information from pathology reports in a timelier way, thus facilitating the use of the data for clinical pathway reporting and screening for clinical trials. Furthermore, CTRs only abstract the subset of pathology reports associated with the cancer diagnosis and first-course treatment. caBERTnet could be used to extract information from pathology reports associated with subsequent biopsies and surgeries that would never be manually curated by the CTRs. To facilitate these use cases, we plan to extract tumor site and histology information close to real time and link these values to other patient data stored in our analytics platform. These data can be incorporated into real-time dashboards and data sets for a wide range of decision support and research applications.</p>
        <p>We do not believe that caBERTnet will replace CTRs at cancer clinics. Many complex, difficult, and rare cases require intuition and information outside of the pathology report to determine the correct coding. These cases are beyond the scope of an NLP tool. However, caBERTnet may help simplify and accelerate MCR workflows. For example, caBERTnet could preprocess pathology reports to identify the top 5 site and histology ICD-O-3 codes and their corresponding phrases. The phrases could then be highlighted within the report body. In addition, two pull-down menus could be prepopulated with top 5 code predictions: one for site and the other for histology. The CTR could then quickly choose a code from either pull-down menu. If the correct code was not among the top 5, then the CTR would resort to their current workflow, entering this information by hand.</p>
      </sec>
      <sec>
        <title>Limitations of Approach</title>
        <p>Although there are immediate applications of the caBERTnet model within our internal workflows, there are a number of aspects of our modeling approach that limit the application of caBERTnet to other use cases. In this section, we provide an overview of several important limitations that we believe should be considered before caBERTnet implementation.</p>
        <p>The most critical limitations correspond to issues related to our data curation and preprocessing approaches. In the curation of our Moffitt fine-tuning data set, we restricted the available reports to only those containing a single primary tumor diagnosis. Although this was partly imposed by the nature of the SQuAD Q&#38;A task (which expects a single answer to each question asked of each input), it nonetheless limits the generalizability of our model to reports containing multiple (or zero) positive diagnoses. We are exploring methods of mitigating these issues within the current setup by adjusting the likelihood thresholds for output predictions that could be used to screen out reports with no diagnosis.
Another limitation inherent to a Q&#38;A system is the necessity of knowing the precise span of text corresponding to the answer to each question asked of each report. Owing to the sheer number of reports in our fine-tuning data set, it was infeasible to manually curate answer labels. To circumvent this issue, we chose to create our own automated system to determine the answer text in each report. Any such automated preprocessing necessarily leaves a fingerprint on downstream tasks. The drawbacks of our approach in particular relate to the restriction of answers from a predetermined list of possibilities for each site and histology. Although these phrase sets were diverse (and our approach even allowed for permutations of phrasing within these terms), this process nevertheless limits the allowed terminology and is necessarily incomplete.</p>
        <p>In addition, a single pathology report may reference several tissue samples (which can be from related or distant sites from the actual site of diagnosis). Although we limited our automated answer-search preprocessing algorithm to only find terms associated with the known diagnosis label for each report, it is possible that the answer found in the report text by the preprocessing corresponds to a false answer from a different sample in the report; this is particularly relevant for cases with multiple samples from related sites (eg, upper outer quadrant breast vs upper inner quadrant breast). When the 2 samples are in the same site (or histology) group, this issue is avoided by outputting the group code; however, this does not help when the 2 samples belong to different groups altogether.</p>
      </sec>
      <sec>
        <title>Future Directions</title>
        <p>We plan to continue development of caBERTnet. Of particular interest is extending the system with additional questions and MCR-derived ground truth labels to train it to extract additional tumor characteristics. These include grade, size, involvement of lymph nodes, primary or metastatic status, presence or absence of molecular markers, and others. caBERTnet could also be customized to extract information on hematological malignancies.</p>
        <p>A caBERTnet-assisted MCR abstraction tool could also be used for active [<xref ref-type="bibr" rid="ref72">72</xref>] or human-in-the-loop [<xref ref-type="bibr" rid="ref73">73</xref>] learning. Briefly, this approach uses human-labeled data to improve the performance of machine learning algorithms over time. It is particularly useful when the subject matter expert (a CTR in our case) provides labels for cases with low-confidence predictions by the machine learning algorithm. However, it would require careful engineering to avoid common pitfalls and ensure seamless operation [<xref ref-type="bibr" rid="ref74">74</xref>].</p>
        <p>The system’s accuracy on rare sites and histologies could be improved with additional training data. A potential option may be to collaborate with other academic cancer centers; distributing caBERTnet for training on local pathology reports at other sites based on ground truth labeling of training and test data sets from highly standardized registry data or the application of federated multitask learning [<xref ref-type="bibr" rid="ref75">75</xref>] that distributes copies of a central model to multiple spoke sites for tuning of the central model on local data could allow vast improvements in caBERTnet accuracy. Using these methods, information learned at local sites (eg, model weights) is transmitted back to the central node where the information is combined in a pluralistic way that avoids the need to impose consensus on the data distributions at the spoke sites. This allows for both heterogeneity in local data and broad generalizability of the central model. That said, the most effective way to protect private health information when sharing such models remains an unsolved problem, and these kinds of expansions would depend on the development of validated privacy schemes specific to BERT.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>Our new NLP system, caBERTnet, is built around a network of 3 cooperating BERT instances. On a sequestered test data set, it produced top 5 accuracies of 92.95% (1794/1930) and 96.01% (1853/1930) for fine-grained ICD-O-3 site and histology codes, respectively. This level of accuracy is on par with existing systems in the literature, while also being accurate over a broader range of site and histology groups.</p>
        <p>Pathology report abstraction systems such as caBERTnet cannot be expected to achieve performance on par with CTRs who abstract data ultimately incorporated into population-based cancer registries [<xref ref-type="bibr" rid="ref76">76</xref>], given the vast amount of ancillary data from within the electronic health record that is required for thorough abstraction and the subtle nuances associated with coding guidelines. However, caBERTnet could expedite access to timely pathology data needed for disease surveillance, cohort identification, and clinical trial matching. Furthermore, it can improve existing workflows, serving as a valuable step toward the ultimate goal of a mostly automated abstraction system.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Additional information on the CancerBERT and CancerBERT network model training and performance as well as an elaboration on the acceptable phrases used to curate labels for pathology reports in the question-and-answer stage.</p>
        <media xlink:href="jmir_v24i3e27210_app1.docx" xlink:title="DOCX File , 360 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">BERT</term>
          <def>
            <p>bidirectional encoder representations from transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BioASQ</term>
          <def>
            <p>Biomedical Semantic Indexing and Question Answering</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">caBERTnet</term>
          <def>
            <p>CancerBERT network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CTR</term>
          <def>
            <p>certified tumor registrar</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">EGJ</term>
          <def>
            <p>esophagogastric junction</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">EMR</term>
          <def>
            <p>electronic medical record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">ICD-O-3</term>
          <def>
            <p>International Classification of Diseases for Oncology, Third Edition</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">MCR</term>
          <def>
            <p>Moffitt Cancer Registry</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">MIMIC-III</term>
          <def>
            <p>Medical Information Mart for Intensive Care, version 3</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">NOS</term>
          <def>
            <p>not otherwise specified</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">Q&#38;A</term>
          <def>
            <p>question and answer</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">SEER</term>
          <def>
            <p>Surveillance, Epidemiology, and End Results</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">SQuAD</term>
          <def>
            <p>Stanford Question Answering Dataset</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Murdoch</surname>
              <given-names>TB</given-names>
            </name>
            <name name-style="western">
              <surname>Detsky</surname>
              <given-names>AS</given-names>
            </name>
          </person-group>
          <article-title>The inevitable application of big data to health care</article-title>
          <source>JAMA</source>
          <year>2013</year>
          <month>04</month>
          <day>03</day>
          <volume>309</volume>
          <issue>13</issue>
          <fpage>1351</fpage>
          <lpage>2</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.2013.393</pub-id>
          <pub-id pub-id-type="medline">23549579</pub-id>
          <pub-id pub-id-type="pii">1674245</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pratt</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thomas</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>An information processing system for pathology data</article-title>
          <source>Pathol Annul</source>
          <year>1966</year>
          <volume>1</volume>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dunham</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Pacak</surname>
              <given-names>MG</given-names>
            </name>
            <name name-style="western">
              <surname>Pratt</surname>
              <given-names>AW</given-names>
            </name>
          </person-group>
          <article-title>Automatic indexing of pathology data</article-title>
          <source>J Am Soc Inf Sci</source>
          <year>1978</year>
          <month>03</month>
          <volume>29</volume>
          <issue>2</issue>
          <fpage>81</fpage>
          <lpage>90</lpage>
          <pub-id pub-id-type="doi">10.1002/asi.4630290207</pub-id>
          <pub-id pub-id-type="medline">10318395</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Meystre</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Haug</surname>
              <given-names>PJ</given-names>
            </name>
          </person-group>
          <article-title>Natural language processing to extract medical problems from electronic clinical documents: performance evaluation</article-title>
          <source>J Biomed Inform</source>
          <year>2006</year>
          <month>12</month>
          <volume>39</volume>
          <issue>6</issue>
          <fpage>589</fpage>
          <lpage>99</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(05)00114-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2005.11.004</pub-id>
          <pub-id pub-id-type="medline">16359928</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(05)00114-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Murff</surname>
              <given-names>HJ</given-names>
            </name>
            <name name-style="western">
              <surname>FitzHenry</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Matheny</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Gentry</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Kotter</surname>
              <given-names>KL</given-names>
            </name>
            <name name-style="western">
              <surname>Crimin</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Dittus</surname>
              <given-names>RS</given-names>
            </name>
            <name name-style="western">
              <surname>Rosen</surname>
              <given-names>AK</given-names>
            </name>
            <name name-style="western">
              <surname>Elkin</surname>
              <given-names>PL</given-names>
            </name>
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>SH</given-names>
            </name>
            <name name-style="western">
              <surname>Speroff</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Automated identification of postoperative complications within an electronic medical record using natural language processing</article-title>
          <source>JAMA</source>
          <year>2011</year>
          <month>08</month>
          <day>24</day>
          <volume>306</volume>
          <issue>8</issue>
          <fpage>848</fpage>
          <lpage>55</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.2011.1204</pub-id>
          <pub-id pub-id-type="medline">21862746</pub-id>
          <pub-id pub-id-type="pii">306/8/848</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Datta</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Si</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Soni</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Xiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Deep learning in clinical natural language processing: a methodical review</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2020</year>
          <month>03</month>
          <day>01</day>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>457</fpage>
          <lpage>70</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/31794016"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocz200</pub-id>
          <pub-id pub-id-type="medline">31794016</pub-id>
          <pub-id pub-id-type="pii">5651084</pub-id>
          <pub-id pub-id-type="pmcid">PMC7025365</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Datta</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bernstam</surname>
              <given-names>EV</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>A frame semantic overview of NLP-based information extraction for cancer-related EHR notes</article-title>
          <source>J Biomed Inform</source>
          <year>2019</year>
          <month>12</month>
          <volume>100</volume>
          <fpage>103301</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(19)30221-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2019.103301</pub-id>
          <pub-id pub-id-type="medline">31589927</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(19)30221-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Napolitano</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Fox</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Middleton</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Connolly</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Pattern-based information extraction from pathology reports for cancer registration</article-title>
          <source>Cancer Causes Control</source>
          <year>2010</year>
          <month>11</month>
          <volume>21</volume>
          <issue>11</issue>
          <fpage>1887</fpage>
          <lpage>94</lpage>
          <pub-id pub-id-type="doi">10.1007/s10552-010-9616-4</pub-id>
          <pub-id pub-id-type="medline">20652738</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soysal</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Warner</surname>
              <given-names>JL</given-names>
            </name>
            <name name-style="western">
              <surname>Denny</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Identifying metastases-related information from pathology reports of lung cancer patients</article-title>
          <source>AMIA Jt Summits Transl Sci Proc</source>
          <year>2017</year>
          <volume>2017</volume>
          <fpage>268</fpage>
          <lpage>77</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/28815141"/>
          </comment>
          <pub-id pub-id-type="medline">28815141</pub-id>
          <pub-id pub-id-type="pmcid">PMC5543353</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>H-J</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>S-B</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>S-H</given-names>
            </name>
            <name name-style="western">
              <surname>Seo</surname>
              <given-names>J-W</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Choi</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Automated extraction of Biomarker information from pathology reports</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2018</year>
          <month>05</month>
          <day>21</day>
          <volume>18</volume>
          <issue>1</issue>
          <fpage>29</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-018-0609-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12911-018-0609-7</pub-id>
          <pub-id pub-id-type="medline">29783980</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12911-018-0609-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC5963015</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Fearn</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Tourassi</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Deep learning for automated extraction of primary sites from cancer pathology reports</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2018</year>
          <month>01</month>
          <volume>22</volume>
          <issue>1</issue>
          <fpage>244</fpage>
          <lpage>51</lpage>
          <pub-id pub-id-type="doi">10.1109/JBHI.2017.2700722</pub-id>
          <pub-id pub-id-type="medline">28475069</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Young</surname>
              <given-names>MT</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>JX</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>H-J</given-names>
            </name>
            <name name-style="western">
              <surname>Christian</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Fearn</surname>
              <given-names>PA</given-names>
            </name>
            <name name-style="western">
              <surname>Tourassi</surname>
              <given-names>GD</given-names>
            </name>
            <name name-style="western">
              <surname>Ramanthan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Hierarchical attention networks for information extraction from cancer pathology reports</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2018</year>
          <month>03</month>
          <day>01</day>
          <volume>25</volume>
          <issue>3</issue>
          <fpage>321</fpage>
          <lpage>30</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29155996"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocx131</pub-id>
          <pub-id pub-id-type="medline">29155996</pub-id>
          <pub-id pub-id-type="pii">4636780</pub-id>
          <pub-id pub-id-type="pmcid">PMC7282502</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pattison</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Denhalter</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Patterson</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>DuVall</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Seigne</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sirovich</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Goodney</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Robertson</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Schroeck</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>MP31-14 leveraging bladder cancer pathology reports for research: gleaning meaning despite widely variable language</article-title>
          <source>J Urol</source>
          <year>2016</year>
          <month>04</month>
          <volume>195</volume>
          <issue>4S</issue>
          <fpage>e424</fpage>
          <lpage>5</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.juro.2016.02.1267"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.juro.2016.02.1267</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Buckley</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Coopey</surname>
              <given-names>SB</given-names>
            </name>
            <name name-style="western">
              <surname>Sharko</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Polubriaginof</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Drohan</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Belli</surname>
              <given-names>AK</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Garber</surname>
              <given-names>JE</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>BL</given-names>
            </name>
            <name name-style="western">
              <surname>Gadd</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Specht</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Roche</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Gudewicz</surname>
              <given-names>TM</given-names>
            </name>
            <name name-style="western">
              <surname>Hughes</surname>
              <given-names>KS</given-names>
            </name>
          </person-group>
          <article-title>The feasibility of using natural language processing to extract clinical information from breast pathology reports</article-title>
          <source>J Pathol Inform</source>
          <year>2012</year>
          <volume>3</volume>
          <fpage>23</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jpathinformatics.org/article.asp?issn=2153-3539;year=2012;volume=3;issue=1;spage=23;epage=23;aulast=Buckley"/>
          </comment>
          <pub-id pub-id-type="doi">10.4103/2153-3539.97788</pub-id>
          <pub-id pub-id-type="medline">22934236</pub-id>
          <pub-id pub-id-type="pii">JPI-3-23</pub-id>
          <pub-id pub-id-type="pmcid">PMC3424662</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Ouyang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Griffin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Taghian</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yala</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Barzilay</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Hughes</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Machine learning to parse breast pathology reports in Chinese</article-title>
          <source>Breast Cancer Res Treat</source>
          <year>2018</year>
          <month>06</month>
          <volume>169</volume>
          <issue>2</issue>
          <fpage>243</fpage>
          <lpage>50</lpage>
          <pub-id pub-id-type="doi">10.1007/s10549-018-4668-3</pub-id>
          <pub-id pub-id-type="medline">29380208</pub-id>
          <pub-id pub-id-type="pii">10.1007/s10549-018-4668-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Coden</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Savova</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Sominsky</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Tanenblatt</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Masanz</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Schuler</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Cooper</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Guan</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>de Groen</surname>
              <given-names>PC</given-names>
            </name>
          </person-group>
          <article-title>Automatically extracting cancer disease characteristics from pathology reports into a Disease Knowledge Representation Model</article-title>
          <source>J Biomed Inform</source>
          <year>2009</year>
          <month>10</month>
          <volume>42</volume>
          <issue>5</issue>
          <fpage>937</fpage>
          <lpage>49</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(08)00158-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2008.12.005</pub-id>
          <pub-id pub-id-type="medline">19135551</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(08)00158-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Structured pathology reporting for cancer from free text: lung cancer case study</article-title>
          <source>Electron J Health Informatics</source>
          <year>2012</year>
          <volume>7</volume>
          <issue>1</issue>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/TELSKS.2001.955803.S.Datta,etal.Journal"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>H-J</given-names>
            </name>
            <name name-style="western">
              <surname>Robinson</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Christian</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>JX</given-names>
            </name>
            <name name-style="western">
              <surname>Tourassi</surname>
              <given-names>GD</given-names>
            </name>
          </person-group>
          <article-title>Filter pruning of Convolutional Neural Networks for text classification: a case study of cancer pathology report comprehension</article-title>
          <source>Proceedings of the 2018 IEEE EMBS International Conference on Biomedical &#38; Health Informatics (BHI)</source>
          <year>2018</year>
          <conf-name>2018 IEEE EMBS International Conference on Biomedical &#38; Health Informatics (BHI)</conf-name>
          <conf-date>Mar 4-7,  2018</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bhi.2018.8333439</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schroeck</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Patterson</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Alba</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>DuVall</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sirovich</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Robertson</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Seigne</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Goodney</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>MP32-12 harnessing full text pathology data from the electronic health record to advance bladder cancer care – development of a natural language processing system to generate longitudinal pathology data</article-title>
          <source>J Urol</source>
          <year>2017</year>
          <month>04</month>
          <volume>197</volume>
          <issue>4S</issue>
          <fpage>e413</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.juro.2017.02.987"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.juro.2017.02.987</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schadow</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>McDonald</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Extracting structured information from free text pathology reports</article-title>
          <source>AMIA Annu Symp Proc</source>
          <year>2003</year>
          <fpage>584</fpage>
          <lpage>8</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/14728240"/>
          </comment>
          <pub-id pub-id-type="medline">14728240</pub-id>
          <pub-id pub-id-type="pii">D030003509</pub-id>
          <pub-id pub-id-type="pmcid">PMC1480213</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Burger</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Abu-Hanna</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>de Keizer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Cornet</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Natural language processing in pathology: a scoping review</article-title>
          <source>J Clin Pathol</source>
          <year>2016</year>
          <month>07</month>
          <day>22</day>
          <fpage>jclinpath-2016-203872</fpage>
          <pub-id pub-id-type="doi">10.1136/jclinpath-2016-203872</pub-id>
          <pub-id pub-id-type="medline">27451435</pub-id>
          <pub-id pub-id-type="pii">jclinpath-2016-203872</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Névéol</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zweigenbaum</surname>
              <given-names>P</given-names>
            </name>
            <collab>Section Editors for the IMIA Yearbook Section on Clinical Natural Language Processing</collab>
          </person-group>
          <article-title>Expanding the diversity of texts and applications: findings from the section on clinical natural language processing of the international medical informatics association yearbook</article-title>
          <source>Yearb Med Inform</source>
          <year>2018</year>
          <month>08</month>
          <volume>27</volume>
          <issue>1</issue>
          <fpage>193</fpage>
          <lpage>8</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.thieme-connect.com/DOI/DOI?10.1055/s-0038-1667080"/>
          </comment>
          <pub-id pub-id-type="doi">10.1055/s-0038-1667080</pub-id>
          <pub-id pub-id-type="medline">30157523</pub-id>
          <pub-id pub-id-type="pmcid">PMC6115241</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          <source>Proceedings of NAACL-HLT 2019</source>
          <year>2019</year>
          <conf-name>Proceedings of NAACL-HLT 2019</conf-name>
          <conf-date>Jun 2 - 7, 2019</conf-date>
          <conf-loc>Minneapolis, Minnesota</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://aclanthology.org/N19-1423.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Michael</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hill</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Levy</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Bowman</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>GLUE: a multi-task benchmark and analysis platform for natural language understanding</article-title>
          <source>Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP</source>
          <year>2018</year>
          <conf-name>Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP</conf-name>
          <conf-date>Nov 1, 2018</conf-date>
          <conf-loc>Brussels, Belgium</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/w18-5446</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Radford</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Narasimhan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Salimans</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Improving language understanding by generative pre-training</article-title>
          <source>Glue Benchmark</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf">https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Radford</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Child</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Luan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Amodei</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Language models are unsupervised multitask learners</article-title>
          <source>OpenAI Blog</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.ceid.upatras.gr/webpages/faculty/zaro/teaching/alg-ds/PRESENTATIONS/PAPERS/2019-Radford-et-al_Language-Models-Are-Unsupervised-Multitask-%20Learners.pdf">https://www.ceid.upatras.gr/webpages/faculty/zaro/teaching/alg-ds/PRESENTATIONS/PAPERS/2019-Radford-et-al_Language-Models-Are-Unsupervised-Multitask-%20Learners.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Carbonell</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Transformer-XL: attentive language models beyond a fixed-length context</article-title>
          <source>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          <year>2019</year>
          <conf-name>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>Jul 28 - Aug 2, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://aclanthology.org/P19-1285.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/p19-1285</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Carbonell</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>XLNet: generalized autoregressive pretraining for language understanding</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online January 2, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1906.08237"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lan</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Goodman</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gimpel</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Sharma</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Soricut</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>ALBERT: a Lite BERT for self-supervised learning of language representations</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online February 9, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1909.11942"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ott</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Goyal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Joshi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>RoBERTa: a robustly optimized BERT pretraining approach</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online July 26, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1907.11692"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sanh</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Debut</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Chaumond</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wolf</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online March 1, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1910.01108"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lample</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Conneau</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Cross-lingual language model pretraining</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online January 22, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1901.07291"/>
          </comment>
          <pub-id pub-id-type="doi">10.1090/mbk/121/79</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raffel</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Narang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Matena</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online July 28, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1910.10683"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ju</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Technical report on conversational question answering</article-title>
          <source>ArXiv</source>
          <comment>Preprint posted online Septmber 29, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1909.10772"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Han</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chiu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Qin</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>ContextNet: improving convolutional neural networks for automatic speech recognition with global context</article-title>
          <source>Proceedings of the INTERSPEECH 2020</source>
          <year>2020</year>
          <conf-name>INTERSPEECH 2020</conf-name>
          <conf-date>Oct 25-29, 2020</conf-date>
          <conf-loc>Shanghai, China</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/2005.03191"/>
          </comment>
          <pub-id pub-id-type="doi">10.21437/interspeech.2020-2059</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Long</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>BiteNet: bidirectional temporal encoder network to predict medical outcomes</article-title>
          <source>Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM)</source>
          <year>2020</year>
          <conf-name>2020 IEEE International Conference on Data Mining (ICDM)</conf-name>
          <conf-date>Nov 17- 20, 2020</conf-date>
          <conf-loc>Sorrento, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icdm50108.2020.00050</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Rawat</surname>
              <given-names>BP</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Fine-tuning Bidirectional Encoder Representations From Transformers (BERT)-based models on large-scale electronic health record notes: an empirical study</article-title>
          <source>JMIR Med Inform</source>
          <year>2019</year>
          <month>09</month>
          <day>12</day>
          <volume>7</volume>
          <issue>3</issue>
          <fpage>e14830</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://medinform.jmir.org/2019/3/e14830/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/14830</pub-id>
          <pub-id pub-id-type="medline">31516126</pub-id>
          <pub-id pub-id-type="pii">v7i3e14830</pub-id>
          <pub-id pub-id-type="pmcid">PMC6746103</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Altosaar</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ranganath</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>ClinicalBERT: modeling clinical notes and predicting hospital readmission</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online November 29, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1904.05342"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tahayori</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Chini-Foroush</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Akhlaghi</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Advanced natural language processing technique to predict patient disposition based on emergency triage notes</article-title>
          <source>Emerg Med Australas</source>
          <year>2020</year>
          <month>10</month>
          <day>11</day>
          <comment>(forthcoming)</comment>
          <pub-id pub-id-type="doi">10.1111/1742-6723.13656</pub-id>
          <pub-id pub-id-type="medline">33043570</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Blinov</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Avetisian</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kokh</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Umerenkov</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Tuzhilin</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Predicting clinical diagnosis from patients electronic health records using BERT-based neural networks</article-title>
          <source>Artificial Intelligence in Medicine</source>
          <year>2020</year>
          <publisher-loc>Cham</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Gopale</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Begoli</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Bethard</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Unified Medical Language System resources improve sieve-based generation and Bidirectional Encoder Representations from Transformers (BERT)-based ranking for concept normalization</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2020</year>
          <month>10</month>
          <day>01</day>
          <volume>27</volume>
          <issue>10</issue>
          <fpage>1510</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/32719838"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocaa080</pub-id>
          <pub-id pub-id-type="medline">32719838</pub-id>
          <pub-id pub-id-type="pii">5876963</pub-id>
          <pub-id pub-id-type="pmcid">PMC7566510</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kandpal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Jasnani</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Raut</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bhorge</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Contextual chatbot for healthcare purposes (using Deep Learning)</article-title>
          <source>Proceedings of the 2020 Fourth World Conference on Smart Trends in Systems, Security and Sustainability (WorldS4)</source>
          <year>2020</year>
          <conf-name>2020 Fourth World Conference on Smart Trends in Systems, Security and Sustainability (WorldS4)</conf-name>
          <conf-date>Jul 27-28, 2020</conf-date>
          <conf-loc>London, UK</conf-loc>
          <pub-id pub-id-type="doi">10.1109/worlds450073.2020.9210351</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Fan</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rao</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Cong</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Sentiment analysis methods for HPV vaccines related tweets based on transfer learning</article-title>
          <source>Healthcare (Basel)</source>
          <year>2020</year>
          <month>08</month>
          <day>28</day>
          <volume>8</volume>
          <issue>3</issue>
          <fpage>307</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=healthcare8030307"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/healthcare8030307</pub-id>
          <pub-id pub-id-type="medline">32872330</pub-id>
          <pub-id pub-id-type="pii">healthcare8030307</pub-id>
          <pub-id pub-id-type="pmcid">PMC7551482</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Pre-training of graph augmented transformers for medication recommendation</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online November 27, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1906.00346"/>
          </comment>
          <pub-id pub-id-type="doi">10.24963/ijcai.2019/825</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gadepalli</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Human-centric metric for accelerating pathology reports annotation</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online November 12, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1911.01226"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Grishman</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Hirschman</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Question answering from natural language medical data bases</article-title>
          <source>Artificial Intelligence</source>
          <year>1978</year>
          <month>08</month>
          <volume>11</volume>
          <issue>1-2</issue>
          <fpage>25</fpage>
          <lpage>43</lpage>
          <pub-id pub-id-type="doi">10.1016/0004-3702(78)90011-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Paszke</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gross</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Massa</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Lerer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bradbury</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chanan</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Killeen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Gimelshein</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Antiga</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Desmaison</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Köpf</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>DeVito</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Raison</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tejani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Chilamkurthy</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Steiner</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bai</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chintala</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>PyTorch: an imperative style, high-performance deep learning library</article-title>
          <source>Proceedings of the 33rd International Conference on Neural Information Processing Systems</source>
          <year>2019</year>
          <conf-name>Proceedings of the 33rd International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>Dec 8-14, 2019</conf-date>
          <conf-loc>Vancouver, Canada</conf-loc>
          <pub-id pub-id-type="doi">10.5555/3454287.3455008</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wolf</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Debut</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sanh</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Chaumond</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Delangue</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Moi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cistac</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Rault</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Louf</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Funtowicz</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Davison</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shleifer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>von Platen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Jernite</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Plu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Scao</surname>
              <given-names>TL</given-names>
            </name>
            <name name-style="western">
              <surname>Gugger</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Drame</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lhoest</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Rush</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Transformers: state-of-the-art natural language processing</article-title>
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          <year>2020</year>
          <conf-name>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</conf-name>
          <conf-date>Nov 16-20, 2020</conf-date>
          <conf-loc>-</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/2020.emnlp-demos.6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jain</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Meenachi</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Venkatraman</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>NukeBERT: a pre-trained language model for low resource nuclear domain</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online August 19, 2020.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/2003.13821"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Neumann</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>King</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Beltagy</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Ammar</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>ScispaCy: fast and robust models for biomedical natural language processing</article-title>
          <source>Proceedings of the 18th BioNLP Workshop and Shared Task</source>
          <year>2019</year>
          <conf-name>Proceedings of the 18th BioNLP Workshop and Shared Task</conf-name>
          <conf-date>Aug 1, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/w19-5034</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaswani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Parmar</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Uszkoreit</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gomez</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Attention is all you need</article-title>
          <source>Proceedings of the 31st Conference on Neural Information Processing Systems</source>
          <year>2017</year>
          <conf-name>31st Conference on Neural Information Processing Systems</conf-name>
          <conf-date>Dec 4 - 9, 2017</conf-date>
          <conf-loc>Long Beach, CA, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Khandelwal</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Levy</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Manning</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>What does BERT look at? An Analysis of BERT’s Attention</article-title>
          <source>Proceedings of the Second BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP</source>
          <year>2019</year>
          <conf-name>Proceedings of the Second BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP</conf-name>
          <conf-date>Aug 1, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/w19-4828</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Sabuncu</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Generalized cross entropy loss for training deep neural networks with noisy labels</article-title>
          <source>Proceedings of the 32nd International Conference on Neural Information Processing Systems</source>
          <year>2018</year>
          <conf-name>Proceedings of the 32nd International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>Dec 3 - 8, 2018</conf-date>
          <conf-loc>Montréal Canada</conf-loc>
          <pub-id pub-id-type="doi">10.5555/3327546.3327555</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>So</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>BioBERT: a pre-trained biomedical language representation model for biomedical text mining</article-title>
          <source>Bioinformatics</source>
          <year>2020</year>
          <month>02</month>
          <day>15</day>
          <volume>36</volume>
          <issue>4</issue>
          <fpage>1234</fpage>
          <lpage>40</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/31501885"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/bioinformatics/btz682</pub-id>
          <pub-id pub-id-type="medline">31501885</pub-id>
          <pub-id pub-id-type="pii">5566506</pub-id>
          <pub-id pub-id-type="pmcid">PMC7703786</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alsentzer</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Boag</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Naumann</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Publicly available clinical BERT embeddings</article-title>
          <source>Proceedigs of the 2nd Clinical Natural Language Processing Workshop</source>
          <year>2019</year>
          <conf-name>2nd Clinical Natural Language Processing Workshop</conf-name>
          <conf-date>Jun 7, 2019</conf-date>
          <conf-loc>Minneapolis, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1904.03323"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/w19-1909</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>AE</given-names>
            </name>
            <name name-style="western">
              <surname>Pollard</surname>
              <given-names>TJ</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ghassemi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Celi</surname>
              <given-names>LA</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>MIMIC-III, a freely accessible critical care database</article-title>
          <source>Sci Data</source>
          <year>2016</year>
          <month>05</month>
          <day>24</day>
          <volume>3</volume>
          <fpage>160035</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/sdata.2016.35"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/sdata.2016.35</pub-id>
          <pub-id pub-id-type="medline">27219127</pub-id>
          <pub-id pub-id-type="pii">sdata201635</pub-id>
          <pub-id pub-id-type="pmcid">PMC4878278</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rajpurkar</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lopyrev</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>SQuAD: 100,000+ questions for machine comprehension of text</article-title>
          <source>Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          <year>2016</year>
          <conf-name>Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP)</conf-name>
          <conf-date>Nov 1–5, 2016</conf-date>
          <conf-loc>Austin, Texas, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1606.05250"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/d16-1264</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tsatsaronis</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Balikas</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Malakasiotis</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Partalas</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Zschunke</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Alvers</surname>
              <given-names>MR</given-names>
            </name>
            <name name-style="western">
              <surname>Weissenborn</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Krithara</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Petridis</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Polychronopoulos</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Almirantis</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Pavlopoulos</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Baskiotis</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Gallinari</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Artiéres</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Ngomo</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Heino</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Gaussier</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Barrio-Alvers</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Schroeder</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Androutsopoulos</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Paliouras</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>An overview of the BIOASQ large-scale biomedical semantic indexing and question answering competition</article-title>
          <source>BMC Bioinformatics</source>
          <year>2015</year>
          <month>04</month>
          <day>30</day>
          <volume>16</volume>
          <fpage>138</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcbioinformatics.biomedcentral.com/articles/10.1186/s12859-015-0564-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12859-015-0564-6</pub-id>
          <pub-id pub-id-type="medline">25925131</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12859-015-0564-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC4450488</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Pre-trained language model for biomedical question answering</article-title>
          <source>ArXiv.</source>
          <comment>Preprint posted online September 18, 2019.
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1909.08229"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/978-3-030-43887-6_64</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hawhee</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Stewart</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lucia</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Rollison</surname>
              <given-names>DE</given-names>
            </name>
          </person-group>
          <article-title>Quality assurance and continuing education: a cyclic approach for maintaining high quality data in a high volume cancer registry</article-title>
          <source>Cancer Control</source>
          <year>2020</year>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>1073274820946794</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://journals.sagepub.com/doi/10.1177/1073274820946794?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub%3dpubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/1073274820946794</pub-id>
          <pub-id pub-id-type="medline">32755235</pub-id>
          <pub-id pub-id-type="pmcid">PMC7791467</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref61">
        <label>61</label>
        <nlm-citation citation-type="book">
          <source>International Classification of Diseases for Oncology</source>
          <year>2013</year>
          <publisher-loc>Geneva</publisher-loc>
          <publisher-name>World Health Organization</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref62">
        <label>62</label>
        <nlm-citation citation-type="web">
          <article-title>ICD-O-3 Coding Materials</article-title>
          <source>NIH National Cancer institute</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://seer.cancer.gov/icd-o-3/">https://seer.cancer.gov/icd-o-3/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref63">
        <label>63</label>
        <nlm-citation citation-type="web">
          <article-title>Site-specific modules</article-title>
          <source>NIH National Cancer Institute</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://training.seer.cancer.gov/modules_site_spec.html">https://training.seer.cancer.gov/modules_site_spec.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref64">
        <label>64</label>
        <nlm-citation citation-type="web">
          <article-title>International classification of diseases for oncology (ICD-O)</article-title>
          <source>International Agency for Research on Cancer</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.iacr.com.fr/index.php?option=com_content&#38;view=category&#38;layout=blog&#38;id=100&#38;Itemid=577">http://www.iacr.com.fr/index.php?option=com_content&#38;view=category&#38;layout=blog&#38;id=100&#38;Itemid=577</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref65">
        <label>65</label>
        <nlm-citation citation-type="web">
          <article-title>ICD-O-3.1-NCIt Mapping Files</article-title>
          <source>National Institutes of Health</source>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://evs.nci.nih.gov/ftp1/NCI_Thesaurus/Mappings/ICD-O-3_Mappings/About.html">https://evs.nci.nih.gov/ftp1/NCI_Thesaurus/Mappings/ICD-O-3_Mappings/About.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref66">
        <label>66</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Krizhevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>GE</given-names>
            </name>
          </person-group>
          <article-title>ImageNet classification with deep convolutional neural networks</article-title>
          <source>Commun ACM</source>
          <year>2017</year>
          <month>05</month>
          <day>24</day>
          <volume>60</volume>
          <issue>6</issue>
          <fpage>84</fpage>
          <lpage>90</lpage>
          <pub-id pub-id-type="doi">10.1145/3065386</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref67">
        <label>67</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alawad</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Tourassi</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Coarse-to-fine multi-task training of convolutional neural networks for automated information extraction from cancer pathology reports</article-title>
          <source>Proceedings of the 2018 IEEE EMBS International Conference on Biomedical &#38; Health Informatics (BHI)</source>
          <year>2018</year>
          <conf-name>IEEE EMBS International Conference on Biomedical Health Informatics (BHI). ieeexplore.ieee.org; . pp. 218?221</conf-name>
          <conf-date>Mar 4-7, 2018</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bhi.2018.8333408</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref68">
        <label>68</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tenney</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Das</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Pavlick</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>BERT rediscovers the classical NLP pipeline</article-title>
          <source>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (2019)</source>
          <year>2019</year>
          <conf-name>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (2019)</conf-name>
          <conf-date>Jul 28-Aug 2, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1905.05950"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/p19-1452</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref69">
        <label>69</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pan</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>A survey on transfer learning</article-title>
          <source>IEEE Trans Knowl Data Eng</source>
          <year>2010</year>
          <month>10</month>
          <volume>22</volume>
          <issue>10</issue>
          <fpage>1345</fpage>
          <lpage>59</lpage>
          <pub-id pub-id-type="doi">10.1109/TKDE.2009.191</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref70">
        <label>70</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Moore</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>O'Dwyer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Philpot</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Assessing the utility of automatic cancer registry notifications data extraction from free-text pathology reports</article-title>
          <source>AMIA Annu Symp Proc</source>
          <year>2015</year>
          <volume>2015</volume>
          <fpage>953</fpage>
          <lpage>62</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/26958232"/>
          </comment>
          <pub-id pub-id-type="medline">26958232</pub-id>
          <pub-id pub-id-type="pmcid">PMC4765645</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref71">
        <label>71</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Amin</surname>
              <given-names>MB</given-names>
            </name>
            <name name-style="western">
              <surname>Edge</surname>
              <given-names>SB</given-names>
            </name>
          </person-group>
          <source>AJCC Cancer Staging Manual</source>
          <year>2017</year>
          <publisher-loc>Cham</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref72">
        <label>72</label>
        <nlm-citation citation-type="web">
          <article-title>Active learning literature survey</article-title>
          <source>University of Wisconsin-Madison Department of Computer Sciences</source>
          <year>2009</year>
          <access-date>2022-02-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://minds.wisconsin.edu/handle/1793/60660">https://minds.wisconsin.edu/handle/1793/60660</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref73">
        <label>73</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Holzinger</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Interactive machine learning for health informatics: when do we need the human-in-the-loop?</article-title>
          <source>Brain Inform</source>
          <year>2016</year>
          <month>06</month>
          <volume>3</volume>
          <issue>2</issue>
          <fpage>119</fpage>
          <lpage>31</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://link.springer.com/article/10.1007/s40708-016-0042-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s40708-016-0042-6</pub-id>
          <pub-id pub-id-type="medline">27747607</pub-id>
          <pub-id pub-id-type="pii">10.1007/s40708-016-0042-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC4883171</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref74">
        <label>74</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Settles</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>From theories to queries: active learning in practice</article-title>
          <source>JMLR Workshop Conference Proc</source>
          <year>2011</year>
          <volume>16</volume>
          <fpage>1</fpage>
          <lpage>18</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://proceedings.mlr.press/v16/settles11a/settles11a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref75">
        <label>75</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Glicksberg</surname>
              <given-names>BS</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Walker</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Bian</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Federated learning for healthcare informatics</article-title>
          <source>J Healthc Inform Res</source>
          <year>2020</year>
          <month>11</month>
          <day>12</day>
          <fpage>1</fpage>
          <lpage>19</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/33204939"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s41666-020-00082-4</pub-id>
          <pub-id pub-id-type="medline">33204939</pub-id>
          <pub-id pub-id-type="pii">82</pub-id>
          <pub-id pub-id-type="pmcid">PMC7659898</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref76">
        <label>76</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thoburn</surname>
              <given-names>KK</given-names>
            </name>
            <name name-style="western">
              <surname>German</surname>
              <given-names>RR</given-names>
            </name>
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Nichols</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Ahmed</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Jackson-Thompson</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Case completeness and data accuracy in the Centers for Disease Control and Prevention's National Program of Cancer Registries</article-title>
          <source>Cancer</source>
          <year>2007</year>
          <month>04</month>
          <day>15</day>
          <volume>109</volume>
          <issue>8</issue>
          <fpage>1607</fpage>
          <lpage>16</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1002/cncr.22566"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/cncr.22566</pub-id>
          <pub-id pub-id-type="medline">17343277</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
