<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v26i1e48443</article-id>
      <article-id pub-id-type="pmid">38271060</article-id>
      <article-id pub-id-type="doi">10.2196/48443</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Unlocking the Secrets Behind Advanced Artificial Intelligence Language Models in Deidentifying Chinese-English Mixed Clinical Text: Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Castonguay</surname>
            <given-names>Alexandre</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhu</surname>
            <given-names>Lingxuan</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Chrimes</surname>
            <given-names>Dillon</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Taira</surname>
            <given-names>Ricky</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>You-Qian</given-names>
          </name>
          <degrees>ME</degrees>
          <xref rid="aff01" ref-type="aff">1</xref>
          <xref rid="aff02" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0419-0919</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Ching-Tai</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff03" ref-type="aff">3</xref>
          <xref rid="aff04" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5695-1962</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Chien-Chang</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff05" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-0103-3089</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>Chung-Hong</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff06" ref-type="aff">6</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4178-5388</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Peitsz</given-names>
          </name>
          <degrees>BS</degrees>
          <xref rid="aff07" ref-type="aff">7</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0007-2241-7244</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Wu</surname>
            <given-names>Chi-Shin</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff08" ref-type="aff">8</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2762-7295</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Dai</surname>
            <given-names>Hong-Jie</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff02" ref-type="aff">2</xref>
          <address>
            <institution>Intelligent System Laboratory, Department of Electrical Engineering</institution>
            <institution>College of Electrical Engineering and Computer Science</institution>
            <institution>National Kaohsiung University of Science and Technology</institution>
            <addr-line>No. 415, Jiangong Road, Sanmin District</addr-line>
            <addr-line>Kaohsiung, 80778</addr-line>
            <country>Taiwan</country>
            <phone>886 73814526 ext 15510</phone>
            <email>hjdai@nkust.edu.tw</email>
          </address>
          <xref rid="aff09" ref-type="aff">9</xref>
          <xref rid="aff10" ref-type="aff">10</xref>
          <xref rid="aff11" ref-type="aff">11</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1516-7255</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff01">
        <label>1</label>
        <institution>Dialogue System Technical Department</institution>
        <institution>Intelligent Robot</institution>
        <institution>Asustek Computer Inc</institution>
        <addr-line>Taipei</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff02">
        <label>2</label>
        <institution>Intelligent System Laboratory, Department of Electrical Engineering</institution>
        <institution>College of Electrical Engineering and Computer Science</institution>
        <institution>National Kaohsiung University of Science and Technology</institution>
        <addr-line>Kaohsiung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff03">
        <label>3</label>
        <institution>Department of Bioinformatics and Medical Engineering</institution>
        <institution>Asia University</institution>
        <addr-line>Taichung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff04">
        <label>4</label>
        <institution>Center for Precision Health Research</institution>
        <institution>Asia University</institution>
        <addr-line>Taichung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff05">
        <label>5</label>
        <institution>Electromagnetic Sensing Control and AI Computing System Laboratory, Department of Electrical Engineering</institution>
        <institution>College of Electrical Engineering and Computer Science</institution>
        <institution>National Kaohsiung University of Science and Technology</institution>
        <addr-line>Kaohsiung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff06">
        <label>6</label>
        <institution>Knowledge Discovery and Data Mining Lab, Department of Electrical Engineering</institution>
        <institution>College of Electrical Engineering and Computer Science</institution>
        <institution>National Kaohsiung University of Science and Technology</institution>
        <addr-line>Kaohsiung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff07">
        <label>7</label>
        <institution>Department of Chemical Engineering</institution>
        <institution>Feng Chia University</institution>
        <addr-line>Taichung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff08">
        <label>8</label>
        <institution>National Center for Geriatrics and Welfare Research</institution>
        <institution>National Health Research Institutes</institution>
        <addr-line>Zhunan</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff09">
        <label>9</label>
        <institution>National Institute of Cancer Research</institution>
        <institution>National Health Research Institutes</institution>
        <addr-line>Tainan</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff10">
        <label>10</label>
        <institution>School of Post-Baccalaureate Medicine</institution>
        <institution>College of Medicine</institution>
        <institution>Kaohsiung Medical University</institution>
        <addr-line>Kaohsiung</addr-line>
        <country>Taiwan</country>
      </aff>
      <aff id="aff11">
        <label>11</label>
        <institution>Center for Big Data Research</institution>
        <institution>Kaohsiung Medical University</institution>
        <addr-line>Kaohsiung</addr-line>
        <country>Taiwan</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Hong-Jie Dai <email>hjdai@nkust.edu.tw</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>25</day>
        <month>1</month>
        <year>2024</year>
      </pub-date>
      <volume>26</volume>
      <elocation-id>e48443</elocation-id>
      <history>
        <date date-type="received">
          <day>24</day>
          <month>4</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>2</day>
          <month>10</month>
          <year>2023</year>
        </date>
        <date date-type="rev-recd">
          <day>27</day>
          <month>10</month>
          <year>2023</year>
        </date>
        <date date-type="accepted">
          <day>5</day>
          <month>12</month>
          <year>2023</year>
        </date>
      </history>
      <copyright-statement>©You-Qian Lee, Ching-Tai Chen, Chien-Chang Chen, Chung-Hong Lee, Peitsz Chen, Chi-Shin Wu, Hong-Jie Dai. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 25.01.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2024/1/e48443" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>The widespread use of electronic health records in the clinical and biomedical fields makes the removal of protected health information (PHI) essential to maintain privacy. However, a significant portion of information is recorded in unstructured textual forms, posing a challenge for deidentification. In multilingual countries, medical records could be written in a mixture of more than one language, referred to as code mixing. Most current clinical natural language processing techniques are designed for monolingual text, and there is a need to address the deidentification of code-mixed text.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The aim of this study was to investigate the effectiveness and underlying mechanism of fine-tuned pretrained language models (PLMs) in identifying PHI in the code-mixed context. Additionally, we aimed to evaluate the potential of prompting large language models (LLMs) for recognizing PHI in a zero-shot manner.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We compiled the first clinical code-mixed deidentification data set consisting of text written in Chinese and English. We explored the effectiveness of fine-tuned PLMs for recognizing PHI in code-mixed content, with a focus on whether PLMs exploit naming regularity and mention coverage to achieve superior performance, by probing the developed models’ outputs to examine their decision-making process. Furthermore, we investigated the potential of prompt-based in-context learning of LLMs for recognizing PHI in code-mixed text.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The developed methods were evaluated on a code-mixed deidentification corpus of 1700 discharge summaries. We observed that different PHI types had preferences in their occurrences within the different types of language-mixed sentences, and PLMs could effectively recognize PHI by exploiting the learned name regularity. However, the models may exhibit suboptimal results when regularity is weak or mentions contain unknown words that the representations cannot generate well. We also found that the availability of code-mixed training instances is essential for the model’s performance. Furthermore, the LLM-based deidentification method was a feasible and appealing approach that can be controlled and enhanced through natural language prompts.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>The study contributes to understanding the underlying mechanism of PLMs in addressing the deidentification process in the code-mixed context and highlights the significance of incorporating code-mixed training instances into the model training phase. To support the advancement of research, we created a manipulated subset of the resynthesized data set available for research purposes. Based on the compiled data set, we found that the LLM-based deidentification method is a feasible approach, but carefully crafted prompts are essential to avoid unwanted output. However, the use of such methods in the hospital setting requires careful consideration of data security and privacy concerns. Further research could explore the augmentation of PLMs and LLMs with external knowledge to improve their strength in recognizing rare PHI.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>code mixing</kwd>
        <kwd>electronic health record</kwd>
        <kwd>deidentification</kwd>
        <kwd>pretrained language model</kwd>
        <kwd>large language model</kwd>
        <kwd>ChatGPT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>In the clinical and biomedical fields, electronic health records (EHRs) have become valuable resources in recent years [<xref ref-type="bibr" rid="ref1">1</xref>]. In order to protect the privacy of patients whose data are used for secondary purposes, regulations or laws have been established requiring the removal of protected health information (PHI) from records before they can be disseminated. In settings, such as research, obtaining explicit consent to access personal data may be impractical or impossible. As a result, the deidentification step becomes a critical data processing step to protect the privacy of individuals. However, the study by Cannon and Lucci [<xref ref-type="bibr" rid="ref2">2</xref>] indicated that up to 65% of important clinical information is recorded in unstructured texts in medical reports written by medical personnel. Compared with structured data, which can be deidentified by encrypting private patient information fields, the deidentification of unstructured data is more challenging.</p>
        <p>Because manual deidentification of large volumes of EHRs is time-consuming and error-prone, automated methods are needed for large-scale deidentification of unstructured clinical data. Compared with Asian countries, such as Japan and China, where the majority of medical records are written monolingually in their native languages, medical records in Taiwan are frequently written in a mixture of Chinese and English. The example sentence “前夫mk1300309 married &#62; mk136從婆家搬出來 &#62; mk139離婚” (got married to my ex-husband on 2141/03/09 &#62; moved out from the ex-husband’s family’s home in 2147 &#62; divorced in 2150) demonstrates how a physician wrote in English and suddenly switched to Chinese. The sentence also includes a transliteration (“mk”) of “民國,” which refers to the Republic of China calendar. This phenomenon is referred to as code alternation, which occurs when a bilingual speaker uses more than one language in a single utterance [<xref ref-type="bibr" rid="ref3">3</xref>]. Code switching refers to code alternation at or above clause level, while code mixing refers to code alteration below clause level. Some examples of code-switched and code-mixed narratives in discharge summaries are shown in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. Throughout this work, we will use code mixing to refer to the above phenomenon.</p>
        <p>In addition to our specific context, it is important to note that code mixing in clinical records is observed in various multilingual health care settings globally. For instance, Alqurashi [<xref ref-type="bibr" rid="ref4">4</xref>] conducted research on the use of code-mixed language in private hospitals in Saudi Arabia, highlighting the prevalence of this practice in health care communication. Furthermore, Dowlagar and Mamidi [<xref ref-type="bibr" rid="ref5">5</xref>] demonstrated that the use of code-mixed utterances, where native languages are blended with English, is a common and natural occurrence among doctors and patients. Beyond this, code mixing in clinical notes extends beyond English-dominant countries. For example, Keresztes [<xref ref-type="bibr" rid="ref6">6</xref>] investigated how Hungarian physicians are influenced by the English language in their writing of cardiology discharge reports, while Karuthan [<xref ref-type="bibr" rid="ref7">7</xref>] showed that nurses in Malaysia tend to use “Manglish,” a mix of Malayalam and English, when writing nursing documents.</p>
        <p>This global prevalence of code mixing highlights the importance of addressing narratives written in this manner. Unfortunately, most existing clinical natural language processing (NLP) techniques are primarily designed for monolingual texts, mainly due to limited research resources. While our previous work [<xref ref-type="bibr" rid="ref8">8</xref>] has highlighted the potential of pretrained language models (PLMs) in mitigating issues related to code mixing during deidentification, the precise mechanism at play remains incompletely understood. To bridge this knowledge gap, there is a pressing need for a code-mixed deidentification data set that can comprehensively evaluate the performance of state-of-the-art NLP models and unravel their underlying mechanisms.</p>
      </sec>
      <sec>
        <title>Goal of This Study</title>
        <p>Previous studies [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>] have demonstrated that state-of-the-art neural network models can effectively exploit strong name regularity and high mention coverage to achieve superior performance. Name regularity refers to the consistency and predictability of the structure and pronunciation of a named entity type within a particular language. For example, person names typically adhere to the “FirstName LastName” format. A high mention coverage denotes that a large proportion of mentions in the test set have been previously observed in the training set.</p>
        <p>Our study aims to explore the impact of these properties on the performance of PLMs in recognizing PHI within code-mixed sentences. We hypothesize that the decision-making processes of the models are influenced by these factors, even when dealing with code-mixed text. However, they may underperform when faced with weak regularity, such as unseen mentions or out-of-vocabulary (OOV) words. Additionally, they may solely memorize popular mentions instead of learning generalization knowledge, which could limit their generality and performance. To investigate the decision-making process of these models and gain insights into their ability to recognize code-mixed PHI, we curated a novel code-mixed deidentification data set, trained and evaluated the performance of state-of-the-art PLMs on the unique data set, and employed methods to interpret their results to validate our hypothesis and gain a deeper understanding of how these models recognize code-mixed PHI.</p>
        <p>In summary, this study has the following 3 major contributions:</p>
        <list list-type="order">
          <list-item>
            <p>Unique code-mixed deidentification data set: We significantly extended our original corpus compiled in our previous work [<xref ref-type="bibr" rid="ref8">8</xref>] by incorporating an additional 900 discharge summaries. Furthermore, we created a manipulated subset of the resynthesized data set available for research purposes. Importantly, our work resulted in the creation of the very first clinical code-mixed data set that encompasses both traditional Chinese and English. This unique data set [<xref ref-type="bibr" rid="ref11">11</xref>] represents a valuable resource for researchers, offering a fresh perspective on issues related to code mixing in the context of health care data and contributing to advancing research in code-mixed clinical text processing.</p>
          </list-item>
          <list-item>
            <p>Enhancing the understanding of PLMs in code-mixed texts: The curated data set enables an in-depth exploration of the effectiveness and underlying mechanism of fine-tuned PLMs in recognizing code-mixed PHI. Through thorough analyses, we gained insights into the mechanism used by these models for this specific task.</p>
          </list-item>
          <list-item>
            <p>Unprecedented exploration of large language models (LLMs) on code-mixed clinical text: The emergence of LLMs has raised a question about their potential application in the deidentification process. This study focused on the feasibility of employing prompt-based, zero-shot, in-context learning with LLMs to recognize PHI. More specifically, we leveraged ChatGPT to recognize PHI within code-mixed text that was previously challenging for the developed PLMs. To the best of our knowledge, this is the first attempt to use prompt-based in-context learning of LLMs for deidentification in code-mixed text.</p>
          </list-item>
        </list>
      </sec>
      <sec>
        <title>Related Work</title>
        <sec>
          <title>Definition of PHI Types</title>
          <p>The Health Insurance Portability and Accountability Act (HIPAA) is a federal law in the United States enacted in 1996 to protect the privacy and security of the PHI of individuals [<xref ref-type="bibr" rid="ref12">12</xref>]. The Safe Harbor method under the HIPAA includes the removal of 18 specific identifiers from PHI, such as name, date of birth, and social security number. If all 18 identifiers are removed, the information is considered deidentified under the method, and the data can be reused for secondary purposes [<xref ref-type="bibr" rid="ref13">13</xref>]. Given the sensitive nature of the clinical text used in this study, we expanded the list of PHI types to 20 fine-grained categories within 6 coarse-grained categories. The details are presented in <xref ref-type="table" rid="table1">Table 1</xref>.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>The coarse-grained and fine-grained categories of protected health information defined in the annotation guideline used in this study.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="250"/>
              <col width="750"/>
              <thead>
                <tr valign="top">
                  <td>Coarse-grained PHI<sup>a</sup></td>
                  <td>Fine-grained PHI</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Date</td>
                  <td>Date</td>
                </tr>
                <tr valign="top">
                  <td>Age</td>
                  <td>Age</td>
                </tr>
                <tr valign="top">
                  <td>Name</td>
                  <td>Patient, person, doctor</td>
                </tr>
                <tr valign="top">
                  <td>Location</td>
                  <td>Named location, nationality, region, country, city, hospital, department, room, number, school, generic location, market</td>
                </tr>
                <tr valign="top">
                  <td>Profession</td>
                  <td>Profession</td>
                </tr>
                <tr valign="top">
                  <td>ID</td>
                  <td>ID number, medical record</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table1fn1">
                <p><sup>a</sup>PHI: protected health information.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Publicly Available Deidentification or Code-Mixed Data Sets</title>
          <p>A code-mixed named entity recognition competition was held in 2018 [<xref ref-type="bibr" rid="ref14">14</xref>], in which twitter data with the mixture of English and Spanish, and Arabic and Egyptian were provided. The former consisted of 67,223 tweets, and the latter consisted of 12,334 tweets. To the best of our knowledge, there is only 1 publicly available Chinese-mixed corpus, which is a simple Chinese-English code-mixed data set for the task of emotion detection [<xref ref-type="bibr" rid="ref15">15</xref>].</p>
          <p>On the other hand, the MIMIC (Medical Information Mart for Intensive Care) II/III database [<xref ref-type="bibr" rid="ref16">16</xref>] is one of the largest deidentified clinical databases on patients admitted to a medical center in the United States, which offers clinical text in English for researchers to access and use under a data access agreement. To facilitate automated deidentification tasks, large annotated corpora of unstructured text with PHI entities are required. However, the development of such corpora is complex and limited in number. To address this need, the Informatics for Integrating Biology and the Bedside (i2b2) has organized several clinical deidentified NLP competitions [<xref ref-type="bibr" rid="ref17">17</xref>], such as the deidentified competition in 2014, which released a corpus of 1304 medical records of 296 patients with diabetes [<xref ref-type="bibr" rid="ref18">18</xref>]. Another is the CEGS N-GRID deidentification corpus [<xref ref-type="bibr" rid="ref19">19</xref>], containing 1000 psychiatric notes from the United States. Alla et al [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>] compiled the first open available deidentification corpus from Australia, which contained 2100 pathology reports for the purpose of automatic deidentification. However, deidentification corpora in languages other than English are scarce. A Norwegian synthetic clinical corpus released by Bråthen et al [<xref ref-type="bibr" rid="ref22">22</xref>] is an example, and it includes 477 sentences. To the best of our knowledge, there is currently no openly available code-mixed deidentification corpus.</p>
        </sec>
        <sec>
          <title>Deidentification Methods and Approaches for Tackling Code-Mixing Challenges</title>
          <p>Manual deidentification is costly, in terms of both finances and time. For example, deidentification of 50,000 patient visit records in the MIMIC-III data set can cost around US $500,000 and 5000 hours. To overcome these challenges, automated deidentification systems have been developed. Initially, rule-based systems using lexical dictionaries, regular expressions, and simple heuristics were introduced [<xref ref-type="bibr" rid="ref23">23</xref>]. However, they have limitations in generalization across different data sets. Subsequently, machine learning–based approaches [<xref ref-type="bibr" rid="ref23">23</xref>-<xref ref-type="bibr" rid="ref25">25</xref>] were proposed, offering better generalization and prediction accuracy. Nonetheless, these methods still rely on handcrafted features. More recently, deep learning techniques have become mainstream, eliminating the need for manually crafted features [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. Neural networks are advantageous because they can be initialized with PLMs acquired from extensive unlabeled data, resulting in faster optimization and superior performance. BERT (Bidirectional Encoder Representations from Transformers) pretrained on English corpora (EN-BERT) [<xref ref-type="bibr" rid="ref28">28</xref>] is one such example of a monolingual transformer model pretrained on the BookCorpus [<xref ref-type="bibr" rid="ref29">29</xref>] and English Wikipedia in a self-supervised fashion, which has achieved exceptional precision in various NLP tasks including the deidentification task [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref30">30</xref>].</p>
          <p>On the other hand, the rise of code-mixed data sets has spurred the development of approaches designed to address the unique challenges posed by code-mixed content. For instance, in the work by Winata et al [<xref ref-type="bibr" rid="ref31">31</xref>], a bidirectional long-short term memory model was used to leverage both character and word information, complemented by pretrained FastText embeddings [<xref ref-type="bibr" rid="ref32">32</xref>], to improve the identification of OOV words in an English-Spanish mixed data set. Another approach by Trivedi et al [<xref ref-type="bibr" rid="ref33">33</xref>] incorporated FastText word vectors [<xref ref-type="bibr" rid="ref34">34</xref>] trained from Spanish Wikipedia pages and pretrained word embeddings sourced from a substantial collection of tweets [<xref ref-type="bibr" rid="ref35">35</xref>] to tackle the English-Spanish code-mixing problem. They applied the singular value decomposition [<xref ref-type="bibr" rid="ref36">36</xref>] method to align word embeddings across different languages, facilitating a unified vector space representation. Devlin et al [<xref ref-type="bibr" rid="ref28">28</xref>] introduced pretrained models beyond monolingual EN-BERT, including BERT pretrained on simplified and traditional Chinese corpora (CH-BERT) and BERT pretrained on Wikipedia corpora from 104 languages (M-BERT). Researchers have been exploring the use of M-BERT to tackle the code-mixing problem owing to its ability to generalize cross-lingually through its multilingual representation [<xref ref-type="bibr" rid="ref37">37</xref>]. For example, Tang et al [<xref ref-type="bibr" rid="ref15">15</xref>] fine-tuned M-BERT on their English-simplified Chinese social media data set for multi-label sentiment analysis, and obtained an F-score of 0.69, which was 15% higher than that for the model without M-BERT. While limited research exists on using BERT-based models or other LLMs for code-mixing deidentification in clinical data sets, our previous work [<xref ref-type="bibr" rid="ref38">38</xref>] suggested the potential benefits of incorporating M-BERT to disambiguate PHI categories in code-mixed sentences. In this study, we focused on comprehending how BERT-family PLMs handle Chinese-English mixed issues that arise in actual clinical text and assessed the feasibility of using state-of-the-art LLMs to recognize PHI from sampled clinical text.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Data Sources and Corpus Construction</title>
        <p>We obtained discharge summaries sampled from the psychiatry section of a medical center, which were used to extract depressive symptoms with text mining approaches [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. From these data, we compiled a code-mixed deidentification corpus of 1700 discharge summaries. Four annotators were enlisted to annotate the entire corpus, and they followed the same annotation procedure as previously described in our work [<xref ref-type="bibr" rid="ref8">8</xref>]. To ensure the quality of the annotation, the annotation process began with the annotators individually annotating an identical set of 200 randomly sampled records, using the annotation guidelines provided. Subsequently, a meeting was organized to facilitate discussions among the annotators, addressing any issues or concerns that arose during the annotation process. This iterative process continued until the annotators achieved a strong level of agreement [<xref ref-type="bibr" rid="ref41">41</xref>]. The interannotation agreement, as measured by κ, was found to be 0.85 after the above procedure. Once this agreement threshold was reached, the remaining unlabeled EHRs were evenly distributed among the 4 annotators for labeling. The training set comprised 1500 discharge summaries along with 297,621 sentences. The test set comprised 200 discharge summaries along with 60,632 sentences. Finally, the principle-based resynthesis method proposed in our previous work [<xref ref-type="bibr" rid="ref38">38</xref>] was used to generate surrogates, and the entire corpus was rechecked by one of the senior annotators (PTC) to ensure a high level of data consistency and correctness.</p>
      </sec>
      <sec>
        <title>Ethics Approval</title>
        <p>The study has been approved by the ethics committee for medical research of the National Institutes of Health (number: EC1090212-E).</p>
      </sec>
      <sec>
        <title>Implementation of Deidentification Methods</title>
        <p>After preprocessing the collected data set, we approached the deidentification task by treating it as a named entity recognition problem in which the target entity types are PHI types defined in <xref ref-type="table" rid="table1">Table 1</xref>. To establish a baseline, we used a dictionary-based method [<xref ref-type="bibr" rid="ref42">42</xref>,<xref ref-type="bibr" rid="ref43">43</xref>] and developed 4 BERT-based models for comparison. The dictionary-based approach (DBA) relies on predefined dictionaries and a lookup method. In our implementation, we compiled a dictionary consisting of tokens for all PHI types collected from the training set. The most frequent tag associated with each token estimated on the training set was set as the token’s tag. For the BERT-based models, the recognition task was formulated as a sequential tagging problem by using the BILOU tagging schema. This schema categorizes the tokens as either the beginning (B), within (I), or last (L) of multi-word PHI, as well as identifies non-PHI (O) and single-word PHI (U) tokens. For more in-depth information about the preprocessing steps and the development of the DBA, please refer to <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p>
        <sec>
          <title>BERT-Based Approach</title>
          <p>The BERT-family PLMs, including EN-BERT, CH-BERT, and M-BERT, were used in this study. In our implementation, an additional fully connected layer was added that takes the token embeddings from the top layer of the underling PLM as input. This layer is employed to predict the probabilities of the tags for each token. When the WordPiece tokenizer splits a token into multiple pieces, we take the representation of the first piece to represent the token [<xref ref-type="bibr" rid="ref44">44</xref>]. Details of the procedure and hyperparameters used for training our BERT-based models are described in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p>
          <p>In addition to fine-tuning the above BERT-based models on the original code-mixed corpus, we translated sentences in our English-Chinese corpus into either English or Chinese and then fine-tuned the M-BERT model on the translated deidentification corpus (TM-BERT) by determining the degree of code mixing present in a given sentence using the code-mixing index (CMI) [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref45">45</xref>]. The flowchart of the translation process for compiling the training set is shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>.</p>
          <p>We determined the degree of code mixing present in a given sentence <italic>s</italic> using the following formula proposed by Gambäck and Das [<xref ref-type="bibr" rid="ref45">45</xref>]:</p>
          <disp-formula>
            <graphic xlink:href="jmir_v26i1e48443_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>where <italic>W<sub>i</sub></italic> is the number of tokens of the most frequent language in the sentence <italic>s</italic>, <italic>n</italic> is the total number of tokens, and <italic>u</italic> is the quantity of tokens in <italic>s</italic> that are independent of any particular language, such as numeric words and punctuation marks.</p>
          <p>By definition, the CMI is always smaller than 0.5 for code mixing of 2 languages. A larger CMI (close to 0.5) indicates serious code mixing, while a smaller CMI (close to 0) suggests that 1 language dominates the majority of tokens. The CMI is set as zero if the entire sentence is either monolingual or composed of numeric words or punctuations. In cases where the CMI of a sentence is not zero, the number of tokens belonging to English and Chinese in the sentence is calculated. If the majority of tokens belong to English, the sentence is translated to English using a machine translation model [<xref ref-type="bibr" rid="ref46">46</xref>]. Conversely, if Chinese tokens dominate, we opt for translation to traditional Chinese [<xref ref-type="bibr" rid="ref47">47</xref>]. Subsequently, we proceeded with the identical fine-tuning process on the translated corpus. In the prediction phase, we employed identical procedures to transform the input sentences into monolingual sentences based on the dominating language and subsequently used TM-BERT to detect PHI.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>The procedure to generate the training and test sets for TM-BERT. CMI: code-mixing index; M-BERT: BERT pretrained on Wikipedia corpora from 104 languages; TM-BERT: M-BERT fine-tuned on the translated deidentification corpus.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e48443_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>ChatGPT-Based Deidentification Framework</title>
          <p><xref rid="figure2" ref-type="fig">Figure 2</xref> presents our framework to use ChatGPT for the recognition of PHI, which is based on recent research suggesting that LLMs can be adapted to downstream tasks through the use of prompts [<xref ref-type="bibr" rid="ref48">48</xref>]. A prompt is a set of instructions that can customize an LLM’s capabilities and affect its subsequent interactions and outputs. To optimize our prompt, we performed prompt engineering based on the PHI definitions. The objective was to ensure that the extracted candidate PHI text could be matched with the original input text represented by the “&#60;SENTENCE&#62;” placeholder. Following the prompt design procedure [<xref ref-type="bibr" rid="ref49">49</xref>], we organized our prompt into 4 main parts:</p>
          <list list-type="order">
            <list-item>
              <p>Overall task instruction: This part provides general instructions for the task. We declared it as “Extracting Private Information.”</p>
            </list-item>
            <list-item>
              <p>Sentence introduction: This part defines the placeholder for the input text (“&#60;SENTENCE&#62;”) and the PHI that we focus on.</p>
            </list-item>
            <list-item>
              <p>Constraint: As illustrated in <xref rid="figure2" ref-type="fig">Figure 2</xref>, the text following the statement “Extract PHI from the given text based on the above definitions” forms the constraint part. The inclusion of this constraint is based on our observation that ChatGPT has the capability to directly extract various PHI types from text and sometimes translate extracted Chinese text into English; therefore, we constrained it to the defined types and requested that it adhere to the regulations in order to avoid translation. This enabled us to match the extracted candidate PHI with the input text and determine their spans for the purpose of performance evaluation.</p>
            </list-item>
            <list-item>
              <p>Retrieval message: We provided the text “PHIs:” to instruct the model to generate results to complete the task.</p>
            </list-item>
          </list>
          <p>The experiment was conducted from February 20, 2023, to April 30, 2023, using the GPT-3.5 model (version from February 13, 2023) with temperature and top_<italic>P</italic> values set at 0.5. As we used the zero-shot in-context learning setting, we observed that ChatGPT could generate different responses for the same prompt. Thus, we sent the same prompt 3 times to collect all PHI candidates and retained the PHI that appeared at least twice as the final results. The decision to send the prompt 3 times aligns with previous work [<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref51">51</xref>], which recommended this practice to ensure the stability of the responses.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>The developed framework for the recognition of protected health information mentioned in code-mixed text based on ChatGPT. EHR: electronic health record; PHI: protected health information.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e48443_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
      <sec>
        <title>Evaluation Metrics and Analysis Methods Used in This Study</title>
        <sec>
          <title>Deidentification Performance Evaluation</title>
          <p>Performance of the developed methods was measured by standard metrics used in previous deidentification tasks [<xref ref-type="bibr" rid="ref18">18</xref>] including recall and precision. The overall performance was assessed using the micro-F<sub>1</sub> measure, a weighted harmonic mean of precision and recall. We further report the macro-F<sub>1</sub> measure as the unweighted mean of the F-scores calculated per PHI type. For the evaluation of the results of the proposed ChatGPT-based deidentification framework, we applied the approximate match criterion [<xref ref-type="bibr" rid="ref52">52</xref>] to determine whether the recognized boundary matched with the human annotated span.</p>
          <p>In addition, to gain more insights into the impact of the code-mixing level on PHI recognition, we further categorized sentences according to their CMI levels and estimated the corresponding sentence error rates, which were computed by dividing the total number of wrongly predicted sentences by the total number of sentences in each CMI range. A sentence was considered incorrectly predicted if any of its tokens were labeled incorrectly.</p>
        </sec>
        <sec>
          <title>Code-Mixing Level Evaluation</title>
          <p>To measure the level of complexity and mixing in our corpus, we applied the CMI defined in equation (1) to calculate the average CMI for a specific language <italic>L</italic>, termed AvgCMI<italic><sub>L</sub></italic>. In this study, a sentence <italic>s</italic> was classified as either English-dominated or Chinese-dominated according to the dominating tokens present in the sentence. AvgCMI<italic><sub>L</sub></italic> was defined as follows:</p>
          <disp-formula>
            <graphic xlink:href="jmir_v26i1e48443_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>where <italic>L</italic> is either the English or Chinese language and <italic>K</italic> indicates the set of <italic>L</italic>-dominated sentences in our corpus. Note that the calculation of AvgCMI<italic><sub>L</sub></italic> includes the sentences with a CMI of zero. Finally, the CMI for the corpus (CMIC) was defined as follows:</p>
          <disp-formula>
            <graphic xlink:href="jmir_v26i1e48443_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
        </sec>
        <sec>
          <title>Methods for Diagnosing and Interpreting the Outputs of the Developed Models</title>
          <p>According to our experimental findings, the primary factor affecting PHI recognition performance was the problem of ambiguity. This problem arises when the identified words or phrases can be categorized under more than one PHI type. For example, “八里 (Bali)” is a suburban district in northwestern New Taipei, Taiwan, but it can also refer to “八里療養院 (Bali psychiatric center)” in sentences such as “Informant: Discharge note from 八里.” To interpret the produced representations and analyze the ambiguity issues encountered, we used the t-Distributed Stochastic Neighbor Embedding (t-SNE) [<xref ref-type="bibr" rid="ref53">53</xref>] method to project the hidden representations of PHI tokens from the last transformer layer of the developed models.</p>
          <p>In our subsequent experiments, we noticed that EN-BERT can recognize the target PHI even if the information is described in Chinese characters or is mentioned in sentences that consist of English-Chinese or Chinese characters only. Moreover, we found that all BERT-based models could recognize a target PHI as a candidate, even if its encoding is replaced with “UNK” (unknown word). For instance, in the sentence “... follow-up in Dr. 張家明’s OPD,” where “張家明” is recognized as “Doctor,” replacing the encoding of “張家明” in the sentence with “UNK” did not affect the ability of BERT-based models to recognize it as candidate PHI of “Doctor.”</p>
          <p>To comprehend the decision-making mechanism, we probed the results of the developed models by applying the input reduction method [<xref ref-type="bibr" rid="ref54">54</xref>] and interpreted it using the attention-head view (AHV) to visualize the attention patterns produced by the models’ attention heads in the last transformer layer. The input reduction method removes as many words as possible based on the calculated gradient values, without altering a tag’s prediction. The AHV visualization was generated by the BERTViz tool [<xref ref-type="bibr" rid="ref55">55</xref>].</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>PHI Distribution and Code-Mixing Level Estimation</title>
        <p>Given the code-mixing characteristics of our corpus, we categorized sentences in our data set based on their code-mixing characteristics into 4 groups: English only, Chinese only, a mixture of Chinese and English, and numeric/symbolic characters. The statistical information for PHI across the 4 groups is presented in <xref ref-type="table" rid="table2">Table 2</xref>. Overall, it is apparent that the distributions of different PHI types are significantly imbalanced. Without considering numeric/symbolic sentences, the most frequent type of sentences was English only, followed by code-mixed and Chinese only. The “Date” type was most frequently observed in numeric/symbolic sentences, while the “Age” type was most commonly noted in English-only sentences. The “Name” and “Profession” types were most frequently observed in code-mixed sentences, and the “Location” type was evenly distributed across English-only, Chinese-only, and code-mixed sentences. Finally, the “ID” type was most commonly found in Chinese-only sentences.</p>
        <p>The AvgCMI<sub>English</sub>, AvgCMI<sub>Chinese</sub>, and CMIC estimated for the training set, test set, and entire corpus are presented in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>, which highlight the writing convention in Taiwan’s medical records, with the majority of records being written in English. Another interesting writing convention observed was that code mixing was more frequent in Chinese-dominated sentences than in English-dominated sentences. The significant difference between AvgCMI<sub>English</sub> and AvgCMI<sub>Chinese</sub> highlighted that the frequency of Chinese tokens occurring in English-dominated sentences is much less than the frequency of English tokens occurring in Chinese-dominated sentences. The CMIC of our corpus was 22.21, which is significantly higher than the CMIC of English-Bangla (approximately 5.15) and Dutch-Turkish (approximately 4.13) data sets [<xref ref-type="bibr" rid="ref45">45</xref>].</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Protected health information statistics for the 4 sentence groups (English-only sentences, Chinese-only sentences, mixed Chinese-English sentences, and numeric or symbolic sentences).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="110"/>
            <col width="130"/>
            <col width="60"/>
            <col width="130"/>
            <col width="60"/>
            <col width="100"/>
            <col width="60"/>
            <col width="100"/>
            <col width="60"/>
            <col width="100"/>
            <col width="60"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Protected health information type</td>
                <td colspan="2">English-only sentences, n</td>
                <td colspan="2">Chinese-only sentences, n</td>
                <td colspan="2">Chinese-English mixed sentences, n</td>
                <td colspan="2">Numeric or symbolic sentences, n</td>
                <td colspan="2">Entire corpus, n</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <break/>
                </td>
                <td>Training set</td>
                <td>Test set</td>
                <td>Training set</td>
                <td>Test set</td>
                <td>Training set</td>
                <td>Test set</td>
                <td>Training set</td>
                <td>Test set</td>
                <td>Training set</td>
                <td>Test set</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">Date<sup>a</sup></td>
                <td>12,436</td>
                <td>2346</td>
                <td>2146</td>
                <td>458</td>
                <td>4530</td>
                <td>919</td>
                <td>19,850</td>
                <td>3724</td>
                <td>38,962</td>
                <td>7447</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Age<sup>a</sup></td>
                <td>2444</td>
                <td>533</td>
                <td>21</td>
                <td>5</td>
                <td>639</td>
                <td>128</td>
                <td>0</td>
                <td>0</td>
                <td>3104</td>
                <td>666</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>Name<sup>a</sup></bold>
                </td>
                <td>100</td>
                <td>18</td>
                <td>273</td>
                <td>44</td>
                <td>1339</td>
                <td>240</td>
                <td>0</td>
                <td>0</td>
                <td>1712</td>
                <td>302</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Patient</td>
                <td>9</td>
                <td>9</td>
                <td>106</td>
                <td>12</td>
                <td>42</td>
                <td>11</td>
                <td>0</td>
                <td>0</td>
                <td>157</td>
                <td>32</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Person</td>
                <td>8</td>
                <td>3</td>
                <td>6</td>
                <td>1</td>
                <td>128</td>
                <td>32</td>
                <td>0</td>
                <td>0</td>
                <td>142</td>
                <td>36</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Doctor</td>
                <td>83</td>
                <td>6</td>
                <td>161</td>
                <td>31</td>
                <td>1169</td>
                <td>197</td>
                <td>0</td>
                <td>0</td>
                <td>1413</td>
                <td>234</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>Location<sup>a</sup></bold>
                </td>
                <td>5030</td>
                <td>1014</td>
                <td>4812</td>
                <td>1108</td>
                <td>5330</td>
                <td>1132</td>
                <td>4</td>
                <td>3</td>
                <td>15,176</td>
                <td>3257</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Named location</td>
                <td>17</td>
                <td>10</td>
                <td>11</td>
                <td>3</td>
                <td>271</td>
                <td>69</td>
                <td>0</td>
                <td>0</td>
                <td>299</td>
                <td>82</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Nationality</td>
                <td>21</td>
                <td>7</td>
                <td>3</td>
                <td>1</td>
                <td>17</td>
                <td>10</td>
                <td>0</td>
                <td>0</td>
                <td>41</td>
                <td>18</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Region</td>
                <td>6</td>
                <td>11</td>
                <td>0</td>
                <td>0</td>
                <td>10</td>
                <td>1</td>
                <td>0</td>
                <td>0</td>
                <td>16</td>
                <td>12</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Country</td>
                <td>355</td>
                <td>52</td>
                <td>6</td>
                <td>5</td>
                <td>171</td>
                <td>44</td>
                <td>0</td>
                <td>0</td>
                <td>532</td>
                <td>101</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>City</td>
                <td>196</td>
                <td>39</td>
                <td>26</td>
                <td>6</td>
                <td>510</td>
                <td>101</td>
                <td>0</td>
                <td>0</td>
                <td>732</td>
                <td>146</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Hospital</td>
                <td>1552</td>
                <td>266</td>
                <td>73</td>
                <td>11</td>
                <td>1915</td>
                <td>393</td>
                <td>0</td>
                <td>0</td>
                <td>3540</td>
                <td>670</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Department</td>
                <td>1616</td>
                <td>391</td>
                <td>2278</td>
                <td>542</td>
                <td>762</td>
                <td>111</td>
                <td>0</td>
                <td>0</td>
                <td>4656</td>
                <td>1044</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Room</td>
                <td>1007</td>
                <td>182</td>
                <td>1200</td>
                <td>267</td>
                <td>248</td>
                <td>53</td>
                <td>4</td>
                <td>3</td>
                <td>2459</td>
                <td>505</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Number</td>
                <td>0</td>
                <td>0</td>
                <td>1162</td>
                <td>256</td>
                <td>25</td>
                <td>1</td>
                <td>0</td>
                <td>0</td>
                <td>1187</td>
                <td>257</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>School</td>
                <td>28</td>
                <td>6</td>
                <td>15</td>
                <td>2</td>
                <td>642</td>
                <td>159</td>
                <td>0</td>
                <td>0</td>
                <td>685</td>
                <td>167</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Generic location</td>
                <td>218</td>
                <td>45</td>
                <td>37</td>
                <td>15</td>
                <td>708</td>
                <td>173</td>
                <td>0</td>
                <td>0</td>
                <td>963</td>
                <td>233</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Market</td>
                <td>12</td>
                <td>3</td>
                <td>1</td>
                <td>0</td>
                <td>50</td>
                <td>17</td>
                <td>0</td>
                <td>0</td>
                <td>63</td>
                <td>20</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Profession<sup>a</sup></td>
                <td>575</td>
                <td>107</td>
                <td>82</td>
                <td>9</td>
                <td>1566</td>
                <td>449</td>
                <td>0</td>
                <td>0</td>
                <td>2223</td>
                <td>565</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>ID<sup>a</sup></bold>
                </td>
                <td>54</td>
                <td>0</td>
                <td>192</td>
                <td>26</td>
                <td>197</td>
                <td>5</td>
                <td>7</td>
                <td>0</td>
                <td>449</td>
                <td>31</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>ID number</td>
                <td>43</td>
                <td>0</td>
                <td>98</td>
                <td>11</td>
                <td>170</td>
                <td>2</td>
                <td>7</td>
                <td>0</td>
                <td>318</td>
                <td>13</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Medical record</td>
                <td>1</td>
                <td>0</td>
                <td>92</td>
                <td>14</td>
                <td>19</td>
                <td>3</td>
                <td>0</td>
                <td>0</td>
                <td>112</td>
                <td>17</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Total number of sentences</td>
                <td>20,639</td>
                <td>4018</td>
                <td>7526</td>
                <td>1650</td>
                <td>13,604</td>
                <td>2873</td>
                <td>19,861</td>
                <td>3727</td>
                <td>61,630</td>
                <td>12,268</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>Coarse-grained protected health information.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Deidentification Performance Comparison</title>
        <p>The performance comparison of the developed deidentification methods is shown in <xref ref-type="table" rid="table3">Table 3</xref>, where the DBA indicates the dictionary-based baseline. The remaining configurations include monolingual BERT models (EN-BERT and CH-BERT) and multilingual BERT models (M-BERT and TM-BERT).</p>
        <p>The results showed a significantly lower performance of the DBA, highlighting that the dictionary collected from the training set is insufficient to provide reliable deidentification results. To be specific, the micro- and macro-F-scores of the DBA were 0.13 and 0.16, respectively, while the micro- and macro-F-scores of BERT-based methods ranged from 0.90 to 0.94 and 0.63 to 0.79, respectively, which were significantly higher than those of the DBA. Among all BERT-based approaches, M-BERT outperformed the others in terms of both micro- and macro-F-scores. CH-BERT demonstrated a performance comparable to that of M-BERT on various PHI types such as “Name” and “Location.” The micro- and macro-F-scores of TM-BERT were higher than those of EN-BERT but lower than those of CH-BERT and M-BERT, suggesting that translating code-mixed sentences into monolingual text does not provide advantages. TM-BERT’s lower performance was mainly due to translation errors that can cause syntactic or semantic issues during fine-tuning. Some words like abbreviations (“NKUST”) and Chinese names that are PHI mentions cannot be accurately translated. Even context words could alter their meaning, for example, the text “occupation: 英語補習班老師” (Profession: English tutoring teacher) was translated as “佔領: 英語補習班老師” (Domination: English tutoring teacher).</p>
        <p>To gain a deeper understanding of the ability of PLMs to recognize different types of PHI mentioned in sentences with different levels of language mixing, we categorized the sentences in our corpus into 3 distinct sentence categories: English-only, Chinese-only, and Chinese-English mixed. We then evaluated the performance of the developed methods for each category. The results are presented in <xref ref-type="table" rid="table4">Table 4</xref>, focusing on the comparison between EN-BERT, CH-BERT, and M-BERT only. We observed that the distribution of different PHI types varied across the 3 sentence categories, suggesting that the performance of the developed models may be influenced by the prevalence of certain PHI types in their respective sentence categories. Specifically, we found that numeric PHI types, such as “Date” and “Age,” were more prevalent in English-only sentences, while nonnumeric PHI types, including “Name” and “Profession,” were more frequent in code-mixed sentences. This finding explains why EN-BERT performed similarly to multilingual and translation models on numeric PHI types but much worse on nonnumeric types. Additionally, the generally high performance for numeric PHI types indicates that mixing Chinese and English in the same sentence has little effect on their recognitions. Detailed results can be found in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>.</p>
        <p>In contrast, the “ID” PHI type occurred most frequently in the Chinese-only sentences, but all BERT-based models performed well in recognizing them. The “Location” PHI type was equally present in all 3 categories, but certain fine-grained types, such as “Region” and “Nationality,” were more challenging due to their limited training instances. Overall, the results showed that M-BERT outperformed the other models in the 3 sentence categories based on the metric of micro-F-score, while CH-BERT had the best performance in terms of macro-F-score in Chinese-only and code-mixed sentences. EN-BERT exhibited comparable performance with M-BERT in English-only sentences, but its performance was inferior to that of CH-BERT and M-BERT in Chinese-only and code-mixed sentences, indicating that these types of sentences are more challenging for EN-BERT. In contrast, both CH-BERT and M-BERT exhibited less effect of the code-mixing issue in their recognition of PHI in code-mixed sentences.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Performance comparison for different methods on the test set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="100"/>
            <col width="0"/>
            <col width="50"/>
            <col width="0"/>
            <col width="50"/>
            <col width="0"/>
            <col width="50"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <col width="0"/>
            <col width="60"/>
            <thead>
              <tr valign="top">
                <td colspan="3">Protected health information type</td>
                <td colspan="6">DBA<sup>a</sup></td>
                <td colspan="7">EN-BERT<sup>b,c</sup></td>
                <td colspan="7">CH-BERT<sup>d</sup></td>
                <td colspan="7">M-BERT<sup>e</sup></td>
                <td colspan="5">TM-BERT<sup>f</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="3">
                  <break/>
                </td>
                <td colspan="2">P<sup>g</sup></td>
                <td colspan="2">R<sup>h</sup></td>
                <td colspan="2">F<sup>i</sup></td>
                <td colspan="2">P</td>
                <td colspan="2">R</td>
                <td colspan="2">F</td>
                <td colspan="3">P</td>
                <td colspan="2">R</td>
                <td colspan="2">F</td>
                <td colspan="3">P</td>
                <td colspan="2">R</td>
                <td colspan="2">F</td>
                <td colspan="3">P</td>
                <td colspan="2">R</td>
                <td>F</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="3">Date</td>
                <td colspan="2">0.07</td>
                <td colspan="2">0.19</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.97</td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="3">0.97</td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="2">0.97</td>
                <td colspan="3">0.98<sup>j</sup></td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="3">0.97</td>
                <td colspan="2">0.97</td>
                <td>0.97</td>
              </tr>
              <tr valign="top">
                <td colspan="3">Age</td>
                <td colspan="2">0.14</td>
                <td colspan="2">0.16</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.95<sup>j</sup></td>
                <td colspan="2">0.97</td>
                <td colspan="2">0.96<sup>j</sup></td>
                <td colspan="3">0.94</td>
                <td colspan="2">0.98<sup>j</sup></td>
                <td colspan="2">0.96<sup>j</sup></td>
                <td colspan="3">0.94</td>
                <td colspan="2">0.97</td>
                <td colspan="2">0.95</td>
                <td colspan="3">0.93</td>
                <td colspan="2">0.97</td>
                <td>0.95</td>
              </tr>
              <tr valign="top">
                <td colspan="3">
                  <bold>Name</bold>
                </td>
                <td colspan="2">0.06</td>
                <td colspan="2">0.14</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.62</td>
                <td colspan="2">0.47</td>
                <td colspan="2">0.51</td>
                <td colspan="3">0.82</td>
                <td colspan="2">0.70<sup>j</sup></td>
                <td colspan="2">0.74</td>
                <td colspan="3">0.87<sup>j</sup></td>
                <td colspan="2">0.68</td>
                <td colspan="2">0.75<sup>j</sup></td>
                <td colspan="3">0.73</td>
                <td colspan="2">0.59</td>
                <td>0.64</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Patient</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.05</td>
                <td colspan="2">0.59</td>
                <td colspan="2">0.41</td>
                <td colspan="2">0.48</td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">0.59<sup>j</sup></td>
                <td colspan="2">0.75<sup>j</sup></td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">0.59<sup>j</sup></td>
                <td colspan="2">0.75<sup>j</sup></td>
                <td colspan="3">0.94</td>
                <td colspan="2">0.59</td>
                <td colspan="2">0.72</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Person</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.06</td>
                <td colspan="2">0.36</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.17</td>
                <td colspan="3">0.54</td>
                <td colspan="2">0.58<sup>j</sup></td>
                <td colspan="2">0.56</td>
                <td colspan="3">0.66<sup>j</sup></td>
                <td colspan="2">0.53</td>
                <td colspan="2">0.56<sup>j</sup></td>
                <td colspan="3">0.30</td>
                <td colspan="2">0.25</td>
                <td colspan="2">0.27</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Doctor</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.22</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.92</td>
                <td colspan="2">0.92</td>
                <td colspan="2">0.92</td>
                <td colspan="3">0.94</td>
                <td colspan="2">0.92</td>
                <td colspan="2">0.93</td>
                <td colspan="3">0.95<sup>j</sup></td>
                <td colspan="2">0.93<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="3">
                  <bold>Location</bold>
                </td>
                <td colspan="2">0.14</td>
                <td colspan="2">0.20</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.64</td>
                <td colspan="2">0.56</td>
                <td colspan="2">0.58</td>
                <td colspan="3">0.79<sup>j</sup></td>
                <td colspan="2">0.69</td>
                <td colspan="2">0.72</td>
                <td colspan="3">0.78</td>
                <td colspan="2">0.74<sup>j</sup></td>
                <td colspan="2">0.75<sup>j</sup></td>
                <td colspan="3">0.67</td>
                <td colspan="2">0.71</td>
                <td>0.69</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Named location</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.34</td>
                <td colspan="2">0.26</td>
                <td colspan="2">0.29</td>
                <td colspan="3">0.64</td>
                <td colspan="2">0.37</td>
                <td colspan="2">0.47</td>
                <td colspan="3">0.61<sup>j</sup></td>
                <td colspan="2">0.57<sup>j</sup></td>
                <td colspan="2">0.59<sup>j</sup></td>
                <td colspan="3">0.34</td>
                <td colspan="2">0.34</td>
                <td colspan="2">0.34</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Nationality</td>
                <td colspan="2">0.43</td>
                <td colspan="2">0.14</td>
                <td colspan="2">0.21</td>
                <td colspan="2">0.67<sup>j</sup></td>
                <td colspan="2">0.19</td>
                <td colspan="2">0.30</td>
                <td colspan="3">0.44</td>
                <td colspan="2">0.52</td>
                <td colspan="2">0.48</td>
                <td colspan="3">0.53</td>
                <td colspan="2">0.48</td>
                <td colspan="2">0.50</td>
                <td colspan="3">0.48</td>
                <td colspan="2">0.53<sup>j</sup></td>
                <td colspan="2">0.50<sup>j</sup></td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Region</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.67<sup>j</sup></td>
                <td colspan="2">0.33<sup>j</sup></td>
                <td colspan="2">0.44<sup>j</sup></td>
                <td colspan="3">0.67<sup>j</sup></td>
                <td colspan="2">0.17</td>
                <td colspan="2">0.27</td>
                <td colspan="3">0.29</td>
                <td colspan="2">0.17</td>
                <td colspan="2">0.23</td>
                <td colspan="3">0.20</td>
                <td colspan="2">0.17</td>
                <td colspan="2">0.18</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Country</td>
                <td colspan="2">0.25</td>
                <td colspan="2">0.33</td>
                <td colspan="2">0.28</td>
                <td colspan="2">0.76</td>
                <td colspan="2">0.75</td>
                <td colspan="2">0.75</td>
                <td colspan="3">0.89</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.91<sup>j</sup></td>
                <td colspan="2">0.90<sup>j</sup></td>
                <td colspan="2">0.91<sup>j</sup></td>
                <td colspan="3">0.87</td>
                <td colspan="2">0.86</td>
                <td colspan="2">0.87</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>City</td>
                <td colspan="2">0.21</td>
                <td colspan="2">0.52</td>
                <td colspan="2">0.30</td>
                <td colspan="2">0.58</td>
                <td colspan="2">0.59</td>
                <td colspan="2">0.58</td>
                <td colspan="3">0.86</td>
                <td colspan="2">0.74</td>
                <td colspan="2">0.79</td>
                <td colspan="3">0.86<sup>j</sup></td>
                <td colspan="2">0.77<sup>j</sup></td>
                <td colspan="2">0.82<sup>j</sup></td>
                <td colspan="3">0.77</td>
                <td colspan="2">0.73</td>
                <td colspan="2">0.75</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Hospital</td>
                <td colspan="2">0.35</td>
                <td colspan="2">0.47</td>
                <td colspan="2">0.40</td>
                <td colspan="2">0.84</td>
                <td colspan="2">0.85</td>
                <td colspan="2">0.84</td>
                <td colspan="3">0.89</td>
                <td colspan="2">0.94</td>
                <td colspan="2">0.91</td>
                <td colspan="3">0.93<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
                <td colspan="3">0.90</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.90</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Department</td>
                <td colspan="2">0.24</td>
                <td colspan="2">0.35</td>
                <td colspan="2">0.29</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.87</td>
                <td colspan="2">0.88</td>
                <td colspan="3">0.92</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.90</td>
                <td colspan="3">0.93<sup>j</sup></td>
                <td colspan="2">0.90<sup>j</sup></td>
                <td colspan="2">0.92<sup>j</sup></td>
                <td colspan="3">0.90</td>
                <td colspan="2">0.87</td>
                <td colspan="2">0.89</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Room</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.06</td>
                <td colspan="2">0.87</td>
                <td colspan="2">0.92</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.89</td>
                <td colspan="2">0.94</td>
                <td colspan="2">0.91</td>
                <td colspan="3">0.91<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
                <td colspan="2">0.92<sup>j</sup></td>
                <td colspan="3">0.77</td>
                <td colspan="2">0.94</td>
                <td colspan="2">0.85</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Number</td>
                <td colspan="2">0.05</td>
                <td colspan="2">0.08</td>
                <td colspan="2">0.06</td>
                <td colspan="2">0.99</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.99</td>
                <td colspan="3">0.99</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.99</td>
                <td colspan="3">0.99</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>School</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.01</td>
                <td colspan="2">0.64</td>
                <td colspan="2">0.66</td>
                <td colspan="2">0.65</td>
                <td colspan="3">0.91<sup>j</sup></td>
                <td colspan="2">0.82</td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.85</td>
                <td colspan="2">0.87<sup>j</sup></td>
                <td colspan="2">0.86<sup>j</sup></td>
                <td colspan="3">0.76</td>
                <td colspan="2">0.82</td>
                <td colspan="2">0.79</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>General business</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.16</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.42</td>
                <td colspan="2">0.31</td>
                <td colspan="2">0.36</td>
                <td colspan="3">0.66</td>
                <td colspan="2">0.64</td>
                <td colspan="2">0.65</td>
                <td colspan="3">0.70<sup>j</sup></td>
                <td colspan="2">0.73<sup>j</sup></td>
                <td colspan="2">0.71<sup>j</sup></td>
                <td colspan="3">0.61</td>
                <td colspan="2">0.61</td>
                <td colspan="2">0.61</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Market</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.11</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="3">0.67</td>
                <td colspan="2">0.40</td>
                <td colspan="2">0.50</td>
                <td colspan="3">0.79<sup>j</sup></td>
                <td colspan="2">0.55</td>
                <td colspan="2">0.65<sup>j</sup></td>
                <td colspan="3">0.45</td>
                <td colspan="2">0.75<sup>j</sup></td>
                <td colspan="2">0.57</td>
              </tr>
              <tr valign="top">
                <td colspan="3">Profession</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.22</td>
                <td colspan="2">0.12</td>
                <td colspan="2">0.46</td>
                <td colspan="2">0.43</td>
                <td colspan="2">0.44</td>
                <td colspan="3">0.66</td>
                <td colspan="2">0.73</td>
                <td colspan="2">0.69</td>
                <td colspan="3">0.78<sup>j</sup></td>
                <td colspan="2">0.76<sup>j</sup></td>
                <td colspan="2">0.77<sup>j</sup></td>
                <td colspan="3">0.59</td>
                <td colspan="2">0.72</td>
                <td>0.65</td>
              </tr>
              <tr valign="top">
                <td colspan="3">
                  <bold>ID</bold>
                </td>
                <td colspan="2">0.34</td>
                <td colspan="2">0.31</td>
                <td colspan="2">0.32</td>
                <td colspan="2">0.75</td>
                <td colspan="2">0.96</td>
                <td colspan="2">0.83</td>
                <td colspan="3">0.79</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.81<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.90<sup>j</sup></td>
                <td colspan="3">0.77</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td>0.85</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>ID number</td>
                <td colspan="2">0.67</td>
                <td colspan="2">0.62</td>
                <td colspan="2">0.64</td>
                <td colspan="2">0.55</td>
                <td colspan="2">0.92</td>
                <td colspan="2">0.69</td>
                <td colspan="3">0.57</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.72</td>
                <td colspan="3">0.65<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.79<sup>j</sup></td>
                <td colspan="3">0.53</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.70</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Medical record</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.00</td>
                <td colspan="2">0.94</td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">0.97</td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="3">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
                <td colspan="2">1.00<sup>j</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="3">Micro-Average</td>
                <td colspan="2">0.09</td>
                <td colspan="2">0.22</td>
                <td colspan="2">0.13</td>
                <td colspan="2">0.91</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.90</td>
                <td colspan="3">0.93</td>
                <td colspan="2">0.93</td>
                <td colspan="2">0.93</td>
                <td colspan="3">0.94<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
                <td colspan="2">0.94<sup>j</sup></td>
                <td colspan="3">0.90</td>
                <td colspan="2">0.92</td>
                <td>0.91</td>
              </tr>
              <tr valign="top">
                <td colspan="3">Macro-Average</td>
                <td colspan="2">0.15</td>
                <td colspan="2">0.20</td>
                <td colspan="2">0.16</td>
                <td colspan="2">0.67</td>
                <td colspan="2">0.62</td>
                <td colspan="2">0.63</td>
                <td colspan="3">0.80</td>
                <td colspan="2">0.75</td>
                <td colspan="2">0.76</td>
                <td colspan="3">0.81<sup>j</sup></td>
                <td colspan="2">0.78<sup>j</sup></td>
                <td colspan="2">0.79<sup>j</sup></td>
                <td colspan="3">0.71</td>
                <td colspan="2">0.75</td>
                <td> 0.72</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>DBA: dictionary-based approach.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>EN-BERT: BERT pretrained on English corpora.</p>
            </fn>
            <fn id="table3fn4">
              <p><sup>d</sup>CH-BERT: BERT pretrained on simplified and traditional Chinese corpora.</p>
            </fn>
            <fn id="table3fn5">
              <p><sup>e</sup>M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
            </fn>
            <fn id="table3fn6">
              <p><sup>f</sup>TM-BERT: M-BERT fine-tuned on the translated deidentification corpus.</p>
            </fn>
            <fn id="table3fn7">
              <p><sup>g</sup>P refers to precision.</p>
            </fn>
            <fn id="table3fn8">
              <p><sup>h</sup>R refers to recall.</p>
            </fn>
            <fn id="table3fn9">
              <p><sup>i</sup>F refers to F1-score.</p>
            </fn>
            <fn id="table3fn10">
              <p><sup>j</sup>The highest score among different methods.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Comparison of performance between EN-BERT, CH-BERT, and M-BERT in English-only, Chinese-only, and Chinese-English mixed sentences.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="110"/>
            <col width="70"/>
            <col width="70"/>
            <col width="70"/>
            <col width="80"/>
            <col width="0"/>
            <col width="70"/>
            <col width="70"/>
            <col width="70"/>
            <col width="90"/>
            <col width="0"/>
            <col width="60"/>
            <col width="60"/>
            <col width="60"/>
            <col width="90"/>
            <thead>
              <tr valign="top">
                <td colspan="2">PHI<sup>a</sup> type</td>
                <td colspan="5">English-only sentences</td>
                <td colspan="5">Chinese-only sentences</td>
                <td colspan="4">Chinese-English mixed sentences</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  <break/>
                </td>
                <td>EN-BERT<sup>b,c</sup></td>
                <td>CH-BERT<sup>d</sup></td>
                <td>M-BERT<sup>e</sup></td>
                <td>PHI, n<sup>f</sup></td>
                <td colspan="2">EN-BERT</td>
                <td>CH-BERT</td>
                <td>M-BERT</td>
                <td>PHI, n<sup>f</sup></td>
                <td colspan="2">EN-BERT</td>
                <td>CH-BERT</td>
                <td>M-BERT</td>
                <td>PHI, n<sup>f</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="2">Date<sup>g</sup></td>
                <td>0.96<sup>h</sup></td>
                <td>0.94</td>
                <td>0.95<sup>h</sup></td>
                <td>2346<sup>h</sup></td>
                <td colspan="2">0.98</td>
                <td>0.97</td>
                <td>0.99<sup>h</sup></td>
                <td>458<sup>h</sup></td>
                <td colspan="2">0.93</td>
                <td>0.93</td>
                <td>0.94<sup>h</sup></td>
                <td>919<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="2">Age<sup>g</sup></td>
                <td>0.97<sup>h</sup></td>
                <td>0.96</td>
                <td>0.96</td>
                <td>533<sup>h</sup></td>
                <td colspan="2">0.67</td>
                <td>0.91<sup>h</sup></td>
                <td>0.73</td>
                <td>5<sup>h</sup></td>
                <td colspan="2">0.94</td>
                <td>0.97<sup>h</sup></td>
                <td>0.95</td>
                <td>128<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>Name<sup>g</sup></bold>
                </td>
                <td>0.24<sup>h</sup></td>
                <td>0.24<sup>h</sup></td>
                <td>0.24<sup>h</sup></td>
                <td>18<sup>h</sup></td>
                <td colspan="2">0.38</td>
                <td>0.54<sup>h</sup></td>
                <td>0.53</td>
                <td>44<sup>h</sup></td>
                <td colspan="2">0.54</td>
                <td>0.80<sup>h</sup></td>
                <td>0.78</td>
                <td>240<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Patient</td>
                <td>0.00</td>
                <td>0.00</td>
                <td>0.00</td>
                <td>9</td>
                <td colspan="2">0.64</td>
                <td>1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>12</td>
                <td colspan="2">0.50</td>
                <td>0.78</td>
                <td>0.88<sup>h</sup></td>
                <td>11</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Person</td>
                <td>N/A<sup>i</sup></td>
                <td>N/A</td>
                <td>N/A</td>
                <td>0</td>
                <td colspan="2">0.00</td>
                <td>0.00</td>
                <td>0.00</td>
                <td>1</td>
                <td colspan="2">0.18</td>
                <td>0.64<sup>h</sup></td>
                <td>0.48</td>
                <td>32</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Doctor</td>
                <td>0.94<sup>h</sup></td>
                <td>0.94<sup>h</sup></td>
                <td>0.94<sup>h</sup></td>
                <td>9</td>
                <td colspan="2">0.49</td>
                <td>0.61<sup>h</sup></td>
                <td>0.60</td>
                <td>31</td>
                <td colspan="2">0.94</td>
                <td>0.97</td>
                <td>0.98<sup>h</sup></td>
                <td>197</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>Location<sup>g</sup></bold>
                </td>
                <td>0.66</td>
                <td>0.61</td>
                <td>0.70<sup>h</sup></td>
                <td>1014<sup>h</sup></td>
                <td colspan="2">0.36</td>
                <td>0.87<sup>h</sup></td>
                <td>0.83</td>
                <td>1108<sup>h</sup></td>
                <td colspan="2">0.37</td>
                <td>0.66<sup>h</sup></td>
                <td>0.65</td>
                <td>1132<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Named location</td>
                <td>0.47</td>
                <td>0.25</td>
                <td>0.62<sup>h</sup></td>
                <td>10</td>
                <td colspan="2">0.00</td>
                <td>0.40<sup>h</sup></td>
                <td>0.40<sup>h</sup></td>
                <td>3</td>
                <td colspan="2">0.28</td>
                <td>0.53<sup>h</sup></td>
                <td>0.52</td>
                <td>69</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Nationality</td>
                <td>0.67<sup>h</sup></td>
                <td>0.38</td>
                <td>0.67<sup>h</sup></td>
                <td>7</td>
                <td colspan="2">0.00</td>
                <td>1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>1</td>
                <td colspan="2">0.00</td>
                <td>0.86<sup>h</sup></td>
                <td>0.78</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Region</td>
                <td>0.47</td>
                <td>0.55<sup>h</sup></td>
                <td>0.50</td>
                <td>11</td>
                <td colspan="2">N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">0.00</td>
                <td>0.00</td>
                <td>0.00</td>
                <td>1</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Country</td>
                <td>0.87</td>
                <td>0.85</td>
                <td>0.95<sup>h</sup></td>
                <td>52</td>
                <td colspan="2">0.25</td>
                <td>1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>5</td>
                <td colspan="2">0.63</td>
                <td>0.94<sup>h</sup></td>
                <td>0.88</td>
                <td>44</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>City</td>
                <td>0.74</td>
                <td>0.72</td>
                <td>0.77<sup>h</sup></td>
                <td>39</td>
                <td colspan="2">0.22</td>
                <td>1.00<sup>h</sup></td>
                <td>0.92</td>
                <td>6</td>
                <td colspan="2">0.53</td>
                <td>0.85<sup>h</sup></td>
                <td>0.82</td>
                <td>101</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Hospital</td>
                <td>0.93<sup>h</sup></td>
                <td>0.92</td>
                <td>0.93<sup>h</sup></td>
                <td>266</td>
                <td colspan="2">0.21</td>
                <td>0.80</td>
                <td>0.83<sup>h</sup></td>
                <td>11</td>
                <td colspan="2">0.79</td>
                <td>0.92</td>
                <td>0.93<sup>h</sup></td>
                <td>393</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Department</td>
                <td>0.82</td>
                <td>0. 81</td>
                <td>0.82<sup>h</sup></td>
                <td>391</td>
                <td colspan="2">0.95</td>
                <td>0.96<sup>h</sup></td>
                <td>0.96<sup>h</sup></td>
                <td>542</td>
                <td colspan="2">0.68</td>
                <td>0.76</td>
                <td>0.77<sup>h</sup></td>
                <td>111</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Room</td>
                <td>0.85</td>
                <td>0.83</td>
                <td>0.85<sup>h</sup></td>
                <td>182</td>
                <td colspan="2">0.97</td>
                <td>0.98<sup>h</sup></td>
                <td>0.97</td>
                <td>267</td>
                <td colspan="2">0.68</td>
                <td>0.76<sup>h</sup></td>
                <td>0.76</td>
                <td>53</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Number</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">0.99</td>
                <td>1.00<sup>h</sup></td>
                <td>0.99</td>
                <td>256</td>
                <td colspan="2">0.00</td>
                <td>0.00</td>
                <td>0.00</td>
                <td>1</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>School</td>
                <td>0.71</td>
                <td>0.25</td>
                <td>0.83<sup>h</sup></td>
                <td>6</td>
                <td colspan="2">0.00</td>
                <td>0.67<sup>h</sup></td>
                <td>0.50</td>
                <td>2</td>
                <td colspan="2">0.65</td>
                <td>0.92<sup>h</sup></td>
                <td>0.90</td>
                <td>159</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Generic business</td>
                <td>0.72<sup>h</sup></td>
                <td>0.59</td>
                <td>0.70</td>
                <td>45</td>
                <td colspan="2">0.00</td>
                <td>0.89<sup>h</sup></td>
                <td>0.72</td>
                <td>15</td>
                <td colspan="2">0.27</td>
                <td>0.66</td>
                <td>0.71<sup>h</sup></td>
                <td>173</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Market</td>
                <td>0.00</td>
                <td>0.50<sup>h</sup></td>
                <td>0.00</td>
                <td>3</td>
                <td colspan="2">N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">0.00</td>
                <td>0.67</td>
                <td>0.73<sup>h</sup></td>
                <td>17</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Profession<sup>g</sup></td>
                <td>0.78</td>
                <td>0.71</td>
                <td>0.81<sup>h</sup></td>
                <td>107<sup>h</sup></td>
                <td colspan="2">0.17</td>
                <td>0.67<sup>h</sup></td>
                <td>0. 61</td>
                <td>9<sup>h</sup></td>
                <td colspan="2">0.36</td>
                <td>0.70</td>
                <td>0.75<sup>h</sup></td>
                <td>449<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>ID<sup>g</sup></bold>
                </td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">0.90</td>
                <td>0.90</td>
                <td>0.91<sup>h</sup></td>
                <td>26<sup>h</sup></td>
                <td colspan="2">0.60</td>
                <td>0.90<sup>h</sup></td>
                <td>0.75</td>
                <td>5<sup>h</sup></td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>ID number</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">0.79</td>
                <td>0.79</td>
                <td>0.81<sup>h</sup></td>
                <td>11</td>
                <td colspan="2">0.33</td>
                <td>0.80<sup>h</sup></td>
                <td>0.50</td>
                <td>2</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Medical record</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td>N/A</td>
                <td colspan="2">1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>14</td>
                <td colspan="2">0.86</td>
                <td>1.00<sup>h</sup></td>
                <td>1.00<sup>h</sup></td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Micro-F-score</td>
                <td>0.92</td>
                <td>0.90</td>
                <td>0.92<sup>h</sup></td>
                <td>4018</td>
                <td colspan="2">0.93</td>
                <td>0.96<sup>h</sup></td>
                <td>0.96<sup>h</sup></td>
                <td>1650</td>
                <td colspan="2">0.71</td>
                <td>0.85</td>
                <td>0.87<sup>h</sup></td>
                <td>2873</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Macro-F-score</td>
                <td>0.60<sup>h</sup></td>
                <td>0.54</td>
                <td>0.60<sup>h</sup></td>
                <td>4018</td>
                <td colspan="2">0.47</td>
                <td>0.83<sup>h</sup></td>
                <td>0.79</td>
                <td>1650</td>
                <td colspan="2">0.48</td>
                <td>0.74<sup>h</sup></td>
                <td>0.72</td>
                <td>2873</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>PHI: protected health information.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>EN-BERT: BERT pretrained on English corpora.</p>
            </fn>
            <fn id="table4fn4">
              <p><sup>d</sup>CH-BERT: BERT pretrained on simplified and traditional Chinese corpora.</p>
            </fn>
            <fn id="table4fn5">
              <p><sup>e</sup>M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
            </fn>
            <fn id="table4fn6">
              <p><sup>f</sup>The number of training instances of the PHI type.</p>
            </fn>
            <fn id="table4fn7">
              <p><sup>g</sup>Coarse-grained PHI.</p>
            </fn>
            <fn id="table4fn8">
              <p><sup>h</sup>The highest F-score among the 3 methods.</p>
            </fn>
            <fn id="table4fn9">
              <p><sup>i</sup>N/A: not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Impact of the CMI Level on Deidentification Performance</title>
        <p>In this subsection, we report sentence error rates by dividing sentences in our corpus into 6 groups according to their CMI levels, as illustrated in <xref rid="figure3" ref-type="fig">Figure 3</xref>, to gain insights into the impact of the code-mixing level on PHI recognition. It is evident from the figure that EN-BERT has the highest error rate for all CMI ranges. This aligns with our earlier finding in the previous section demonstrating that EN-BERT struggles with recognizing PHI in code-mixed sentences. Although CH-BERT and M-BERT exhibited lower sentence error rates than EN-BERT, the error rates were still considerably higher for code-mixed sentences (CMI &#62;0), indicating that the recognition of PHI in code-mixed sentences remains a challenging task for the 2 PLMs.</p>
        <p>Further evidence supporting this conclusion can be found in <xref ref-type="table" rid="table5">Table 5</xref>, where we fine-tuned M-BERT on training instances from 1 sentence category and evaluated it on the test set categorized into the 3 categories. From the results, we observed that fine-tuning M-BERT with training instances from a specific sentence category can significantly improve its performance on that category. However, fine-tuning on the code-mixed training instances yielded the best results, which further enabled the model to effectively recognize PHI in both Chinese and English sentences. This finding emphasizes the significance of incorporating code-mixed training instances into the model’s training data.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Sentence error rate in each code-mixing index range for the test set. The numbers in parentheses represent the number of incorrect examples for the corresponding code-mixing index range. CH-BERT: BERT pretrained on simplified and traditional Chinese corpora; EN-BERT: BERT pretrained on English corpora; M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>The “zero-shot” performance of BERT pretrained on Wikipedia corpora from 104 languages (M-BERT) for the 3 sentence categories in terms of micro-F-scores.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="320"/>
            <col width="180"/>
            <col width="180"/>
            <col width="320"/>
            <thead>
              <tr valign="top">
                <td>Evaluation/Fine-tuning</td>
                <td>English-only</td>
                <td>Chinese-only</td>
                <td>Chinese-English mixed</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>English-only</td>
                <td>0.93</td>
                <td>0.42</td>
                <td>0.57</td>
              </tr>
              <tr valign="top">
                <td>Chinese-only</td>
                <td>0.59</td>
                <td>0.96</td>
                <td>0.65</td>
              </tr>
              <tr valign="top">
                <td>Chinese-English mixed</td>
                <td>0.80</td>
                <td>0.94</td>
                <td>0.87</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>Performance Comparison With ChatGPT</title>
        <p>To assess the effectiveness of the proposed ChatGPT-based framework for recognizing PHI in our code-mixed corpus, we collected a subset of the original data that included sentences containing one or more PHI sets that were not recognized by any of the BERT-based models we developed. This subset was chosen to represent the most challenging cases of the compiled corpus while minimizing the risk of sensitive information exposure. <xref rid="figure4" ref-type="fig">Figure 4</xref> shows the results, with M-BERT used for comparison.</p>
        <p>Compared with M-BERT, the proposed ChatGPT-based framework achieved slightly better recall, but lower precision, leading to a lower micro-F-score in the zero-shot setting. Nonetheless, our framework presents an appealing approach to use LLMs for the deidentification task over standard-sized models like BERT. LLMs are notably easier to use and can be controlled through natural language prompts, requiring little to no machine learning expertise. For example, we found only 1 error case for the “Doctor” PHI that both BERT-based models and ChatGPT failed to recognize: “follow-up: (ydh/ptz).” However, we can enhance the prompt shown in <xref rid="figure2" ref-type="fig">Figure 2</xref> by including a statement following the “DOCTOR NAME” definition that lists the abbreviation names of physicians in a hospital, directing ChatGPT to recognize “ptz” as a doctor name: “Here is a list of abbreviated physician names of our hospital. Remember and recognize them from the given text: 1. hjd, 2. ptz, 3. ctc.” This exemplifies the adaptability and flexibility of LLMs for deidentification tasks, making them particularly well-suited for real hospital environments. While the potential for using ChatGPT in deidentification is notable, the model is currently only accessible through an online application programming interface. This renders it unsuitable for use in a hospital setting, where patient data cannot be stored or transmitted to unauthorized external parties.</p>
        <p>On the other hand, it crucially highlights the significance of carefully crafted prompts when working with ChatGPT. The constraint statements presented in <xref rid="figure1" ref-type="fig">Figure 1</xref> play a pivotal role in guiding ChatGPT’s responses. When these constraints were removed, we observed that ChatGPT can generate responses that may be unpredictable and go beyond the desired control. Take the text “he was brought to Dr. 莊凱傑’s opd on 3/13 and haldol + anxicam 1 amp was given” as an example. Here, “莊凱傑” represents a Chinese name, which can be translated to “KAI-JIE ZHUANG.” Using the developed prompts without adding the constraint part, both the original and recently released ChatGPT 3.5 models (version from September 27, 2023) could potentially return responses like “MEDICATION: haldol, anxicam” and “DOSE: 1 amp.”</p>
        <p>Another issue to consider is the translation of recognized PHI into a language other than that mentioned in the given text. The issue is more prominent in Chinese-majority code-mixed sentences, where the recognized Chinese PHI might be translated into the English counterparts. For example, in the sentence “education: 高中讀中和高中, 僅唸數個月就因跟不上而休學, 大學念正修夜間部商學” (education: high school-Zhonghe senior high school, dropped out after only a few months; College-Business in the night school at Cheng Shiu university), the Chinese PHI could be extracted as translated forms in English, which can lead to a problem in surrogate generation and affect the performance evaluation procedure.</p>
        <p>Lastly, similar to the observations in previous work [<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref51">51</xref>], we found that the practice of requesting ChatGPT to generate responses 3 times resulted in the improvement of the overall precision, recall, and F1-score by 0.055, 0.068, and 0.064, respectively. The strategy can effectively filter out noisy outputs generated by ChatGPT. For instance, even with the inclusion of constraint statements in our prompts, we noticed that ChatGPT occasionally generated responses that violated the defined constraints, introducing undesired entity types such as “MEDICATION,” “DOSE,” or “DIAGNOSIS.” By implementing the 3-query strategy, we could successfully eliminate these cases, ensuring the quality and relevance for our guideline of the generated content. Further details regarding the limitations for this strategy and its implications are provided in the limitations section.</p>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Performance comparison with the ChatGPT-based deidentification framework and M-BERT on the sampled test subset. M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Ambiguity Issue Among PHI</title>
        <p>We explored the issue of ambiguity in the classification of fine-grained PHI at the coarse-grained level. When fine-grained PHI is misclassified, it may belong to the same coarse-grained PHI or different ones. Misclassifying it under the same coarse-grained PHI is generally more acceptable since fine-grained PHI belonging to the same coarse-grained category tends to share similar semantic characteristics. To quantify the extent of ambiguity, we calculated the micro-F-scores at the coarse-grained level, treating an incorrect recognition of a fine-grained PHI type as a true positive if it belongs to the same coarse-grained PHI type as the corresponding gold annotation. Our results (detailed in <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>) showed that intratype ambiguities are more prevalent than intertype ambiguities in “Name” and “Location” PHI types.</p>
        <p>To gain insights into the level of ambiguity among the “Location” fine-grained types, we visualized the representations generated by M-BERT for different “Location” PHI types through t-SNE. The resulting plot (<xref rid="figure5" ref-type="fig">Figure 5</xref>) revealed that the clusters of “Nationality,” “Country,” and “Region” are intertwined, while the cluster of “ID” is notably separated from the others. The high ambiguity may be due to the limited number of training instances, as the number of training instances is imbalanced among PHI as illustrated in <xref ref-type="table" rid="table2">Table 2</xref>. As a result, all BERT-based models yielded better F-scores for “Country” than for “Nationality” and “Region.”</p>
        <p>On the other hand, the “ID” PHI type had a perfect mapped recall, indicating that all of the coarse-grained PHI could be recognized by the developed models. The confusion matrix shown in Figure S1 in <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref> reveals only 1 ambiguous case where M-BERT misclassified the “Location” PHI type as the “ID” PHI type. This occurred in the sentence “She visited psychiatric clinic of 804 then but have neither regular follow-up nor fair drug compliance,” where “804” refers to the Military Taoyuan General Hospital (804 is the army number assigned to the hospital), but the model recognized it as “ID” PHI. This example demonstrates one of the most challenging cases where the model failed at recognition, even with the recall-oriented ChatGPT framework. To correctly classify the instance, the model requires additional domain knowledge. We attempted to use ChatGPT to determine whether it knows that “804 hospital” refers to the military hospital, but it replied that “804 hospital” was not a common name or official name for the hospital. However, a quick Google search revealed that the website of the hospital was the first suggested link, indicating the importance of external knowledge sources for accurate classification [<xref ref-type="bibr" rid="ref56">56</xref>].</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>t-Distributed Stochastic Neighbor Embedding visualization of the 5 fine-grained “Location” protected health information (PHI) types and the “ID” PHI type for M-BERT. M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>OOV Case Study to Interpret the Models’ Decisions</title>
        <p>This section delves into how the BERT-based models use regularity and mention coverage, as outlined in the Introduction section, during their recognition process. First, we estimated the OOV rates on the test set to determine the mention coverage rate of our corpus. The estimated rates were 0.356, 0.953, 0.686, 0.876, 0.696, and 0.290 for the “Date,” “Age,” “Name,” “Location,” “Profession,” and “ID” categories, respectively. OOV was defined as a mention containing at least one word not present in the vocabulary compiled from the training data set. The inadequate performance of the DBA along with the high OOV rate suggests that the effect of the mention coverage in our corpus is limited. To analyze the effect, we re-estimated the performance of coarse-grained PHI for tokens belonging to in-vocabulary (IV) and OOV using EN-BERT, M-BERT, and TM-BERT in <xref rid="figure6" ref-type="fig">Figure 6</xref>. More detailed results are available in <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>.</p>
        <p>The presented results reveal that the developed models have different levels of tolerance for OOV issues when recognizing different PHI types. Interestingly, we observed that the EN-BERT model achieved indistinguishable high F-scores for IV and OOV tokens for the “Date” PHI type. This is notable since not all Chinese characters used in the “Data” PHI are presented in the EN-BERT vocabulary. For example, the numeric value “六” (six) is not available as shown in <xref ref-type="table" rid="table6">Table 6</xref>. We believe that the presence of those characters in the extensive training data enabled EN-BERT to leverage the naming regularity of the “Date” PHI and improve its ability to identify occurrences of these entities in Chinese-only and code-mixed sentences. <xref rid="figure7" ref-type="fig">Figure 7</xref> illustrates how EN-BERT’s attention mechanism can use specific patterns in text, such as “年月日,” for recognizing “Date” PHI, even when the majority of the Chinese tokens are mapped to [UNK] (unknown word) by the EN-BERT tokenizer. The model’s attention was primarily drawn to the pattern words “年” and “月,” allowing it to successfully assign labels to the respective unknown words.</p>
        <p>However, the model’s overreliance on pattern words and strong name regularities could lead to misclassification. <xref rid="figure8" ref-type="fig">Figure 8</xref> illustrates an example of the false-positive case generated by EN-BERT. The model has been trained to recognize the first likely token for “Date” PHI by considering its surrounding tokens, including previous unidentified tokens, the forward slash (/), and the final token representing the end of the “Date” PHI. However, this mechanism caused the model to incorrectly extract the text “5/4/3” as “Date” PHI. The third AHV displayed in <xref rid="figure4" ref-type="fig">Figure 4</xref> illustrates that when the previous tokens are not unknown, the model focuses more on the preceding token (a number) and gives less attention to the pattern tokens (“/”), which helps prevent the occurrence of false positives.</p>
        <p>On the other hand, the “Age,” “Location,” and “Profession” PHI types exhibited higher IV F-scores compared to their OOV scores. Specifically, the “Age” PHI type demonstrated an almost perfect IV F-score and slightly lower OOV score across all 3 models. After examining the outcomes of the developed models, we arrived at a conclusion similar to that of the “Date” PHI. The models could distinguish the “Age” PHI type in various code-mixing–level sentences because of the consistent naming pattern. There were only a few cases in which the developed models failed. For instance, in the sentence, “occupation: 會計18y-&#62;美容師3y-&#62;保險電話銷售人員-&#62;未上市股票電話銷售員-&#62;賣雞排 (2223.3-2224.9),” only CH-BERT was able to correctly avoid classifying “18” as “Age” PHI, while the other BERT-based models and the developed ChatGPT framework recognized “18” as “Age” PHI.</p>
        <p>In contrast, the “Location” and “Profession” PHI types exhibited lower IV F-scores and apparently lower OOV F-scores. To investigate the discrepancies, we probed the results of the developed model by applying input reduction. We took the input sentence “travel history: [PHI] last week” as an example to summarize the observations as follows. First, we noticed that for EN-BERT, the presence of unknown tokens in the input sentence resulted in heavy reliance on the regularity of entity naming and contextual patterns for effective generalization over unseen mentions or mentions that contain unknown words. In the left part of <xref rid="figure9" ref-type="fig">Figure 9</xref>, the AHV represents the reduced interpretation of the 2 “City” PHI inputs “台北” (Taipei) and “恆春” (Hengchun) that contain unknown words. From the view, we can see that the preceding token, “travel,” provides a strong indication for EN-BERT to recognize “City” PHI. The word “台北” (“[UNK]北” for EN-BERT) was considered as “Hospital” PHI, indicating that the model learned to memorize the most frequent label assigned to the IV word, such as “署/[UNK]北” hospitals, in the training set.</p>
        <p>Second, in addition to the naming regularity and contextual patterns, CH-BERT and M-BERT can leverage their representation for the given token sequence to assign the corresponding labels better. This is evident from the right part of <xref rid="figure5" ref-type="fig">Figure 5</xref> in which we can see after fine-tuning, M-BERT was able to improve its representation of “Hospital” PHI entities, such as “ntuh,” “台大” (National Taiwan University Hospital), and “署北” (Taipei Hospital), as well as “Date” PHI entities, such as “九月” (September) and “mk87,” by clustering them in closer proximity to each other. Despite being OOV entities, OOV PHI, such as “mk99,” “菜園” (vegetable garden), “軍醫院” (military hospital), and “小診所” (small clinic), were still represented in close proximity to the corresponding IV PHI, such as “mk87,” “clothing factory,” and “療養院” (nursing home). However, “Profession” PHI and some of the fine-grained “Location” PHI types, such as “Generic location,” lacked the regularity of their naming conventions, which posed a challenge for accurate recognition. Moreover, informal language, typographical errors, ambiguity issues, and limited training instances for fine-grained PHI, such as “Region” and “Market,” further complicate the task of correctly identifying these OOV PHI mentions.</p>
        <p>To mitigate the aforementioned OOV issue, we highlight several promising research directions. The first direction is data augmentation. During the pretraining phases, BERT-based models can benefit from automatically generating synthetic training examples that include PHI. This can be accomplished by leveraging LLMs to create data of reasonable quality [<xref ref-type="bibr" rid="ref57">57</xref>]. This approach, similar to data augmentation, should help improve the model’s robustness when handling different contexts and OOV PHI. The second direction is vocabulary expansion. For specific categories like “Location” and “Profession,” where we observed lower F-scores on OOV PHI, traditional dictionary expansion methods [<xref ref-type="bibr" rid="ref58">58</xref>] can be applied to augment BERT’s vocabulary. This approach empowers the model to recognize and process OOV and UNK words more effectively. The third direction is knowledge graph integration [<xref ref-type="bibr" rid="ref59">59</xref>]. This approach enriches the contextual information available to the model, providing additional context and semantics to assist in recognizing OOV or UNK PHI and thereby improving overall performance.</p>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>F-scores of coarse-grained protected health information (PHI) for tokens belonging to in-vocabulary (IV) and out-of-vocabulary (OOV) in the test set. EN-BERT: BERT pretrained on English corpora; M-BERT: BERT pretrained on Wikipedia corpora from 104 languages; TM-BERT: M-BERT fine-tuned on the translated deidentification corpus.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Examples of Chinese characters related to the “Date” protected health information listed in the BERT pretrained on English corpora (EN-BERT) vocabulary.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Chinese character</td>
                <td>Example “Date” PHI<sup>a</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>清、明、中、秋</td>
                <td>清明節、中秋節</td>
              </tr>
              <tr valign="top">
                <td>一、二、三、四、五、八、十、月、日、年、上、下、年、西、星、民、國</td>
                <td>三月八日、上周五、下周一、西元2321年、星期五、民國一一二年、今年十一月</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>PHI: protected health information.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>The attention-head view for the following code-mixed sentence: “she tried to hang herself but failed due to drunk (大約前年五月開始不斷有死的念頭)” with the tokens “[UNK]” before/after “年 (year)” selected. The English translation for the Chinese text is “she had suicidal thoughts since around May of two years ago.” [UNK] is the token for unknown word. [CLS] is a classification token and is added to the beginning of every sentence. [SEP] is a separation token inserted in between 2 sentences.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure8" position="float">
          <label>Figure 8</label>
          <caption>
            <p>The first and second attention-head views have been generated for EN-BERT when applied to process the following sentences: “他在 78/12/15 求診” (sampled from the training set; the English translation is “He went to see a doctor on 78/12/15”; “78/12/15” is annotated as “Date” PHI) and “3. ocd (不時想到數字5/4/3/8)” (sampled from the test set; the English translation is “sometimes thinks of the numbers 5/4/3/8”). The third view has been generated for EN-BERT when processing the translated text of the second sentence. [UNK] is the token for unknown word. [CLS] is a classification token and is added to the beginning of every sentence. [SEP] is a separation token inserted in between 2 sentences. EN-BERT: BERT pretrained on English corpora; PHI: protected health information.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure9" position="float">
          <label>Figure 9</label>
          <caption>
            <p>The left part shows the attention-head views (AHVs) of the input reduction probing for the input text “travel history: [PHI] last week.” The first column of the AHVs comprises 3 rows, wherein the PHI includes “台北” and “恆春.” We also present the result of the input “台北” in the last row of the first column in the view. The right part shows the t-Distributed Stochastic Neighbor Embedding visualizations of EN-BERT, M-BERT without fine-tuning, and M-BERT with fine-tuning from top to bottom for randomly sampled in-vocabulary (IV) and out-of-vocabulary (OOV) mentions of the 3 fine-grained “Location” PHI types including “Named location,” “Generic location,” and “Hospital.” For comparison, we also include 5 “Date” PHI instances including “過年” (New Year), “November,” “九月” (September), “mk87,” and “mk99” (where “mk” is a transliteration of “民國,” which refers to the Republic of China calendar). [UNK] is the token for unknown word. [CLS] is a classification token and is added to the beginning of every sentence. [SEP] is a separation token inserted in between 2 sentences. PHI: protected health information; EN-BERT: BERT pretrained on English corpora; M-BERT: BERT pretrained on Wikipedia corpora from 104 languages.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e48443_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>While our study has made significant contributions to the field of clinical text deidentification in code-mixed languages, it is important to acknowledge some limitations. First, despite the successes achieved, M-BERT, our best-performing model, has some limitations. Notably, it exhibits intratype ambiguities in certain PHI, such as “Location,” likely due to the limited number of training instances. Second, M-BERT yields decreased performance for OOV terms compared to IV terms in specific PHI categories, such as “Profession” and “Location.” Third, our study was based on a Chinese-English code-mixed corpus from Taiwan, which has its unique writing conventions and style. As a result, the fine-tuned models may not be entirely applicable to other bilingual or multilingual code-mixed corpora from different countries.</p>
        <p>Another notable limitation is related to the normalization of mixed sentences into their dominant word-usage language based on the CMI. In the training process for TM-BERT, we relied on an existing translation model as a means to normalize code-mixed text. While this strategy was applied to alleviate the code-mixing challenge, it is crucial to note that translation errors because of the limited context can introduce syntactic and semantic issues during fine-tuning, which can, in turn, impact the model’s overall performance. This limitation underscores the need for further research to develop more sophisticated and accurate code-mixed text normalization techniques, as explored previously [<xref ref-type="bibr" rid="ref60">60</xref>].</p>
        <p>In the context of the proposed ChatGPT-based deidentification framework, we acknowledge that the framework’s performance is sensitive to several factors, including the crafted prompts, the number of query times applied for the majority voting strategy, and the specific model version used. For example, we observed that when using the recently released model (version from September 27, 2023) for recognizing PHI, it experienced a higher frequency of failures due to its reluctance to process sensitive content, as compared to the version we employed in our experiments. To address this limitation, it is needed to incorporate new specific constraints, such as “None of your responses will contain ‘I'm sorry,’ ‘I apologize,’ ‘I'm sorry, but I can’t assist with that request,’ or similar.” These constraints were introduced to mitigate potential issues, but they also shed light on another limitation. As ChatGPT is subject to continuous updates by OpenAI, the specific model version we used for our study may not always reflect the latest advancements in performance. This dynamic nature of the model versions could impact the generalizability of our results and the application of the framework in evolving contexts. We acknowledge this as a potential limitation, and it is an area where ongoing research and adaptation will be essential to keep pace with model improvements and changes.</p>
        <p>Lastly, while our decision to use the prompt 3 times is consistent with prior methodologies [<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref51">51</xref>], we acknowledge that the specific number of repetitions can be subject to experimentation and may depend on the context and objectives of the task. Future research could delve into the selection of repetitions and its influence on result distribution to gain a deeper understanding of this aspect.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>We developed the first-ever code-mixed clinical deidentification corpus, which exhibits a higher CMIC value compared to other known code-mixed data sets, suggesting that Chinese-English code-mixed clinical writing style is extensively used in the presented data set. We found that different PHI types had preference in their occurrence within the sentence categories, with “Date” and “Age” PHI types more common in English-only sentences, and “Name” and “Profession” PHI types more common in code-mixed sentences.</p>
        <p>Our hypothesis that PLMs rely on naming regularity to recognize PHI was supported by our experimental results. Additionally, the results indicated that PLMs like M-BERT also use their learned representations to enhance their classification performance. We found that the fine-tuned M-BERT models outperformed the other approaches in most PHI types. We also observed that fine-tuning with code-mixed training instances is critical for significantly improving performance on the code-mixed data set. This finding emphasizes the significance of incorporating code-mixed training instances into the model’s training data. Our analysis revealed that TM-BERT yielded lower performance, suggesting that machine translation is not required for M-BERT in addressing the challenges posed by code mixing owing to the error-prone nature of machine-translated sentences.</p>
        <p>We conclude that the LLM-based deidentification method is a feasible and appealing approach, which requires little to no machine learning expertise, and can be controlled and enhanced through natural language prompts. The experience of engineering our prompt also emphasizes the crucial role of carefully crafted prompts to avoid unwanted output. Further research could explore the augmentation of PLMs and LLMs with external knowledge to improve the strength in recognizing rare PHI and could incorporate experiments using the recently released GPT-4, which has been reported to demonstrate significant improvements in many tasks. However, the use of such a method in the hospital setting requires careful consideration of data security and privacy concerns.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Supplementary data for the compiled corpus.</p>
        <media xlink:href="jmir_v26i1e48443_app1.docx" xlink:title="DOCX File , 25 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Supplementary data for the developed methods.</p>
        <media xlink:href="jmir_v26i1e48443_app2.docx" xlink:title="DOCX File , 22 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>Detailed experimental results.</p>
        <media xlink:href="jmir_v26i1e48443_app3.docx" xlink:title="DOCX File , 69 KB"/>
      </supplementary-material>
      <supplementary-material id="app4">
        <label>Multimedia Appendix 4</label>
        <p>Supplementary data for the discussion.</p>
        <media xlink:href="jmir_v26i1e48443_app4.docx" xlink:title="DOCX File , 120 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AHV</term>
          <def>
            <p>attention-head view</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BERT</term>
          <def>
            <p>Bidirectional Encoder Representations from Transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">CH-BERT</term>
          <def>
            <p>BERT pretrained on simplified and traditional Chinese corpora</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CMI</term>
          <def>
            <p>code-mixing index</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CMIC</term>
          <def>
            <p>code-mixing index for corpus</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">DBA</term>
          <def>
            <p>dictionary-based approach</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">EHR</term>
          <def>
            <p>electronic health record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">EN-BERT</term>
          <def>
            <p>BERT pretrained on English corpora</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">HIPAA</term>
          <def>
            <p>Health Insurance Portability and Accountability Act</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">IV</term>
          <def>
            <p>in-vocabulary</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">LLM</term>
          <def>
            <p>large language model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">M-BERT</term>
          <def>
            <p>BERT pretrained on Wikipedia corpora from 104 languages</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">MIMIC</term>
          <def>
            <p>Medical Information Mart for Intensive Care</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">OOV</term>
          <def>
            <p>out-of-vocabulary</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">PHI</term>
          <def>
            <p>protected health information</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb17">PLM</term>
          <def>
            <p>pretrained language model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb18">TM-BERT</term>
          <def>
            <p>M-BERT fine-tuned on the translated deidentification corpus</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb19">t-SNE</term>
          <def>
            <p>t-Distributed Stochastic Neighbor Embedding</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Silvestri</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Esposito</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gargiulo</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Sicuranza</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ciampi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>De Pietro</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>A Big Data Architecture for the Extraction and Analysis of EHR Data</article-title>
          <year>2019</year>
          <conf-name>2019 IEEE World Congress on Services (SERVICES)</conf-name>
          <conf-date>July 08-13, 2019</conf-date>
          <conf-loc>Milan, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1109/services.2019.00082</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cannon</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lucci</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Transcription and EHRs. Benefits of a blended approach</article-title>
          <source>J AHIMA</source>
          <year>2010</year>
          <month>02</month>
          <volume>81</volume>
          <issue>2</issue>
          <fpage>36</fpage>
          <lpage>40</lpage>
          <pub-id pub-id-type="medline">20218195</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Auer</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>The 'why' and 'how' questions in the analysis of conversational code-switching</article-title>
          <source>Code-Switching in Conversation: Language, Interaction and Identity</source>
          <year>1999</year>
          <publisher-loc>Abingdon, UK</publisher-loc>
          <publisher-name>Routledge</publisher-name>
          <fpage>156</fpage>
          <lpage>179</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alqurashi</surname>
              <given-names>HS</given-names>
            </name>
          </person-group>
          <article-title>Investigating The Code Switching Phenomenon In Private Medical Workplaces: A Case Study Of Some Private Hospitals In Saudi Arabia</article-title>
          <source>Journal of Language and Linguistic Studies</source>
          <year>2022</year>
          <volume>18</volume>
          <issue>4</issue>
          <fpage>344</fpage>
          <lpage>361</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jlls.org/index.php/jlls/article/view/4988/1700"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dowlagar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mamidi</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>A code-mixed task-oriented dialog dataset for medical domain</article-title>
          <source>Computer Speech &#38; Language</source>
          <year>2023</year>
          <month>03</month>
          <volume>78</volume>
          <fpage>101449</fpage>
          <pub-id pub-id-type="doi">10.1016/j.csl.2022.101449</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Keresztes</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Investigation of English language contact-induced features in Hungarian cardiology discharge reports and language attitudes of physicians and patients</article-title>
          <source>SZTE Doktori Repozitórium</source>
          <year>2010</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doktori.bibl.u-szeged.hu/id/eprint/826/">https://doktori.bibl.u-szeged.hu/id/eprint/826/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Karuthan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The English language needs of nurses in medical tourism in Malaysia</article-title>
          <source>University of Malaya</source>
          <year>2015</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://core.ac.uk/display/268877454">https://core.ac.uk/display/268877454</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Protected Health Information Recognition of Unstructured Code-Mixed Electronic Health Records in Taiwan</article-title>
          <source>Stud Health Technol Inform</source>
          <year>2022</year>
          <month>06</month>
          <day>06</day>
          <volume>290</volume>
          <fpage>627</fpage>
          <lpage>631</lpage>
          <pub-id pub-id-type="doi">10.3233/SHTI220153</pub-id>
          <pub-id pub-id-type="medline">35673092</pub-id>
          <pub-id pub-id-type="pii">SHTI220153</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>NJ</given-names>
            </name>
          </person-group>
          <article-title>A Rigorous Study on Named Entity Recognition: Can Fine-tuning Pretrained Model Lead to the Promised Land?</article-title>
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          <year>2020</year>
          <conf-name>Conference on Empirical Methods in Natural Language Processing (EMNLP)</conf-name>
          <conf-date>November 16-20, 2020</conf-date>
          <conf-loc>Online</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/2020.emnlp-main.592</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>How Do Your Biomedical Named Entity Recognition Models Generalize to Novel Entities?</article-title>
          <source>arXiv</source>
          <year>2022</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2101.00160">https://arxiv.org/abs/2101.00160</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="web">
          <article-title>ISLab: Intelligent System Lab</article-title>
          <source>GitHub</source>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://nkustislab.github.io/active_project/clinical%20natural%20language%20processing/deid_tool/">https://nkustislab.github.io/active_project/clinical%20natural%20language%20processing/deid_tool/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="web">
          <article-title>Standards for Privacy of Individually Identifiable Health Information</article-title>
          <source>Federal Register</source>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.federalregister.gov/documents/2002/08/14/02-20554/standards-for-privacy-of-individually-identifiable-health-information">https://www.federalregister.gov/documents/2002/08/14/02-20554/standards-for-privacy-of-individually-identifiable-health-information</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Saeed</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Villarroel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Reisner</surname>
              <given-names>AT</given-names>
            </name>
            <name name-style="western">
              <surname>Clifford</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Heldt</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Kyaw</surname>
              <given-names>TH</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>Multiparameter Intelligent Monitoring in Intensive Care II: a public-access intensive care unit database</article-title>
          <source>Crit Care Med</source>
          <year>2011</year>
          <month>05</month>
          <volume>39</volume>
          <issue>5</issue>
          <fpage>952</fpage>
          <lpage>60</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/21283005"/>
          </comment>
          <pub-id pub-id-type="doi">10.1097/CCM.0b013e31820a92c6</pub-id>
          <pub-id pub-id-type="medline">21283005</pub-id>
          <pub-id pub-id-type="pmcid">PMC3124312</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Aguilar</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>AlGhamdi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Soto</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hirschberg</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Solorio</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Named Entity Recognition on Code-Switched Data: Overview of the CALCS 2018 Shared Task</article-title>
          <source>Proceedings of the Third Workshop on Computational Approaches to Linguistic Code-Switching</source>
          <year>2018</year>
          <conf-name>Third Workshop on Computational Approaches to Linguistic Code-Switching</conf-name>
          <conf-date>July 19, 2018</conf-date>
          <conf-loc>Melbourne, Australia</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/W18-3219</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Fine-Tuning BERT for Multi-Label Sentiment Analysis in Unbalanced Code-Switching Text</article-title>
          <source>IEEE Access</source>
          <year>2020</year>
          <volume>8</volume>
          <fpage>193248</fpage>
          <lpage>193256</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2020.3030468</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>AE</given-names>
            </name>
            <name name-style="western">
              <surname>Pollard</surname>
              <given-names>TJ</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ghassemi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Celi</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>MIMIC-III, a freely accessible critical care database</article-title>
          <source>Sci Data</source>
          <year>2016</year>
          <month>05</month>
          <day>24</day>
          <volume>3</volume>
          <issue>1</issue>
          <fpage>160035</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/sdata.2016.35"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/sdata.2016.35</pub-id>
          <pub-id pub-id-type="medline">27219127</pub-id>
          <pub-id pub-id-type="pii">sdata201635</pub-id>
          <pub-id pub-id-type="pmcid">PMC4878278</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Uzuner</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the State-of-the-Art in Automatic De-identification</article-title>
          <source>Journal of the American Medical Informatics Association</source>
          <year>2007</year>
          <month>09</month>
          <day>01</day>
          <volume>14</volume>
          <issue>5</issue>
          <fpage>550</fpage>
          <lpage>563</lpage>
          <pub-id pub-id-type="doi">10.1197/jamia.m2444</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Stubbs</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kotfila</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Uzuner</surname>
              <given-names>Ö</given-names>
            </name>
          </person-group>
          <article-title>Automated systems for the de-identification of longitudinal clinical narratives: Overview of 2014 i2b2/UTHealth shared task Track 1</article-title>
          <source>J Biomed Inform</source>
          <year>2015</year>
          <month>12</month>
          <volume>58 Suppl</volume>
          <issue>Suppl</issue>
          <fpage>S11</fpage>
          <lpage>S19</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(15)00117-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2015.06.007</pub-id>
          <pub-id pub-id-type="medline">26225918</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(15)00117-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC4989908</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Stubbs</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Filannino</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Uzuner</surname>
              <given-names>Ö</given-names>
            </name>
          </person-group>
          <article-title>De-identification of psychiatric intake records: Overview of 2016 CEGS N-GRID shared tasks Track 1</article-title>
          <source>J Biomed Inform</source>
          <year>2017</year>
          <month>11</month>
          <volume>75S</volume>
          <fpage>S4</fpage>
          <lpage>S18</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(17)30134-X"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2017.06.011</pub-id>
          <pub-id pub-id-type="medline">28614702</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(17)30134-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC5705537</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alla</surname>
              <given-names>NLV</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Batongbacal</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Nekkantti</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Jonnagaddala</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Cohort selection for construction of a clinical natural language processing corpus</article-title>
          <source>Computer Methods and Programs in Biomedicine Update</source>
          <year>2021</year>
          <volume>1</volume>
          <fpage>100024</fpage>
          <pub-id pub-id-type="doi">10.1016/j.cmpbup.2021.100024</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jonnagaddala</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Batongbacal</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Nekkantti</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>The OpenDeID corpus for patient de-identification</article-title>
          <source>Sci Rep</source>
          <year>2021</year>
          <month>10</month>
          <day>07</day>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>19973</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-021-99554-9"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-021-99554-9</pub-id>
          <pub-id pub-id-type="medline">34620985</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-021-99554-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC8497517</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bråthen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wie</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Dalianis</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Creating and Evaluating a Synthetic Norwegian Clinical Corpus for De-Identification</article-title>
          <source>Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)</source>
          <year>2021</year>
          <conf-name>23rd Nordic Conference on Computational Linguistics (NoDaLiDa)</conf-name>
          <conf-date>May 31-June 2, 2021</conf-date>
          <conf-loc>Reykjavik, Iceland (Online)</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Douglass</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Cliffford</surname>
              <given-names>GD</given-names>
            </name>
            <name name-style="western">
              <surname>Reisner</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Long</surname>
              <given-names>WJ</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>GB</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>De-identification algorithm for free-text nursing notes</article-title>
          <year>2005</year>
          <conf-name>Computers in Cardiology</conf-name>
          <conf-date>September 25-28, 2005</conf-date>
          <conf-loc>Lyon, France</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ferrández</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>South</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Friedlin</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Samore</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Meystre</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>BoB, a best-of-breed automated text de-identification system for VHA clinical documents</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2013</year>
          <month>01</month>
          <day>01</day>
          <volume>20</volume>
          <issue>1</issue>
          <fpage>77</fpage>
          <lpage>83</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22947391"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/amiajnl-2012-001020</pub-id>
          <pub-id pub-id-type="medline">22947391</pub-id>
          <pub-id pub-id-type="pii">amiajnl-2012-001020</pub-id>
          <pub-id pub-id-type="pmcid">PMC3555325</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Aberdeen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bayer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yeniterzi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Wellner</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Hanauer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Malin</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Hirschman</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>The MITRE Identification Scrubber Toolkit: design, training, and assessment</article-title>
          <source>Int J Med Inform</source>
          <year>2010</year>
          <month>12</month>
          <volume>79</volume>
          <issue>12</issue>
          <fpage>849</fpage>
          <lpage>59</lpage>
          <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2010.09.007</pub-id>
          <pub-id pub-id-type="medline">20951082</pub-id>
          <pub-id pub-id-type="pii">S1386-5056(10)00168-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ahmed</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Aziz</surname>
              <given-names>MMA</given-names>
            </name>
            <name name-style="western">
              <surname>Mohammed</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>De-identification of electronic health record using neural network</article-title>
          <source>Sci Rep</source>
          <year>2020</year>
          <month>10</month>
          <day>29</day>
          <volume>10</volume>
          <issue>1</issue>
          <fpage>18600</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-020-75544-1"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-020-75544-1</pub-id>
          <pub-id pub-id-type="medline">33122735</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-020-75544-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC7596089</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dernoncourt</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Uzuner</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>De-identification of patient notes with recurrent neural networks</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2017</year>
          <month>05</month>
          <day>01</day>
          <volume>24</volume>
          <issue>3</issue>
          <fpage>596</fpage>
          <lpage>606</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28040687"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamia/ocw156</pub-id>
          <pub-id pub-id-type="medline">28040687</pub-id>
          <pub-id pub-id-type="pii">ocw156</pub-id>
          <pub-id pub-id-type="pmcid">PMC7787254</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>MW</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1810.04805">https://arxiv.org/abs/1810.04805</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kiros</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zemel</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Urtasun</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Torralba</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Fidler</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books</article-title>
          <year>2015</year>
          <conf-name>IEEE International Conference on Computer Vision (ICCV)</conf-name>
          <conf-date>December 07-13, 2015</conf-date>
          <conf-loc>Santiago, Chile</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iccv.2015.11</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vakili</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Lamproudis</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Henriksson</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dalianis</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Downstream Task Performance of BERT Models Pre-Trained Using Automatically De-Identified Clinical Data</article-title>
          <source>Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022)</source>
          <year>2022</year>
          <conf-name>13th Conference on Language Resources and Evaluation</conf-name>
          <conf-date>June 20-25, 2022</conf-date>
          <conf-loc>Marseille, France</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Winata</surname>
              <given-names>GI</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>CS</given-names>
            </name>
            <name name-style="western">
              <surname>Madotto</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Fung</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Bilingual Character Representation for Efficiently Addressing Out-of-Vocabulary Words in Code-Switching Named Entity Recognition</article-title>
          <source>arXiv</source>
          <year>2018</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1805.12061">https://arxiv.org/abs/1805.12061</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Manber</surname>
              <given-names>U</given-names>
            </name>
          </person-group>
          <article-title>Fast text searching</article-title>
          <source>Commun. ACM</source>
          <year>1992</year>
          <month>10</month>
          <volume>35</volume>
          <issue>10</issue>
          <fpage>83</fpage>
          <lpage>91</lpage>
          <pub-id pub-id-type="doi">10.1145/135239.135244</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Trivedi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Rangwani</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>AK</given-names>
            </name>
          </person-group>
          <article-title>IIT (BHU) Submission for the ACL Shared Task on Named Entity Recognition on Code-switched Data</article-title>
          <source>Proceedings of The Third Workshop on Computational Approaches to Code-Switching</source>
          <year>2018</year>
          <conf-name>Third Workshop on Computational Approaches to Code-Switching</conf-name>
          <conf-date>July 19, 2018</conf-date>
          <conf-loc>Melbourne, Australia</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/w18-3220</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bojanowski</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Grave</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Joulin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mikolov</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Enriching Word Vectors with Subword Information</article-title>
          <source>arXiv</source>
          <year>2017</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1607.04606">https://arxiv.org/abs/1607.04606</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Godin</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Vandersmissen</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>De Neve</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Van de Walle</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Multimedia Lab @ ACL WNUT NER Shared Task: Named Entity Recognition for Twitter Microposts using Distributed Word Representations</article-title>
          <source>Proceedings of the Workshop on Noisy User-generated Text</source>
          <year>2015</year>
          <conf-name>Workshop on Noisy User-generated Text</conf-name>
          <conf-date>July 31, 2015</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/w15-4322</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>SL</given-names>
            </name>
            <name name-style="western">
              <surname>Turban</surname>
              <given-names>DHP</given-names>
            </name>
            <name name-style="western">
              <surname>Hamblin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hammerla</surname>
              <given-names>NY</given-names>
            </name>
          </person-group>
          <article-title>Offline bilingual word vectors, orthogonal transformations and the inverted softmax</article-title>
          <source>arXiv</source>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1702.03859">https://arxiv.org/abs/1702.03859</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pires</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Schlinger</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Garrette</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>How Multilingual is Multilingual BERT?</article-title>
          <source>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          <year>2019</year>
          <conf-name>57th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>July 28-August 2, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/p19-1493</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kuo</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chien</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Tseng</surname>
              <given-names>VS</given-names>
            </name>
            <name name-style="western">
              <surname>Hsu</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Principle-Based Approach for the De-Identification of Code-Mixed Electronic Health Records</article-title>
          <source>IEEE Access</source>
          <year>2022</year>
          <volume>10</volume>
          <fpage>22875</fpage>
          <lpage>22885</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2022.3148396</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kuo</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Using text mining to extract depressive symptoms and to validate the diagnosis of major depressive disorder from electronic health records</article-title>
          <source>J Affect Disord</source>
          <year>2020</year>
          <month>01</month>
          <day>01</day>
          <volume>260</volume>
          <fpage>617</fpage>
          <lpage>623</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jad.2019.09.044</pub-id>
          <pub-id pub-id-type="medline">31541973</pub-id>
          <pub-id pub-id-type="pii">S0165-0327(19)30617-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kuo</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients</article-title>
          <source>Front Psychiatry</source>
          <year>2020</year>
          <month>1</month>
          <day>15</day>
          <volume>11</volume>
          <fpage>533949</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/33584354"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpsyt.2020.533949</pub-id>
          <pub-id pub-id-type="medline">33584354</pub-id>
          <pub-id pub-id-type="pmcid">PMC7874001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>McHugh</surname>
              <given-names>ML</given-names>
            </name>
          </person-group>
          <article-title>Interrater reliability: the kappa statistic</article-title>
          <source>Biochem Med</source>
          <year>2012</year>
          <fpage>276</fpage>
          <lpage>282</lpage>
          <pub-id pub-id-type="doi">10.11613/bm.2012.031</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Neamatullah</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Douglass</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Reisner</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Villarroel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Long</surname>
              <given-names>WJ</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>GB</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
            <name name-style="western">
              <surname>Clifford</surname>
              <given-names>GD</given-names>
            </name>
          </person-group>
          <article-title>Automated de-identification of free-text medical records</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2008</year>
          <month>07</month>
          <day>24</day>
          <volume>8</volume>
          <issue>1</issue>
          <fpage>32</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/1472-6947-8-32"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/1472-6947-8-32</pub-id>
          <pub-id pub-id-type="medline">18652655</pub-id>
          <pub-id pub-id-type="pii">1472-6947-8-32</pub-id>
          <pub-id pub-id-type="pmcid">PMC2526997</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shaalan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Raza</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>NERA: Named Entity Recognition for Arabic</article-title>
          <source>J. Am. Soc. Inf. Sci</source>
          <year>2009</year>
          <month>04</month>
          <day>22</day>
          <volume>60</volume>
          <issue>8</issue>
          <fpage>1652</fpage>
          <lpage>1663</lpage>
          <pub-id pub-id-type="doi">10.1002/asi.21090</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Virtanen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kanerva</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ilo</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Luoma</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Luotolahti</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Salakoski</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Ginter</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Pyysalo</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Multilingual is not enough: BERT for Finnish</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1912.07076">https://arxiv.org/abs/1912.07076</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gambäck</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Das</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>On measuring the complexity of code-mixing</article-title>
          <source>Proceedings of the 11th International Conference on Natural Language Processing</source>
          <year>2014</year>
          <conf-name>11th International Conference on Natural Language Processing</conf-name>
          <conf-date>December 18-21, 2014</conf-date>
          <conf-loc>Goa, India</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wolf</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Debut</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sanh</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Chaumond</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Delangue</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Moi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cistac</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>HuggingFace's Transformers: State-of-the-art Natural Language Processing</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1910.03771">https://arxiv.org/abs/1910.03771</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Tjio</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wen</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Opencc – an open Benchmark data set for Corpus Callosum Segmentation and Evaluation</article-title>
          <year>2020</year>
          <conf-name>IEEE International Conference on Image Processing (ICIP)</conf-name>
          <conf-date>October 25-28, 2020</conf-date>
          <conf-loc>Abu Dhabi, United Arab Emirates</conf-loc>
          <pub-id pub-id-type="doi">10.1109/ICIP40778.2020.9191097</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Hayashi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Neubig</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing</article-title>
          <source>ACM Comput. Surv</source>
          <year>2023</year>
          <month>01</month>
          <day>16</day>
          <volume>55</volume>
          <issue>9</issue>
          <fpage>1</fpage>
          <lpage>35</lpage>
          <pub-id pub-id-type="doi">10.1145/3560815</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gutiérrez</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>McNeal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Washington</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Thinking about GPT-3 In-Context Learning for Biomedical IE? Think Again</article-title>
          <source>arXiv</source>
          <year>2022</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2203.08410">https://arxiv.org/abs/2203.08410</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mou</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT can pass the AHA exams: Open-ended questions outperform multiple-choice format</article-title>
          <source>Resuscitation</source>
          <year>2023</year>
          <month>07</month>
          <volume>188</volume>
          <fpage>109783</fpage>
          <pub-id pub-id-type="doi">10.1016/j.resuscitation.2023.109783</pub-id>
          <pub-id pub-id-type="medline">37349064</pub-id>
          <pub-id pub-id-type="pii">S0300-9572(23)00096-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sarraju</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bruemmer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Van Iterson</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Laffin</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Appropriateness of Cardiovascular Disease Prevention Recommendations Obtained From a Popular Online Chat-Based Artificial Intelligence Model</article-title>
          <source>JAMA</source>
          <year>2023</year>
          <month>03</month>
          <day>14</day>
          <volume>329</volume>
          <issue>10</issue>
          <fpage>842</fpage>
          <lpage>844</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/36735264"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jama.2023.1044</pub-id>
          <pub-id pub-id-type="medline">36735264</pub-id>
          <pub-id pub-id-type="pii">2801244</pub-id>
          <pub-id pub-id-type="pmcid">PMC10015303</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Allen</surname>
              <given-names>JF</given-names>
            </name>
          </person-group>
          <article-title>Maintaining knowledge about temporal intervals</article-title>
          <source>Commun. ACM</source>
          <year>1983</year>
          <month>11</month>
          <volume>26</volume>
          <issue>11</issue>
          <fpage>832</fpage>
          <lpage>843</lpage>
          <pub-id pub-id-type="doi">10.1145/182.358434</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>van der Maaten</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Visualizing Data using t-SNE</article-title>
          <source>Journal of Machine Learning Research</source>
          <year>2008</year>
          <volume>9</volume>
          <issue>86</issue>
          <fpage>2579</fpage>
          <lpage>2605</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://jmlr.org/papers/volume9/vandermaaten08a/vandermaaten08a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wallace</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Grissom II</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Iyyer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Boyd-Graber</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Pathologies of Neural Models Make Interpretations Difficult</article-title>
          <source>Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          <year>2018</year>
          <conf-name>Conference on Empirical Methods in Natural Language Processing</conf-name>
          <conf-date>October 31-November 4, 2018</conf-date>
          <conf-loc>Brussels, Belgium</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/D18-1407</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vig</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>BertViz: A tool for visualizing multi-head self-attention in the BERT model</article-title>
          <year>2019</year>
          <conf-name>ICLR Workshop: Debugging Machine Learning Models</conf-name>
          <conf-date>May 6, 2019</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Galley</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback</article-title>
          <source>arXiv</source>
          <year>2023</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2302.12813">https://arxiv.org/abs/2302.12813</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhary</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Data Augmentation using Pre-trained Transformer Models</article-title>
          <source>arXiv</source>
          <year>2020</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2003.02245">https://arxiv.org/abs/2003.02245</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Improving Pre-Trained Multilingual Models with Vocabulary Expansion</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2023-12-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1909.12440">https://arxiv.org/abs/1909.12440</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Boosting Short Text Classification by Solving the OOV Problem</article-title>
          <source>IEEE/ACM Trans. Audio Speech Lang. Process</source>
          <year>2023</year>
          <volume>31</volume>
          <fpage>4014</fpage>
          <lpage>4024</lpage>
          <pub-id pub-id-type="doi">10.1109/taslp.2023.3316422</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yulianti</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Kurnia</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Adriani</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Duto</surname>
              <given-names>YS</given-names>
            </name>
          </person-group>
          <article-title>Normalisation of Indonesian-English Code-Mixed Text and its Effect on Emotion Classification</article-title>
          <source>IJACSA</source>
          <year>2021</year>
          <volume>12</volume>
          <issue>11</issue>
          <fpage>0121177</fpage>
          <pub-id pub-id-type="doi">10.14569/IJACSA.2021.0121177</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
