<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v27i1e73516</article-id><article-id pub-id-type="doi">10.2196/73516</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>Performance Evaluation of Deep Learning for the Detection and Segmentation of Thyroid Nodules: Systematic Review and Meta-Analysis</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Ni</surname><given-names>Jiayu</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>You</surname><given-names>Yue</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Wu</surname><given-names>Xiaohe</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Chen</surname><given-names>Xueke</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wang</surname><given-names>Jiaying</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Li</surname><given-names>Yuan</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff5">5</xref></contrib></contrib-group><aff id="aff1"><institution>Departement of Otolaryngology-Head and Neck Surgery, Affiliated Hospital of Hangzhou Normal University</institution><addr-line>No. 126, Wenzhou Road</addr-line><addr-line>Hangzhou</addr-line><country>China</country></aff><aff id="aff2"><institution>Departement of Ultrasound, The First Affiliated Hospital of Xinjiang Medical University</institution><addr-line>Urumqi</addr-line><country>China</country></aff><aff id="aff3"><institution>Departement of Ultrasound, Yangming Hospital Affiliated to Ningbo University</institution><addr-line>Yuyao</addr-line><country>China</country></aff><aff id="aff4"><institution>Departement of Otolaryngology-Head and Neck Surgery, Hangzhou Normal University</institution><addr-line>Hangzhou</addr-line><country>China</country></aff><aff id="aff5"><institution>Department of Otorhinolaryngology, Deqing Hospital of Hangzhou Normal University (The Third People's Hospital of Deqing)</institution><addr-line>Huzhou</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Sarvestan</surname><given-names>Javad</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Jafarizadeh</surname><given-names>Ali</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Panboonyuen</surname><given-names>Teerapong</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Yuan Li, MMed, Departement of Otolaryngology-Head and Neck Surgery, Affiliated Hospital of Hangzhou Normal University, No. 126, Wenzhou Road, Hangzhou, 310015, China, 86 15005812373; <email>liyuan81629@163.com</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>14</day><month>8</month><year>2025</year></pub-date><volume>27</volume><elocation-id>e73516</elocation-id><history><date date-type="received"><day>06</day><month>03</month><year>2025</year></date><date date-type="rev-recd"><day>09</day><month>05</month><year>2025</year></date><date date-type="accepted"><day>16</day><month>05</month><year>2025</year></date></history><copyright-statement>&#x00A9; Jiayu Ni, Yue You, Xiaohe Wu, Xueke Chen, Jiaying Wang, Yuan Li. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 14.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2025/1/e73516"/><abstract><sec><title>Background</title><p>Thyroid cancer is one of the most common endocrine malignancies. Its incidence has steadily increased in recent years. Distinguishing between benign and malignant thyroid nodules (TNs) is challenging due to their overlapping imaging features. The rapid advancement of artificial intelligence (AI) in medical image analysis, particularly deep learning (DL) algorithms, has provided novel solutions for automated TN detection. However, existing studies exhibit substantial heterogeneity in diagnostic performance. Furthermore, no systematic evidence-based research comprehensively assesses the diagnostic performance of DL models in this field.</p></sec><sec><title>Objective</title><p>This study aimed to execute a systematic review and meta-analysis to appraise the performance of DL algorithms in diagnosing TN malignancy, identify key factors influencing their diagnostic efficacy, and compare their accuracy with that of clinicians in image-based diagnosis.</p></sec><sec sec-type="methods"><title>Methods</title><p>We systematically searched multiple databases, including PubMed, Cochrane, Embase, Web of Science, and IEEE, and identified 41 eligible studies for systematic review and meta-analysis. Based on the task type, studies were categorized into segmentation (n=14) and detection (n=27) tasks. The pooled sensitivity, specificity, and the area under the receiver operating characteristic curve (AUC) were calculated for each group. Subgroup analyses were performed to examine the impact of transfer learning and compare model performance against clinicians.</p></sec><sec sec-type="results"><title>Results</title><p>For segmentation tasks, the pooled sensitivity, specificity, and AUC were 82% (95% CI 79%&#x2010;84%), 95% (95% CI 92%&#x2010;96%), and 0.91 (95% CI 0.89&#x2010;0.94), respectively. For detection tasks, the pooled sensitivity, specificity, and AUC were 91% (95% CI 89%&#x2010;93%), 89% (95% CI 86%&#x2010;91%), and 0.96 (95% CI 0.93&#x2010;0.97), respectively. Some studies demonstrated that DL models could achieve diagnostic performance comparable with, or even exceeding, that of clinicians in certain scenarios. The application of transfer learning contributed to improved model performance.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>DL algorithms exhibit promising diagnostic accuracy in TN imaging, highlighting their potential as auxiliary diagnostic tools. However, current studies are limited by suboptimal methodological design, inconsistent image quality across datasets, and insufficient external validation, which may introduce bias. Future research should enhance methodological standardization, improve model interpretability, and promote transparent reporting to facilitate the sustainable clinical translation of DL-based solutions.</p></sec><sec><title>Trial Registration</title><p>PROSPERO CRD42024599495; https://www.crd.york.ac.uk/PROSPERO/view/CRD42024599495</p></sec></abstract><kwd-group><kwd>thyroid imaging</kwd><kwd>artificial intelligence</kwd><kwd>diagnostic performance</kwd><kwd>sensitivity and specificity</kwd><kwd>systematic review</kwd><kwd>PRISMA</kwd><kwd>Preferred Reporting Items for Systematic reviews and Meta-Analyses</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Thyroid cancer (TC) is the leading type of malignant tumor in the endocrine system. Over the past 3 decades, the global incidence of TC has steadily risen. Between 1980 and 1997, the prevalence was about 2.4%, while by 2009, it increased to 6.6% [<xref ref-type="bibr" rid="ref1">1</xref>]. In clinical settings, the prevalence of TC ranges from approximately 19% to 68%. Furthermore, according to Bray et al [<xref ref-type="bibr" rid="ref2">2</xref>], the global prevalence of TC ranks ninth, while its mortality is positioned sixth. This elevation may be closely tied to the development of diagnostic technologies and the improved rates of early disease detection. Evaluating the risk of TC in patients with thyroid nodules (TNs) is clinically important and helps to reduce health care costs and patient suffering. Among various diagnostic methods available, ultrasound imaging has emerged as the preferred diagnostic tool due to its simplicity, rapidity, and strong reproducibility. However, its interpretation is heavily dependent on the experience of radiologists, potentially leading to variability among various observers.</p><p>In order to address the above limitations, artificial intelligence (AI) is extensively applied in medical imaging today [<xref ref-type="bibr" rid="ref3">3</xref>]. As a crucial branch of AI, machine learning (ML) technologies, particularly deep learning (DL) frameworks, have been rapidly developed, offering significant application potential and technical support for automated medical imaging tasks, like segmentation, detection, and classification [<xref ref-type="bibr" rid="ref4">4</xref>]. DL enhances diagnostic accuracy and efficiency while fully and accurately capturing lesion information. It outperforms traditional segmentation methods in terms of feature extraction, generalization, and handling complex structures [<xref ref-type="bibr" rid="ref5">5</xref>]. Nevertheless, due to the presence of high noise, the quality of ultrasound elastography images is relatively low, making automated segmentation and detection a challenging task.</p><p>As radiomics research gains increasing attention, a noticeable number of original studies [<xref ref-type="bibr" rid="ref6">6</xref>-<xref ref-type="bibr" rid="ref8">8</xref>] and meta-analyses [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>] have been published across various medical fields, particularly in the field of thyroid disease. Despite being the standard imaging method for diagnosing TN and TC, ultrasound has been confirmed to have some limitations. However, radiomics shows the potential to offer more accurate and precise results in TN and TC diagnoses, with promising application prospects [<xref ref-type="bibr" rid="ref12">12</xref>]. Despite the growing number of studies on DL-based methods for thyroid image analysis, there is still considerable variation in study design, dataset quality, model architecture, and performance evaluation metrics. In addition, many studies are limited by small sample sizes, insufficient external validation, and inadequate reporting transparency, which may reduce reproducibility and overestimate the diagnostic performance. Given these limitations, comprehensively assessing the diagnostic performance of DL algorithms is needed to offer an evidence-based understanding of their clinical use.</p><p>This meta-analysis thoroughly appraises the performance of DL models in the segmentation and detection of TC and TN images. The reasons for heterogeneity among studies are explored, and potential sources are discussed. The impact of dataset size, network architecture, and external validation on model performance has also been explored. In addition, the limitations of the included studies are discussed separately, providing guidance for deep investigations and promoting the advancement of DL in the clinical application of this disease.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Search Strategy</title><p>For this study, searches were carried out in PubMed, Cochrane, Embase, Web of Science, and IEEE databases, with the search timeframe extending from the inception of the databases to December 2024. All articles from the search were imported into EndNote for management. Duplicate records were excluded. The search was limited to articles published in English. Studies published earlier than 2018, reviews, conference abstracts, editorial reviews, and studies related to animal experiments were excluded. The complete search strategy for each database was created by a team of experienced clinicians and medical investigators. The detailed search strategy pertaining to the keywords and concepts included &#x201C;Thyroid Nodule,&#x201D; &#x201C;Thyroid Cancer,&#x201D; &#x201C;Thyroid Lesion,&#x201D; &#x201C;Thyroid Tumor,&#x201D; &#x201C;Thyroid Neoplasm,&#x201D; &#x201C;Thyroid Carcinoma,&#x201D; &#x201C;Machine learning (ML),&#x201D; &#x201C;Deep learning (DL),&#x201D; &#x201C;Artificial Intelligence,&#x201D; &#x201C;Artificial Neural Network,&#x201D; &#x201C;External Validation,&#x201D; and &#x201C;Convolutional Neural Network.&#x201D; We combined each concept&#x2019;s medical subject headings and keywords with &#x201C;OR&#x201D; and then joined the concepts with &#x201C;AND.&#x201D; Specific search strategies were tailored for each database. <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> provides a summary of the search strategy used in each database. This study was conducted in line with the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines (the PRISMA 2020 checklist is provided in <xref ref-type="supplementary-material" rid="app12">Checklist 1</xref>).</p></sec><sec id="s2-2"><title>Inclusion and Exclusion Criteria</title><p>The original studies were first screened by 2 independent investigators (JN and YY) using titles and abstracts. They reviewed the entire text afterward, following the inclusion and exclusion criteria. Any disagreements or differing opinions would be discussed and resolved with a third party (YL). Randomized controlled trials, cohort studies, case-control studies, and cross-sectional studies were included. We focused on studies that assessed the diagnostic performance of DL for TN detection and segmentation. Studies that reported diagnostic outcomes, like the area under the receiver operating characteristic curve (AUC) of summary receiver operating characteristics (SROC), concordance index, accuracy, pooled sensitivity, and specificity, were included. Imaging techniques used for TN and TC diagnoses, like ultrasound, computed tomography (CT), and magnetic resonance imaging, were included. Reviews, conference abstracts, case reports, letters to editors, comments, and unpublished gray studies were excluded. Studies that were not relevant to the inclusion criteria and were published in languages other than English were also excluded.</p><p>Reviews, conference abstracts, case reports, letters to editors, comments, and unpublished gray studies were excluded. Studies that were not relevant to the inclusion criteria and were published in languages other than English were also excluded.</p></sec><sec id="s2-3"><title>Data Extraction</title><p>The following data were extracted by 2 independent investigators (JN and YY): first author, publication year, sample size (including training and testing set sizes), mean or median age, indicator definition, algorithm, feature extraction, and selection details. In case of discrepancies, discussions with a third party (YL) were held to resolve them. Binary data for diagnostic accuracy were extracted directly into contingency tables, which included true-positives, false-positives, true-negatives, and false-negatives. These were then used to calculate pooled sensitivity, specificity, and other metrics. If a study presented multiple contingency tables for the same or various DL algorithms, they were assumed independent of each other.</p></sec><sec id="s2-4"><title>Quality Assessment</title><p>Two independent investigators leveraged the quality assessment of diagnostic accuracy studies using AI (QUADAS-AI) [<xref ref-type="bibr" rid="ref13">13</xref>] and Review Manager (version 5.4) to appraise study quality. Four domains are appraised in the QUADAS-AI tool: (1) patient selection, (2) index test, (3) reference standard, and (4) flow and timing. Each domain was used to assess the risk of bias (ROB). Furthermore, the first 3 domains were also used to evaluate concerns about applicability.</p></sec><sec id="s2-5"><title>Statistical Analysis</title><p>The meta-analysis was implemented by means of the meta-analysis of diagnostic accuracy studies module in STATA (version 17). The pooled sensitivity and specificity, along with their 95% CIs, were appraised to quantify the predictive accuracy of radiomics. In addition, an SROC curve and AUC were generated to summarize diagnostic accuracy. We plotted the corresponding combined 95% CI and 95% prediction intervals around the mean sensitivity, specificity, and AUC estimates in the SROC plot.</p><p>To examine heterogeneity, a forest plot was created to display the pooled sensitivity and specificity, while the <italic>I</italic><sup>2</sup> and Q values were calculated. The <italic>I</italic><sup>2</sup> values were categorized as follows: 0%~25%, 25%~50%, 50%~75%, and &#x003E;75%, indicating very low, low, moderate, and high heterogeneity between studies, correspondingly. A random-effects model was leveraged to pool the effect sizes from each study, addressing potential heterogeneity in true effect distributions. The model was specifically designed to aggregate sensitivity, specificity, and AUC values from a variety of studies. Its strength lies in its ability to effectively manage the differences between these metrics while recognizing their interconnections. In addition, we executed detailed subgroup analyses, including whether transfer learning (TL) and DL or ML algorithms were applied, to explore how different features and conditions affected the diagnostic performance of DL models.</p></sec><sec id="s2-6"><title>Ethical Considerations</title><p>The study was registered with the PROSPERO (International Prospective Register of Systematic Reviews; CRD42024599495). It followed the preferred reporting items for systematic reviews and meta-analyses guidelines [<xref ref-type="bibr" rid="ref14">14</xref>]. For this study, no ethical approval or informed consent was needed.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Literature Selection</title><p>From the databases, 5280 articles were totally retrieved. Out of these articles, 2663 were reviewed based on their titles and abstracts after removing duplicates. Among these studies, 2576 were deleted for not fulfilling the inclusion criteria. Finally, 41 studies were included. Among these, 14 studies [<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref28">28</xref>] centered on segmentation tasks, while 27 [<xref ref-type="bibr" rid="ref29">29</xref>-<xref ref-type="bibr" rid="ref55">55</xref>] studies focused on detection tasks (<xref ref-type="fig" rid="figure1">Figure 1</xref>).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Study selection process following Preferred Reporting Items for Systematic reviews and Meta-Analyses guidelines. PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig01.png"/></fig></sec><sec id="s3-2"><title>Study Characteristics</title><p>Before presenting the results of the meta-analysis, we briefly summarized the characteristics of the included studies. The 41 studies were published between 2018 and 2024, all of which were retrospective. Out of the 14 studies on segmentation tasks, 4 used data from private sources, while in studies on detection tasks, 2 used data from public sources. In terms of algorithm selection, all studies on segmentation tasks used DL algorithms, whereas in studies on detection tasks, 16 studies used DL algorithms, and 11 studies leveraged ML algorithms. In terms of medical imaging modalities, all studies on segmentation tasks extracted TN features from ultrasound images. Among studies on detection tasks, 2 studies used CT images for TN feature extraction, while one study used both ultrasound and shear wave elastography images. Regarding TL, 5 studies on segmentation tasks used TL. Ten studies on detection tasks also used TL, while the remaining studies on detection tasks merely mentioned it. Furthermore, none of the studies on segmentation tasks reported information on image quality. However, in studies on detection tasks, 13 studies excluded low-quality images (Tables S1-S3 in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendices 2</xref><xref ref-type="supplementary-material" rid="app3"/>-<xref ref-type="supplementary-material" rid="app4">4</xref>).</p></sec><sec id="s3-3"><title>Algorithm Performance</title><sec id="s3-3-1"><title>Pooled Analysis</title><p>The 14 studies on segmentation tasks all provided sufficient data to create a contingency table for diagnostic performance. The hierarchical SROC curves for these studies (48 contingency tables) are depicted in <xref ref-type="fig" rid="figure2">Figure 2</xref>A . For all algorithms, the pooled findings indicated that the sensitivity and specificity were 82% (95% CI 79%&#x2010;84%) and 95% (95% CI 92%&#x2010;96%), and the AUC was 0.91 (95% CI 0.89&#x2010;0.94).</p><p>Since most studies on segmentation tasks used multiple algorithms to appraise diagnostic performance, the highest accuracy of these algorithms was appraised across 18 contingency tables. The pooled results demonstrated that the sensitivity and specificity were 87% (95% CI 83%&#x2010;90%) and 96% (95% CI 93%&#x2010;98%), and the AUC was 0.95 (95% CI 0.93&#x2010;0.97). Further details can be found in <xref ref-type="fig" rid="figure2">Figure 2B</xref>.</p><p>In 26 studies on detection tasks, sufficient data were offered to generate a contingency table for diagnostic performance. <xref ref-type="fig" rid="figure3">Figure 3A</xref> illustrates the hierarchical SROC curves for these studies (61 contingency tables). The pooled results for all algorithms revealed that the sensitivity and specificity were 91% (95% CI 89%&#x2010;93%) and 89% (95% CI 86%&#x2010;91%), and the AUC was 0.96 (95% CI 0.93&#x2010;0.97).</p><p>The highest accuracy of various algorithms for detection tasks was appraised across 26 contingency tables. The pooled findings demonstrated that the sensitivity and specificity were 93% (95% CI 90%&#x2010;95%) and 90% (95% CI 84%&#x2010;93%), and the AUC was 0.97 (95% CI 0.95&#x2010;0.98). More details are available in <xref ref-type="fig" rid="figure3">Figure 3B</xref>.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Pooled overall performance of algorithms: (A) Receiver operator characteristic curves of all studies on segmentation tasks (14 studies with 48 tables) and (B) receiver operator characteristic curves of studies on segmentation tasks reporting the highest accuracy (14 studies with 18 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig02.png"/></fig><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Pooled overall performance of algorithms; (A) Receiver operator characteristic curves of studies on all detection tasks (26 studies with 61 tables) and (B) Receiver operator characteristic curves of studies on detection tasks reporting the highest accuracy (26 studies with 26 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig03.png"/></fig></sec><sec id="s3-3-2"><title>Subgroup Analysis</title><sec id="s3-3-2-1"><title>Transfer Learning</title><p>Four studies used TL for segmentation tasks, with 12 contingency tables. The pooled results indicated that the sensitivity and specificity were 86% (95% CI 86%&#x2010;86%) and 95% (95% CI 95%&#x2010;95%), correspondingly, with an AUC of 0.93 (95% CI 0.90&#x2010;0.95; <xref ref-type="fig" rid="figure4">Figure 4A</xref>). Ten studies on segmentation tasks did not mention the use of TL, with 36 contingency tables. According to the pooled results, the sensitivity and specificity were 80% (95% CI 77%&#x2010;83%) and 95% (95% CI 92%&#x2010;97%), and the AUC was 0.91 (95% CI 0.88&#x2010;0.93). Details can be found in <xref ref-type="fig" rid="figure4">Figure 4B</xref>.</p><p>Ten studies used TL for detection tasks, with 17 contingency tables. The pooled findings implied that the sensitivity and specificity were 91% (95% CI 86%&#x2010;94%) and 85% (95% CI 81%&#x2010;89%), correspondingly, with an AUC of 0.94 (95% CI 0.91&#x2010;0.96; <xref ref-type="fig" rid="figure5">Figure 5A</xref>). 16 studies on detection tasks did not mention the use of TL, with 44 contingency tables. The pooled results indicated that the sensitivity and specificity were 91% (95% CI 88%&#x2010;93%) and 90% (95% CI 86%&#x2010;93%), and the AUC was 0.96 (95% CI 0.94&#x2010;0.97). Details are available in <xref ref-type="fig" rid="figure5">Figure 5B</xref>.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Pooled performance of algorithms with or without transfer learning: (A) Receiver operator characteristic curves of studies on segmentation tasks with transfer learning (4 studies with 12 tables) and (B) Receiver operator characteristic curves of studies on segmentation tasks without transfer learning (10 studies with 36 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig04.png"/></fig><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Pooled performance of algorithms with or without transfer learning: (A) Receiver operator characteristic curves of studies on detection tasks with transfer learning (10 studies with 17 tables) and (B) receiver operator characteristic curves of studies on detection tasks without transfer learning (16 studies with 44 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig05.png"/></fig></sec><sec id="s3-3-2-2"><title>DL Algorithms Versus Non-DL or ML Algorithms</title><p>In 26 studies on detection tasks, the diagnostic performance of DL algorithms was compared with non-DL or ML algorithms, with 33 contingency tables for DL algorithms and 28 for non-DL or ML algorithms. According to the pooled results, the sensitivity was 93% (95% CI 91%&#x2010;95%) for DL algorithms and 88% (95% CI 84%&#x2010;91%) for non-DL or ML algorithms. The specificity was 93% (95% CI 89%&#x2010;95%) for DL algorithms and 82% (95% CI 78%&#x2010;86%) for non-DL or ML algorithms. The AUC was 0.97 (95% CI 0.95&#x2010;0.98) for DL algorithms and 0.91 (95% CI 0.89&#x2010;0.94) for non-DL or ML algorithms (<xref ref-type="fig" rid="figure6">Figures 6A and 6B</xref>).</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Pooled performance of deep learning algorithms or non-deep learning/machine learning algorithms: (A) Receiver operator characteristic curves for studies on detection tasks with deep learning algorithms (15 studies with 33 tables) and (B) receiver operator characteristic curves for studies on detection tasks with non-deep learning/machine learning algorithms (12 studies with 28 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig06.png"/></fig></sec><sec id="s3-3-2-3"><title>Algorithms Versus Human Clinicians</title><p>Five studies on detection tasks compared diagnostic performance between DL or ML algorithms and human clinicians using the same dataset, with 14 contingency tables for human clinicians and 9 for DL or ML algorithms. The pooled sensitivity was 86% (95% CI 79%&#x2010;91%) for algorithms and 87% (95% CI 82%&#x2010;91%) for human clinicians. The pooled specificity was 80% (95% CI 71%&#x2010;87%) for algorithms and 68% (95% CI 58%&#x2010;76%) for human clinicians. The AUC was 0.90 (95% CI 0.87&#x2010;0.93) for algorithms and 0.86 (95% CI 0.83&#x2010;0.89) for human clinicians (<xref ref-type="fig" rid="figure7">Figures 7A and 7B</xref>).</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Pooled performance of algorithms versus human clinicians and human clinicians using the same sample: (A) Receiver operator characteristic curves of studies on detection tasks with algorithms (5 studies with 9 tables) and (B) receiver operator characteristic curves of studies on detection tasks with human clinicians (5 studies with 14 tables).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig07.png"/></fig></sec></sec></sec><sec id="s3-4"><title>Heterogeneity Analysis</title><p>All included studies demonstrated that DL or ML algorithms were beneficial for TN segmentation and detection using medical imaging, in comparison with histopathological analysis. Nevertheless, considerable heterogeneity was noted. For studies on segmentation tasks, both sensitivity (<italic>I</italic><sup>2</sup>=99.33%) and specificity (<italic>I</italic><sup>2</sup>=99.67%) exhibited high heterogeneity (<italic>P</italic>&#x003C;.0001; <xref ref-type="fig" rid="figure8">Figure 8A</xref>). For studies on detection tasks, sensitivity (<italic>I</italic><sup>2</sup>=97.29%) and specificity (<italic>I</italic><sup>2</sup>=98.55%) showed notable heterogeneity (<italic>P</italic>&#x003C;.0001; <xref ref-type="fig" rid="figure8">Figure 8B</xref>).</p><p>Deek&#x2019;s funnel plots generated using STATA 17.0 were used to assess publication bias. No publication bias was noted in studies on segmentation tasks (<italic>P</italic>=.09) and detection tasks (<italic>P</italic>=.50), even though the studies were widely distributed around the regression line (Figure S1a-b in <xref ref-type="supplementary-material" rid="app5">Multimedia Appendix 5</xref>). To determine the sources of the extreme heterogeneity, subgroup analyses were conducted.</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Summary estimate of pooled performance using forest plot: (A) Forest plot of studies on segmentation tasks (14 studies) and (B) forest plot of studies on detection tasks (27 studies). For a higher-resolution version of this figure, see <xref ref-type="supplementary-material" rid="app6">Multimedia Appendix 6</xref>.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e73516_fig08.png"/></fig><sec id="s3-4-1"><title>Transfer L</title><p>The results of heterogeneity analysis for the subgroup analysis based on the application of TL were as follows: studies on segmentation tasks with TL (sensitivity: <italic>I</italic><sup>2</sup>=99.05%, specificity: <italic>I</italic><sup>2</sup>=99.59, <italic>P</italic>&#x003C;.0001), studies on segmentation tasks without TL (sensitivity: <italic>I</italic><sup>2</sup>=99.32%, specificity: <italic>I</italic><sup>2</sup>=99.65, <italic>P</italic>&#x003C;.0001), studies on detection tasks with TL (sensitivity: <italic>I</italic><sup>2</sup>=98.38%, specificity: <italic>I</italic><sup>2</sup>=95.09, <italic>P</italic>&#x003C;.0001), and studies on detection tasks without TL (sensitivity: <italic>I</italic><sup>2</sup>=96.08%, specificity: <italic>I</italic><sup>2</sup>=98.89, <italic>P</italic>&#x003C;.0001; Figure S2-S3 in <xref ref-type="supplementary-material" rid="app7">Multimedia Appendices 7</xref> and <xref ref-type="supplementary-material" rid="app8">8</xref>).</p></sec><sec id="s3-4-2"><title>DL Algorithms Versus Non-DL or ML Algorithms</title><p>The results of heterogeneity analysis for the subgroup analysis based on the application of DL algorithms were as follows: studies on detection tasks with DL algorithms (sensitivity: <italic>I</italic><sup>2</sup>=98.17%, specificity: <italic>I</italic><sup>2</sup>=98.24, <italic>P</italic>&#x003C;.0001), and studies on detection tasks with non-DL/ML algorithms (sensitivity: <italic>I</italic><sup>2</sup>=96.72%, specificity: <italic>I</italic><sup>2</sup>=98.25, <italic>P</italic>&#x003C;.0001; Figure S4 in <xref ref-type="supplementary-material" rid="app9">Multimedia Appendix 9</xref>). Nevertheless, the source of heterogeneity did not stem from specific subgroups, as <italic>I</italic><sup>2</sup> values remained high. Therefore, we could not infer whether TL and algorithm models likely influenced the performance of algorithms for segmenting and detecting TN.</p></sec></sec><sec id="s3-5"><title>Quality Assessment</title><p>The quality of the included studies was appraised by means of the QUADAS-AI (Figure S5a-b in <xref ref-type="supplementary-material" rid="app10">Multimedia Appendix 10</xref>). A thorough evaluation of each item, based on the ROB domain and applicability concerns, is presented in Figure S6a-b in <xref ref-type="supplementary-material" rid="app11">Multimedia Appendix 11</xref>.</p><sec id="s3-5-1"><title>Studies on Segmentation Tasks</title><p>For the patient selection domain, 3 studies were rated as unclear ROB due to unreported inclusion or exclusion criteria or improper exclusions. Regarding the index test domain, only 1 study was classified as having high or unclear ROB due to the absence of a predefined threshold, while the others were deemed to have low ROB. Three studies were deemed to have unclear ROB due to inconsistencies in reference standards. There was no mention of whether the threshold was determined in advance and whether blinding was implemented. For the flow and timing domain, 5 studies were considered to have high or unclear ROB as their authors did not mention whether an appropriate time gap was maintained or whether the same gold standard was used.</p></sec><sec id="s3-5-2"><title>Studies on Detection Tasks</title><p>Regarding the patient selection domain, 9 studies were considered to have high or unclear ROB unreported inclusion or exclusion criteria or improper exclusions. In terms of the index test domain, 6 studies were deemed to have high or unclear ROB due to the absence of a predefined threshold, while the remaining studies were considered to have low ROB. Only 1 study was rated as unclear ROB due to inconsistencies in the reference standard. The predetermination of the threshold and the implementation of blinding were not mentioned. Regarding the flow and timing domain, 11 studies were classified as high or unclear ROB because their authors did not specify whether an appropriate time gap was maintained or if the same gold standard was leveraged.</p></sec></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>This meta-analysis evaluates the performance of DL models in the segmentation and detection of TC and TN images. The results uncover that the pooled sensitivity, specificity, and AUC for segmentation tasks are 86%, 95%, and 0.93, respectively. For detection tasks, the combined sensitivity, specificity, and AUC are 91%, 85%, and 0.94, respectively. Some of the studies also compare the performance of DL models with that of the clinicians in image interpretation. The results reveal that the 2 are closer in terms of accuracy. This implies that AI technologies might assist in TN diagnosis. DL has high diagnostic accuracy for recognizing benign and malignant TN in imaging.</p><p>TN is frequently observed in clinical settings. The prevalence of TC has been rising steadily on a global scale in recent years [<xref ref-type="bibr" rid="ref56">56</xref>]. In clinical settings, accurately identifying the few malignant nodules with clinical significance among the many benign TNs is challenging. This is crucial for determining which patients require biopsy or surgical removal, ultimately reducing health care costs and patient suffering. Hence, a reliable and noninvasive approach to assess TN is urgently needed. In clinical settings, radiologists preliminarily rely on visual standards for diagnosis, like size ratio, size, calcification, structure (single or multiple), borders, and echogenic characteristics (hyperechoic, isoechoic, or hypoechoic). Furthermore, due to differences in technical expertise, subjective experience, and physical condition, radiologists may interpret thyroid ultrasound images differently [<xref ref-type="bibr" rid="ref57">57</xref>]</p><p>Through image recognition technology, AI can support physicians in making fast, precise, and efficient clinical decisions [<xref ref-type="bibr" rid="ref58">58</xref>,<xref ref-type="bibr" rid="ref59">59</xref>]. For example, DL models can recognize tumor boundaries and even predict the type and growth rate of tumors by learning from extensive image data. In addition, lymph node metastasis is closely linked to the local recurrence, distant metastasis, and staging of TC, providing remarkable guidance for the development of the surgical plan. Thus, the integration of AI into clinical practice demonstrates favorable performance in modern healthcare. Convolutional neural networks (CNNs) are regarded as one of the most advanced algorithms, applied in segmentation [<xref ref-type="bibr" rid="ref60">60</xref>], detection [<xref ref-type="bibr" rid="ref61">61</xref>], and classification [<xref ref-type="bibr" rid="ref62">62</xref>] of TN. Ma et al [<xref ref-type="bibr" rid="ref63">63</xref>] have used a CNN model for TN segmentation. Furthermore, Li et al [<xref ref-type="bibr" rid="ref64">64</xref>] have developed a more improved Faster R-CNN based on CNN for TN detection. However, given the vastness and complexity of biomedical data, it is crucial to conduct rigorous testing on it [<xref ref-type="bibr" rid="ref65">65</xref>].</p><p>After carefully selecting studies on related topics, it is found that ML algorithms exhibit excellent performance in medical image&#x2013;based segmentation and detection of TN, demonstrating comparable or even superior performance to human clinicians. This study appraises the performance of distinct algorithm types (including DL or ML) based on different task types, considering the use of TL, as well as the performance under various levels of ROB. Furthermore, potential sources of heterogeneity between studies are identified based on the above subgroups. More importantly, study quality and ROB are critically assessed using the adapted QUADAS-AI [<xref ref-type="bibr" rid="ref13">13</xref>] assessment tool. This is the strength of this study, providing better guidance to future related studies. This study seeks to identify accurate and reliable detection methods in the segmentation and diagnostic detection of TN.</p><p>By systematically searching the relevant studies, 4 systematic reviews and meta-analyses on ML algorithms for TN in medical imaging are found. Cleere et al [<xref ref-type="bibr" rid="ref66">66</xref>] focus on the application of radiomics in TN diagnosis. Their study does not explicitly analyze the symmetry of the funnel plots and may miss studies with negative results, leading to an overestimation of the performance of imaging histology. The accuracy of imaging histology is highly dependent on ultrasound image quality and segmentation accuracy. Nevertheless, image standardization or quality control measures are not discussed in detail in their paper. Two studies investigate the accuracy of DL algorithms in diagnosing the benign and malignant characteristics of TN through ultrasound imaging. According to Zhu et al [<xref ref-type="bibr" rid="ref11">11</xref>] and Zhong et al [<xref ref-type="bibr" rid="ref9">9</xref>], the VGGNet (a CNN) model and S-Detect both demonstrate high sensitivity and specificity in differentiating between benign and malignant TN. Nonetheless, the greater level of heterogeneity and the relatively low quality of the samples render their results less persuasive. Besides, Zhao et al [<xref ref-type="bibr" rid="ref67">67</xref>] are the first to appraise the diagnostic performance of the computer-aided diagnosis system for TN. However, their study only appraises computer-aided design (CAD) systems and does not cover a wider range of imaging histology methods. Furthermore, it fails to provide an in-depth discussion of the algorithmic differences between CAD systems. Based on the results of the above studies, this study has conducted a targeted comparative analysis and optimized the above deficiencies. Next, a detailed explanation of the comparison between DL algorithms and non-DL or ML algorithms will be provided, aiming to offer more substantial support and references for theoretical development and practical applications in this field.</p><p>This study reveals that DL algorithms are capable of segmenting and detecting TN using medical images. The 6 studies on detection tasks included mention comparisons between ML algorithms and clinicians, as well as comparisons between ML algorithms and clinicians working in conjunction with ML algorithms. The results indicate that DL algorithms demonstrate performance comparable with that of clinical physicians, and in certain respects, they may even exhibit superior capabilities. Nevertheless, it is essential to critically assess some problems of this evidence. In fact, both the judgments made solely by ML and those made by clinicians are subject to certain avoidable research biases. Comparing the diagnostic performance between AI and human clinicians is challenging. AI systems may have lower sensitivity and even higher error rates. Thus, we should not hastily conclude that AI has outpaced clinicians, as both have their respective advantages. Hence, it is more feasible to combine ML with clinicians and use ML as a supportive tool for diagnostic decision-making in clinical research. With continuous development and improvements, AI is expected to have an even greater impact on TN diagnosis in the future by optimizing algorithms and increasing training data.</p><p>The studies included in this article are all retrospective, leading to notable methodological flaws. In clinical settings, accurately obtaining test data is crucial for interpreting model performance. In the 41 included studies, only Koh et al [<xref ref-type="bibr" rid="ref49">49</xref>] conducted external validation using multicenter data. Most studies on detection tasks are conducted in single centers without external validation, limiting the generalization ability of algorithm models. The risk of overfitting has also increased, leading to decreased reproducibility and affecting the reliability of the study. The ability of models to generalize is a key consideration in practical clinical applications, especially in environments with high data heterogeneity. Thus, we cannot adequately assess the performance of models in different populations and imaging sources. Most included studies conduct cross-validation internally, either through random or nonrandom methods. Using internal datasets to validate the model is more likely to be homogeneous and may lead to an overestimation of diagnostic performance, especially in private datasets where investigators may remove images that are difficult to detect. Strict external validation is required when designing AI-related diagnostic studies. Furthermore, in the 14 studies on segmentation tasks, only 4 are based on non&#x2013;open access datasets. Public datasets are beneficial for reducing health care costs and making it easier to compare the performance of various algorithms and models, but there may be discrepancies in image quality, such as resolution, noise levels, and the accuracy of annotations. These differences may also have an impact on the generalization ability of models and performance outcomes. In addition, studies using public datasets generally do not specify inclusion and exclusion criteria, potentially leading to images with limited relevance and representativeness, increasing heterogeneity between studies, and affecting the reliability of the results. Furthermore, although 41 studies meet the inclusion criteria for the study, only half of the studies could be used to generate the specified contingency tables. Numerous studies use evaluation metrics like the Dice similarity coefficient, <italic>F</italic><sub>1</sub>-score, and Jaccard index. However, these metrics are not comprehensive and may provide insufficient information to fully construct a contingency table when used alone. Therefore, in certain conditions, it is necessary to compute, supplement, or derive the missing components of the confusion matrix to ensure a comprehensive and accurate evaluation. In future studies, clearly defined metrics should also be carefully considered [<xref ref-type="bibr" rid="ref68">68</xref>].</p><p>The sources and types of medical images are diverse, encompassing clinical laboratory reports, clinical images, and information derived from medical devices. The quality of the images notably affects the training and prediction capabilities of DL. In practical applications, factors such as image resolution, noise, and annotation quality should be considered, and appropriate preprocessing and augmentation measures should be taken to improve the performance and generalization ability of models. Due to the limited number of public datasets for TN, the public datasets used in the studies included are relatively homogeneous, such as the DDTI dataset [<xref ref-type="bibr" rid="ref69">69</xref>] and the TN3K dataset [<xref ref-type="bibr" rid="ref70">70</xref>]. Despite conducting an Egger linear regression test based on data extracted from the 41 studies, no evidence of publication bias is noted. However, the absence of prospective studies and the presence of negative results in studies may introduce potential biases. Therefore, there is a need for more high-quality studies, like prospective studies and clinical trials, to strengthen the existing evidence base [<xref ref-type="bibr" rid="ref71">71</xref>]. It has been suggested by investigators that using synthetic data to augment experiments can overcome the limitations posed by restricted data [<xref ref-type="bibr" rid="ref72">72</xref>].</p><p>Although DL algorithms have demonstrated promising diagnostic performance in the detection and segmentation of TNs, certain limitations persist within the included studies. First, most of the studies do not provide sufficient detailed information on model parameters or fine-tuning strategies, limiting our ability to evaluate the robustness, reproducibility, and generalizability of the models across different clinical scenarios. Second, few studies have reported on the computational cost, especially in terms of computational resources and processing time in the inference phase. These are especially critical for the deployment of models in real clinical settings, as processing speed and hardware efficiency directly affect their usability. In the absence of information on inference elapsed time or hardware requirements, it is difficult to determine whether these models are suitable for embedding in routine diagnostic processes. In addition, some of the studies exhibit potential biases, including selection bias and validation bias. These biases may arise from the inclusion of data from only a specific institution, a specific image quality, or a single population, which may limit the model&#x2019;s ability to generalize to a wide range of populations. At the same time, insufficient external data validation further affects the judgment of its clinical applicability. This is in line with the retrospective data issues mentioned by Chu et al [<xref ref-type="bibr" rid="ref73">73</xref>] in their meta-analysis of retinopathy of prematurity diagnosis.</p><p>The inclusion criteria for this study cover a wide range of study designs (like randomized controlled trials, cohort studies, case-control studies, and cross-sectional studies). It is worth noting that all the studies ultimately included are retrospective, which reduces methodological heterogeneity to a certain extent. Nevertheless, it also precludes us from carrying out subgroup analyses or adjustments with respect to study type. Consequently, it limits our ability to perform a subgroup analysis or adjustments for the performance of the model across different study contexts of a comprehensive assessment. In addition, retrospective studies are inherently more susceptible to selection bias and information bias, which may interfere with the estimation of model diagnostic performance. Therefore, caution should be exercised when interpreting the combined results and emphasizing the need for more prospective, high-quality studies in the future to validate the robustness and generalizability of the current findings.</p><p>We preliminarily believe that DL algorithms are capable of automatically segmenting and detecting TN, demonstrating high sensitivity and specificity comparable to that of clinical clinicians. Furthermore, these algorithms possess noticeable potential in the segmentation and detection of TN based on medical imaging. Nonetheless, it should also be noted that this finding comes from studies with relatively low methodological quality, which inevitably leads to an overestimation of the accuracy of the algorithms. The study design of ML-based segmentation and detection of TN still needs further refinement. In addition, AI application in medical diagnosis also raises important ethical and social issues, like transparency of algorithms, attribution of responsibility in case of diagnostic errors, and privacy protection of patient data [<xref ref-type="bibr" rid="ref74">74</xref>,<xref ref-type="bibr" rid="ref75">75</xref>]. Future research should pay more attention to these aspects in order to realize the responsible application of DL models in the clinic.</p></sec></body><back><ack><p>This work was supported by the National Natural Science Foundation of Zhejiang province, China (Y2100578 and Y2090486), the Medical and health research project of Zhejiang Province, China (2017RC011), the Project of Medical and health science and technology of Hangzhou, China (B20220098), and the Project of Public welfare application research of Huzhou municipal science and Technology Bureau, China (2021GY44).</p></ack><notes><sec><title>Data Availability</title><p>The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>All authors contributed to the study conception and design. JN contributed to writing&#x2014;original draft. YL handled writing&#x2014;review and editing. JN and YY managed conceptualization. JN, YY, and XW were responsible for methodology. JN and XC conducted formal analysis and investigation. XW and XC managed resources. JW handled supervision. All authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">AUC</term><def><p>area under the receiver operating characteristic curve</p></def></def-item><def-item><term id="abb3">CAD</term><def><p>computer-aided design</p></def></def-item><def-item><term id="abb4">CNN</term><def><p>convolutional neural network</p></def></def-item><def-item><term id="abb5">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb6">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb7">ML</term><def><p>machine learning</p></def></def-item><def-item><term id="abb8">PRISMA</term><def><p>Preferred Reporting Items for Systematic reviews and Meta-Analyses</p></def></def-item><def-item><term id="abb9">PROSPERO</term><def><p>International Prospective Register of Systematic Reviews</p></def></def-item><def-item><term id="abb10">QUADAS-AI</term><def><p>quality assessment of diagnostic accuracy studies using AI</p></def></def-item><def-item><term id="abb11">ROB</term><def><p>risk of bias</p></def></def-item><def-item><term id="abb12">SROC</term><def><p>summary receiver operating characteristics</p></def></def-item><def-item><term id="abb13">TC</term><def><p>thyroid cancer</p></def></def-item><def-item><term id="abb14">TL</term><def><p>transfer learning</p></def></def-item><def-item><term id="abb15">TN</term><def><p>thyroid nodule</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pellegriti</surname><given-names>G</given-names> </name><name name-style="western"><surname>Frasca</surname><given-names>F</given-names> </name><name name-style="western"><surname>Regalbuto</surname><given-names>C</given-names> </name><name name-style="western"><surname>Squatrito</surname><given-names>S</given-names> </name><name name-style="western"><surname>Vigneri</surname><given-names>R</given-names> </name></person-group><article-title>Worldwide increasing incidence of thyroid cancer: update on epidemiology and risk factors</article-title><source>J Cancer Epidemiol</source><year>2013</year><volume>2013</volume><fpage>965212</fpage><pub-id pub-id-type="doi">10.1155/2013/965212</pub-id><pub-id pub-id-type="medline">23737785</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bray</surname><given-names>F</given-names> </name><name name-style="western"><surname>Ferlay</surname><given-names>J</given-names> </name><name name-style="western"><surname>Soerjomataram</surname><given-names>I</given-names> </name><name name-style="western"><surname>Siegel</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Torre</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Jemal</surname><given-names>A</given-names> </name></person-group><article-title>Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries</article-title><source>CA Cancer J Clin</source><year>2018</year><month>11</month><volume>68</volume><issue>6</issue><fpage>394</fpage><lpage>424</lpage><pub-id pub-id-type="doi">10.3322/caac.21492</pub-id><pub-id pub-id-type="medline">30207593</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Giger</surname><given-names>ML</given-names> </name></person-group><article-title>Machine learning in medical imaging</article-title><source>J Am Coll Radiol</source><year>2018</year><month>03</month><volume>15</volume><issue>3 Pt B</issue><fpage>512</fpage><lpage>520</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2017.12.028</pub-id><pub-id pub-id-type="medline">29398494</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chartrand</surname><given-names>G</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>PM</given-names> </name><name name-style="western"><surname>Vorontsov</surname><given-names>E</given-names> </name><etal/></person-group><article-title>Deep learning: a primer for radiologists</article-title><source>Radiographics</source><year>2017</year><volume>37</volume><issue>7</issue><fpage>2113</fpage><lpage>2131</lpage><pub-id pub-id-type="doi">10.1148/rg.2017170077</pub-id><pub-id pub-id-type="medline">29131760</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chan</surname><given-names>HP</given-names> </name><name name-style="western"><surname>Hadjiiski</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Samala</surname><given-names>RK</given-names> </name></person-group><article-title>Computer-aided diagnosis in the era of deep learning</article-title><source>Med Phys</source><year>2020</year><month>06</month><volume>47</volume><issue>5</issue><fpage>e218</fpage><lpage>e227</lpage><pub-id pub-id-type="doi">10.1002/mp.13764</pub-id><pub-id pub-id-type="medline">32418340</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wildman-Tobriner</surname><given-names>B</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Allen</surname><given-names>BC</given-names> </name><name name-style="western"><surname>Ho</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Mazurowski</surname><given-names>MA</given-names> </name></person-group><article-title>Simplifying risk stratification for thyroid nodules on ultrasound: validation and performance of an artificial intelligence thyroid imaging reporting and data system</article-title><source>Curr Probl Diagn Radiol</source><year>2024</year><volume>53</volume><issue>6</issue><fpage>695</fpage><lpage>699</lpage><pub-id pub-id-type="doi">10.1067/j.cpradiol.2024.07.006</pub-id><pub-id pub-id-type="medline">39033064</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zheng</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>A segmentation-based algorithm for classification of benign and malignancy thyroid nodules with multi-feature information</article-title><source>Biomed Eng Lett</source><year>2024</year><month>07</month><volume>14</volume><issue>4</issue><fpage>785</fpage><lpage>800</lpage><pub-id pub-id-type="doi">10.1007/s13534-024-00375-2</pub-id><pub-id pub-id-type="medline">38946824</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>QY</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>CJ</given-names> </name><etal/></person-group><article-title>Predicting malignancy in thyroid nodules based on conventional ultrasound and elastography: the value of predictive models in a multi-center study</article-title><source>Endocrine</source><year>2023</year><month>04</month><volume>80</volume><issue>1</issue><fpage>111</fpage><lpage>123</lpage><pub-id pub-id-type="doi">10.1007/s12020-022-03271-w</pub-id><pub-id pub-id-type="medline">36495391</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name></person-group><article-title>Diagnostic accuracy of S-Detect in distinguishing benign and malignant thyroid nodules: a meta-analysis</article-title><source>PLoS ONE</source><year>2022</year><volume>17</volume><issue>8</issue><fpage>e0272149</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0272149</pub-id><pub-id pub-id-type="medline">35930525</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Deng</surname><given-names>C</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Tang</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Application of CT and MRI images based on artificial intelligence to predict lymph node metastases in patients with oral squamous cell carcinoma: a subgroup meta-analysis</article-title><source>Front Oncol</source><year>2024</year><volume>14</volume><fpage>1395159</fpage><pub-id pub-id-type="doi">10.3389/fonc.2024.1395159</pub-id><pub-id pub-id-type="medline">38957322</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhu</surname><given-names>PS</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>YR</given-names> </name><name name-style="western"><surname>Ren</surname><given-names>JY</given-names> </name><etal/></person-group><article-title>Ultrasound-based deep learning using the VGGNet model for the differentiation of benign and malignant thyroid nodules: a meta-analysis</article-title><source>Front Oncol</source><year>2022</year><volume>12</volume><fpage>944859</fpage><pub-id pub-id-type="doi">10.3389/fonc.2022.944859</pub-id><pub-id pub-id-type="medline">36249056</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>HajiEsmailPoor</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Kargar</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Tabnak</surname><given-names>P</given-names> </name></person-group><article-title>Radiomics diagnostic performance in predicting lymph node metastasis of papillary thyroid carcinoma: a systematic review and meta-analysis</article-title><source>Eur J Radiol</source><year>2023</year><month>11</month><volume>168</volume><fpage>111129</fpage><pub-id pub-id-type="doi">10.1016/j.ejrad.2023.111129</pub-id><pub-id pub-id-type="medline">37820522</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sounderajah</surname><given-names>V</given-names> </name><name name-style="western"><surname>Ashrafian</surname><given-names>H</given-names> </name><name name-style="western"><surname>Rose</surname><given-names>S</given-names> </name><etal/></person-group><article-title>A quality assessment tool for artificial intelligence-centered diagnostic test accuracy studies: QUADAS-AI</article-title><source>Nat Med</source><year>2021</year><month>10</month><volume>27</volume><issue>10</issue><fpage>1663</fpage><lpage>1665</lpage><pub-id pub-id-type="doi">10.1038/s41591-021-01517-0</pub-id><pub-id pub-id-type="medline">34635854</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McInnes</surname><given-names>MDF</given-names> </name><name name-style="western"><surname>Moher</surname><given-names>D</given-names> </name><name name-style="western"><surname>Thombs</surname><given-names>BD</given-names> </name><etal/></person-group><article-title>Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: the PRISMA-DTA statement</article-title><source>JAMA</source><year>2018</year><month>01</month><day>23</day><volume>319</volume><issue>4</issue><fpage>388</fpage><lpage>396</lpage><pub-id pub-id-type="doi">10.1001/jama.2017.19163</pub-id><pub-id pub-id-type="medline">29362800</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shankarlal</surname><given-names>B</given-names> </name><name name-style="western"><surname>Sathya</surname><given-names>PD</given-names> </name><name name-style="western"><surname>Sakthivel</surname><given-names>VP</given-names> </name></person-group><article-title>Computer-aided detection and diagnosis of thyroid nodules using machine and deep learning classification algorithms</article-title><source>IETE J Res</source><year>2023</year><month>02</month><day>17</day><volume>69</volume><issue>2</issue><fpage>995</fpage><lpage>1006</lpage><pub-id pub-id-type="doi">10.1080/03772063.2020.1844083</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><name name-style="western"><surname>Nie</surname><given-names>X</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>H-Net: A dual-decoder enhanced FCNN for automated biomedical image diagnosis</article-title><source>Inf Sci (Ny)</source><year>2022</year><month>10</month><volume>613</volume><fpage>575</fpage><lpage>590</lpage><pub-id pub-id-type="doi">10.1016/j.ins.2022.09.019</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>P</given-names> </name></person-group><article-title>Research on thyroid nodule segmentation using an improved U-Net network</article-title><source>RIMNI</source><year>2024</year><volume>40</volume><issue>2</issue><pub-id pub-id-type="doi">10.23967/j.rimni.2024.05.012</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>R</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>P</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>Y</given-names> </name></person-group><article-title>A multiscale attentional unet model for automatic segmentation in medical ultrasound images</article-title><source>Ultrason Imaging</source><year>2023</year><month>07</month><volume>45</volume><issue>4</issue><fpage>159</fpage><lpage>174</lpage><pub-id pub-id-type="doi">10.1177/01617346231169789</pub-id><pub-id pub-id-type="medline">37114669</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Usman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rehman</surname><given-names>A</given-names> </name><name name-style="western"><surname>Masood</surname><given-names>S</given-names> </name><name name-style="western"><surname>Khan</surname><given-names>TM</given-names> </name><name name-style="western"><surname>Qadir</surname><given-names>J</given-names> </name></person-group><article-title>Intelligent healthcare system for IoMT-integrated sonography: leveraging multi-scale self-guided attention networks and dynamic self-distillation</article-title><source>Internet of Things</source><year>2024</year><month>04</month><volume>25</volume><fpage>101065</fpage><pub-id pub-id-type="doi">10.1016/j.iot.2024.101065</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>T</given-names> </name></person-group><article-title>GLFNet: global-local fusion network for the segmentation in ultrasound images</article-title><source>Comput Biol Med</source><year>2024</year><month>03</month><volume>171</volume><fpage>108103</fpage><pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108103</pub-id><pub-id pub-id-type="medline">38335822</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>T</given-names> </name></person-group><article-title>CRSANet: class representations self-attention network for the segmentation of thyroid nodules</article-title><source>Biomed Signal Process Control</source><year>2024</year><month>05</month><volume>91</volume><fpage>105917</fpage><pub-id pub-id-type="doi">10.1016/j.bspc.2023.105917</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>K</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>W</given-names> </name><etal/></person-group><article-title>Learning contextual representations with copula function for medical image segmentation</article-title><source>Biomed Signal Process Control</source><year>2023</year><month>08</month><volume>85</volume><fpage>104900</fpage><pub-id pub-id-type="doi">10.1016/j.bspc.2023.104900</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>S</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>Y</given-names> </name></person-group><article-title>A weakly supervised deep active contour model for nodule segmentation in thyroid ultrasound images</article-title><source>Pattern Recognit Lett</source><year>2023</year><month>01</month><volume>165</volume><fpage>128</fpage><lpage>137</lpage><pub-id pub-id-type="doi">10.1016/j.patrec.2022.12.015</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>G</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>R</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>K</given-names> </name><name name-style="western"><surname>Geng</surname><given-names>C</given-names> </name><name name-style="western"><surname>Lyu</surname><given-names>L</given-names> </name></person-group><article-title>Fusing enhanced transformer and large kernel CNN for malignant thyroid nodule segmentation</article-title><source>Biomed Signal Process Control</source><year>2023</year><month>05</month><volume>83</volume><fpage>104636</fpage><pub-id pub-id-type="doi">10.1016/j.bspc.2023.104636</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>H</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>C</given-names> </name><etal/></person-group><article-title>FDE-net: frequency-domain enhancement network using dynamic-scale dilated convolution for thyroid nodule segmentation</article-title><source>Comput Biol Med</source><year>2023</year><month>02</month><volume>153</volume><fpage>106514</fpage><pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.106514</pub-id><pub-id pub-id-type="medline">36628913</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ali</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>M</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>J</given-names> </name></person-group><article-title>CIL-Net: densely connected context information learning network for boosting thyroid nodule segmentation using ultrasound images</article-title><source>Cogn Comput</source><year>2024</year><month>05</month><volume>16</volume><issue>3</issue><fpage>1176</fpage><lpage>1197</lpage><pub-id pub-id-type="doi">10.1007/s12559-024-10289-x</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ajilisa</surname><given-names>OA</given-names> </name><name name-style="western"><surname>Jagathy Raj</surname><given-names>VP</given-names> </name><name name-style="western"><surname>Sabu</surname><given-names>MK</given-names> </name></person-group><article-title>Segmentation of thyroid nodules from ultrasound images using convolutional neural network architectures</article-title><source>IFS</source><year>2022</year><volume>43</volume><issue>1</issue><fpage>687</fpage><lpage>705</lpage><pub-id pub-id-type="doi">10.3233/JIFS-212398</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rehman</surname><given-names>HAU</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>CY</given-names> </name><name name-style="western"><surname>Su</surname><given-names>SF</given-names> </name></person-group><article-title>Deep learning based fast screening approach on ultrasound images for thyroid nodules diagnosis</article-title><source>Diagnostics (Basel)</source><year>2021</year><month>11</month><day>26</day><volume>11</volume><issue>12</issue><fpage>2209</fpage><pub-id pub-id-type="doi">10.3390/diagnostics11122209</pub-id><pub-id pub-id-type="medline">34943444</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fu</surname><given-names>CP</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Fuh</surname><given-names>CS</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>RF</given-names> </name></person-group><article-title>Stratifying high-risk thyroid nodules using a novel deep learning system</article-title><source>Exp Clin Endocrinol Diabetes</source><year>2023</year><month>10</month><volume>131</volume><issue>10</issue><fpage>508</fpage><lpage>514</lpage><pub-id pub-id-type="doi">10.1055/a-2122-5585</pub-id><pub-id pub-id-type="medline">37604165</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Agustin</surname><given-names>S</given-names> </name><name name-style="western"><surname>S</surname><given-names>S</given-names> </name><name name-style="western"><surname>James</surname><given-names>A</given-names> </name><name name-style="western"><surname>Simon</surname><given-names>P</given-names> </name></person-group><article-title>Residual U-Net approach for thyroid nodule detection and classification from thyroid ultrasound images</article-title><source>Automatika</source><year>2024</year><month>07</month><day>2</day><volume>65</volume><issue>3</issue><fpage>726</fpage><lpage>737</lpage><pub-id pub-id-type="doi">10.1080/00051144.2024.2316503</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Buda</surname><given-names>M</given-names> </name><name name-style="western"><surname>Wildman-Tobriner</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hoang</surname><given-names>JK</given-names> </name><etal/></person-group><article-title>Management of thyroid nodules seen on US images: deep learning may match performance of radiologists</article-title><source>Radiology</source><year>2019</year><month>09</month><volume>292</volume><issue>3</issue><fpage>695</fpage><lpage>701</lpage><pub-id pub-id-type="doi">10.1148/radiol.2019181343</pub-id><pub-id pub-id-type="medline">31287391</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>H</given-names> </name><name name-style="western"><surname>Pan</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>ThyroidNet: a deep learning network for localization and classification of thyroid nodules</article-title><source>Comput Model Eng Sci</source><year>2023</year><month>12</month><day>30</day><volume>139</volume><issue>1</issue><fpage>361</fpage><lpage>382</lpage><pub-id pub-id-type="doi">10.32604/cmes.2023.031229</pub-id><pub-id pub-id-type="medline">38566835</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gong</surname><given-names>ZJ</given-names> </name><name name-style="western"><surname>Xin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yin</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Diagnostic value of artificial intelligence-assistant diagnostic system combined with contrast-enhanced ultrasound in thyroid TI-RADS 4 nodules</article-title><source>J Ultrasound Med</source><year>2023</year><month>07</month><volume>42</volume><issue>7</issue><fpage>1527</fpage><lpage>1535</lpage><pub-id pub-id-type="doi">10.1002/jum.16170</pub-id><pub-id pub-id-type="medline">36723397</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>X</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>BJ</given-names> </name><name name-style="western"><surname>Lei</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Thyroid gland delineation in noncontrast-enhanced CTs using deep convolutional neural networks</article-title><source>Phys Med Biol</source><year>2021</year><month>02</month><day>16</day><volume>66</volume><issue>5</issue><fpage>055007</fpage><pub-id pub-id-type="doi">10.1088/1361-6560/abc5a6</pub-id><pub-id pub-id-type="medline">33590826</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>M</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><etal/></person-group><article-title>SDA-Net: self-distillation driven deformable attentive aggregation network for thyroid nodule identification in ultrasound images</article-title><source>Artif Intell Med</source><year>2023</year><month>12</month><volume>146</volume><fpage>102699</fpage><pub-id pub-id-type="doi">10.1016/j.artmed.2023.102699</pub-id><pub-id pub-id-type="medline">38042598</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><name name-style="western"><surname>Yan</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Comparison of diagnostic accuracy and utility of artificial intelligence&#x2013;optimized ACR TI-RADS and original ACR TI-RADS: a multi-center validation study based on 2061 thyroid nodules</article-title><source>Eur Radiol</source><year>2022</year><month>11</month><volume>32</volume><issue>11</issue><fpage>7733</fpage><lpage>7742</lpage><pub-id pub-id-type="doi">10.1007/s00330-022-08827-y</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Si</surname><given-names>CF</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Cui</surname><given-names>YY</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>YJ</given-names> </name><name name-style="western"><surname>Cui</surname><given-names>KF</given-names> </name></person-group><article-title>Diagnostic and therapeutic performances of three score-based thyroid imaging reporting and data systems after application of equal size thresholds</article-title><source>Quant Imaging Med Surg</source><year>2023</year><month>04</month><day>1</day><volume>13</volume><issue>4</issue><fpage>2109</fpage><lpage>2118</lpage><pub-id pub-id-type="doi">10.21037/qims-22-592</pub-id><pub-id pub-id-type="medline">37064344</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Srivastava</surname><given-names>R</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>P</given-names> </name></person-group><article-title>GSO-CNN-based model for the identification and classification of thyroid nodule in medical USG images</article-title><source>Netw Model Anal Health Inform Bioinforma</source><year>2022</year><month>12</month><volume>11</volume><issue>1</issue><fpage>1</fpage><lpage>14</lpage><pub-id pub-id-type="doi">10.1007/s13721-022-00388-w</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Srivastava</surname><given-names>R</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>P</given-names> </name></person-group><article-title>Optimizing CNN based model for thyroid nodule classification using data augmentation, segmentation and boundary detection techniques</article-title><source>Multimed Tools Appl</source><year>2023</year><month>11</month><volume>82</volume><issue>26</issue><fpage>41037</fpage><lpage>41072</lpage><pub-id pub-id-type="doi">10.1007/s11042-023-15068-8</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tong</surname><given-names>WJ</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>SH</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>MQ</given-names> </name><etal/></person-group><article-title>Integration of artificial intelligence decision aids to reduce workload and enhance efficiency in thyroid nodule management</article-title><source>JAMA Netw Open</source><year>2023</year><month>05</month><day>1</day><volume>6</volume><issue>5</issue><fpage>e2313674</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2023.13674</pub-id><pub-id pub-id-type="medline">37191957</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>W</given-names> </name><name name-style="western"><surname>Jia</surname><given-names>X</given-names> </name><name name-style="western"><surname>Mei</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Generalizability and diagnostic performance of AI models for thyroid US</article-title><source>Radiology</source><year>2023</year><month>06</month><volume>307</volume><issue>5</issue><fpage>e221157</fpage><pub-id pub-id-type="doi">10.1148/radiol.221157</pub-id><pub-id pub-id-type="medline">37338356</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>CK</given-names> </name><name name-style="western"><surname>Ren</surname><given-names>TT</given-names> </name><name name-style="western"><surname>Yin</surname><given-names>YF</given-names> </name><etal/></person-group><article-title>A comparative analysis of two machine learning-based diagnostic patterns with thyroid imaging reporting and data system for thyroid nodules: diagnostic performance and unnecessary biopsy rate</article-title><source>Thyroid</source><year>2021</year><month>03</month><volume>31</volume><issue>3</issue><fpage>470</fpage><lpage>481</lpage><pub-id pub-id-type="doi">10.1089/thy.2020.0305</pub-id><pub-id pub-id-type="medline">32781915</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zheng</surname><given-names>LL</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Diagnostic performance of artificial intelligence-based computer-aided diagnosis system in longitudinal and transverse ultrasonic views for differentiating thyroid nodules</article-title><source>Front Endocrinol (Lausanne)</source><year>2023</year><volume>14</volume><fpage>1137700</fpage><pub-id pub-id-type="doi">10.3389/fendo.2023.1137700</pub-id><pub-id pub-id-type="medline">36864838</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zheng</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>L</given-names> </name><name name-style="western"><surname>Qiu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>A</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>P</given-names> </name><name name-style="western"><surname>Xue</surname><given-names>Z</given-names> </name></person-group><article-title>Automated detection and recognition of thyroid nodules in ultrasound images using Improve Cascade Mask R-CNN</article-title><source>Multimed Tools Appl</source><year>2022</year><month>04</month><volume>81</volume><issue>10</issue><fpage>13253</fpage><lpage>13273</lpage><pub-id pub-id-type="doi">10.1007/s11042-021-10939-4</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Aided diagnosis of thyroid nodules based on an all-optical diffraction neural network</article-title><source>Quant Imaging Med Surg</source><year>2023</year><month>09</month><day>1</day><volume>13</volume><issue>9</issue><fpage>5713</fpage><lpage>5726</lpage><pub-id pub-id-type="doi">10.21037/qims-23-98</pub-id><pub-id pub-id-type="medline">37711804</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>LT</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>FJ</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>DZ</given-names> </name><etal/></person-group><article-title>A comparison of the performances of artificial intelligence system and radiologists in the ultrasound diagnosis of thyroid nodules</article-title><source>Curr Med Imaging</source><year>2022</year><volume>18</volume><issue>13</issue><fpage>1369</fpage><lpage>1377</lpage><pub-id pub-id-type="doi">10.2174/1573405618666220422132251</pub-id><pub-id pub-id-type="medline">35466880</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Heng</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Automated diagnosis and management of follicular thyroid nodules based on the devised small-dataset interpretable foreground optimization network deep learning: a multicenter diagnostic study</article-title><source>Int J Surg</source><year>2023</year><month>09</month><day>1</day><volume>109</volume><issue>9</issue><fpage>2732</fpage><lpage>2741</lpage><pub-id pub-id-type="doi">10.1097/JS9.0000000000000506</pub-id><pub-id pub-id-type="medline">37204464</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>YQ</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>TT</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>AX</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>Y</given-names> </name></person-group><article-title>Applying machine-learning models to differentiate benign and malignant thyroid nodules classified as C-TIRADS 4 based on 2D-ultrasound combined with five contrast-enhanced ultrasound key frames</article-title><source>Front Endocrinol (Lausanne)</source><year>2024</year><volume>15</volume><fpage>1299686</fpage><pub-id pub-id-type="doi">10.3389/fendo.2024.1299686</pub-id><pub-id pub-id-type="medline">38633756</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Koh</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>E</given-names> </name><name name-style="western"><surname>Han</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Diagnosis of thyroid nodules on ultrasonography by a deep convolutional neural network</article-title><source>Sci Rep</source><year>2020</year><month>09</month><day>17</day><volume>10</volume><issue>1</issue><fpage>15245</fpage><pub-id pub-id-type="doi">10.1038/s41598-020-72270-6</pub-id><pub-id pub-id-type="medline">32943696</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>The value of the computer-aided diagnosis system for thyroid lesions based on computed tomography images</article-title><source>Quant Imaging Med Surg</source><year>2019</year><month>04</month><volume>9</volume><issue>4</issue><fpage>642</fpage><lpage>653</lpage><pub-id pub-id-type="doi">10.21037/qims.2019.04.01</pub-id><pub-id pub-id-type="medline">31143655</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>C</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>Q</given-names> </name><etal/></person-group><article-title>Evaluation of a deep learning-based computer-aided diagnosis system for distinguishing benign from malignant thyroid nodules in ultrasound images</article-title><source>Med Phys</source><year>2020</year><month>09</month><volume>47</volume><issue>9</issue><fpage>3952</fpage><lpage>3960</lpage><pub-id pub-id-type="doi">10.1002/mp.14301</pub-id><pub-id pub-id-type="medline">32473030</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ye</surname><given-names>H</given-names> </name><name name-style="western"><surname>Hang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><etal/></person-group><article-title>An intelligent platform for ultrasound diagnosis of thyroid nodules</article-title><source>Sci Rep</source><year>2020</year><month>08</month><day>6</day><volume>10</volume><issue>1</issue><fpage>13223</fpage><pub-id pub-id-type="doi">10.1038/s41598-020-70159-y</pub-id><pub-id pub-id-type="medline">32764673</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>B</given-names> </name><name name-style="western"><surname>Tian</surname><given-names>J</given-names> </name><name name-style="western"><surname>Pei</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Machine learning-assisted system for thyroid nodule diagnosis</article-title><source>Thyroid</source><year>2019</year><month>06</month><volume>29</volume><issue>6</issue><fpage>858</fpage><lpage>867</lpage><pub-id pub-id-type="doi">10.1089/thy.2018.0380</pub-id><pub-id pub-id-type="medline">30929637</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>W</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>A new classification method in ultrasound images of benign and malignant thyroid nodules based on transfer learning and deep convolutional neural network</article-title><source>Complexity</source><year>2021</year><month>01</month><volume>2021</volume><issue>1</issue><fpage>1</fpage><lpage>6296811</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://onlinelibrary.wiley.com/toc/8503/2021/1">https://onlinelibrary.wiley.com/toc/8503/2021/1</ext-link></comment><pub-id pub-id-type="doi">10.1155/2021/6296811</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kwon</surname><given-names>SW</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Kang</surname><given-names>JY</given-names> </name><name name-style="western"><surname>Jang</surname><given-names>WI</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>GH</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>MC</given-names> </name></person-group><article-title>Ultrasonographic thyroid nodule classification using a deep convolutional neural network with surgical pathology</article-title><source>J Digit Imaging</source><year>2020</year><month>10</month><volume>33</volume><issue>5</issue><fpage>1202</fpage><lpage>1208</lpage><pub-id pub-id-type="doi">10.1007/s10278-020-00362-w</pub-id><pub-id pub-id-type="medline">32705433</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><collab>American Thyroid Association (ATA) Guidelines Taskforce on Thyroid Nodules and Differentiated Thyroid Cancer</collab><name name-style="western"><surname>Cooper</surname><given-names>DS</given-names> </name><name name-style="western"><surname>Doherty</surname><given-names>GM</given-names> </name><etal/></person-group><article-title>Revised American Thyroid Association management guidelines for patients with thyroid nodules and differentiated thyroid cancer</article-title><source>Thyroid</source><year>2009</year><month>11</month><volume>19</volume><issue>11</issue><fpage>1167</fpage><lpage>1214</lpage><pub-id pub-id-type="doi">10.1089/thy.2009.0110</pub-id><pub-id pub-id-type="medline">19860577</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Singh Ospina</surname><given-names>N</given-names> </name><name name-style="western"><surname>Maraka</surname><given-names>S</given-names> </name><name name-style="western"><surname>Espinosa DeYcaza</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Diagnostic accuracy of thyroid nodule growth to predict malignancy in thyroid nodules with benign cytology: systematic review and meta&#x2010;analysis</article-title><source>Clin Endocrinol (Oxf)</source><year>2016</year><month>07</month><volume>85</volume><issue>1</issue><fpage>122</fpage><lpage>131</lpage><pub-id pub-id-type="doi">10.1111/cen.12975</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>WL</surname><given-names>VCB</given-names> </name></person-group><article-title>Machine learning in medicine</article-title><source>N Engl J Med</source><year>2019</year><month>06</month><day>27</day><volume>380</volume><issue>26</issue><fpage>2588</fpage><lpage>2590</lpage><pub-id pub-id-type="doi">10.1056/NEJMc1906060</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Goecks</surname><given-names>J</given-names> </name><name name-style="western"><surname>Jalili</surname><given-names>V</given-names> </name><name name-style="western"><surname>Heiser</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Gray</surname><given-names>JW</given-names> </name></person-group><article-title>How machine learning will transform biomedicine</article-title><source>Cell</source><year>2020</year><month>04</month><day>2</day><volume>181</volume><issue>1</issue><fpage>92</fpage><lpage>101</lpage><pub-id pub-id-type="doi">10.1016/j.cell.2020.03.022</pub-id><pub-id pub-id-type="medline">32243801</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abdelhafiz</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bi</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ammar</surname><given-names>R</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Nabavi</surname><given-names>S</given-names> </name></person-group><article-title>Convolutional neural network for automated mass segmentation in mammography</article-title><source>BMC Bioinformatics</source><year>2020</year><month>12</month><day>9</day><volume>21</volume><issue>Suppl 1</issue><fpage>192</fpage><pub-id pub-id-type="doi">10.1186/s12859-020-3521-y</pub-id><pub-id pub-id-type="medline">33297952</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kooi</surname><given-names>T</given-names> </name><name name-style="western"><surname>Litjens</surname><given-names>G</given-names> </name><name name-style="western"><surname>van Ginneken</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Large scale deep learning for computer aided detection of mammographic lesions</article-title><source>Med Image Anal</source><year>2017</year><month>01</month><volume>35</volume><fpage>303</fpage><lpage>312</lpage><pub-id pub-id-type="doi">10.1016/j.media.2016.07.007</pub-id><pub-id pub-id-type="medline">27497072</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>BK</given-names> </name><name name-style="western"><surname>Madusanka</surname><given-names>N</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>HK</given-names> </name><etal/></person-group><article-title>Convolutional neural network-based MR image analysis for Alzheimer&#x2019;s disease classification</article-title><source>Curr Med Imaging Rev</source><year>2020</year><volume>16</volume><issue>1</issue><fpage>27</fpage><lpage>35</lpage><pub-id pub-id-type="doi">10.2174/1573405615666191021123854</pub-id><pub-id pub-id-type="medline">31989891</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>T</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Kong</surname><given-names>D</given-names> </name></person-group><article-title>Ultrasound image-based thyroid nodule automatic segmentation using convolutional neural networks</article-title><source>Int J Comput Assist Radiol Surg</source><year>2017</year><month>11</month><volume>12</volume><issue>11</issue><fpage>1895</fpage><lpage>1910</lpage><pub-id pub-id-type="doi">10.1007/s11548-017-1649-7</pub-id><pub-id pub-id-type="medline">28762196</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>H</given-names> </name><name name-style="western"><surname>Weng</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>An improved deep learning approach for detection of thyroid papillary cancer in ultrasound images</article-title><source>Sci Rep</source><year>2018</year><month>04</month><day>26</day><volume>8</volume><issue>1</issue><fpage>6600</fpage><pub-id pub-id-type="doi">10.1038/s41598-018-25005-7</pub-id><pub-id pub-id-type="medline">29700427</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Leeflang</surname><given-names>MMG</given-names> </name><name name-style="western"><surname>Allerberger</surname><given-names>F</given-names> </name></person-group><article-title>How to: evaluate a diagnostic test</article-title><source>Clin Microbiol Infect</source><year>2019</year><month>01</month><volume>25</volume><issue>1</issue><fpage>54</fpage><lpage>59</lpage><pub-id pub-id-type="doi">10.1016/j.cmi.2018.06.011</pub-id><pub-id pub-id-type="medline">29906592</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cleere</surname><given-names>EF</given-names> </name><name name-style="western"><surname>Davey</surname><given-names>MG</given-names> </name><name name-style="western"><surname>O&#x2019;Neill</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Radiomic detection of malignancy within thyroid nodules using ultrasonography-a systematic review and meta-analysis</article-title><source>Diagnostics (Basel)</source><year>2022</year><month>03</month><day>24</day><volume>12</volume><issue>4</issue><fpage>794</fpage><pub-id pub-id-type="doi">10.3390/diagnostics12040794</pub-id><pub-id pub-id-type="medline">35453841</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>WJ</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>LR</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>ZM</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>JQ</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>BY</given-names> </name></person-group><article-title>Effectiveness evaluation of computer-aided diagnosis system for the diagnosis of thyroid nodules on ultrasound: a systematic review and meta-analysis</article-title><source>Medicine (Baltimore)</source><year>2019</year><month>08</month><volume>98</volume><issue>32</issue><fpage>e16379</fpage><pub-id pub-id-type="doi">10.1097/MD.0000000000016379</pub-id><pub-id pub-id-type="medline">31393347</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>PHC</given-names> </name><name name-style="western"><surname>Krause</surname><given-names>J</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>L</given-names> </name></person-group><article-title>How to read articles that use machine learning: users&#x2019; guides to the medical literature</article-title><source>JAMA</source><year>2019</year><month>11</month><day>12</day><volume>322</volume><issue>18</issue><fpage>1806</fpage><lpage>1816</lpage><pub-id pub-id-type="doi">10.1001/jama.2019.16489</pub-id><pub-id pub-id-type="medline">31714992</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Lea</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vargas</surname><given-names>C</given-names> </name><name name-style="western"><surname>Narv&#x00E1;ez</surname><given-names>F</given-names> </name><name name-style="western"><surname>Dur&#x00E1;n</surname><given-names>O</given-names> </name><name name-style="western"><surname>Mu&#x00F1;oz</surname><given-names>E</given-names> </name><name name-style="western"><surname>Romero</surname><given-names>E</given-names> </name></person-group><article-title>An open access thyroid ultrasound image database</article-title><conf-name>Tenth International Symposium on Medical Information Processing and Analysis</conf-name><conf-date>2014</conf-date><conf-loc>Cartagena de Indias, Colombia</conf-loc><pub-id pub-id-type="doi">10.1117/12.2073532</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Gong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>G</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Multi-task learning for thyroid nodule segmentation with thyroid region prior</article-title><conf-name>2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI)</conf-name><conf-date>2021</conf-date><conf-loc>Nice, France</conf-loc><fpage>257</fpage><lpage>261</lpage><pub-id pub-id-type="doi">10.1109/ISBI48211.2021.9434087</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nagendran</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lovejoy</surname><given-names>CA</given-names> </name><etal/></person-group><article-title>Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies</article-title><source>BMJ</source><year>2020</year><month>03</month><day>25</day><volume>368</volume><fpage>m689</fpage><pub-id pub-id-type="doi">10.1136/bmj.m689</pub-id><pub-id pub-id-type="medline">32213531</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hoffmann</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bar-Sinai</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>LM</given-names> </name><etal/></person-group><article-title>Machine learning in a data-limited regime: augmenting experiments with synthetic data uncovers order in crumpled sheets</article-title><source>Sci Adv</source><year>2019</year><month>04</month><volume>5</volume><issue>4</issue><fpage>eaau6792</fpage><pub-id pub-id-type="doi">10.1126/sciadv.aau6792</pub-id><pub-id pub-id-type="medline">31032399</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Image analysis-based machine learning for the diagnosis of retinopathy of prematurity: a meta-analysis and systematic review</article-title><source>Ophthalmol Retina</source><year>2024</year><month>07</month><volume>8</volume><issue>7</issue><fpage>678</fpage><lpage>687</lpage><pub-id pub-id-type="doi">10.1016/j.oret.2024.01.013</pub-id><pub-id pub-id-type="medline">38237772</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Price</surname><given-names>WN</given-names>  <suffix>II</suffix></name><name name-style="western"><surname>Cohen</surname><given-names>IG</given-names> </name></person-group><article-title>Privacy in the age of medical big data</article-title><source>Nat Med</source><year>2019</year><month>01</month><volume>25</volume><issue>1</issue><fpage>37</fpage><lpage>43</lpage><pub-id pub-id-type="doi">10.1038/s41591-018-0272-7</pub-id><pub-id pub-id-type="medline">30617331</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>RJ</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>MY</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>TY</given-names> </name><name name-style="western"><surname>Williamson</surname><given-names>DFK</given-names> </name><name name-style="western"><surname>Mahmood</surname><given-names>F</given-names> </name></person-group><article-title>Synthetic data in machine learning for medicine and healthcare</article-title><source>Nat Biomed Eng</source><year>2021</year><month>06</month><volume>5</volume><issue>6</issue><fpage>493</fpage><lpage>497</lpage><pub-id pub-id-type="doi">10.1038/s41551-021-00751-8</pub-id><pub-id pub-id-type="medline">34131324</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Search terms and search strategy.</p><media xlink:href="jmir_v27i1e73516_app1.docx" xlink:title="DOCX File, 331 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Study design and basic demographics.</p><media xlink:href="jmir_v27i1e73516_app2.xlsx" xlink:title="XLSX File, 15 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Methods of model training and validation.</p><media xlink:href="jmir_v27i1e73516_app3.xlsx" xlink:title="XLSX File, 11 KB"/></supplementary-material><supplementary-material id="app4"><label>Multimedia Appendix 4</label><p>Indicators, algorithms, and data sources.</p><media xlink:href="jmir_v27i1e73516_app4.xlsx" xlink:title="XLSX File, 15 KB"/></supplementary-material><supplementary-material id="app5"><label>Multimedia Appendix 5</label><p>Publication bias.</p><media xlink:href="jmir_v27i1e73516_app5.png" xlink:title="PNG File, 71 KB"/></supplementary-material><supplementary-material id="app6"><label>Multimedia Appendix 6</label><p>Summary estimate of pooled performance using forest plot: (A) Forest plot of studies on segmentation tasks (14 studies) and (B) forest plot of studies on detection tasks (27 studies).</p><media xlink:href="jmir_v27i1e73516_app6.docx" xlink:title="DOCX File, 12051 KB"/></supplementary-material><supplementary-material id="app7"><label>Multimedia Appendix 7</label><p>Summary estimate of pooled performance using forest plot.</p><media xlink:href="jmir_v27i1e73516_app7.png" xlink:title="PNG File, 219 KB"/></supplementary-material><supplementary-material id="app8"><label>Multimedia Appendix 8</label><p>Summary estimate of pooled performance using forest plot.</p><media xlink:href="jmir_v27i1e73516_app8.png" xlink:title="PNG File, 274 KB"/></supplementary-material><supplementary-material id="app9"><label>Multimedia Appendix 9</label><p>Summary estimate of pooled performance using forest plot.</p><media xlink:href="jmir_v27i1e73516_app9.png" xlink:title="PNG File, 277 KB"/></supplementary-material><supplementary-material id="app10"><label>Multimedia Appendix 10</label><p>Quality assessment of diagnostic accuracy studies-2 summary plot.</p><media xlink:href="jmir_v27i1e73516_app10.png" xlink:title="PNG File, 177 KB"/></supplementary-material><supplementary-material id="app11"><label>Multimedia Appendix 11</label><p>Risk of bias and concern of applicability for each item in included studies.</p><media xlink:href="jmir_v27i1e73516_app11.png" xlink:title="PNG File, 442 KB"/></supplementary-material><supplementary-material id="app12"><label>Checklist 1</label><p>PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 checklist.</p><media xlink:href="jmir_v27i1e73516_app12.pdf" xlink:title="PDF File, 413 KB"/></supplementary-material></app-group></back></article>