<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v26i1e54047</article-id>
      <article-id pub-id-type="pmid">39753218</article-id>
      <article-id pub-id-type="doi">10.2196/54047</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Slit Lamp Report Generation and Question Answering: Development and Validation of a Multimodal Transformer Model with Large Language Model Integration</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Yoo</surname>
            <given-names>Tae Keun</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>McRoy</surname>
            <given-names>Susan</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Zhao</surname>
            <given-names>Ziwei</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-4551-348X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>Weiyi</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-2780-9121</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Xiaolan</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1581-5045</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Song</surname>
            <given-names>Fan</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-8134-2402</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Gunasegaram</surname>
            <given-names>James</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5517-986X</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Huang</surname>
            <given-names>Wenyong</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4894-0937</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Shi</surname>
            <given-names>Danli</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-6094-137X</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>He</surname>
            <given-names>Mingguang</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6912-2810</ext-link>
        </contrib>
        <contrib id="contrib9" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Liu</surname>
            <given-names>Na</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff6" ref-type="aff">6</xref>
          <address>
            <institution>Guangzhou Cadre and Talent Health Management Center</institution>
            <addr-line>No. 109 Changling Road</addr-line>
            <addr-line>Huangpu District</addr-line>
            <addr-line>Guangzhou, 510700</addr-line>
            <country>China</country>
            <phone>86 18701985445</phone>
            <email>1256695904@qq.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-4492-229X</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>School of Optometry</institution>
        <institution>The Hong Kong Polytechnic University</institution>
        <addr-line>Hong Kong</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Monash University</institution>
        <addr-line>Victoria</addr-line>
        <country>Australia</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Zhongshan Ophthalmic Center</institution>
        <institution>Sun Yat-sen University</institution>
        <addr-line>Guangzhou</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Research Centre for SHARP Vision</institution>
        <institution>The Hong Kong Polytechnic University</institution>
        <addr-line>Hong Kong</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>Centre for Eye and Vision Research (CEVR)</institution>
        <addr-line>Hong Kong</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff6">
        <label>6</label>
        <institution>Guangzhou Cadre and Talent Health Management Center</institution>
        <addr-line>Guangzhou</addr-line>
        <country>China</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Na Liu <email>1256695904@qq.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>30</day>
        <month>12</month>
        <year>2024</year>
      </pub-date>
      <volume>26</volume>
      <elocation-id>e54047</elocation-id>
      <history>
        <date date-type="received">
          <day>27</day>
          <month>10</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>3</day>
          <month>1</month>
          <year>2024</year>
        </date>
        <date date-type="rev-recd">
          <day>24</day>
          <month>2</month>
          <year>2024</year>
        </date>
        <date date-type="accepted">
          <day>5</day>
          <month>9</month>
          <year>2024</year>
        </date>
      </history>
      <copyright-statement>©Ziwei Zhao, Weiyi Zhang, Xiaolan Chen, Fan Song, James Gunasegaram, Wenyong Huang, Danli Shi, Mingguang He, Na Liu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 30.12.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2024/1/e54047" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Large language models have shown remarkable efficacy in various medical research and clinical applications. However, their skills in medical image recognition and subsequent report generation or question answering (QA) remain limited.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>We aim to finetune a multimodal, transformer-based model for generating medical reports from slit lamp images and develop a QA system using Llama2. We term this entire process slit lamp–GPT.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>Our research used a dataset of 25,051 slit lamp images from 3409 participants, paired with their corresponding physician-created medical reports. We used these data, split into training, validation, and test sets, to finetune the Bootstrapping Language-Image Pre-training framework toward report generation. The generated text reports and human-posed questions were then input into Llama2 for subsequent QA. We evaluated performance using qualitative metrics (including BLEU [bilingual evaluation understudy], CIDEr [consensus-based image description evaluation], ROUGE-L [Recall-Oriented Understudy for Gisting Evaluation—Longest Common Subsequence], SPICE [Semantic Propositional Image Caption Evaluation], accuracy, sensitivity, specificity, precision, and <italic>F</italic><sub>1</sub>-score) and the subjective assessments of two experienced ophthalmologists on a 1-3 scale (1 referring to high quality).</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>We identified 50 conditions related to diseases or postoperative complications through keyword matching in initial reports. The refined slit lamp–GPT model demonstrated BLEU scores (1-4) of 0.67, 0.66, 0.65, and 0.65, respectively, with a CIDEr score of 3.24, a ROUGE (Recall-Oriented Understudy for Gisting Evaluation) score of 0.61, and a Semantic Propositional Image Caption Evaluation score of 0.37. The most frequently identified conditions were cataracts (22.95%), age-related cataracts (22.03%), and conjunctival concretion (13.13%). Disease classification metrics demonstrated an overall accuracy of 0.82 and an <italic>F</italic><sub>1</sub>-score of 0.64, with high accuracies (≥0.9) observed for intraocular lens, conjunctivitis, and chronic conjunctivitis, and high <italic>F</italic><sub>1</sub>-scores (≥0.9) observed for cataract and age-related cataract. For both report generation and QA components, the two evaluating ophthalmologists reached substantial agreement, with κ scores between 0.71 and 0.84. In assessing 100 generated reports, they awarded scores of 1.36 for both completeness and correctness; 64% (64/100) were considered “entirely good,” and 93% (93/100) were “acceptable.” In the evaluation of 300 generated answers to questions, the scores were 1.33 for completeness, 1.14 for correctness, and 1.15 for possible harm, with 66.3% (199/300) rated as “entirely good” and 91.3% (274/300) as “acceptable.”</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This study introduces the slit lamp–GPT model for report generation and subsequent QA, highlighting the potential of large language models to assist ophthalmologists and patients.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>large language model</kwd>
        <kwd>slit lamp</kwd>
        <kwd>medical report generation</kwd>
        <kwd>question answering</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>The slit lamp, a cornerstone in ophthalmology, allows for detailed examination of the eye’s anterior segment [<xref ref-type="bibr" rid="ref1">1</xref>]. Using an illuminated, narrow beam, this noninvasive method facilitates the evaluation of abnormalities by depth and size. While instrumental in diagnosing common eye diseases such as keratitis, conjunctivitis, conjunctival concretions, and cataracts, interpreting slit lamp results can be challenging for primary care physicians due to the need for specialized training. This can result in overlooked abnormalities or misdiagnosis. Furthermore, ophthalmologists are tasked with interpreting, documenting, and effectively communicating these results to patients, a time and effort-intensive process. The scarcity of experienced ophthalmologists, particularly in rural areas, further exacerbates the situation [<xref ref-type="bibr" rid="ref2">2</xref>].</p>
      <p>Artificial intelligence (AI) and large language models (LLMs) have made significant strides in the medical field, enhancing the capabilities of health care professionals in interpreting, and analyzing medical imagery. For instance, AI has been instrumental in advancing the analysis of x-rays [<xref ref-type="bibr" rid="ref3">3</xref>], magnetic resonance images [<xref ref-type="bibr" rid="ref4">4</xref>], ultrasounds [<xref ref-type="bibr" rid="ref5">5</xref>], and dermatological images [<xref ref-type="bibr" rid="ref6">6</xref>]. Generative pretrained transformers (GPT) models such as ChatGPT [<xref ref-type="bibr" rid="ref7">7</xref>] and Llama2 [<xref ref-type="bibr" rid="ref8">8</xref>], have showcased remarkable capabilities in problem-solving scenarios across a spectrum of medical applications. These AI models are instrumental in streamlining clinical documentation [<xref ref-type="bibr" rid="ref9">9</xref>], refining patient communication [<xref ref-type="bibr" rid="ref10">10</xref>], aiding administrative tasks [<xref ref-type="bibr" rid="ref11">11</xref>], enriching textual data [<xref ref-type="bibr" rid="ref12">12</xref>], and bolstering evidence-based decision-making [<xref ref-type="bibr" rid="ref13">13</xref>]. Their versatility extends to comprehensive patient assessments [<xref ref-type="bibr" rid="ref14">14</xref>], precise disease diagnostics [<xref ref-type="bibr" rid="ref15">15</xref>], informed treatment proposals [<xref ref-type="bibr" rid="ref16">16</xref>], meticulous medical writing [<xref ref-type="bibr" rid="ref17">17</xref>], innovative teaching methodologies [<xref ref-type="bibr" rid="ref18">18</xref>], and robust question answering (QA) systems [<xref ref-type="bibr" rid="ref19">19</xref>], embodying a multifaceted impact on the health care industry.</p>
      <p>Deep learning strategies currently used to transform images into high-quality features include convolutional neural networks (CNNs), recurrent neural networks (RNNs), transformer networks, and their variants such as long short-term memory (LSTM) and gated recurrent units (GRUs). CNNs are often combined with other networks such as RNNs to generate text [<xref ref-type="bibr" rid="ref20">20</xref>]. RNNs and their variants, recognized for their prowess in handling sequential data, account for element dependencies within sequences. Despite their effectiveness, RNNs face challenges with extended sequences and potential gradient issues which are mitigated by long short-term memories and GRUs through a gate mechanism. Transformer networks, proposed in 2017, use self-attention mechanisms to manage long sequences and parallel computations, thus boasting swift training speed at the cost of substantial computational resources [<xref ref-type="bibr" rid="ref21">21</xref>]. Bootstrapping Language-Image Pre-training (BLIP), a hybrid approach leveraging transformer networks’ architecture and amalgamating natural language processing and computer vision, enhances model performance via pretraining. BLIP’s principal strength lies in its multimodal capacity to concurrently handle image and text data, allowing it to excel in specific tasks such as image description generation.</p>
      <p>In the specific context of slit lamp imaging augmented with AI, research has primarily concentrated on individual disease detection and grading, such as in the case of cataracts [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref23">23</xref>], pterygium [<xref ref-type="bibr" rid="ref24">24</xref>], and infectious keratitis [<xref ref-type="bibr" rid="ref25">25</xref>]. However, there is a noticeable lack of a unified system that uses slit lamp images for the generation of systematic anterior segment reports and QA. While the advent of OpenAI’s GPT-4V offered the possibility of image-based AI medical dialogue, its direct clinical application has been limited by inaccuracies and the generation of unreliable information, which was termed “hallucinations” [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. Additionally, due to its closed-source nature, there is a constraint on the fine-tuning ability, which is paramount for medical applications. In response to this, our study has used Llama2, an open-source model, to harness the anticipated benefits of a specialized LLM tool that ensures enhanced control and reliability in the subsequent QA scenarios. Based on our experience in ophthalmic QA tasks and LLMs, including fundus fluorescein angiography and indocyanine green angiography QA [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref29">29</xref>] we aim to extend these methodologies to slit lamp imaging by developing a novel slit lamp–GPT system, using BLIP and LLMs specifically tailored for ophthalmology, with dual objectives: to generate reports and to facilitate QA.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Overview</title>
        <p>The flow of our study is outlined in <xref rid="figure1" ref-type="fig">Figure 1</xref>.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Flow diagram of this study.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e54047_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Dataset</title>
        <p>We collected data from a Chinese physical examination center for this retrospective study, which included both essential clinical information and annual slit lamp images. We included slit lamp photographs with corresponding medical reports, excluding any of inadequate quality. This study used data collected from a previous study [<xref ref-type="bibr" rid="ref30">30</xref>], all participants’ information was deidentified per the Declaration of Helsinki’s guidelines. All slit lamp images, captured via a Haag-Streit BQ-900 at a 2048×1536-pixel resolution, included at least 4 images per participant showcasing the pupil, upper eyelid, and lower eyelid. Initial reports, written by ophthalmologists in Chinese, contained disease diagnoses, recommendations, or detailed descriptions of ocular signs. A subset of representative reports from the dataset was selected for translation into English to form a bilingual dataset.</p>
      </sec>
      <sec>
        <title>Model Construction</title>
        <p>Similar to other studies [<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref29">29</xref>], we initially trained and tested the BLIP [<xref ref-type="bibr" rid="ref28">28</xref>] network for report generation. Subsequently, the generated reports from the test set were input into Llama2 for QA validation, further evaluating the quality and practicality of the reports.</p>
        <p>During the report generation phase, we used the BLIP framework, a multimodal transformer model skilled at aligning visual interpretation with text generation. The model filtered out noisy data during training and generated slit lamp reports from paired images and text inputs. Our design incorporated a vision transformer [<xref ref-type="bibr" rid="ref31">31</xref>] and BERT [<xref ref-type="bibr" rid="ref32">32</xref>] as the image and language encoder and decoder, respectively. The vision transformer converts an image into encoded patch sequences, while BERT, trained on extensive unlabeled text data, enables deep contextualized representation learning. The pretrained BLIP model was fine-tuned using slit lamp images and associated reports, with each case providing at least four images during training, resized to 224×224 pixels. We applied the AdamW optimizer (the University of Freiburg), using an initial learning rate of 0.00002, a weight decay of 0.05, and a cosine learning rate schedule, across 50 epochs on one NVIDIA Tesla V100 GPU (NVIDIA Corp). The model with the highest BLEU1 (bilingual evaluation understudy) score (detailed in the performance evaluation part) on the validation set was selected for testing.</p>
        <p>For the question and answering phase, we created a question set related to slit lamp examination and reporting based on prior studies [<xref ref-type="bibr" rid="ref33">33</xref>] and our clinical expertise. These questions, along with the corresponding reports, were seamlessly input into the Llama2 model. This integration allowed for QA without the need for fine-tuning, while enhancing the interpretation of the generated reports. The process involved instructing the model using a specific prompt: “Answer based on: [slit lamp report content here].”</p>
      </sec>
      <sec>
        <title>Performance Evaluation</title>
        <p>We used both language-based and disease classification metrics for quantitative evaluations of report quality, supplemented by manual assessments for report generation and QA.</p>
        <p>For language-based metrics, we used BLEU [<xref ref-type="bibr" rid="ref34">34</xref>], CIDEr [<xref ref-type="bibr" rid="ref35">35</xref>], ROUGE-L [<xref ref-type="bibr" rid="ref36">36</xref>]), and Semantic Propositional Image Caption Evaluation [<xref ref-type="bibr" rid="ref37">37</xref>], each with its strengths. However, traditional language metrics may be less dependable for medical conditions due to the infrequent occurrence of disease-related keywords in reports. To address this, we introduced a classification evaluation procedure that used a manually curated dictionary to identify disease-related conditions or postoperative statuses from both original and generated reports. Disease classification metrics, such as specificity, accuracy, precision, sensitivity, and the <italic>F</italic><sub>1</sub>-score, provided a comprehensive performance review of the model.</p>
        <p>Considering the complexity of medical terminology and the potential harm of inaccurate reporting, manual assessment remains crucial. For report generation, 100 test set cases were randomly selected and independently evaluated by 2 ophthalmologists (ZZ and FS) using a 3-point scale, focusing on “completeness” (how well the generated reports matched the ground truth conditions) and “correctness” (the accuracy of diagnosis and condition descriptions). Scores ranged from 1 (excellent) to 3 (poor), with 2 representing an acceptable rating. The final score was the average of the scores from the 2 evaluators. For QA, 20 prepared human-posed questions and the translated report were put into Llama2 to generate answers, which were evaluated based on “completeness,” “correctness,” and “possible harm.” Scores ranged from 1 (recommendable to patients) to 3 (not recommendable for patients), with 2 indicating that minor adjustments could make the answer suitable for recommendation. The average score was also used as the final score. For detailed scoring criteria in these 2 sections, refer to Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>This study used data collected from a previous study [<xref ref-type="bibr" rid="ref30">30</xref>]. All patient data were anonymized and de-identified following the Declaration of Helsinki. Individual consent was waived due to the retrospective nature and the thorough anonymization process of the study. The Institutional Review Board of the Hong Kong Polytechnic University approved the study (HSEARS20240301004).</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Data</title>
        <p>Our final dataset includes 25,051 slit-lamp images and 3409 reports. Most images (12,496, 49.89%) focus on the cornea, with 32.74% (n=8202) on the upper eyelid and 17.38% (n=4353) on the lower eyelid. The median age of participants is 65, with an IQR of 60 to 72 years, and the majority (2009/3409, 58.93%) are male. The demographics and image types are similar across all sets.</p>
        <p>The distribution of images across years is as follows: 1257 (5.02%) from 2013, 12,206 (48.72%) from 2015, and 11,588 (46.26%) from 2016. The 2013 and 2015 images form the training set, while the 2016 images are partitioned evenly into validation and testing sets. There were no significant differences in demographic characteristics and positioning type between these datasets. <xref ref-type="table" rid="table1">Table 1</xref> provides a comprehensive overview of the dataset characteristics.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Slit lamp images: dataset characteristics.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="30"/>
            <col width="0"/>
            <col width="170"/>
            <col width="0"/>
            <col width="170"/>
            <col width="0"/>
            <col width="180"/>
            <col width="170"/>
            <col width="0"/>
            <col width="180"/>
            <col width="0"/>
            <col width="0"/>
            <col width="70"/>
            <thead>
              <tr valign="top">
                <td colspan="5">
                  <break/>
                </td>
                <td colspan="2">Total</td>
                <td>Train</td>
                <td colspan="2">Validation</td>
                <td colspan="2">Test</td>
                <td colspan="2"><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="14">
                  <bold>Participants</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="4">Number</td>
                <td colspan="2">3409</td>
                <td>1846</td>
                <td colspan="2">781</td>
                <td colspan="2">782</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="4">Age, median (Q1<sup>a</sup>, Q3<sup>b</sup>)</td>
                <td colspan="2">65.46 (60.52, 72.47)</td>
                <td>65.31 (60.12, 72.13)</td>
                <td colspan="2">62.04 (59.58, 65.9)</td>
                <td colspan="2">71.04 (65.47, 77.03)</td>
                <td colspan="2">&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="12">
                  <bold>Sex, n (%)</bold>
                </td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="2">
                  <break/>
                </td>
                <td>Male</td>
                <td colspan="2">2009 (58.93)</td>
                <td colspan="2">1101 (59.64)</td>
                <td>420 (53.78)</td>
                <td colspan="2">488 (62.4)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="2">
                  <break/>
                </td>
                <td>Female</td>
                <td colspan="2">1400 (41.07)</td>
                <td colspan="2">745 (40.36)</td>
                <td>361 (46.22)</td>
                <td colspan="2">294 (37.6)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="14">
                  <bold>Slit lamp images</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="4">Number</td>
                <td colspan="2">25,051</td>
                <td>13,463</td>
                <td colspan="2">5987</td>
                <td colspan="2">5601</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="12">
                  <bold>Position, n (%)</bold>
                </td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td colspan="2">Upper eyelid</td>
                <td colspan="2">8202 (32.74)</td>
                <td colspan="2">4315 (32.05)</td>
                <td>2046 (34.2)</td>
                <td colspan="2">1841 (32.9)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td colspan="2">Lower eyelid</td>
                <td colspan="2">4353 (17.38)</td>
                <td colspan="2">2423 (18)</td>
                <td>951 (15.9)</td>
                <td colspan="2">979 (17.5)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td colspan="2">Cornea</td>
                <td colspan="2">12,496 (49.89)</td>
                <td colspan="2">6725 (49.95)</td>
                <td>2990 (49.9)</td>
                <td colspan="2">2781 (49.7)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>Q1: first quartile.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>Q3: third quartile.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>We used a custom dictionary to extract diagnoses and physical signs by keyword matching from the Chinese reports. We identified 50 conditions, including age-related cataracts (478/2170, 22.03%), cataracts (498/2170, 22.95%), conjunctival concretion (285/2170, 13.13%), after intraocular lens implantation (151/2170, 6.96%), pterygium (144/2170, 6.64%), conjunctivitis (97/2170, 4.47%), chronic conjunctivitis (93/2170, 4.29%), and other eye conditions with lower proportions. This led to 1377 Chinese reports primarily featuring diagnostic terms or descriptions of ocular signs.</p>
      </sec>
      <sec>
        <title>Quantitative Model Performance</title>
        <p>Language-based metrics are provided in <xref ref-type="table" rid="table2">Table 2</xref>, with BLEU (1-4) scores (0.67, 0.66, 0.65, and 0.65) indicating good lexical accuracy and a ROUGE-L score of 0.61 highlighting effective content retention. The CIDEr score of 3.24 reflects its ability to align closely with human judgment on sentence quality, while a SPICE score of 0.37 demonstrates moderate success in capturing complex semantic relationships. For disease classification metrics (see <xref ref-type="table" rid="table3">Table 3</xref>), our model achieved a weighted accuracy of 0.82 and a weighted <italic>F</italic><sub>1</sub>-score of 0.64. However, performance varied across diseases. It was highly accurate (≥0.9) for conditions of intraocular lens, conjunctivitis, and chronic conjunctivitis, and had high <italic>F</italic><sub>1</sub>-scores (≥0.9) for cataracts and age-related cataracts. The model demonstrated excellent accuracy for positive cases of cataracts and age-related cataracts. Despite high accuracy, specificity, and precision for postoperative intraocular lens implantation, sensitivity was relatively low: a clinically acceptable trade-off. However, for conjunctival concretions, conjunctivitis, and chronic conjunctivitis, the model’s overall predictive capacity fell short.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Language-based metrics of report generation in the test set (5601 images from 782 participants).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="150"/>
            <col width="150"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <thead>
              <tr valign="top">
                <td>BLEU_1<sup>a</sup></td>
                <td>BLEU_2<sup>a</sup></td>
                <td>BLEU_3<sup>a</sup></td>
                <td>BLEU_4<sup>a</sup></td>
                <td>CIDEr<sup>b</sup></td>
                <td>ROUGE<sup>c</sup></td>
                <td>SPICE<sup>d</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>0.67</td>
                <td>0.66</td>
                <td>0.65</td>
                <td>0.65</td>
                <td>3.24</td>
                <td>0.61</td>
                <td>0.37</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>BLEU: bilingual evaluation understudy.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>CIDEr: consensus-based image description evaluation.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>ROUGE: Recall-Oriented Understudy for Gisting Evaluation.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>SPICE: Semantic Propositional Image Caption Evaluation.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Disease classification metrics of report generation in the test set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="350"/>
            <col width="130"/>
            <col width="130"/>
            <col width="130"/>
            <col width="130"/>
            <col width="130"/>
            <thead>
              <tr valign="top">
                <td>Condition</td>
                <td>Specificity</td>
                <td>Accuracy</td>
                <td>Precision</td>
                <td>Sensitivity</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Age-related cataract</td>
                <td>0.6</td>
                <td>0.79</td>
                <td>0.9</td>
                <td>0.8</td>
                <td>0.84</td>
              </tr>
              <tr valign="top">
                <td>Cataract</td>
                <td>0.58</td>
                <td>0.78</td>
                <td>0.9</td>
                <td>0.79</td>
                <td>0.84</td>
              </tr>
              <tr valign="top">
                <td>After intraocular lens implantation</td>
                <td>0.93</td>
                <td>0.94</td>
                <td>0.96</td>
                <td>0.48</td>
                <td>0.64</td>
              </tr>
              <tr valign="top">
                <td>Conjunctival concretion</td>
                <td>0.83</td>
                <td>0.7</td>
                <td>0.37</td>
                <td>0.44</td>
                <td>0.4</td>
              </tr>
              <tr valign="top">
                <td>Chronic conjunctivitis</td>
                <td>0.92</td>
                <td>0.9</td>
                <td>0.34</td>
                <td>0.15</td>
                <td>0.2</td>
              </tr>
              <tr valign="top">
                <td>Conjunctivitis</td>
                <td>0.92</td>
                <td>0.9</td>
                <td>0.34</td>
                <td>0.14</td>
                <td>0.2</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>Qualitative Model Performance</title>
        <sec>
          <title>Overview</title>
          <p>The score distribution is depicted in Figure S1 in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p>
        </sec>
        <sec>
          <title>Report Generation</title>
          <p>Two ophthalmologists scored the model highly for completeness (mean 1.36, SD 0.61, κ=0.84) and correctness (mean 1.36, SD 0.59, κ=0.72). Reports that received a score of 1 for both completeness and correctness were defined as entirely good and constituted 64% (64/100) of the evaluated reports. Reports that scored either 1 or 2 by both reviewers for both completeness and correctness were deemed acceptable, representing 93% (93/100) of the reports. These scores primarily corresponded to reports detailing specific conditions such as cataracts, age-related cataracts, and negative findings. However, 7% (7/100) of reports scored a 3, indicating deficiencies.</p>
          <p>We discovered that lower scores were linked to issues such as limited sample sizes for specific diseases, difficulties in clearly identifying lesions, and challenges in interpreting diseases or signs from images due to the unique aspects of slit lamp photography. These complications were common in conditions such as xanthomas, trichiasis, after-glaucoma surgery, lagophthalmos, and some small conjunctival concretions. Additionally, images not focused on the cornea made it difficult to detect corneal lesions.</p>
          <p>Through our hands-on evaluation, we noticed that the model sometimes added diagnoses that were not in the original reports but were still acceptable based on the images. For example, it sometimes diagnosed mild cataracts even when the images did not show apparent lens abnormalities. We considered these decisions acceptable when considering the challenge faced by an ophthalmologist in making a precise distinction based solely on images.</p>
        </sec>
        <sec>
          <title>About QA</title>
          <p>Our constructed questionnaires included 20 items, addressing a breadth of topics such as diagnosis, pathologic localization, severity grading, visual impairment, prognosis, associated complications, therapeutic recommendations, suggested further examinations, preventive advice, and scientific education pertinent to slit lamp examination (Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p>
          <p>We selectively curated 15 representative English reports on conditions including cataracts, conjunctival concretions, conjunctivitis, postintraocular lens implantation, and pterygium, as well as their mixed states. Each report contributed 20 questions, culminating in a total of 300 questions.</p>
          <p>Our model scored well on completeness (1.33, κ=0.84), correctness (1.14, κ=0.71), and possible harm (1.15, κ=0.82). Similarly, QA responses that scored a 1 in completeness, correctness, and possible harm were defined as entirely good, representing 66.3% (199/300) of the QA responses. Responses scoring either 1 or 2 across these categories were considered acceptable, comprising 91.3% (274/300). Less than 9% (26/300) of the 300 questions scored a 3 in any category. These were typically related to reports focusing more on physical signs than diagnoses or conditions and statements about binocular intraocular lenses. Figure S2 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref> provides examples of generated answers with different scores.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>Our study introduces a novel method for analyzing slit lamp images through the integration of a multimodal transformer with an LLM. This approach has enabled the accurate identification of common anterior segment eye diseases and supports a QA system that directly addresses symptoms, diagnosis, and treatment options, as illustrated in <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Demonstration of the question-answering system. (A) Input image, ground truth, and model prediction. (B) Question answering. Blue highlight: corresponds to accurate diagnosis matches. Yellow highlight: supplementary predicted information (not in the manual report but correct).</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e54047_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>LLMs represent a breakthrough in AI with large knowledge bases and strong logical reasoning abilities. They have exhibited efficacy across various natural language processing tasks, including text generation, summarization, translation, and QA. However, in the medical realm, the quality of these answers warrants further scrutiny. Previous research has shown mixed results for the ability of LLMs to pass ophthalmology examinations. The study of Kung et al [<xref ref-type="bibr" rid="ref38">38</xref>] indicates that ChatGPT can pass the United States Medical Licensing Examination without any specialized training or reinforcement. However, Thirunavukarasu’s [<xref ref-type="bibr" rid="ref39">39</xref>] attempt to assess ChatGPT’s proficiency in the FRCOphth (Fellowship of the Royal College of Ophthalmologists) examination showed subpar performance, thereby underscoring the inability of LLMs to replace physicians in highly specialized fields. Conversely, in advising patients about symptoms or ongoing conditions—tasks less demanding of expertise—ChatGPT seems to demonstrate competence. Many patients turn to the internet for self-diagnosis before consulting a health care professional [<xref ref-type="bibr" rid="ref40">40</xref>]. The use of LLMs for medical consultations can increase patient independence and potentially aid in accurate diagnosis. The release of GPT4V represents an innovative leap in the realm of LLM integration with computer vision, with promising prospects for extensive application in the medical field. Wu et al [<xref ref-type="bibr" rid="ref26">26</xref>] assessed images from eight modalities across 17 human body systems and concluded that while GPT4V excels at identifying image modalities and anatomical structures, it encounters significant challenges in disease diagnosis and comprehensive report generation. In another study, we used a similar 1-3 evaluation scale to assess GPT4V’s performance on ophthalmology-related tasks, including image interpretation and QA [<xref ref-type="bibr" rid="ref27">27</xref>]. The model performed best in analyzing slit lamp images; however, it only reached 42% (42/100) in accuracy, 38.5% (34.7/90) in usability, and 68.5% (61.7/90) in safety of the responses. These results are significantly lower than the “entirely good” rates we reported previously—64% (64/100) for report generation and 66.3% (199/300) for QA. This discrepancy underscores the need for models tailored to ophthalmology to ensure high-quality outcomes. To address this gap, we implemented an experimental model, slit lamp–GPT, harnessing the BLIP and Llama2 frameworks. This initiative represents merely the first step in a broader journey toward refining AI applications in ophthalmology.</p>
        <p>The model demonstrated proficiency in identifying and reporting common anterior segment eye diseases within our dataset. However, its performance on rare conditions highlighted a critical area for improvement, suggesting that its effectiveness is closely tied to the diversity and representation of conditions in the training data. Per report generation, suboptimal performance was linked to specific diseases such as trichiasis, postglaucoma surgery complications, and corneal pathologies. Given our dataset’s origin in routine health examination data, these conditions were underrepresented, likely contributing to the poor performance. Another hypothesis considers the dynamic nature of slit lamp examinations in clinical settings, where ophthalmologists manually focus to obtain the best diagnostic view, a process not fully captured by static images. Instances of misdiagnosed keratitis, where images did not focus precisely on the cornea, support this assumption. Integrating our model with a broader spectrum of ophthalmic imaging techniques—such as indocyanine green angiography, fundus fluorescein angiography, ocular ultrasound, optical coherence tomography, and fundus photography—may enhance diagnostic alignment with actual clinical observations and further improve overall performance.</p>
        <p>The current results suggest potential applicability in cataract screening, particularly in regions with a shortage of ophthalmologists. Previous studies have primarily focused on applying deep learning to the diagnosis and grading of cataracts, fundamentally using classification models. In contrast, our model is a natural language processing system capable of generating free-text reports. It not only provides descriptive insights but also achieves cataract classification accuracy similar to existing models [<xref ref-type="bibr" rid="ref41">41</xref>,<xref ref-type="bibr" rid="ref42">42</xref>]. Beyond this, our model could function as an educational tool for patients. In bustling eye clinics, patients may lack sufficient time to fully comprehend their examination reports and medical conditions. As demonstrated in this study, the slit lamp–GPT can provide patients with basic clinical explanations and recommendations concerning causes, abnormalities, treatment, and follow-up, indicating its potential to reduce medical consultation expenditure and bolster the use of remote health care services.</p>
        <p>The manual evaluation suggests that slit lamp–GPT exhibits a promising capacity to assist participants with minimal risk. During the QA stage, 89.3% (268/300) of the responses were deemed completely harmless, surpassing the performance of GPT4V. However, the potential risks of using LLMs are yet to be thoroughly understood. A common problem with LLMs is that they sometimes generate inaccuracies and false statements, which are often referred to as “hallucinations” in the field [<xref ref-type="bibr" rid="ref43">43</xref>]. These incorrect assertions can appear to be true, which could harm patients. This was reflected in our study, where the model sometimes created content. For example, the Llama2 model wrongly identified a binocular intraocular lens as a disease instead of a postoperative condition, creating the nonexistent “binocular intraocular lens syndrome.” This led to poor scores on the related 20 questions, highlighting the need for specialized fine-tuned LLM and knowledge-based generation [<xref ref-type="bibr" rid="ref44">44</xref>]. Nonetheless, it is important to recognize that LLMs should serve as adjuncts or supplements in the clinical diagnosis and treatment process, not as fully trusted entities devoid of physician oversight. As LLM technology evolves, it is incumbent on stakeholders to collaboratively establish best practice standards to ensure patient safety.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>This study has a few limitations. First, the dataset used is skewed, coming mainly from routine health checks of healthy people. The small sample size for certain diseases might affect the effectiveness of classification. Using datasets from high-quality outpatient clinics could lead to better results. Second, as with other language models, our model sometimes produces repetitive text, and the accuracy of the responses it generates can be inconsistent. At times, the model’s answers show logical errors. For instance, it diagnosed both a postintraocular lens implantation status and a senile cataract in the same eye. These issues might be addressed by incorporating expert knowledge and fine-tuning LLMs. There are also notable concerns about bias, as a single mistake in report generation can lead to multiple errors during the question-and-answer process. This highlights the need for further improvements to increase the accuracy and completeness of report generation. Lastly, creating a standardized manual evaluation process for these types of models is challenging [<xref ref-type="bibr" rid="ref45">45</xref>,<xref ref-type="bibr" rid="ref46">46</xref>]. This study was limited to slit lamp anterior segment images, indicating a need for future research to include diverse datasets. This will help evaluate the model’s applicability across various types of imaging.</p>
      </sec>
      <sec>
        <title>Conclusion</title>
        <p>This research underscores the effectiveness and potential of using LLMs for slit lamp image report generation and QA tasks, showcasing their viability in ophthalmic medical image analysis.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Supplementary tables with data.</p>
        <media xlink:href="jmir_v26i1e54047_app1.docx" xlink:title="DOCX File , 23 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Distribution of scores in the manual assessment of report generation and question-answering. RG: report generation; QA: question answering.</p>
        <media xlink:href="jmir_v26i1e54047_app2.png" xlink:title="PNG File , 78 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>Examples of generated answers with different scores. Red bold text indicates incorrect information.</p>
        <media xlink:href="jmir_v26i1e54047_app3.png" xlink:title="PNG File , 918 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BLEU</term>
          <def>
            <p>bilingual evaluation understudy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BLIP</term>
          <def>
            <p>Bootstrapping Language-Image Pre-training</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CIDEr</term>
          <def>
            <p>consensus-based image description evaluation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CNNs</term>
          <def>
            <p>convolutional neural networks</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">GRUs</term>
          <def>
            <p>gated recurrent units</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">LLM</term>
          <def>
            <p>large language model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">QA</term>
          <def>
            <p>question answering</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">ROUGE-L</term>
          <def>
            <p>Recall-Oriented Understudy for Gisting Evaluation—Longest Common Subsequence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">RNN</term>
          <def>
            <p>recurrent neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">SPICE</term>
          <def>
            <p>Semantic Propositional Image Caption Evaluation</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This study was supported by the Global STEM Professorship Scheme (P0046113) and Henry G. Leong Endowed Professorship in Elderly Vision Health. The sponsor or funding organization had no role in the design or conduct of this research. We thank the InnoHK Hong Kong special administrative region government for providing valuable supports.</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>NL is the primary corresponding author of this article, while MH and DS are co-corresponding author. DS and MH conceived this study. NL provided data. DS and WZ built the deep learning model. DS and ZZ did the literature search and analyzed the data. DS, ZZ, XC, FS, and MH contributed to key data interpretation. ZZ and FS did the manual evaluation. ZZ wrote this paper. JG edited and reviewed this paper. All authors commented and critically revised this paper.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Doggart</surname>
              <given-names>JH</given-names>
            </name>
          </person-group>
          <article-title>SLIT-LAMP Microscopy</article-title>
          <source>Br J Ophthalmol</source>
          <year>1948</year>
          <volume>32</volume>
          <issue>4</issue>
          <fpage>232</fpage>
          <lpage>247</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/18170442"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bjo.32.4.232</pub-id>
          <pub-id pub-id-type="medline">18170442</pub-id>
          <pub-id pub-id-type="pmcid">PMC510816</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Magyezi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Arunga</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Eye care where there are no ophthalmologists: the Uganda experience</article-title>
          <source>Community Eye Health</source>
          <year>2020</year>
          <volume>33</volume>
          <issue>110</issue>
          <fpage>48</fpage>
          <lpage>50</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34007108"/>
          </comment>
          <pub-id pub-id-type="medline">34007108</pub-id>
          <pub-id pub-id-type="pii">jceh_33_110_048</pub-id>
          <pub-id pub-id-type="pmcid">PMC8115714</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Niehoff</surname>
              <given-names>JH</given-names>
            </name>
            <name name-style="western">
              <surname>Kalaitzidis</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kroeger</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Schoenbeck</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Borggrefe</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Michael</surname>
              <given-names>AE</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of the clinical performance of an AI-based application for the automated analysis of chest X-rays</article-title>
          <source>Sci Rep</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>1</issue>
          <fpage>3680</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-023-30521-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-023-30521-2</pub-id>
          <pub-id pub-id-type="medline">36872333</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-023-30521-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC9985819</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sheth</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Giger</surname>
              <given-names>ML</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence in the interpretation of breast cancer on MRI</article-title>
          <source>J Magn Reson Imaging</source>
          <year>2020</year>
          <volume>51</volume>
          <issue>5</issue>
          <fpage>1310</fpage>
          <lpage>1324</lpage>
          <pub-id pub-id-type="doi">10.1002/jmri.26878</pub-id>
          <pub-id pub-id-type="medline">31343790</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence in ultrasound</article-title>
          <source>Eur J Radiol</source>
          <year>2021</year>
          <volume>139</volume>
          <fpage>109717</fpage>
          <pub-id pub-id-type="doi">10.1016/j.ejrad.2021.109717</pub-id>
          <pub-id pub-id-type="medline">33962110</pub-id>
          <pub-id pub-id-type="pii">S0720-048X(21)00197-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jain</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Eng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Way</surname>
              <given-names>DH</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Bui</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Kanada</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>de Oliveira Marinho</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Gallegos</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gabriele</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gupta</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Natarajan</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Hofmann-Wellenhof</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Corrado</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Webster</surname>
              <given-names>DR</given-names>
            </name>
            <name name-style="western">
              <surname>Ai</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Dunn</surname>
              <given-names>RC</given-names>
            </name>
            <name name-style="western">
              <surname>Coz</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>A deep learning system for differential diagnosis of skin diseases</article-title>
          <source>Nat Med</source>
          <year>2020</year>
          <volume>26</volume>
          <issue>6</issue>
          <fpage>900</fpage>
          <lpage>908</lpage>
          <pub-id pub-id-type="doi">10.1038/s41591-020-0842-3</pub-id>
          <pub-id pub-id-type="medline">32424212</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-020-0842-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="web">
          <source>Introducing ChatGPT</source>
          <access-date>2024-10-19</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://openai.com/blog/chatgpt">https://openai.com/blog/chatgpt</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Touvron</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Stone</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Llama 2: Open foundation and fine-tuned chat models</article-title>
          <source>arXiv:2307.09288</source>
          <year>2023</year>
          <fpage>1</fpage>
          <lpage>77</lpage>
          <comment>Preprint published on July 18, 2023</comment>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>SB</given-names>
            </name>
            <name name-style="western">
              <surname>Lam</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT: the future of discharge summaries?</article-title>
          <source>Lancet Digit Health</source>
          <year>2023</year>
          <volume>5</volume>
          <issue>3</issue>
          <fpage>e107</fpage>
          <lpage>e108</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2589-7500(23)00021-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/S2589-7500(23)00021-3</pub-id>
          <pub-id pub-id-type="medline">36754724</pub-id>
          <pub-id pub-id-type="pii">S2589-7500(23)00021-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thirunavukarasu</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ting</surname>
              <given-names>DSJ</given-names>
            </name>
            <name name-style="western">
              <surname>Elangovan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Gutierrez</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>TF</given-names>
            </name>
            <name name-style="western">
              <surname>Ting</surname>
              <given-names>DSW</given-names>
            </name>
          </person-group>
          <article-title>Large language models in medicine</article-title>
          <source>Nat Med</source>
          <year>2023</year>
          <volume>29</volume>
          <issue>8</issue>
          <fpage>1930</fpage>
          <lpage>1940</lpage>
          <pub-id pub-id-type="doi">10.1038/s41591-023-02448-8</pub-id>
          <pub-id pub-id-type="medline">37460753</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-023-02448-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kahambing</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT, public health communication and 'intelligent patient companionship'</article-title>
          <source>J Public Health (Oxf)</source>
          <year>2023</year>
          <volume>45</volume>
          <issue>3</issue>
          <fpage>e590</fpage>
          <pub-id pub-id-type="doi">10.1093/pubmed/fdad028</pub-id>
          <pub-id pub-id-type="medline">37036209</pub-id>
          <pub-id pub-id-type="pii">7111147</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>ChatFFA: an ophthalmic chat system for unified vision-language understanding and question answering for fundus fluorescein angiography</article-title>
          <source>iScience</source>
          <year>2024</year>
          <volume>27</volume>
          <issue>7</issue>
          <fpage>110021</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2589-0042(24)01246-X"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.isci.2024.110021</pub-id>
          <pub-id pub-id-type="medline">39055931</pub-id>
          <pub-id pub-id-type="pii">S2589-0042(24)01246-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC11269310</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Liao</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>AugGPT: leveraging ChatGPT for text data augmentation</article-title>
          <source>arXiv:2302.13007</source>
          <year>2023</year>
          <fpage>1</fpage>
          <lpage>12</lpage>
          <comment>Preprint published on February 25, 2023</comment>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ebrahimian</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Homayounieh</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Rockenbach</surname>
              <given-names>MABC</given-names>
            </name>
            <name name-style="western">
              <surname>Putha</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Raj</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Dayan</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Bizzo</surname>
              <given-names>BC</given-names>
            </name>
            <name name-style="western">
              <surname>Buch</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Digumarthy</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Kalra</surname>
              <given-names>MK</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence matches subjective severity assessment of pneumonia for prediction of patient outcome and need for mechanical ventilation: a cohort study</article-title>
          <source>Sci Rep</source>
          <year>2021</year>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>858</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-020-79470-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-020-79470-0</pub-id>
          <pub-id pub-id-type="medline">33441578</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-020-79470-0</pub-id>
          <pub-id pub-id-type="pmcid">PMC7807029</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Zeng</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Nie</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Ding</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Guan</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Sang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Xiong</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cui</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Lure</surname>
              <given-names>FYM</given-names>
            </name>
            <name name-style="western">
              <surname>Zhan</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence for stepwise diagnosis and monitoring of COVID-19</article-title>
          <source>Eur Radiol</source>
          <year>2022</year>
          <volume>32</volume>
          <issue>4</issue>
          <fpage>2235</fpage>
          <lpage>2245</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34988656"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s00330-021-08334-6</pub-id>
          <pub-id pub-id-type="medline">34988656</pub-id>
          <pub-id pub-id-type="pii">10.1007/s00330-021-08334-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC8731211</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Siewiorek</surname>
              <given-names>DP</given-names>
            </name>
            <name name-style="western">
              <surname>Smailagic</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bernardino</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Badia</surname>
              <given-names>SBBI</given-names>
            </name>
          </person-group>
          <article-title>A human-AI collaborative approach for clinical decision making on rehabilitation assessment</article-title>
          <year>2021</year>
          <conf-name>Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems</conf-name>
          <conf-date>2021 May 07</conf-date>
          <conf-loc>Yokohama, Japan</conf-loc>
          <fpage>1</fpage>
          <lpage>14</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Biswas</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT and the future of medical writing</article-title>
          <source>Radiology</source>
          <year>2023</year>
          <volume>307</volume>
          <issue>2</issue>
          <fpage>e223312</fpage>
          <pub-id pub-id-type="doi">10.1148/radiol.223312</pub-id>
          <pub-id pub-id-type="medline">36728748</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Masters</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence in medical education</article-title>
          <source>Med Teach</source>
          <year>2019</year>
          <volume>41</volume>
          <issue>9</issue>
          <fpage>976</fpage>
          <lpage>980</lpage>
          <pub-id pub-id-type="doi">10.1080/0142159X.2019.1595557</pub-id>
          <pub-id pub-id-type="medline">31007106</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Singhal</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Tu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Gottweis</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sayres</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Wulczyn</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Hou</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Pfohl</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Cole-Lewis</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Neal</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Schaekermann</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Amin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lachgar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mansfield</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Prakash</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Green</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Dominowska</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Arcas</surname>
              <given-names>BAY</given-names>
            </name>
            <name name-style="western">
              <surname>Tomasev</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wong</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Semturs</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Mahdavi</surname>
              <given-names>SS</given-names>
            </name>
            <name name-style="western">
              <surname>Barral</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Webster</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Corrado</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Matias</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Azizi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Karthikesalingam</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Natarajan</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Towards expert-level medical question answering with large language models</article-title>
          <source>arXiv:2305.09617</source>
          <year>2023</year>
          <fpage>1</fpage>
          <lpage>30</lpage>
          <comment>Preprint published on May 16, 2023</comment>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Krizhevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>GE</given-names>
            </name>
          </person-group>
          <article-title>ImageNet classification with deep convolutional neural networks</article-title>
          <source>NeurIPS Proceedings</source>
          <year>2012</year>
          <conf-name>26th Annual Conference on Neural Information Processing Systems 2012</conf-name>
          <conf-date>2012-12-3</conf-date>
          <conf-loc>Lake Tahoe, Nevada, United States</conf-loc>
          <fpage>1</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaswani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Parmar</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Uszkoreit</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gomez</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Kaiser</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Polosukhin</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Attention is all you need</article-title>
          <source>NeurIPS Proceedings</source>
          <year>2017</year>
          <conf-name>the 31st International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>2017 December 04</conf-date>
          <conf-loc>Red Hook, NY, United States</conf-loc>
          <fpage>6000</fpage>
          <lpage>6010</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Son</surname>
              <given-names>KY</given-names>
            </name>
            <name name-style="western">
              <surname>Ko</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>SY</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Chung</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>DH</given-names>
            </name>
          </person-group>
          <article-title>Deep learning-based cataract detection and grading from slit-lamp and retro-illumination photographs: model development and validation study</article-title>
          <source>Ophthalmol Sci</source>
          <year>2022</year>
          <volume>2</volume>
          <issue>2</issue>
          <fpage>100147</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2666-9145(22)00036-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.xops.2022.100147</pub-id>
          <pub-id pub-id-type="medline">36249697</pub-id>
          <pub-id pub-id-type="pii">S2666-9145(22)00036-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC9559082</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Rong</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ding</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Lens opacities classification system III-based artificial intelligence program for automatic cataract grading</article-title>
          <source>J Cataract Refract Surg</source>
          <year>2022</year>
          <volume>48</volume>
          <issue>5</issue>
          <fpage>528</fpage>
          <lpage>534</lpage>
          <pub-id pub-id-type="doi">10.1097/j.jcrs.0000000000000790</pub-id>
          <pub-id pub-id-type="medline">34433780</pub-id>
          <pub-id pub-id-type="pii">02158034-202205000-00003</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Deshmukh</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chee</surname>
              <given-names>ML</given-names>
            </name>
            <name name-style="western">
              <surname>Soh</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Teo</surname>
              <given-names>ZL</given-names>
            </name>
            <name name-style="western">
              <surname>Thakur</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Goh</surname>
              <given-names>JHL</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Husain</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mehta</surname>
              <given-names>JS</given-names>
            </name>
            <name name-style="western">
              <surname>Wong</surname>
              <given-names>TY</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rim</surname>
              <given-names>TH</given-names>
            </name>
            <name name-style="western">
              <surname>Tham</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Deep learning algorithms for automatic detection of pterygium using anterior segment photographs from slit-lamp and hand-held cameras</article-title>
          <source>Br J Ophthalmol</source>
          <year>2022</year>
          <volume>106</volume>
          <issue>12</issue>
          <fpage>1642</fpage>
          <lpage>1647</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://bjo.bmj.com/lookup/pmidlookup?view=long&#38;pmid=34244208"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bjophthalmol-2021-318866</pub-id>
          <pub-id pub-id-type="medline">34244208</pub-id>
          <pub-id pub-id-type="pii">bjophthalmol-2021-318866</pub-id>
          <pub-id pub-id-type="pmcid">PMC9685734</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soleimani</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Esmaili</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Rahdar</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Aminizadeh</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cheraqpour</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Tabatabaei</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Mirshahi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bibak</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Mohammadi</surname>
              <given-names>SF</given-names>
            </name>
            <name name-style="western">
              <surname>Koganti</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yousefi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Djalilian</surname>
              <given-names>AR</given-names>
            </name>
          </person-group>
          <article-title>From the diagnosis of infectious keratitis to discriminating fungal subtypes; a deep learning-based study</article-title>
          <source>Sci Rep</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>1</issue>
          <fpage>22200</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-023-49635-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-023-49635-8</pub-id>
          <pub-id pub-id-type="medline">38097753</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-023-49635-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC10721811</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lei</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Can GPT-4V (ision) serve medical applications? case studies on GPT-4V for multimodal medical diagnosis</article-title>
          <source>arXiv:2310.09909</source>
          <year>2023</year>
          <fpage>1</fpage>
          <lpage>178</lpage>
          <comment>Preprint published October 15 2023</comment>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Unveiling the clinical incapabilities: a benchmarking study of GPT-4V(ision) for ophthalmic multimodal image analysis</article-title>
          <source>Br J Ophthalmol</source>
          <year>2024</year>
          <volume>108</volume>
          <issue>10</issue>
          <fpage>1384</fpage>
          <lpage>1389</lpage>
          <pub-id pub-id-type="doi">10.1136/bjo-2023-325054</pub-id>
          <pub-id pub-id-type="medline">38789133</pub-id>
          <pub-id pub-id-type="pii">bjo-2023-325054</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>FFA-GPT: an automated pipeline for fundus fluorescein angiography interpretation and question-answer</article-title>
          <source>NPJ Digit Med</source>
          <year>2024</year>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>111</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-024-01101-z"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-024-01101-z</pub-id>
          <pub-id pub-id-type="medline">38702471</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-024-01101-z</pub-id>
          <pub-id pub-id-type="pmcid">PMC11068733</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>ICGA-GPT: report generation and question answering for indocyanine green angiography images</article-title>
          <source>Br J Ophthalmol</source>
          <year>2024</year>
          <volume>108</volume>
          <issue>10</issue>
          <fpage>1450</fpage>
          <lpage>1456</lpage>
          <pub-id pub-id-type="doi">10.1136/bjo-2023-324446</pub-id>
          <pub-id pub-id-type="medline">38508675</pub-id>
          <pub-id pub-id-type="pii">bjo-2023-324446</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Niu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Holden</surname>
              <given-names>BA</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>The association of longitudinal trend of fasting plasma glucose with retinal microvasculature in people without established diabetes</article-title>
          <source>Invest Ophthalmol Vis Sci</source>
          <year>2015</year>
          <volume>56</volume>
          <issue>2</issue>
          <fpage>842</fpage>
          <lpage>848</lpage>
          <pub-id pub-id-type="doi">10.1167/iovs.14-15943</pub-id>
          <pub-id pub-id-type="medline">25613941</pub-id>
          <pub-id pub-id-type="pii">iovs.14-15943</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Xiong</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Hoi</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Blip: bootstrapping language-image pre-training for unified vision-language understanding and generation</article-title>
          <year>2022</year>
          <conf-name>Proceedings of the 39th International Conference on Machine Learning</conf-name>
          <conf-date>2022 January 28</conf-date>
          <conf-loc>Baltimore, Maryland, USA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>MW</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          <source>arXiv:1810.04805. Presented October 11, 2018</source>
          <year>2018</year>
          <fpage>1</fpage>
          <lpage>16</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Momenaei</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Wakabayashi</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Shahlaee</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Durrani</surname>
              <given-names>AF</given-names>
            </name>
            <name name-style="western">
              <surname>Pandit</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Mansour</surname>
              <given-names>HA</given-names>
            </name>
            <name name-style="western">
              <surname>Abishek</surname>
              <given-names>RM</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sridhar</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yonekawa</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kuriyan</surname>
              <given-names>AE</given-names>
            </name>
          </person-group>
          <article-title>Appropriateness and readability of ChatGPT-4-generated responses for surgical treatment of retinal diseases</article-title>
          <source>Ophthalmol Retina</source>
          <year>2023</year>
          <volume>7</volume>
          <issue>10</issue>
          <fpage>862</fpage>
          <lpage>868</lpage>
          <pub-id pub-id-type="doi">10.1016/j.oret.2023.05.022</pub-id>
          <pub-id pub-id-type="medline">37277096</pub-id>
          <pub-id pub-id-type="pii">S2468-6530(23)00246-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Papineni</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Roukos</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ward</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>WJ</given-names>
            </name>
          </person-group>
          <article-title>BLEU: a method for automatic evaluation of machine translation</article-title>
          <year>2002</year>
          <conf-name>Proceedings of the 40th Annual Meeting on Association for Computational Linguistics</conf-name>
          <conf-date>2002 July 06</conf-date>
          <conf-loc>United States</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vedantam</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Lawrence</surname>
              <given-names>ZC</given-names>
            </name>
            <name name-style="western">
              <surname>Parikh</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>CIDEr: Consensus-based image description evaluation</article-title>
          <year>2015</year>
          <conf-name>IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>
          <conf-date>2015 June 07-12</conf-date>
          <conf-loc>Boston, MA, USA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>CY</given-names>
            </name>
          </person-group>
          <article-title>Rouge: A package for automatic evaluation summaries</article-title>
          <year>2004</year>
          <conf-name>Proceedings of the Workshop on Text Summarization Branches Out</conf-name>
          <conf-date>2004 July 25–26</conf-date>
          <conf-loc>Barcelona, Spain</conf-loc>
          <fpage>74</fpage>
          <lpage>81</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Anderson</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Fernando</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gould</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>SPICE: semantic propositional image caption evaluation</article-title>
          <year>2016</year>
          <conf-name>14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part V</conf-name>
          <conf-date>2016 October 11-14</conf-date>
          <conf-loc>Amsterdam, The Netherlands</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kung</surname>
              <given-names>TH</given-names>
            </name>
            <name name-style="western">
              <surname>Cheatham</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Medenilla</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sillos</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>De Leon</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Elepaño</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Madriaga</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Aggabao</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Diaz-Candido</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Maningo</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tseng</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models</article-title>
          <source>PLOS Digit Health</source>
          <year>2023</year>
          <volume>2</volume>
          <issue>2</issue>
          <fpage>e0000198</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/36812645"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pdig.0000198</pub-id>
          <pub-id pub-id-type="medline">36812645</pub-id>
          <pub-id pub-id-type="pii">PDIG-D-22-00371</pub-id>
          <pub-id pub-id-type="pmcid">PMC9931230</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thirunavukarasu</surname>
              <given-names>AJ</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT cannot pass FRCOphth examinations: implications for ophthalmology and large language model artificial intelligence</article-title>
          <source>Eye News</source>
          <year>2023</year>
          <access-date>2024-10-24</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.eyenews.uk.com/media/31505/eye-am23-onex-arun-proof-2.pdf">https://www.eyenews.uk.com/media/31505/eye-am23-onex-arun-proof-2.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kuehn</surname>
              <given-names>BM</given-names>
            </name>
          </person-group>
          <article-title>More than one-third of US individuals use the internet to self-diagnose</article-title>
          <source>JAMA</source>
          <year>2013</year>
          <volume>309</volume>
          <issue>8</issue>
          <fpage>756</fpage>
          <lpage>757</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.2013.629</pub-id>
          <pub-id pub-id-type="medline">23443421</pub-id>
          <pub-id pub-id-type="pii">1656251</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shimizu</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Yazu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Aketa</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Tanji</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sakasegawa</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Nakayama</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ishikawa</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yokoiwa</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Sato</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Katayama</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Hanyuda</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sato</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Fukagawa</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Fujishima</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ogawa</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Tsubota</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Innovative artificial intelligence-based cataract diagnostic method uses a slit-lamp video recording device and multiple machine-learning</article-title>
          <source>Invest Ophthalmol Vis Sci</source>
          <year>2021</year>
          <volume>62</volume>
          <issue>8</issue>
          <fpage>1031</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goh</surname>
              <given-names>JHL</given-names>
            </name>
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>ZW</given-names>
            </name>
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Anees</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Nusinovici</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Rim</surname>
              <given-names>TH</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Tham</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence for cataract detection and management</article-title>
          <source>Asia Pac J Ophthalmol (Phila)</source>
          <year>2020</year>
          <volume>9</volume>
          <issue>2</issue>
          <fpage>88</fpage>
          <lpage>95</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2162-0989(23)00214-1"/>
          </comment>
          <pub-id pub-id-type="doi">10.1097/01.APO.0000656988.16221.04</pub-id>
          <pub-id pub-id-type="medline">32349116</pub-id>
          <pub-id pub-id-type="pii">S2162-0989(23)00214-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thirunavukarasu</surname>
              <given-names>AJ</given-names>
            </name>
          </person-group>
          <article-title>Large language models will not replace healthcare professionals: curbing popular fears and hype</article-title>
          <source>J R Soc Med</source>
          <year>2023</year>
          <volume>116</volume>
          <issue>5</issue>
          <fpage>181</fpage>
          <lpage>182</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://journals.sagepub.com/doi/abs/10.1177/01410768231173123?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/01410768231173123</pub-id>
          <pub-id pub-id-type="medline">37199678</pub-id>
          <pub-id pub-id-type="pmcid">PMC10331084</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>EyeGPT: ophthalmic assistant with large language models for patient inquiries and medical education</article-title>
          <source>J Med Internet Res</source>
          <year>2024</year>
          <month>11</month>
          <day>1</day>
          <lpage>6063</lpage>
          <comment>(forthcoming)(forthcoming)</comment>
          <pub-id pub-id-type="doi">10.2196/60063</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Yi</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ye</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>PS</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>A survey on evaluation of large language models</article-title>
          <source>Assoc Comput Mach</source>
          <year>2024</year>
          <volume>15</volume>
          <issue>3</issue>
          <fpage>39</fpage>
          <pub-id pub-id-type="doi">10.1145/364128</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Xiang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Evaluating large language models in medical applications: a survey</article-title>
          <source>arXiv:2405.07468</source>
          <year>2024</year>
          <fpage>1</fpage>
          <lpage>42</lpage>
          <comment>Preprint published on May 13, 2024</comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
