<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v25i1e43110</article-id>
      <article-id pub-id-type="pmid">36927634</article-id>
      <article-id pub-id-type="doi">10.2196/43110</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Viewpoint</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Viewpoint</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>What Does DALL-E 2 Know About Radiology?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Mavragani</surname>
            <given-names>Amaryllis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kapsetaki</surname>
            <given-names>Marianna</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Antani</surname>
            <given-names>Sameer</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Adams</surname>
            <given-names>Lisa C</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5836-4542</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Busch</surname>
            <given-names>Felix</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-9770-8555</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Truhn</surname>
            <given-names>Daniel</given-names>
          </name>
          <degrees>BSc, MD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9605-0728</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Makowski</surname>
            <given-names>Marcus R</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-8778-647X</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Aerts</surname>
            <given-names>Hugo J W L</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff5" ref-type="aff">5</xref>
          <xref rid="aff6" ref-type="aff">6</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2122-2003</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Bressem</surname>
            <given-names>Keno K</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Department of Radiology</institution>
            <institution>Charité Universitätsmedizin Berlin</institution>
            <institution>Corporate Member of Freie Universität Berlin and Humboldt Universität zu Berlin</institution>
            <addr-line>Hindenburgdamm 30</addr-line>
            <addr-line>Berlin, 12203</addr-line>
            <country>Germany</country>
            <phone>49 30 450 527792</phone>
            <email>keno-kyrill.bressem@charite.de</email>
          </address>
          <xref rid="aff5" ref-type="aff">5</xref>
          <xref rid="aff6" ref-type="aff">6</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-9249-8624</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Radiology</institution>
        <institution>Charité Universitätsmedizin Berlin</institution>
        <institution>Corporate Member of Freie Universität Berlin and Humboldt Universität zu Berlin</institution>
        <addr-line>Berlin</addr-line>
        <country>Germany</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Radiology</institution>
        <institution>Stanford University</institution>
        <addr-line>Stanford, CA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Department of Diagnostic and Interventional Radiology</institution>
        <institution>University Hospital Aachen</institution>
        <addr-line>Aachen</addr-line>
        <country>Germany</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Department of Diagnostic and Interventional Radiology</institution>
        <institution>Technical University of Munich</institution>
        <addr-line>Munich</addr-line>
        <country>Germany</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>Artificial Intelligence in Medicine Program (AIM)</institution>
        <institution>Mass General Brigham</institution>
        <institution>Harvard Medical School</institution>
        <addr-line>Boston, MA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff6">
        <label>6</label>
        <institution>Radiology and Nuclear Medicine, CARIM &#38; GROW</institution>
        <institution>Maastricht University</institution>
        <addr-line>Maastricht</addr-line>
        <country>Netherlands</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Keno K Bressem <email>keno-kyrill.bressem@charite.de</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2023</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>16</day>
        <month>3</month>
        <year>2023</year>
      </pub-date>
      <volume>25</volume>
      <elocation-id>e43110</elocation-id>
      <history>
        <date date-type="received">
          <day>30</day>
          <month>9</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>20</day>
          <month>12</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>30</day>
          <month>12</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>27</day>
          <month>1</month>
          <year>2023</year>
        </date>
      </history>
      <copyright-statement>©Lisa C Adams, Felix Busch, Daniel Truhn, Marcus R Makowski, Hugo J W L Aerts, Keno K Bressem. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 16.03.2023.</copyright-statement>
      <copyright-year>2023</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2023/1/e43110" xlink:type="simple"/>
      <related-article related-article-type="companion" id="v9iie46885" ext-link-type="doi" xlink:href="10.2196/46885" vol="9" page="e46885" xlink:title="See also:" xlink:type="simple">https://mededu.jmir.org/2023/1/e46885</related-article>
      <abstract>
        <p>Generative models, such as DALL-E 2 (OpenAI), could represent promising future tools for image generation, augmentation, and manipulation for artificial intelligence research in radiology, provided that these models have sufficient medical domain knowledge. Herein, we show that DALL-E 2 has learned relevant representations of x-ray images, with promising capabilities in terms of zero-shot text-to-image generation of new images, the continuation of an image beyond its original boundaries, and the removal of elements; however, its capabilities for the generation of images with pathological abnormalities (eg, tumors, fractures, and inflammation) or computed tomography, magnetic resonance imaging, or ultrasound images are still limited. The use of generative models for augmenting and generating radiological data thus seems feasible, even if the further fine-tuning and adaptation of these models to their respective domains are required first.</p>
      </abstract>
      <kwd-group>
        <kwd>DALL-E</kwd>
        <kwd>creating images from text</kwd>
        <kwd>image creation</kwd>
        <kwd>image generation</kwd>
        <kwd>transformer language model</kwd>
        <kwd>machine learning</kwd>
        <kwd>generative model</kwd>
        <kwd>radiology</kwd>
        <kwd>x-ray</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>medical imaging</kwd>
        <kwd>text-to-image</kwd>
        <kwd>diagnostic imaging</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec>
      <title>The Potential Impact of Generative Models in Radiology</title>
      <p>Artificial intelligence is often seen as a potential transformer of medicine [<xref ref-type="bibr" rid="ref1">1</xref>]. DALL-E 2 is a novel deep learning model for text-to-image generation that was first introduced by OpenAI in April 2022 [<xref ref-type="bibr" rid="ref2">2</xref>]. The model has recently gained widespread public interest due to its ability to create photorealistic images solely from short written inputs [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref5">5</xref>]. Trained on billions of text-image pairs extracted from the internet, DALL-E 2 learned a wide range of representations that it can recombine to create novel images, thereby exceeding the variability it has seen in the training data, even when given implausible prompts (eg, “a painting of a pelvic X-ray by Claude Monet”) [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref6">6</xref>-<xref ref-type="bibr" rid="ref8">8</xref>].</p>
      <p>DALL-E 2’s powerful generative capabilities raise the question of whether these can be transferred to the medical domain to create or augment data, as medical data can be sparse and hard to come by [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>]. Given the large number of images on which DALL-E 2 was trained, it is reasonable to assume that the training data for DALL-E 2 included radiological images and that the model may have learned something about the composition and structure of x-ray, cross-sectional, and ultrasound images.</p>
      <p>This viewpoint article describes how we systematically examined the radiological knowledge embedded in DALL-E 2 by creating and manipulating x-ray, computed tomography (CT), magnetic resonance imaging (MRI), and ultrasound images.</p>
    </sec>
    <sec>
      <title>Generating Radiological Images From Short Descriptive Texts</title>
      <p>To investigate the general radiological domain knowledge embedded in DALL-E 2, we first tested how well radiological images could be created from short descriptive texts by using the words “An X-ray of” and a 1-word description of an anatomical area. For the head, chest, shoulder, abdomen, pelvis, hand, knee, and ankle, we let DALL-E 2 synthesize 4 artificial x-ray images, which are displayed in <xref rid="figure1" ref-type="fig">Figure 1</xref>. For each prompt, the model produced 4 outputs, of which all are presented in order to avoid selection bias.</p>
      <p>The coloring of the images and the general structure of the bones appeared realistic, and the overall anatomical propositions were correct, indicating the presence of fundamental concepts of x-ray anatomy. However, upon close inspection, it was noted that the trabecular structure of the bone appeared random and did not follow the course of mechanical stress as it would in real x-rays. Sometimes, smaller bones were missing, like the fibula in multiple x-rays of the knee, or multiple small bones, such as several carpal or tarsal bones, were fused into 1 bone. Nevertheless, the overall structure of the depicted anatomic area was recognizable. In rare cases, additional bones and joints were created, such as an additional index finger in a hand x-ray. The model had the greatest difficulties in generating joints correctly, as joints appeared to be fused, or the joint surfaces were distorted. Finally, the quality of the images was not what one would expect for clinic x-ray images. The collimation was largely not correct, with parts of the organ missing or the relevant regions not being centered. Examples of this are visible in <xref rid="figure1" ref-type="fig">Figure 1</xref>, wherein the tips of the fingers are missing in all hand x-rays, and the outer parts of the pelvis are missing in one image.</p>
      <p>When compared to x-rays, the generation of cross-sectional images from CT or MRI was very limited (<xref rid="figure2" ref-type="fig">Figure 2</xref>). Although the desired anatomical structure was discernible on the images, the overall impression was predominantly nonsensical. Instead of a cross-sectional image, a photograph of an entire CT or MRI examination on x-ray film was often shown. On ultrasound images, the desired anatomic structures were not discernible; instead, all images resembled obstetric images. Nevertheless, the images showed concepts of CT, MRI, and ultrasound images, suggesting that representations of these modalities are present in DALL-E 2.</p>
      <fig id="figure1" position="float">
        <label>Figure 1</label>
        <caption>
          <p>Examples of anatomical structures in x-ray images that were created with DALL-E 2 based on short text descriptions.</p>
        </caption>
        <graphic xlink:href="jmir_v25i1e43110_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <fig id="figure2" position="float">
        <label>Figure 2</label>
        <caption>
          <p>Examples of text-to-image–generated anatomical structures in CT, MRI, and ultrasound images created with DALL-E 2. CT: computed tomography; MRI: magnetic resonance imaging.</p>
        </caption>
        <graphic xlink:href="jmir_v25i1e43110_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
    </sec>
    <sec>
      <title>Changing Areas in Radiological Images While Maintaining the Anatomical Structure</title>
      <p>To investigate the extent of DALL-E 2’s radiological knowledge, we tested how well the modal can reconstruct missing areas in a radiological image (inpainting). To do this, we selected radiographs of the pelvis, ankle, chest, shoulder, knee, wrist, and thoracic spine and erased specific areas before handing the remnants to DALL-E 2 for reconstruction (<xref rid="figure3" ref-type="fig">Figure 3</xref>). The accompanying prompts were identical to the those used for generating the text-based images described in the <italic>Generating Radiological Images From Short Descriptive Texts</italic> section. DALL-E 2 provided realistic replacement images that were nearly indistinguishable from the original radiographs of the pelvis, thorax, and thoracic spine. However, the results were not as convincing when a joint was included in the image area. In the ankle and wrist images, the number of tarsal bones and the structure varied greatly from those in the original radiographs and deviated from realistic representations. For the shoulder images, DALL-E 2 failed to reconstruct the glenoid cavity and articular surface of the humerus. In one image, a foreign body was inserted into the shoulder that remotely resembled a prosthesis. Further, when reconstructing the knee image, the model omitted the patella but retained the bicondylar structure of the femur.</p>
      <fig id="figure3" position="float">
        <label>Figure 3</label>
        <caption>
          <p>Reconstructed areas of different anatomical locations in x-rays created by using DALL-E 2. The yellow-bordered regions of the original images were erased before providing the remnant images for reconstruction.</p>
        </caption>
        <graphic xlink:href="jmir_v25i1e43110_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
    </sec>
    <sec>
      <title>Extending Radiological Images Beyond Image Borders</title>
      <p>Since the previous two experiments showed that DALL-E 2 could handle the basics of standard radiological images, we also wanted to investigate how well anatomical knowledge is embedded in the model. To do this, we randomly selected radiographic images of the abdomen, chest, pelvis, knee, various spinal regions, wrist bones, and hand and had DALL-E 2 extend these images beyond their boundaries, since the model needs to know both locations and anatomical proportions for this task. The accompanying prompts were selected based on the anatomical regions to be created. Again, the style of the augmented images represented a realistic representation of radiographs (<xref rid="figure4" ref-type="fig">Figure 4</xref>). It was possible to create a complete full-body radiograph by using only 1 image of the knee as a starting point. The greater the distance between the original image and the generated area, the less detailed the images became. Anatomical proportions, such as the length of the femur or the size of the lung, remained realistic, but finer details, such as the number of lumbar vertebrae, were inconsistent. The model generally performed best when creating anterior and posterior views, while the creation of lateral views was more challenging and produced poorer results.</p>
      <fig id="figure4" position="float">
        <label>Figure 4</label>
        <caption>
          <p>Extending x-ray images of different anatomical regions beyond their borders by using DALL-E 2. The original x-rays are shown in yellow boxes, and the areas outside of the yellow boxes were generated by DALL-E 2.</p>
        </caption>
        <graphic xlink:href="jmir_v25i1e43110_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
    </sec>
    <sec>
      <title>Generation of Pathological Images</title>
      <p>The generation of pathological images to, for example, visualize fractures, intracranial hemorrhages, or tumors was limited. We tested this by generating fractures on radiographs, which mostly resulted in distorted images that were similar to those in <xref rid="figure1" ref-type="fig">Figure 1</xref>. In addition, DALL-E 2 has a filter that prevents the generation of harmful content. Therefore, we could not test most pathologies because words such as “bleeding” triggered this filter.</p>
    </sec>
    <sec>
      <title>Conclusion and Outlook</title>
      <p>We were able to show that DALL-E 2 can generate x-ray images that are similar to authentic x-ray images in terms of style and anatomical proportions. Thus, we conclude that relevant representations for x-rays were learned by DALL-E 2 during training. However, it seems that only representations for radiographs without pathologies are available or are allowed by the filter, as the generation of pathological images (eg, images of fractures and tumors) was limited. In addition, DALL-E 2’s generative capabilities were poor for CT, MRI, and ultrasound images.</p>
      <p>Access to data is critical in deep learning, and in general, the larger the data set, the better the performance of the models trained on the data set. However, especially in radiology, there is no single large database from which to create such a data set. Instead, the necessary data are divided among several institutions, and privacy concerns additionally prevent the merging of these data. Synthetic data from generative models, such as DALL-E 2, show promise for addressing these issues by enabling the creation of data sets that are much larger than those that are currently available and greatly accelerating the development of new deep learning tools for radiology [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>].</p>
      <p>Previous approaches to data generation via generative adversarial networks were hampered by instabilities during training and the inability to capture sufficient scatter in the training data [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. In contrast, novel diffusion-based generative models, of which DALL-E 2 is a prominent representative, solved these problems and outperformed generative adversarial networks in image generation [<xref ref-type="bibr" rid="ref15">15</xref>].</p>
      <p>Diffusion models are very interesting for augmenting radiological data due to their control capabilities and flexibility in data generation, as images can be modified, and pathologies can be added or removed simply by using text inputs. Nevertheless, generated images should be subjected to quality control by domain experts to reduce the risk of incorrect information from DALL-E 2 entering a generated data set. This would help to avoid possible negative effects on models trained with these data or on individuals using these data for training.</p>
      <p>The biggest impact of these text-to-image models in radiology could be data generation for research and education. Notably, our experiments showed that DALL-E 2, in its current form, was not able to produce convincing pathological images, probably due to the lack of such images in the training data. Radiological images containing a variety of pathologies are more likely to be provided on professional websites, may be under restricted access, or are stored as specialized data formats (eg, DICOM). Overall, these restrictions likely resulted in only a small number of images being included in the training data. In addition, the images in the data could be distorted. When looking at <xref rid="figure1" ref-type="fig">Figures 1</xref> and <xref rid="figure4" ref-type="fig">4</xref>, the absence of breast shadows is noticeable, leading to the assumption that all images probably show men, which makes gender bias in the model likely. Furthermore, because there is no way to ensure that the images in the training data come only from curated medical websites, incorrect information in these images may cause the model to learn incorrect concepts and thus produce biased results.</p>
      <p>Further work is needed to explore diffusion models for medical data generation and modification [<xref ref-type="bibr" rid="ref16">16</xref>]. Although DALL-E 2 is not publicly available, other models with a similar architecture, such as Stable Diffusion (Stability AI), were recently made available to the public [<xref ref-type="bibr" rid="ref17">17</xref>].</p>
      <p>Future research should focus on fine-tuning these models to medical data and incorporating medical terminology to create powerful models for data generation and augmentation in radiology research.</p>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">CT</term>
          <def>
            <p>computed tomography</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">MRI</term>
          <def>
            <p>magnetic resonance imaging</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>KKB is grateful for his participation in the Berlin Institute of Health Charité Digital Clinician Scientist Program, which is funded by the Charité – Universitätsmedizin Berlin and the Berlin Institute of Health.</p>
    </ack>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>All generated research data are included in the manuscript.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>LCA and KKB conceptualized the experiments and were responsible for the methodology. LCA, FB, and KKB curated the data, conducted the formal analysis, performed the investigation, and wrote the original manuscript draft. KKB was responsible for resources and visualization. LCA, FB, DT, HJWLA, and KKB performed the validation. LCA, FB, DT, MRM, HJWLA, and KKB reviewed and edited the manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sharma</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Savage</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Nair</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Larsson</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Svedberg</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nygren</surname>
              <given-names>JM</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence applications in health care practice: Scoping review</article-title>
          <source>J Med Internet Res</source>
          <year>2022</year>
          <month>10</month>
          <day>05</day>
          <volume>24</volume>
          <issue>10</issue>
          <fpage>e40238</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2022/10/e40238/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/40238</pub-id>
          <pub-id pub-id-type="medline">36197712</pub-id>
          <pub-id pub-id-type="pii">v24i10e40238</pub-id>
          <pub-id pub-id-type="pmcid">PMC9582911</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ramesh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dhariwal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nichol</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Chu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Hierarchical text-conditional image generation with CLIP latents</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on April 13, 2022
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2204.06125.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2204.06125</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kather</surname>
              <given-names>JN</given-names>
            </name>
            <name name-style="western">
              <surname>Laleh</surname>
              <given-names>NG</given-names>
            </name>
            <name name-style="western">
              <surname>Foersch</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Truhn</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Medical domain knowledge in domain-agnostic generative AI</article-title>
          <source>NPJ Digit Med</source>
          <year>2022</year>
          <month>07</month>
          <day>11</day>
          <volume>5</volume>
          <issue>1</issue>
          <fpage>90</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-022-00634-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-022-00634-5</pub-id>
          <pub-id pub-id-type="medline">35817798</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-022-00634-5</pub-id>
          <pub-id pub-id-type="pmcid">PMC9273760</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Conwell</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ullman</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Testing relational understanding in text-guided image generation</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on July 29, 2022
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2208.00005.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2208.00005</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Marcus</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Davis</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Aaronson</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A very preliminary analysis of DALL-E 2</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on April 25, 2022
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/vc/arxiv/papers/2204/2204.13807v1.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2204.13807</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nichol</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dhariwal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Ramesh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shyam</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Mishkin</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>McGrew</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>GLIDE: Towards photorealistic image generation and editing with text-guided diffusion models</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on December 20, 2021
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2112.10741v1.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2112.10741</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lyu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Communication in human–AI co-creation: Perceptual analysis of paintings generated by text-to-image system</article-title>
          <source>Appl Sci (Basel)</source>
          <year>2022</year>
          <month>11</month>
          <day>08</day>
          <volume>12</volume>
          <issue>22</issue>
          <fpage>11312</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/2076-3417/12/22/11312"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/app122211312</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Oppenlaender</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>The creativity of text-to-image generation</article-title>
          <year>2022</year>
          <month>11</month>
          <conf-name>Academic Mindtrek 2022: 25th International Academic Mindtrek conference</conf-name>
          <conf-date>November 16-18, 2022</conf-date>
          <conf-loc>Tampere, Finland</conf-loc>
          <fpage>192</fpage>
          <lpage>202</lpage>
          <pub-id pub-id-type="doi">10.1145/3569219.3569352</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Park</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Digital health care industry ecosystem: Network analysis</article-title>
          <source>J Med Internet Res</source>
          <year>2022</year>
          <month>08</month>
          <day>17</day>
          <volume>24</volume>
          <issue>8</issue>
          <fpage>e37622</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2022/8/e37622/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/37622</pub-id>
          <pub-id pub-id-type="medline">35976690</pub-id>
          <pub-id pub-id-type="pii">v24i8e37622</pub-id>
          <pub-id pub-id-type="pmcid">PMC9434390</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schütte</surname>
              <given-names>AD</given-names>
            </name>
            <name name-style="western">
              <surname>Hetzel</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gatidis</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hepp</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Dietz</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Bauer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Schwab</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Overcoming barriers to data sharing with medical image generation: a comprehensive evaluation</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <month>09</month>
          <day>24</day>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>141</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-021-00507-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-021-00507-3</pub-id>
          <pub-id pub-id-type="medline">34561528</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-021-00507-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC8463544</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jordon</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Szpruch</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Houssiau</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Bottarelli</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cherubin</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Maple</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>SN</given-names>
            </name>
            <name name-style="western">
              <surname>Weller</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Synthetic data -- what, why and how?</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on May 6, 2022
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2205.03257.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2205.03257</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jordon</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wilson</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>van der Schaar</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Synthetic data: Opening the data floodgates to enable faster, more directed development of machine learning methods</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on December 8, 2020
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2012.04580.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2012.04580</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goodfellow</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Pouget-Abadie</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mirza</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Warde-Farley</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Ozair</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Courville</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bengio</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Generative adversarial networks</article-title>
          <source>Commun ACM</source>
          <year>2020</year>
          <month>11</month>
          <volume>63</volume>
          <issue>11</issue>
          <fpage>139</fpage>
          <lpage>144</lpage>
          <pub-id pub-id-type="doi">10.1145/3422622</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ciano</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Andreini</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Mazzierli</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Bianchini</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Scarselli</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>A multi-stage GAN for multi-organ chest x-ray image generation and segmentation</article-title>
          <source>Mathematics (Basel)</source>
          <year>2021</year>
          <month>11</month>
          <day>14</day>
          <volume>9</volume>
          <issue>22</issue>
          <fpage>2896</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/2227-7390/9/22/2896"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/math9222896</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dhariwal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nichol</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Diffusion models beat GANs on image synthesis</article-title>
          <year>2021</year>
          <conf-name>35th Conference on Neural Information Processing Systems (NeurIPS 2021)</conf-name>
          <conf-date>December 6-14, 2021</conf-date>
          <conf-loc>Virtual-only conference</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Summers</surname>
              <given-names>RM</given-names>
            </name>
          </person-group>
          <article-title>A disentangled generative model for disease decomposition in chest x-rays via normal image synthesis</article-title>
          <source>Med Image Anal</source>
          <year>2021</year>
          <month>01</month>
          <volume>67</volume>
          <fpage>101839</fpage>
          <pub-id pub-id-type="doi">10.1016/j.media.2020.101839</pub-id>
          <pub-id pub-id-type="medline">33080508</pub-id>
          <pub-id pub-id-type="pii">S1361-8415(20)30203-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Buehler</surname>
              <given-names>MJ</given-names>
            </name>
          </person-group>
          <article-title>Generating 3D architectured nature-inspired materials and granular media using diffusion models based on language cues</article-title>
          <source>Oxf Open Mater Sci</source>
          <year>2022</year>
          <month>11</month>
          <day>11</day>
          <volume>2</volume>
          <issue>1</issue>
          <fpage>itac010</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/36756638"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/oxfmat/itac010</pub-id>
          <pub-id pub-id-type="medline">36756638</pub-id>
          <pub-id pub-id-type="pmcid">PMC9767007</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
