<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v25i1e46216</article-id>
      <article-id pub-id-type="pmid">37261889</article-id>
      <article-id pub-id-type="doi">10.2196/46216</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Prediction of Sleep Stages Via Deep Learning Using Smartphone Audio Recordings in Home Environments: Model Development and Validation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Mungoli</surname>
            <given-names>Neelesh</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Wall</surname>
            <given-names>Karen</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Tummalacherla</surname>
            <given-names>Meghasyam</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Tran</surname>
            <given-names>Hai Hong</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4006-3909</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Hong</surname>
            <given-names>Jung Kyung</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1893-7868</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Jang</surname>
            <given-names>Hyeryung</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-7314-0739</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Jung</surname>
            <given-names>Jinhwan</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2455-4206</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Jongmok</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0006-5593-9244</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Hong</surname>
            <given-names>Joonki</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-8378-3332</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>Minji</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4962-3869</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Jeong-Whun</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff5" ref-type="aff">5</xref>
          <xref rid="aff6" ref-type="aff">6</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4858-3316</ext-link>
        </contrib>
        <contrib id="contrib9" contrib-type="author">
          <name name-style="western">
            <surname>Kushida</surname>
            <given-names>Clete A</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff7" ref-type="aff">7</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9430-3752</ext-link>
        </contrib>
        <contrib id="contrib10" contrib-type="author">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>Dongheon</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4660-6174</ext-link>
        </contrib>
        <contrib id="contrib11" contrib-type="author">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Daewoo</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2475-6956</ext-link>
        </contrib>
        <contrib id="contrib12" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Yoon</surname>
            <given-names>In-Young</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <address>
            <institution>Department of Psychiatry</institution>
            <institution>Seoul National University Bundang Hospital</institution>
            <addr-line>82 Gumi-ro 173beon-gil, Bundang-gu, Seongnam-si</addr-line>
            <addr-line>Gyeonggi-do, 13620</addr-line>
            <country>Republic of Korea</country>
            <fax>82 31 787 4058</fax>
            <phone>82 31 787 7433</phone>
            <email>iyoon@snu.ac.kr</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3995-8238</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Asleep Inc.</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Psychiatry</institution>
        <institution>Seoul National University College of Medicine</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Department of Artificial Intelligence</institution>
        <institution>Dongguk University</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Department of Psychiatry</institution>
        <institution>Seoul National University Bundang Hospital</institution>
        <addr-line>Gyeonggi-do</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>Department of Otorhinolaryngology</institution>
        <institution>Seoul National University Bundang Hospital</institution>
        <addr-line>Gyeonggi-do</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff6">
        <label>6</label>
        <institution>Department of Otorhinolaryngology</institution>
        <institution>Seoul National University College of Medicine</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff7">
        <label>7</label>
        <institution>Department of Psychiatry and Behavioral Sciences</institution>
        <institution>Stanford University Medical Center</institution>
        <addr-line>Redwood City, CA</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: In-Young Yoon <email>iyoon@snu.ac.kr</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2023</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>1</day>
        <month>6</month>
        <year>2023</year>
      </pub-date>
      <volume>25</volume>
      <elocation-id>e46216</elocation-id>
      <history>
        <date date-type="received">
          <day>2</day>
          <month>2</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>2</day>
          <month>3</month>
          <year>2023</year>
        </date>
        <date date-type="rev-recd">
          <day>23</day>
          <month>3</month>
          <year>2023</year>
        </date>
        <date date-type="accepted">
          <day>31</day>
          <month>3</month>
          <year>2023</year>
        </date>
      </history>
      <copyright-statement>©Hai Hong Tran, Jung Kyung Hong, Hyeryung Jang, Jinhwan Jung, Jongmok Kim, Joonki Hong, Minji Lee, Jeong-Whun Kim, Clete A Kushida, Dongheon Lee, Daewoo Kim, In-Young Yoon. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 01.06.2023.</copyright-statement>
      <copyright-year>2023</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2023/1/e46216" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>The growing public interest and awareness regarding the significance of sleep is driving the demand for sleep monitoring at home. In addition to various commercially available wearable and nearable devices, sound-based sleep staging via deep learning is emerging as a decent alternative for their convenience and potential accuracy. However, sound-based sleep staging has only been studied using in-laboratory sound data. In real-world sleep environments (homes), there is abundant background noise, in contrast to quiet, controlled environments such as laboratories. The use of sound-based sleep staging at homes has not been investigated while it is essential for practical use on a daily basis. Challenges are the lack of and the expected huge expense of acquiring a sufficient size of home data annotated with sleep stages to train a large-scale neural network.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aims to develop and validate a deep learning method to perform sound-based sleep staging using audio recordings achieved from various uncontrolled home environments.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>To overcome the limitation of lacking home data with known sleep stages, we adopted advanced training techniques and combined home data with hospital data. The training of the model consisted of 3 components: (1) the original supervised learning using 812 pairs of hospital polysomnography (PSG) and audio recordings, and the 2 newly adopted components; (2) transfer learning from hospital to home sounds by adding 829 smartphone audio recordings at home; and (3) consistency training using augmented hospital sound data. Augmented data were created by adding 8255 home noise data to hospital audio recordings. Besides, an independent test set was built by collecting 45 pairs of overnight PSG and smartphone audio recording at homes to examine the performance of the trained model.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The accuracy of the model was 76.2% (63.4% for wake, 64.9% for rapid-eye movement [REM], and 83.6% for non-REM) for our test set. The macro F1-score and mean per-class sensitivity were 0.714 and 0.706, respectively. The performance was robust across demographic groups such as age, gender, BMI, or sleep apnea severity (accuracy 73.4%-79.4%). In the ablation study, we evaluated the contribution of each component. While the supervised learning alone achieved accuracy of 69.2% on home sound data, adding consistency training to the supervised learning helped increase the accuracy to a larger degree (+4.3%) than adding transfer learning (+0.1%). The best performance was shown when both transfer learning and consistency training were adopted (+7.0%).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This study shows that sound-based sleep staging is feasible for home use. By adopting 2 advanced techniques (transfer learning and consistency training) the deep learning model robustly predicts sleep stages using sounds recorded at various uncontrolled home environments, without using any special equipment but smartphones only.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>respiratory sounds</kwd>
        <kwd>sleep stages</kwd>
        <kwd>deep learning</kwd>
        <kwd>smartphone</kwd>
        <kwd>home environment</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Growing knowledge that sleep plays a vital role in maintaining well-being and good health, both physical and mental, increases public interest and awareness regarding the importance of sleep to health. Therefore, the demand for knowing and taking care of one’s own sleep increases, so does the demand for sleep monitoring [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. The gold-standard test for monitoring and quantifying sleep is polysomnography (PSG), which typically requires 1 night of sleep at a sleep center with various biosignals recorded, such as the electroencephalogram (EEG; brain wave activity), electrooculogram (EOG; eye movement activity), electromyogram (EMG; muscle activity), electrocardiogram (ECG; heartbeat activity), and respiratory signals. After the overnight recording, the sleep data are reviewed by human experts to score sleep stages, arousals (ie, brief awakening), and respiratory and movement events. While PSG remains the most accurate diagnostic tool for sleep, it is too expensive and inconvenient to be used for a general population on a daily basis. In addition, a standard PSG taken in a laboratory environment may not reflect one’s habitual sleep at home [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. An easy and convenient method is thus required to enable home-based daily sleep monitoring for the general population [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>].</p>
      <p>Various commercial sleep trackers are available (ie, wearable or nearable devices), mostly using accelerometer for activity and movements, ECG or photoplethysmogram for heart rate variability, piezoelectricity or radar for respiratory movements, or EEG for brain activity [<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref13">13</xref>]. However, because of their inconvenience and high cost, people do not vigorously use these devices. Recently, sound-based sleep staging has emerged as a new alternative, relying on recognizing sound patterns of respiratory and body movements [<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref20">20</xref>]. The advantage of using sounds is that sleep can be measured remotely without contact [<xref ref-type="bibr" rid="ref17">17</xref>-<xref ref-type="bibr" rid="ref20">20</xref>]. Among the various studies performed in this regard, a deep learning model (SoundSleepNet) predicted sleep stages using smartphone audio recordings with good accuracy [<xref ref-type="bibr" rid="ref20">20</xref>], showing the potential of sound-based sleep staging using smartphones.</p>
      <p>However, sound-based sleep staging models have been developed and tested only in laboratory environments such as in a hospital [<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref20">20</xref>]. Unfortunately, feasible sleep sounds are mostly recorded during PSG in hospitals, while PSG requires a controlled environment (ie, a quiet and soundproof room where the examinee stays alone). It also remains questionable as to whether a sound-based sleep staging model can work well in home environments, as diverse and dynamic background noise is present in home environments (eg, home appliances, pets, roommates, and outdoor noise such as traffic noise). Therefore, it is more difficult to train a model to predict sleep stages using home sounds, which are full of uncontrolled noise, compared with using hospital sounds. Consequently, a specific training is needed to derive a model to work at home.</p>
      <p>An obstacle in this regard is that deep learning models require thousands of ground truth (ie, PSG for sleep measure) to be trained; however, large amounts of home PSG data are not yet available. Thus, introduction of advanced techniques may help bypass the step of collecting large-sized home PSG data. One technique that may be useful is transfer learning. It allows a model trained by hospital sounds to learn to predict sleep stages for home sounds [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>]. Another useful technique is consistency training, using which a model can be trained with hospital sounds augmented by adding home noise. It makes a model learn to predict sleep stages regardless of the presence of home noise [<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref26">26</xref>].</p>
      <p>Meanwhile, validation of sleep trackers at home is important [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. Most sleep trackers have only been validated in laboratory environments [<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref20">20</xref>] because it requires a lot of effort to prospectively collect home PSG data [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. However, the performance should be addressed specifically in “home environments” to really serve as daily home sleep trackers [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>].</p>
      <p>In this study, we propose a deep learning model adopting advanced training techniques for sound-based sleep staging at home, an uncontrolled environment full of noise. The performance of the proposed model was examined by level 2 PSGs conducted at home.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Sleep Sound Data Sets</title>
        <sec>
          <title>Overview of the Data Sets</title>
          <p>This study used 3 different data sets: a hospital PSG data set (level 1 PSG and audio recording for 812 nights) and a home smartphone data set (smartphone audio recordings without PSG for 829 nights) for training, and a home PSG data set (level 2 PSG and matched smartphone audio recordings for 45 nights) as the test data set.</p>
        </sec>
        <sec>
          <title>Hospital PSG Data Set</title>
          <p>This is a clinical data set from the sleep center of Seoul National University Bundang Hospital (SNUBH) collected between 2019 and 2020, which includes PSG and matched audio data [<xref ref-type="bibr" rid="ref20">20</xref>]. As the data set was retrospectively collected from PSGs previously conducted, additional informed consents were not available. All data were anonymized.</p>
        </sec>
        <sec>
          <title>Home Smartphone Data Set</title>
          <p>Adult volunteers were recruited and screened through an internet survey and audio recordings were collected between June and November 2022. Informed consent was obtained from each participant by an electronic form. Audio recording at night was self-conducted by each participant using his or her own smartphone at home following predefined instructions. Various models of smartphone were used, ranging from Android (OS version later than 8.0; Google LLC/Alphabet Inc) to iOS devices (OS version later than 15; Apple Inc). The participants were asked to place the phone 0.5-1 m from their head. Using phone models owned by the participants for data collection simulates the real scenario, while recordings from various phones help the model to adapt to different microphone settings (see <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for additional information regarding how the participants were selected for the study).</p>
        </sec>
        <sec>
          <title>Home PSG Data Set</title>
          <p>Adult volunteers were recruited at the sleep center of the SNUBH and home PSG tests were performed together with audio recordings between June and November 2022. Written informed consents were obtained from each participant. A portable PSG setup was made by sleep technicians at the center and an iPhone 11 was provided for audio recording. Participants were asked to sleep at home during PSG, with the provided smartphone placed on a side table or mattress, with a 0.5-1.0-m distance from their head.</p>
          <p>The demographics of the participants in each data set are presented in <xref ref-type="table" rid="table1">Table 1</xref>. Additional details of the 3 data sets are described in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Demographics of the 3 data sets.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="30"/>
              <col width="290"/>
              <col width="230"/>
              <col width="260"/>
              <col width="190"/>
              <thead>
                <tr valign="top">
                  <td colspan="2">Demographics</td>
                  <td>Hospital PSG<sup>a</sup> (training) data set (n=812)</td>
                  <td>Home smartphone (training) data set (n=829)</td>
                  <td>Home PSG (testing) data set (n=45)</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td colspan="2">Age<sup>b</sup> (year), mean (SD)</td>
                  <td>52.7 (13.6)</td>
                  <td>36.2 (9.7)</td>
                  <td>44.7 (15.8)</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">Male, n (%)</td>
                  <td>562 (69.2)</td>
                  <td>330 (39.8)</td>
                  <td>19 (42.2)</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">BMI<sup>b</sup> (kg/m<sup>2</sup>), mean (SD)</td>
                  <td>25.9 (4.1)</td>
                  <td>23.2 (4.3)</td>
                  <td>24.0 (3.9)</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">
                    <bold>AHI<sup>b,c</sup>, mean (SD)</bold>
                  </td>
                  <td>23.3 (23.0)</td>
                  <td>—<sup>d</sup></td>
                  <td>11.8 (16.4)</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>AHI&#60;5, n (%)</td>
                  <td>193 (23.8)</td>
                  <td>—</td>
                  <td>22 (48.9)</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>5≤AHI&#60;15, n (%)</td>
                  <td>182 (22.4)</td>
                  <td>—</td>
                  <td>11 (24.4)</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>15≤AHI&#60;30, n (%)</td>
                  <td>207 (25.5)</td>
                  <td>—</td>
                  <td>7 (15.6)</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>30≤AHI, n (%)</td>
                  <td>230 (28.3)</td>
                  <td>—</td>
                  <td>5 (11.1)</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">Rapid-eye movement sleep behavior disorder, n (%)</td>
                  <td>72 (8.9)</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">Restless legs syndrome, n (%)</td>
                  <td>26 (3.2)</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td colspan="2">Insomnia, n (%)</td>
                  <td>220 (27.1)</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table1fn1">
                <p><sup>a</sup>PSG: polysomnography.</p>
              </fn>
              <fn id="table1fn2">
                <p><sup>b</sup>Continuous variable.</p>
              </fn>
              <fn id="table1fn3">
                <p><sup>c</sup>AHI: apnea-hypopnea index.</p>
              </fn>
              <fn id="table1fn4">
                <p><sup>d</sup>Not available.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
      <sec>
        <title>Polysomnography</title>
        <p>For PSG, standard sensors and channels were used (eg, 6-channel EEG, 2-channel EOG, chin EMG, ECG, 2-leg EMGs, respiratory effort, airflow, and oxygen saturation). Level 1 PSG was performed at the hospital under monitoring by sleep technologists. For level 2 PSG, after experienced sleep technologists from SNUBH hooked up participants with recording electrodes and equipment for each test, the participants went home to conduct their test at home. The main difference between level 1 and level 2 PSG is the presence of technologists and real-time monitoring during the recording [<xref ref-type="bibr" rid="ref27">27</xref>]. After the PSG recording, sleep technologists reviewed each study and manually annotated the study for sleep stages, followed by confirmation by a sleep specialist, in accordance with the American Academy of Sleep Medicine scoring manual [<xref ref-type="bibr" rid="ref28">28</xref>]. Each 30-second epoch of PSG was scored as 1 of 5 sleep stages, namely, wake, rapid-eye movement (REM), non-REM (NREM) stage 1 (N1), 2 (N2), and 3 (N3).</p>
      </sec>
      <sec>
        <title>Data Preprocessing</title>
        <p>All audio data were cut into 30-second epochs, preprocessed by adaptive noise reduction and Mel spectrogram conversion, and matched with the corresponding PSG labels to train and verify the model [<xref ref-type="bibr" rid="ref20">20</xref>]. In addition, pitch shifting was applied as a simple data augmentation technique. Ground truth labels were only available for the 2 data sets for which PSGs were conducted concurrently, namely, the hospital PSG data set for training and the home PSG data set for test.</p>
      </sec>
      <sec>
        <title>Deep Neural Network Architecture</title>
        <sec>
          <title>Training Overview</title>
          <p>To fairly demonstrate the effects of the training techniques proposed in this paper, we adopted the SoundSleepNet model and its trained network parameters that performed well in hospital environments [<xref ref-type="bibr" rid="ref20">20</xref>]. The network processed 40 input Mel spectrograms of sound data, each representing one 30-second sleep epoch, and output sleep-stage predictions of the 20 middle epochs (40 to 20). By adopting a well-trained model, performance difference is guaranteed to arise from only the additional training techniques, not from the network architecture.</p>
        </sec>
        <sec>
          <title>Training Components</title>
          <p>The proposed model, dubbed HomeSleepNet, was trained by 3 training components (<xref rid="figure1" ref-type="fig">Figure 1</xref>A).</p>
          <p>The first component was supervised learning [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref30">30</xref>], where the large-sized hospital PSG data set was used to train the HomeSleepNet model to make correct predictions of sleep stages from the input Mel spectrograms in hospital environments.</p>
          <p>The second component was transfer learning [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>], where Mel spectrograms from both hospital and home were used. Using a domain discriminator, the feature extractor was trained to transfer the sleep staging knowledge from the hospital domain to the home domain.</p>
          <p>The third component was consistency training [<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref26">26</xref>], for which 2 augmented hospital sound inputs were needed. Consistency training helps the HomeSleepNet model to perform sleep staging reliably in the presence of home noise.</p>
          <p>The 3 training components were executed concurrently to preserve the effects of each component. The details of each training component are described in the following sections.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>Training and inference of HomeSleepNet. (A) Training phase with 3 training components. On the left side, the data input for each training component was marked; on the right side, the impact of each component to the training is visually explained. Three training components are (1) supervised learning that trained the feature extractor and classifier to correctly predict sleep stages with Mel spectrogram inputs in hospital environments; (2) transfer learning that transferred sleep staging knowledge from hospital to home environments using a domain discriminator; (3) consistency training that helped the model make robust predictions despite the home noise presence. (B) HomeSleepNet in the inference phase after training is completed. All training blocks were removed, and only the feature extractor and classifier remained for the classification task. <italic>L<sub>c</sub></italic>: cross-entropy loss; <italic>L<sub>a</sub></italic>: auxiliary loss; <italic>L<sub>js</sub></italic>: Jensen-Shannon consistency loss; <italic>L<sub>d</sub></italic>: binary cross-entropy loss; NREM: nonrapid-eye movement; PSG: polysomnography; REM: rapid-eye movement.</p>
            </caption>
            <graphic xlink:href="jmir_v25i1e46216_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Supervised Learning for Sleep Staging in Hospital Environments</title>
          <p>The purpose of the supervised learning for HomeSleepNet is to train the network with preprocessed Mel spectrograms and matched sleep-stage ground truths from the hospital PSG data set so that the network can predict sleep stages using the input Mel spectrogram data in hospital environments.</p>
          <p>The supervised learning component was used to train 2 subnetworks: a feature extractor and a classifier (<xref rid="figure1" ref-type="fig">Figure 1</xref>A). The feature extractor uses Mel spectrograms of hospital sound data as input and extracts temporal and frequency features related to respiratory and sleep activity patterns. The classifier receives the features and predicts sleep stages of each Mel spectrogram. As a result, we train both the feature extractor and the classifier by minimizing the cross-entropy loss that measures the difference between the sleep-stage ground truth and the network predictions.</p>
        </sec>
        <sec>
          <title>Transfer Learning via Unsupervised Domain Adaptation</title>
          <p>Transfer learning for HomeSleepNet was executed by unsupervised domain adaptation (UDA) [<xref ref-type="bibr" rid="ref21">21</xref>]. The goal of UDA is to make a model originally trained with a source domain (hospital environments) perform similarly for a target domain (home environments). One popular direction of UDA is to extract common features (ie, <italic>domain-invariant features</italic>) between data from the source domain and the target domain so that the model can perform well regardless of the domain of the input data [<xref ref-type="bibr" rid="ref22">22</xref>-<xref ref-type="bibr" rid="ref24">24</xref>].</p>
          <p>Following Ganin et al [<xref ref-type="bibr" rid="ref22">22</xref>], we added a domain discriminator comprising simple convolutional layers followed by several fully connected layers (<xref rid="figure2" ref-type="fig">Figure 2</xref>). The feature extractor generates features from input Mel spectrograms, and the domain discriminator predicts the original domain of the features (hospital domain or home domain). Hospital PSG sound data and home smartphone sound data were used as inputs from the 2 domains. The training itself, however, does not require sleep-stage labels but only the domain ground truth of each input data. A binary cross-entropy loss, which indicates the difference between the domain prediction and the domain ground truth, was used to train both subnetworks. We used an adversarial training algorithm [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>], in which the domain discriminator is trained to be more accurate (i.e., minimize the loss), while the feature extractor is trained to make the domain discriminator less accurate (i.e., maximize the loss) by extracting features that confuse the domain discriminator. In the end, the domain discriminator can no longer recognize the input domains from the extracted features, which means the extracted features are “domain invariant.” Therefore, the well-trained classifier (<xref rid="figure1" ref-type="fig">Figure 1</xref>) can correctly predict the sleep stages regardless of the original domain.</p>
          <p>In addition to the adversarial training, an auxiliary loss, which consists of conditional entropy and virtual adversarial training [<xref ref-type="bibr" rid="ref24">24</xref>], was applied (<xref rid="figure1" ref-type="fig">Figure 1</xref>A) to reserve the classifying performance after domain adaptation.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Training procedure of transfer learning (unsupervised domain adaptation). Mel spectrograms from hospital and home domains are used as training data. The feature extractor will extract only domain-invariant features when the domain discriminator is not able to classify the input domains. <italic>L<sub>d</sub></italic> is the binary cross-entropy loss function representing the difference between the domain prediction and the domain ground truth. CNN: convolutional neural network; FC: fully-connected layer; PSG: polysomnography.</p>
            </caption>
            <graphic xlink:href="jmir_v25i1e46216_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Consistency Training Using Augmented Data</title>
          <p>In this study, we applied <italic>consistency training</italic> [<xref ref-type="bibr" rid="ref25">25</xref>] to train the model to output consistent predictions on hospital data, regardless of the presence of home noise. <italic>Data augmentation</italic> was used for this purpose by artificially adding home noise to the original hospital data to simulate sounds recorded at home. The feature extractor and classifier were trained through consistency training to predict the same sleep stage for the augmented data as they do for the original data.</p>
          <p>In detail, to create the augmented data, home noise audio was converted into a Mel spectrogram and added into the Mel spectrogram of hospital data with randomly generated phases and signal-to-noise ratio value ranging from –10 dB to 10 dB (<xref rid="figure3" ref-type="fig">Figure 3</xref>A). Noise audios were downloaded from Freesound (Music Technology Group) [<xref ref-type="bibr" rid="ref31">31</xref>], an open database of sounds that can be used for scientific research. We used audio tags about home environments (such as home appliances, room noise, air conditioner, fan) to filter out the unrelated audio files and form a home noise data set. In total, 8255 sound clips, preprocessed identically to the sleep sound audio, were used for this research. More information on how the noise clips were filtered and downloaded is presented in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
          <p>For consistency training (<xref rid="figure3" ref-type="fig">Figure 3</xref>B), we created 2 augmented noisy samples from the original hospital data <italic>x<sub>S</sub></italic> by applying the noise adding process with 2 different types of home noise data. The hospital data <italic>x<sub>S</sub></italic> and the 2 augmented samples were then fed into the feature extractor and classifier to obtain 3 corresponding predictions. We then adapted Jensen-Shannon divergence loss [<xref ref-type="bibr" rid="ref26">26</xref>] as the consistency loss that measures the difference between the 3 predictions. By minimizing this consistency loss, the resulting model is able to generate robust and consistent predictions, even in the presence of home noise.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Framework of consistency training. (A) The process of creating an augmented Mel spectrogram from hospital data and home noise data. The augmented data retain breathing patterns in the clean hospital data and noise features in the home noise data. (B) The consistency training procedure. Consistency loss <italic>L<sub>js</sub></italic> is minimized to make identical predictions on hospital data and augmented data. PSG: polysomnography; SNR: signal-to-noise ratio.</p>
            </caption>
            <graphic xlink:href="jmir_v25i1e46216_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Training Settings</title>
          <p>Pretrained network parameters from SoundSleepNet [<xref ref-type="bibr" rid="ref20">20</xref>] were used to initialize the parameters of HomeSleepNet. The training used the Adam optimizer with a fixed learning rate of 0.0002 within 20 training epochs. The aggregated training algorithm using the 3 components is presented in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. In the inference phase after training, only feature extractor and classifier subnetworks remained for the sleep staging task (<xref rid="figure1" ref-type="fig">Figure 1</xref>B).</p>
        </sec>
      </sec>
      <sec>
        <title>Evaluation Methods</title>
        <p>We evaluated HomeSleepNet in 4 different ways using the home PSG data set.</p>
        <p>First, the main outcome was the sleep staging performance for the 3-stage classification (wake, REM, and NREM) with evaluation metrics of accuracy, Cohen κ, macro <italic>F</italic><sub>1</sub>-score, and mean per-class sensitivity. Accuracy shows the overall quality of the model prediction; Cohen κ evaluates the interrater reliability between HomeSleepNet predictions and PSG sleep stages; macro <italic>F</italic><sub>1</sub>-score evaluates the model while considering the data imbalance; mean per-class sensitivity evaluates the model predictions for each sleep stage. For all 4 metrics, the higher the score, the better the performance. Performance for the 4-stage (wake, light sleep, deep sleep, and REM) and 2-stage (wake and sleep) classifications was also reported. In 4 stages, N1 and N2 were classified as light sleep and N3 was defined as deep sleep. The principal component analysis (PCA) plots [<xref ref-type="bibr" rid="ref32">32</xref>] were presented to show clusters in the feature space of the model. Using the output of the last hidden layer in HomeSleepNet, PCA was used to extract the most representative features of each input data in a 2D format. These extracted 2D features are then illustrated on a 2D coordinate plane. If there appear sleep-stage clusters in the plane, it means that the predictions from HomeSleepNet are reliable.</p>
        <p>Second, multiple sleep metrics were compared between predictions of HomeSleepNet and manual annotations of PSGs. The presented sleep metrics were total sleep time, sleep onset latency, sleep efficiency, wake after sleep onset, REM latency, and portions of each sleep stage, which were all calculated per night. Total sleep time is the total time asleep, calculated by adding all 30-second epochs annotated or predicted as sleep (ie, N1, N2, N3, and REM). Sleep onset latency is the length of time between lights off and the first epoch scored as sleep. Sleep efficiency is calculated as total sleep time divided by the total time spent in the bed (in our case, the recording time). Wake after sleep onset is the total wake time between the first sleep and the last sleep epoch of the night. REM latency is the length of time between the first sleep epoch and the first REM sleep epoch. Portions of each sleep stage were calculated by the sum of each stage divided by the recording time per night. The agreement between the 2 measurements was presented by the Bland-Altman plots.</p>
        <p>Third, to investigate performance according to demographic characteristics, we divided the test data set into groups regarding age, gender, BMI, and apnea-hypopnea index (AHI). Performance of HomeSleepNet was evaluated on each group, respectively.</p>
        <p>Lastly, an ablation study was conducted to show the contribution of each training component in HomeSleepNet. Specifically, from the original SoundSleepNet model, we trained 2 additional variant models: one with added transfer learning only and another with added consistency training only. As a result, our final model HomeSleepNet was compared against its 3 variants: (1) SoundSleepNet, (2) SoundSleepNet with transfer learning, and (3) SoundSleepNet with consistency training. SoundSleepNet was only derived from supervised learning (the first training component) using the hospital PSG data set, without any additional techniques for training or input of home sound data [<xref ref-type="bibr" rid="ref20">20</xref>].</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>The use of the 3 data sets (hospital PSG data set, home smartphone data set, and home PSG data set) was approved by the Institutional Review Board of  Seoul National University Bundang Hospital (SNUBH; approval number B-2205-755-308). All participants signed the written consents before the data recording was performed. All the recorded data were anonymized for privacy and confidentiality protection of the participants.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Sleep Staging Performance</title>
        <p>HomeSleepNet showed a good performance for the 3-stage classification with an overall accuracy of 76.2%. Specifically, it correctly predicted 63.4% of wake, 83.6% of NREM sleep, and 64.9% of REM sleep (<xref rid="figure4" ref-type="fig">Figure 4</xref>). Other metrics also showed a reasonable performance for both macro <italic>F</italic><sub>1</sub>-score (0.714) and mean per-class sensitivity (0.706). Only Cohen κ was not as high, with a value of 0.557 (<xref ref-type="table" rid="table2">Table 2</xref>). For the 2-stage classification, all 4 metrics showed an even better performance. Accuracy of sleep-wake prediction was high, up to 88.5%. Both macro <italic>F</italic><sub>1</sub>-score and mean per-class sensitivity were around 0.8 and Cohen κ increased to 0.610. For the 4-stage classification, the performance was not as good, with an accuracy of 59.4%.</p>
        <p><xref rid="figure5" ref-type="fig">Figure 5</xref> shows the whole-night sleep-stage predictions from the baseline model SoundSleepNet and our proposed HomeSleepNet for 2 participants. The first participant was a 44-year-old male with BMI of 24.1 kg/m<sup>2</sup> and AHI of 47.5, and the second participant was a 65-year-old female with BMI of 23.6 kg/m<sup>2</sup> and AHI of 1.8. According to the analysis on different demographic groups (discussed later), sound-based sleep staging for the first participant is expected to be easier than that for the second participant. Indeed, SoundSleepNet performed reasonably well for the first participant, although it misclassified 2 REM blocks in the middle and in the end of the sleep. However, for the second participant, SoundSleepNet did not perform well, predicting most epochs as wake which were wrong. By contrast, HomeSleepNet was able to successfully predict most sleep stages and captured the sleep transitions for both participants.</p>
        <p>In <xref rid="figure6" ref-type="fig">Figure 6</xref>, we present PCA plots [<xref ref-type="bibr" rid="ref32">32</xref>] from the last hidden layer outputs of our proposed HomeSleepNet and the baseline SoundSleepNet models. We randomly selected 800 sleep epochs from each class for visualization (2400 sleep epochs in total), due to limited computing resources. The feature space is better organized with more clearly divided clusters in HomeSleepNet compared with SoundSleepNet. In the feature space of SoundSleepNet, data points from each class were widely scattered, especially those from the REM stage. This finding also supported the superior sleep staging ability of HomeSleepNet over SoundSleepNet using sounds recorded from home environments.</p>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Confusion matrices showing performance of HomeSleepNet on the home PSG data set: (A) 4-stage classification, (B) 3-stage classification, and (C) 2-stage classification. Light: N1+N2; Deep: N3; NREM: N1+N2+N3. NREM: nonrapid-eye movement; REM: rapid-eye movement; PSG: polysomnography.</p>
          </caption>
          <graphic xlink:href="jmir_v25i1e46216_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Sleep staging performance of HomeSleepNet on the Home PSG<sup>a</sup> data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="230"/>
            <col width="160"/>
            <col width="180"/>
            <col width="290"/>
            <col width="140"/>
            <thead>
              <tr valign="bottom">
                <td>Classification type</td>
                <td>Cohen κ</td>
                <td>Macro <italic>F</italic><sub>1</sub>-score</td>
                <td>Mean per-class sensitivity</td>
                <td>Accuracy, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>4 Stage<sup>b</sup></td>
                <td>0.416</td>
                <td>0.582</td>
                <td>0.610</td>
                <td>59.4</td>
              </tr>
              <tr valign="top">
                <td>3 Stage<sup>c</sup></td>
                <td>0.557</td>
                <td>0.714</td>
                <td>0.706</td>
                <td>76.2</td>
              </tr>
              <tr valign="top">
                <td>2 Stage<sup>d</sup></td>
                <td>0.610</td>
                <td>0.805</td>
                <td>0.789</td>
                <td>88.5</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>PSG: polysomnography.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>Wake, light (N1+N2), deep (N3), and rapid-eye movement sleep.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>Wake, rapid-eye movement, and nonrapid-eye movement (N1+N2+N3) sleep.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>Wake and sleep (N1+N2+N3+rapid-eye movement).</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Comparison of whole-night sleep stage predictions among manual annotations of PSG (top), predictions of HomeSleepNet (middle), and predictions of SoundSleepNet (bottom) for 2 participants: (A) Male, 44 years old, BMI 24.1 kg/m<sup>2</sup>, AHI 47.5; (B) Female, 65 years old, BMI 23.6 kg/m<sup>2</sup>, AHI 1.8. The highlighted red regions indicate different predictions compared with the sleep stages from PSG. NREM: nonrapid-eye movement; PSG: polysomnography; REM: rapid-eye movement.</p>
          </caption>
          <graphic xlink:href="jmir_v25i1e46216_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Feature space visualization using principal component analysis with 2 different models: (A) the baseline SoundSleepNet and (B) the proposed HomeSleepNet. NREM: nonrapid-eye movement; REM: rapid-eye movement.</p>
          </caption>
          <graphic xlink:href="jmir_v25i1e46216_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Sleep Metrics Comparison</title>
        <p>The sleep metrics calculated using the 3-stage predictions from HomeSleepNet were compared with those derived from manual annotations of PSGs (<xref ref-type="table" rid="table3">Table 3</xref>). For most sleep metrics, the mean predicted values were similar to the mean values from PSG, and the differences were relatively small. Bland-Altman plots also showed consistent agreement between the sleep metrics derived from HomeSleepNet and PSG (<xref rid="figure7" ref-type="fig">Figure 7</xref>). The line of equality in all graphs is located within the range of a 95% CI of the mean difference or close to the border, which suggests that there was no significant systematic difference between the 2 methods. Although HomeSleepNet presented lower averaged sleep onset latency compared with PSG, the gap was attributed to incorrect predictions for a few outliers with unusually long sleep onset latencies (eg, 5 hours). When excluding the outliers, the mean predicted sleep onset latency was similar to that of PSG. For more information, please refer to <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Comparison of sleep metrics between HomeSleepNet and a portable PSG<sup>a</sup> device.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="200"/>
            <col width="250"/>
            <col width="290"/>
            <col width="260"/>
            <thead>
              <tr valign="top">
                <td>Sleep metrics</td>
                <td>Portable PSG, mean (SD); 95 percentile confidence interval</td>
                <td>HomeSleepNet, mean (SD); 95 percentile confidence interval</td>
                <td>Difference, mean (SD); 95 percentile confidence interval</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Total sleep time (minutes)</td>
                <td>375.0 (75.0); 353.1 to 396.9</td>
                <td>387.0 (74.9); 365.1 to 408.9</td>
                <td>12.0 (44.1); –0.9 to 24.8</td>
              </tr>
              <tr valign="top">
                <td>Sleep onset latency (minutes)</td>
                <td>26.4 (53.9); 10.6 to 42.1</td>
                <td>12.6 (27.9); 4.4 to 20.7</td>
                <td>–13.8 (36.2); –24.4 to –3.2</td>
              </tr>
              <tr valign="top">
                <td>Sleep efficiency (%)</td>
                <td>81.1 (16.3); 76.3 to 85.8</td>
                <td>83.5 (15.6); 79.0 to 88.1</td>
                <td>2.4 (9.3); –0.3 to 5.2</td>
              </tr>
              <tr valign="top">
                <td>Wake after sleep onset (minutes)</td>
                <td>63.3 (56.9); 46.7 to 79.9</td>
                <td>65.1 (67.3); 45.5 to 84.8</td>
                <td>1.8 (53.7); –13.9 to 17.5</td>
              </tr>
              <tr valign="top">
                <td>REM<sup>b</sup> latency (minutes)</td>
                <td>80.1 (52.7); 64.7 to 95.5</td>
                <td>73.1 (61.6); 55.1 to 91.2</td>
                <td>–7.0 (81.1); –30.7 to 16.7</td>
              </tr>
              <tr valign="top">
                <td>REM (%)</td>
                <td>19.1 (7.3); 17.0 to 21.3</td>
                <td>19.8 (12.6); 16.1 to 23.5</td>
                <td>0.7 (10.3); –2.4 to 3.7</td>
              </tr>
              <tr valign="top">
                <td>NREM<sup>c</sup> (%)</td>
                <td>61.9 (11.6); 58.6 to 65.3</td>
                <td>63.7 (14.9); 59.4 to 68.1</td>
                <td>1.8 (13.2); –2.1 to 5.7</td>
              </tr>
              <tr valign="top">
                <td>Wake (%)</td>
                <td>18.9 (16.3); 14.2 to 23.7</td>
                <td>16.5 (15.6); 11.9 to 21.0</td>
                <td>–2.4 (9.3); –5.2 to 0.3</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>PSG: polysomnography.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>REM: rapid-eye movement.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>b</sup>NREM: nonrapid-eye movement.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>Bland-Altman plots of common sleep metrics: Total sleep time (TST), sleep onset latency (SOL), sleep efficiency (SE), wake after sleep onset (WASO), REM latency (RL), and each sleep stage portion in the 3-class setting. The x-axis represents the mean value of the 2 methods and the y-axis represents the difference values. The solid line indicates mean difference and the dash-single dotted line is the line of equality (y=0). The dashed lines indicate the 95% limit of agreement, and the shaded regions are the 95% CI of the mean difference. NREM: nonrapid-eye movement; PSG: polysomnography; REM: rapid-eye movement.</p>
          </caption>
          <graphic xlink:href="jmir_v25i1e46216_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Performance on Different Demographic Groups</title>
        <p>Among the 3 age groups, performance tended to slightly increase as age reduced (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The performance of HomeSleepNet was better for men than for women. The performance was similar between the high and low BMI groups. Regarding AHI, the performance was better in people with moderate-to-severe sleep apnea (AHI≥15) than in people with no or mild sleep apnea (AHI&#60;15). Overall, HomeSleepNet showed a robust performance in all groups, with all accuracies higher than 73%. More detailed results can be found in Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
      </sec>
      <sec>
        <title>Ablation Study</title>
        <p>The comparison of the sleep staging performance between HomeSleepNet and its 3 variants is presented in <xref ref-type="table" rid="table4">Table 4</xref>. As expected, the original SoundSleepNet model showed the worst performance in all evaluation metrics. Although adding only transfer learning to SoundSleepNet did not significantly improve the accuracy, transfer learning resulted in balancing the predictions between the classes, as the mean per-class sensitivity increased from 0.65 to almost 0.68. By contrast, adding consistency training to SoundSleepNet enhanced accuracy by 4.3%; besides, there were slight improvements in other metrics as well: 0.01 in mean per-class sensitivity, 0.038 in macro <italic>F</italic><sub>1</sub>-score, and 0.05 in Cohen κ. HomeSleepNet, using both transfer learning and consistency training, achieved the best performance with an increased accuracy of 7% from SoundSleepNet as well as considerable improvements in other metrics: around 0.1 in Cohen κ, 0.08 in macro <italic>F</italic><sub>1</sub>-score, and 0.056 in mean per-class sensitivity.</p>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Ablation study comparing HomeSleepNet with its multiple variants.<sup>a</sup></p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="370"/>
            <col width="140"/>
            <col width="160"/>
            <col width="220"/>
            <col width="110"/>
            <thead>
              <tr valign="bottom">
                <td>Model variants</td>
                <td>Cohen κ</td>
                <td>Macro <italic>F</italic><sub>1</sub>-score</td>
                <td>Mean per-class sensitivity</td>
                <td>Accuracy, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>SoundSleepNet</td>
                <td>0.454</td>
                <td>0.632</td>
                <td>0.650</td>
                <td>69.2</td>
              </tr>
              <tr valign="top">
                <td>SoundSleepNet + transfer learning only</td>
                <td>0.477</td>
                <td>0.648</td>
                <td>0.679</td>
                <td>69.3</td>
              </tr>
              <tr valign="top">
                <td>SoundSleepNet + consistency training only</td>
                <td>0.501</td>
                <td>0.670</td>
                <td>0.660</td>
                <td>73.5</td>
              </tr>
              <tr valign="top">
                <td>HomeSleepNet</td>
                <td>0.557</td>
                <td>0.714</td>
                <td>0.706</td>
                <td>76.2</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>All results were based on the 3-stage classification.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>Our finding shows that sound-based sleep staging can perform well not only in the hospital but also in individuals’ home environments. Our proposed deep learning model, HomeSleepNet, was designed specifically for home sleep monitoring by adopting transfer learning and consistency training. We collected numerous home sound data and selected a large variety of home noise data from an open database, and we utilized them together with a labeled hospital sound data set for transfer learning and consistency training, respectively. The significance of HomeSleepNet is that it enables sound-based sleep staging with a good performance in home environments where a lot of background noise exists.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>To the best of our knowledge, this is the first study to tackle the sound-based sleep staging problem in home environments. Early sound-based sleep staging studies were limited by the need for professional recording equipment [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref18">18</xref>] or short recording distances [<xref ref-type="bibr" rid="ref19">19</xref>]. SoundSleepNet was a breakthrough in sound-based sleep staging, using sounds recorded from a distance of 1 m with only smartphone microphones, a more practical approach that still yielded good performance [<xref ref-type="bibr" rid="ref20">20</xref>]. Although SoundSleepNet demonstrated the potential of using smartphone audio recordings for sleep staging, its real-world performance for home use remains unknown. The model was trained and tested on sounds recorded in a hospital environment and its ability to accurately classify sounds in a real-world, nonclinical setting is yet to be fully evaluated.</p>
        <p>Therefore, this study aimed to develop a specifically designed model for home use. HomeSleepNet achieved an overall accuracy of over 75% in differentiating between wake, REM, and non-REM sleep using home sound data. This level of accuracy is similar to previous methods that used hospital sound data, which typically had lower levels of background noise [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>]. For example, SoundSleepNet showed an accuracy of 79.8%, macro <italic>F</italic><sub>1</sub>-score of 0.749, and a mean per-class sensitivity of 0.757 for the 3-stage classification using hospital data [<xref ref-type="bibr" rid="ref20">20</xref>]. However, as can be seen in the ablation study, the same model failed to work well on home-based sounds, showing an accuracy of 69.2%, macro <italic>F</italic><sub>1</sub>-score of 0.632, and a mean per-class sensitivity of 0.650 for the 3-stage classification, showing a significant decrease in performance.</p>
        <p>Even for the gold-standard test, PSG, the interrater agreement of manual scoring between technologists is approximately 82%-83% for 5 sleep stages [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. Regarding other methods, the mean per-class sensitivity for 4 sleep stages was 0.480-0.632 among the commercial sleep trackers and 0.655 for SoundSleepNet [<xref ref-type="bibr" rid="ref20">20</xref>]. The mean per-class sensitivity of 0.610 should thus be considered acceptable, especially when using home sounds that are full of uncontrolled noise.</p>
      </sec>
      <sec>
        <title>Difference Between Hospital and Home Environments</title>
        <p>In the real world, home-based sound data from users will be recorded by their own smartphones. A huge variety of input from home-based sounds is expected, reflecting the diversity of smartphone models, home environments, and background noise. By contrast, the data obtained in hospitals were collected under controlled conditions (noise isolation and designated devices for audio recording). The decreased performance of SoundSleepNet on home data confirms that sounds recorded from hospitals and home should be considered as different data domains. This reinforces the theory that a good performance in a controlled environment may not warrant practical use in the real world [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. Although the performance of HomeSleepNet on home data might not be as reliable as its performance on hospital data, considering the challenges of using home sound data for sleep staging (ie, more noisy data and lack of ground truth), the robust performance of HomeSleepNet using home-based sound data is of great importance.</p>
      </sec>
      <sec>
        <title>Transfer Learning and Consistency Training</title>
        <p>It is important to note that HomeSleepNet was trained without home PSGs. Instead, 2 specific techniques, transfer learning and consistency training, were added to solve the problem of the lack of ground truth for home data. When comparing transfer learning and consistency training, with regard to improvement of performance on home data, adding consistency training to SoundSleepNet showed a greater improvement than adding transfer learning. We have two hypothesis for this phenomenon. First, we allowed participants to use their own smartphones to record home-based sounds to better reflect real-world data. In this study, we broadly grouped all home-based sound data using different devices into 1 domain for these to be distinguishable from hospital data. However, the home smartphone data set itself is actually heterogenous; technically, each type of smartphone used for audio recording can be considered an independent domain because of its own configuration. Grouping heterogeneous data into 1 target domain might impair the domain adaptation training. Second, adapting the model from controlled data (collected by 1 designated microphone in the hospital) to uncontrolled home data (collected through various types of smartphones) might have reduced the performance of the transfer learning. However, adding both transfer learning and consistency training showed a better performance than adding only consistency training, implying that transfer learning also had positive effects on performance. Therefore, it can be concluded that both transfer learning and consistency training contribute to the enhanced performance of sleep staging in HomeSleepNet, with each method employing unique mechanisms to assisting the network training.</p>
        <p>In general, transfer learning and consistency training can be applied to any deep learning model for better performance across various target domains. When either sleep sound data or background noise data from a target domain are sufficiently available, the proposed training methods can be used to benefit a sound-based sleep staging model. Examples of the target domains are sounds recorded in other environments or sounds recorded by other devices (eg, other types of smartphones, smart speakers, or smart televisions).</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>There are several limitations to our research. First, the performance of the proposed model was based on data that were collected with the condition that people slept alone in the room. The presence of multiple people or accompanying pets may cause overlapping or interruption of the sleep sounds and reduce the performance for sleep staging. Second, the majority of home smartphone data were collected from young healthy adults. Third, the test set might not be big enough to test performance under all diverse cases. Fourth, the HomeSleepNet model has difficulty differentiating light and deep sleep, a limitation shared by sound-based sleep staging methods [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>].</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>To the best of our knowledge, this is the first sound-based sleep staging study to utilize sounds recorded in individual home environments. The performance was validated by comparing with PSGs recorded concurrently at home. By adopting the 2 techniques for training — one that transfers learning from a hospital data to home audios, and another that ensures consistency in the presence of home noise — the proposed model is able to accurately predict sleep stages using home sounds. Our proposed model expands the use of sound-based sleep staging from in-laboratory sounds to home sounds full of uncontrolled noise. Daily sleep monitoring using the simple audio recording function of smartphones is feasible. An easy and convenient noncontact sleep tracker may encourage individuals to track their own sleep, which may further modify their awareness of and behaviors for sleep.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Additional information on the 3 sleep data sets used in this study, the HomeSleepNet training algorithm, and further discussion of results.</p>
        <media xlink:href="jmir_v25i1e46216_app1.pdf" xlink:title="PDF File  (Adobe PDF File), 341 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AHI</term>
          <def>
            <p>apnea-hypopnea index</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">NREM</term>
          <def>
            <p>nonrapid-eye movement</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">PSG</term>
          <def>
            <p>polysomnography</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">REM</term>
          <def>
            <p>rapid-eye movement</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">UDA</term>
          <def>
            <p>unsupervised domain adaptation</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>The data used in this study are not publicly available. Because of privacy and security concerns, the audio data are not easily redistributed to researchers other than those engaged in the institutional review board–approved research collaborations with the named medical centers.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>JKH, JH, JJ, DL, DK, and I-YY conceived the study and participated in its design; DL provided the resources and facilitated the project administration; ML and I-YY participated in clinical data acquisition; HHT and JH participated in data acquisition and preprocessing; HHT developed and trained the artificial intelligence model; HHT, HJ, JK, JH, and DK designed the artificial intelligence model and drafted the manuscript about the model; JKH, HJ, JJ, and DK assisted with the statistical analysis and interpretation of the data; HHT and JKH wrote the original manuscript; HJ, JJ, DK, J-WK, CAK, and I-YY revised the manuscript; DK and I-YY supervised the study and contributed equally as cocorresponding authors. All authors read and approved the final manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>CAK is a consultant of Asleep Inc.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kelly</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Strecker</surname>
              <given-names>RE</given-names>
            </name>
            <name name-style="western">
              <surname>Bianchi</surname>
              <given-names>MT</given-names>
            </name>
          </person-group>
          <article-title>Recent developments in home sleep-monitoring devices</article-title>
          <source>ISRN Neurol</source>
          <year>2012</year>
          <volume>2012</volume>
          <fpage>768794</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.5402/2012/768794"/>
          </comment>
          <pub-id pub-id-type="doi">10.5402/2012/768794</pub-id>
          <pub-id pub-id-type="medline">23097718</pub-id>
          <pub-id pub-id-type="pmcid">PMC3477711</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>de Zambotti</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cellini</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Goldstone</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Colrain</surname>
              <given-names>IM</given-names>
            </name>
            <name name-style="western">
              <surname>Baker</surname>
              <given-names>FC</given-names>
            </name>
          </person-group>
          <article-title>Wearable Sleep Technology in Clinical and Research Settings</article-title>
          <source>Med Sci Sports Exerc</source>
          <year>2019</year>
          <month>07</month>
          <volume>51</volume>
          <issue>7</issue>
          <fpage>1538</fpage>
          <lpage>1557</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/30789439"/>
          </comment>
          <pub-id pub-id-type="doi">10.1249/MSS.0000000000001947</pub-id>
          <pub-id pub-id-type="medline">30789439</pub-id>
          <pub-id pub-id-type="pmcid">PMC6579636</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Newell</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mairesse</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Verbanck</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Neu</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Is a one-night stay in the lab really enough to conclude? First-night effect and night-to-night variability in polysomnographic recordings among different clinical population samples</article-title>
          <source>Psychiatry Res</source>
          <year>2012</year>
          <month>12</month>
          <day>30</day>
          <volume>200</volume>
          <issue>2-3</issue>
          <fpage>795</fpage>
          <lpage>801</lpage>
          <pub-id pub-id-type="doi">10.1016/j.psychres.2012.07.045</pub-id>
          <pub-id pub-id-type="medline">22901399</pub-id>
          <pub-id pub-id-type="pii">S0165-1781(12)00397-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bevan</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Grantham-Hill</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bowen</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Clayton</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Grice</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Venditti</surname>
              <given-names>HC</given-names>
            </name>
            <name name-style="western">
              <surname>Stickland</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Hill</surname>
              <given-names>CM</given-names>
            </name>
          </person-group>
          <article-title>Sleep quality and noise: comparisons between hospital and home settings</article-title>
          <source>Arch Dis Child</source>
          <year>2019</year>
          <month>02</month>
          <volume>104</volume>
          <issue>2</issue>
          <fpage>147</fpage>
          <lpage>151</lpage>
          <pub-id pub-id-type="doi">10.1136/archdischild-2018-315168</pub-id>
          <pub-id pub-id-type="medline">30018067</pub-id>
          <pub-id pub-id-type="pii">archdischild-2018-315168</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Tsai</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hsu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Majumdar</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kuan</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Hsu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Comparison of Hospital-Based and Home-Based Obstructive Sleep Apnoea Severity Measurements with a Single-Lead Electrocardiogram Patch</article-title>
          <source>Sensors (Basel)</source>
          <year>2021</year>
          <month>12</month>
          <day>03</day>
          <volume>21</volume>
          <issue>23</issue>
          <fpage>8097</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=s21238097"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/s21238097</pub-id>
          <pub-id pub-id-type="medline">34884101</pub-id>
          <pub-id pub-id-type="pii">s21238097</pub-id>
          <pub-id pub-id-type="pmcid">PMC8659975</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yo</surname>
              <given-names>SW</given-names>
            </name>
            <name name-style="western">
              <surname>Joosten</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Wimaleswaran</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Mansfield</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Thomson</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Landry</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Edwards</surname>
              <given-names>BA</given-names>
            </name>
            <name name-style="western">
              <surname>Hamilton</surname>
              <given-names>GS</given-names>
            </name>
          </person-group>
          <article-title>Body position during laboratory and home polysomnography compared to habitual sleeping position at home</article-title>
          <source>J Clin Sleep Med</source>
          <year>2022</year>
          <month>09</month>
          <day>01</day>
          <volume>18</volume>
          <issue>9</issue>
          <fpage>2103</fpage>
          <lpage>2111</lpage>
          <pub-id pub-id-type="doi">10.5664/jcsm.9990</pub-id>
          <pub-id pub-id-type="medline">35459447</pub-id>
          <pub-id pub-id-type="pii">jcsm.9990</pub-id>
          <pub-id pub-id-type="pmcid">PMC9435326</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Siyahjani</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Garcia Molina</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Barr</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mushtaq</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Performance Evaluation of a Smart Bed Technology against Polysomnography</article-title>
          <source>Sensors (Basel)</source>
          <year>2022</year>
          <month>03</month>
          <day>29</day>
          <volume>22</volume>
          <issue>7</issue>
          <fpage>2605</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=s22072605"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/s22072605</pub-id>
          <pub-id pub-id-type="medline">35408220</pub-id>
          <pub-id pub-id-type="pii">s22072605</pub-id>
          <pub-id pub-id-type="pmcid">PMC9002520</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hsieh</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kuo</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Home-Use and Real-Time Sleep-Staging System Based on Eye Masks and Mobile Devices with a Deep Learning Model</article-title>
          <source>J Med Biol Eng</source>
          <year>2021</year>
          <volume>41</volume>
          <issue>5</issue>
          <fpage>659</fpage>
          <lpage>668</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34512223"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s40846-021-00649-5</pub-id>
          <pub-id pub-id-type="medline">34512223</pub-id>
          <pub-id pub-id-type="pii">649</pub-id>
          <pub-id pub-id-type="pmcid">PMC8418457</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Radha</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Fonseca</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Moreau</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ross</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cerny</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Anderer</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Long</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Aarts</surname>
              <given-names>RM</given-names>
            </name>
          </person-group>
          <article-title>A deep transfer learning approach for wearable sleep stage classification with photoplethysmography</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <month>09</month>
          <day>15</day>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>135</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-021-00510-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-021-00510-8</pub-id>
          <pub-id pub-id-type="medline">34526643</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-021-00510-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC8443610</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>de Zambotti</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rosas</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Colrain</surname>
              <given-names>IM</given-names>
            </name>
            <name name-style="western">
              <surname>Baker</surname>
              <given-names>FC</given-names>
            </name>
          </person-group>
          <article-title>The Sleep of the Ring: Comparison of the ŌURA Sleep Tracker Against Polysomnography</article-title>
          <source>Behav Sleep Med</source>
          <year>2019</year>
          <volume>17</volume>
          <issue>2</issue>
          <fpage>124</fpage>
          <lpage>136</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28323455"/>
          </comment>
          <pub-id pub-id-type="doi">10.1080/15402002.2017.1300587</pub-id>
          <pub-id pub-id-type="medline">28323455</pub-id>
          <pub-id pub-id-type="pmcid">PMC6095823</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ganglberger</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Panneerselvam</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Leone</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Quadri</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Goparaju</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Tesh</surname>
              <given-names>RA</given-names>
            </name>
            <name name-style="western">
              <surname>Akeju</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Thomas</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Westover</surname>
              <given-names>MB</given-names>
            </name>
          </person-group>
          <article-title>Sleep staging from electrocardiography and respiration with deep learning</article-title>
          <source>Sleep</source>
          <year>2020</year>
          <month>07</month>
          <day>13</day>
          <volume>43</volume>
          <issue>7</issue>
          <fpage>zsz306</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31863111"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/sleep/zsz306</pub-id>
          <pub-id pub-id-type="medline">31863111</pub-id>
          <pub-id pub-id-type="pii">5682785</pub-id>
          <pub-id pub-id-type="pmcid">PMC7355395</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Haghayegh</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Khoshnevis</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Smolensky</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Diller</surname>
              <given-names>KR</given-names>
            </name>
            <name name-style="western">
              <surname>Castriotta</surname>
              <given-names>RJ</given-names>
            </name>
          </person-group>
          <article-title>Accuracy of Wristband Fitbit Models in Assessing Sleep: Systematic Review and Meta-Analysis</article-title>
          <source>J Med Internet Res</source>
          <year>2019</year>
          <month>11</month>
          <day>28</day>
          <volume>21</volume>
          <issue>11</issue>
          <fpage>e16273</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2019/11/e16273/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/16273</pub-id>
          <pub-id pub-id-type="medline">31778122</pub-id>
          <pub-id pub-id-type="pii">v21i11e16273</pub-id>
          <pub-id pub-id-type="pmcid">PMC6908975</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mouritzen</surname>
              <given-names>NJ</given-names>
            </name>
            <name name-style="western">
              <surname>Larsen</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Lauritzen</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Kjær</surname>
              <given-names>Troels W</given-names>
            </name>
          </person-group>
          <article-title>Assessing the performance of a commercial multisensory sleep tracker</article-title>
          <source>PLoS One</source>
          <year>2020</year>
          <volume>15</volume>
          <issue>12</issue>
          <fpage>e0243214</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dx.plos.org/10.1371/journal.pone.0243214"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pone.0243214</pub-id>
          <pub-id pub-id-type="medline">33306678</pub-id>
          <pub-id pub-id-type="pii">PONE-D-20-21216</pub-id>
          <pub-id pub-id-type="pmcid">PMC7732119</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kalkbrenner</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Brucher</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Kesztyüs</surname>
              <given-names>Tibor</given-names>
            </name>
            <name name-style="western">
              <surname>Eichenlaub</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rottbauer</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Scharnbeck</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Automated sleep stage classification based on tracheal body sound and actigraphy</article-title>
          <source>Ger Med Sci</source>
          <year>2019</year>
          <volume>17</volume>
          <fpage>Doc02</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/30996721"/>
          </comment>
          <pub-id pub-id-type="doi">10.3205/000268</pub-id>
          <pub-id pub-id-type="medline">30996721</pub-id>
          <pub-id pub-id-type="pii">000268</pub-id>
          <pub-id pub-id-type="pmcid">PMC6449867</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nakano</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Furukawa</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Tanigawa</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Tracheal Sound Analysis Using a Deep Neural Network to Detect Sleep Apnea</article-title>
          <source>J Clin Sleep Med</source>
          <year>2019</year>
          <month>08</month>
          <day>15</day>
          <volume>15</volume>
          <issue>8</issue>
          <fpage>1125</fpage>
          <lpage>1133</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31482834"/>
          </comment>
          <pub-id pub-id-type="doi">10.5664/jcsm.7804</pub-id>
          <pub-id pub-id-type="medline">31482834</pub-id>
          <pub-id pub-id-type="pmcid">PMC6707047</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Montazeri Ghahjaverestan</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Akbarian</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hafezi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Saha</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Gavrilovic</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Taati</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yadollahi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Sleep/Wakefulness Detection Using Tracheal Sounds and Movements</article-title>
          <source>Nat Sci Sleep</source>
          <year>2020</year>
          <volume>12</volume>
          <fpage>1009</fpage>
          <lpage>1021</lpage>
          <pub-id pub-id-type="doi">10.2147/NSS.S276107</pub-id>
          <pub-id pub-id-type="medline">33235534</pub-id>
          <pub-id pub-id-type="pii">276107</pub-id>
          <pub-id pub-id-type="pmcid">PMC7680175</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dafna</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Halevi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ben Or</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Tarasiuk</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zigel</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Estimation of macro sleep stages from whole night audio analysis</article-title>
          <source>Annu Int Conf IEEE Eng Med Biol Soc</source>
          <year>2016</year>
          <month>08</month>
          <volume>2016</volume>
          <fpage>2847</fpage>
          <lpage>2850</lpage>
          <pub-id pub-id-type="doi">10.1109/EMBC.2016.7591323</pub-id>
          <pub-id pub-id-type="medline">28268910</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dafna</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Tarasiuk</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zigel</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Sleep staging using nocturnal sound analysis</article-title>
          <source>Sci Rep</source>
          <year>2018</year>
          <month>09</month>
          <day>07</day>
          <volume>8</volume>
          <issue>1</issue>
          <fpage>13474</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-018-31748-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-018-31748-0</pub-id>
          <pub-id pub-id-type="medline">30194402</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-018-31748-0</pub-id>
          <pub-id pub-id-type="pmcid">PMC6128888</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xue</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Hong</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>DD</given-names>
            </name>
          </person-group>
          <article-title>Non-Contact Sleep Stage Detection Using Canonical Correlation Analysis of Respiratory Sound</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2020</year>
          <month>02</month>
          <volume>24</volume>
          <issue>2</issue>
          <fpage>614</fpage>
          <lpage>625</lpage>
          <pub-id pub-id-type="doi">10.1109/JBHI.2019.2910566</pub-id>
          <pub-id pub-id-type="medline">30990201</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hong</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tran</surname>
              <given-names>HH</given-names>
            </name>
            <name name-style="western">
              <surname>Jung</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hong</surname>
              <given-names>JK</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>End-to-End Sleep Staging Using Nocturnal Sounds from Microphone Chips for Mobile Devices</article-title>
          <source>Nat Sci Sleep</source>
          <year>2022</year>
          <volume>14</volume>
          <fpage>1187</fpage>
          <lpage>1201</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/35783665"/>
          </comment>
          <pub-id pub-id-type="doi">10.2147/NSS.S361270</pub-id>
          <pub-id pub-id-type="medline">35783665</pub-id>
          <pub-id pub-id-type="pii">361270</pub-id>
          <pub-id pub-id-type="pmcid">PMC9241996</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kamath</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Whitaker</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Transfer Learning: Domain Adaptation</article-title>
          <source>Deep Learning for NLP and Speech Recognition</source>
          <year>2019</year>
          <publisher-loc>Berlin, Germany</publisher-loc>
          <publisher-name>Springer International Publishing</publisher-name>
          <fpage>495</fpage>
          <lpage>535</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ganin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ustinova</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Ajakan</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Germain</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Larochelle</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Laviolette</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Marchand</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lempitsky</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Domain-adversarial training of neural networks</article-title>
          <source>Journal of Machine Learning Research</source>
          <year>2016</year>
          <volume>17</volume>
          <issue>59</issue>
          <fpage>2030</fpage>
          <lpage>2096</lpage>
          <pub-id pub-id-type="doi">10.1007/978-3-319-58347-1_10</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Saito</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Watanabe</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ushiku</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Harada</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Maximum classifier discrepancy for unsupervised domain adaptation</article-title>
          <year>2018</year>
          <conf-name>Conference on Computer Vision and Pattern Recognition</conf-name>
          <conf-date>June 18-22, 2018</conf-date>
          <conf-loc>Salt Lake City, UT</conf-loc>
          <fpage>3723</fpage>
          <lpage>3732</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2018.00392</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shu</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bui</surname>
              <given-names>HH</given-names>
            </name>
            <name name-style="western">
              <surname>Narui</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ermon</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A DIRT-T Approach to Unsupervised Domain Adaptation</article-title>
          <year>2018</year>
          <conf-name>International Conference on Learning Representations</conf-name>
          <conf-date>April 30-May 3, 2018</conf-date>
          <conf-loc>Vancouver, BC, Canada</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://openreview.net/forum?id=H1q-TM-AW"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Hovy</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Luong</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Unsupervised data augmentation for consistency training</article-title>
          <year>2020</year>
          <conf-name>Conference on Neural Information Processing Systems</conf-name>
          <conf-date>Dec 6-12, 2020</conf-date>
          <conf-loc>Virtual-only Conference</conf-loc>
          <fpage>6256</fpage>
          <lpage>6268</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hendrycks</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Mu</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Cubuk</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Zoph</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Gilmer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lakshminarayanan</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>AugMix: A Simple Data Processing Method to Improve Robustness and Uncertainty</article-title>
          <year>2020</year>
          <conf-name>International Conference on Learning Representations</conf-name>
          <conf-date>April 26-30, 2020</conf-date>
          <conf-loc>Virtual-only Conference</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://openreview.net/forum?id=S1gmrxHFvB"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ferber</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Millman</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Coppola</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Fleetham</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Murray</surname>
              <given-names>CF</given-names>
            </name>
            <name name-style="western">
              <surname>Iber</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>McCall</surname>
              <given-names>WV</given-names>
            </name>
            <name name-style="western">
              <surname>Nino-Murcia</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Pressman</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sanders</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Portable recording in the assessment of obstructive sleep apnea. ASDA standards of practice</article-title>
          <source>Sleep</source>
          <year>1994</year>
          <month>06</month>
          <volume>17</volume>
          <issue>4</issue>
          <fpage>378</fpage>
          <lpage>92</lpage>
          <pub-id pub-id-type="doi">10.1093/sleep/17.4.378</pub-id>
          <pub-id pub-id-type="medline">7973323</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Iber</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ancoli-Israel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chesson</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Quan</surname>
              <given-names>SF</given-names>
            </name>
          </person-group>
          <source>The AASM Manual for the Scoring of Sleep and Associated Events: Rules, Terminology and Technical Specifications (Vol 1)</source>
          <year>2007</year>
          <publisher-loc>Westchester, IL</publisher-loc>
          <publisher-name>American Academy of Sleep Medicine</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kotsiantis</surname>
              <given-names>SB</given-names>
            </name>
          </person-group>
          <article-title>Supervised Machine Learning: A Review of Classification Techniques</article-title>
          <source>Emerging Artificial Intelligence Applications in Computer Engineering</source>
          <year>2007</year>
          <publisher-loc>Amsterdam, The Netherlands</publisher-loc>
          <publisher-name>IOS Press</publisher-name>
          <fpage>3</fpage>
          <lpage>24</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cunningham</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Cord</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Delany</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Supervised Learning</article-title>
          <source>Machine Learning Techniques for Multimedia</source>
          <year>2008</year>
          <publisher-loc>Berlin, Germany</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>21</fpage>
          <lpage>49</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Font</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Roma</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Serra</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Freesound technical demo</article-title>
          <year>2013</year>
          <conf-name>21st ACM International Conference on Multimedia</conf-name>
          <conf-date>October 21-25, 2013</conf-date>
          <conf-loc>Barcelona, Spain</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>Association for Computing Machinery</publisher-name>
          <fpage>411</fpage>
          <lpage>412</lpage>
          <pub-id pub-id-type="doi">10.1145/2502081.2502245</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wold</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Esbensen</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Geladi</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Principal component analysis</article-title>
          <source>Chemometrics and Intelligent Laboratory Systems</source>
          <year>1987</year>
          <month>8</month>
          <volume>2</volume>
          <issue>1-3</issue>
          <fpage>37</fpage>
          <lpage>52</lpage>
          <pub-id pub-id-type="doi">10.1016/0169-7439(87)80084-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rosenberg</surname>
              <given-names>RS</given-names>
            </name>
            <name name-style="western">
              <surname>Van Hout</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>The American Academy of Sleep Medicine inter-scorer reliability program: sleep stage scoring</article-title>
          <source>J Clin Sleep Med</source>
          <year>2013</year>
          <month>01</month>
          <day>15</day>
          <volume>9</volume>
          <issue>1</issue>
          <fpage>81</fpage>
          <lpage>7</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/23319910"/>
          </comment>
          <pub-id pub-id-type="doi">10.5664/jcsm.2350</pub-id>
          <pub-id pub-id-type="medline">23319910</pub-id>
          <pub-id pub-id-type="pmcid">PMC3525994</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Danker-Hopfe</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Anderer</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zeitlhofer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Boeck</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dorn</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Gruber</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Heller</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Loretz</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Moser</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Parapatics</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Saletu</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Schmidt</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dorffner</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <source>J Sleep Res</source>
          <year>2009</year>
          <month>03</month>
          <volume>18</volume>
          <issue>1</issue>
          <fpage>74</fpage>
          <lpage>84</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://onlinelibrary.wiley.com/doi/10.1111/j.1365-2869.2008.00700.x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/j.1365-2869.2008.00700.x</pub-id>
          <pub-id pub-id-type="medline">19250176</pub-id>
          <pub-id pub-id-type="pii">JSR700</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
