<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v26i1e58572</article-id>
      <article-id pub-id-type="pmid">39324329</article-id>
      <article-id pub-id-type="doi">10.2196/58572</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Automated Speech Analysis for Risk Detection of Depression, Anxiety, Insomnia, and Fatigue: Algorithm Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Coristine</surname>
            <given-names>Andrew</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Yang</surname>
            <given-names>Lulu</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Martin</surname>
            <given-names>Vincent</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Riad</surname>
            <given-names>Rachid</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Callyope</institution>
            <addr-line>5 Parvis Alan Turing</addr-line>
            <addr-line>Paris, 75013</addr-line>
            <country>France</country>
            <phone>33 666522141</phone>
            <email>rachid@callyope.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-7753-1219</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Denais</surname>
            <given-names>Martin</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0000-4297-5517</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>de Gennes</surname>
            <given-names>Marc</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1631-3231</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Lesage</surname>
            <given-names>Adrien</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0001-6886-0475</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Oustric</surname>
            <given-names>Vincent</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-6985-7631</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Cao</surname>
            <given-names>Xuan Nga</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0003-2162-6328</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Mouchabac</surname>
            <given-names>Stéphane</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4226-7956</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author">
          <name name-style="western">
            <surname>Bourla</surname>
            <given-names>Alexis</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1438-5733</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Callyope</institution>
        <addr-line>Paris</addr-line>
        <country>France</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Psychiatry, Saint-Antoine Hospital</institution>
        <institution>Sorbonne University</institution>
        <institution>Assistance publique - Hôpitaux de Paris</institution>
        <addr-line>Paris</addr-line>
        <country>France</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Infrastructure for Clinical Research in Neurosciences</institution>
        <institution>Paris Brain Institute</institution>
        <addr-line>Paris</addr-line>
        <country>France</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Medical Strategy and Innovation Department</institution>
        <institution>Clariane</institution>
        <addr-line>Paris</addr-line>
        <country>France</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>NeuroStim Psychiatry Practice</institution>
        <addr-line>Paris</addr-line>
        <country>France</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Rachid Riad <email>rachid@callyope.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>31</day>
        <month>10</month>
        <year>2024</year>
      </pub-date>
      <volume>26</volume>
      <elocation-id>e58572</elocation-id>
      <history>
        <date date-type="received">
          <day>20</day>
          <month>3</month>
          <year>2024</year>
        </date>
        <date date-type="rev-request">
          <day>8</day>
          <month>4</month>
          <year>2024</year>
        </date>
        <date date-type="rev-recd">
          <day>7</day>
          <month>9</month>
          <year>2024</year>
        </date>
        <date date-type="accepted">
          <day>25</day>
          <month>9</month>
          <year>2024</year>
        </date>
      </history>
      <copyright-statement>©Rachid Riad, Martin Denais, Marc de Gennes, Adrien Lesage, Vincent Oustric, Xuan Nga Cao, Stéphane Mouchabac, Alexis Bourla. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 31.10.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2024/1/e58572" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>While speech analysis holds promise for mental health assessment, research often focuses on single symptoms, despite symptom co-occurrences and interactions. In addition, predictive models in mental health do not properly assess the limitations of speech-based systems, such as uncertainty, or fairness for a safe clinical deployment.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>We investigated the predictive potential of mobile-collected speech data for detecting and estimating depression, anxiety, fatigue, and insomnia, focusing on other factors than mere accuracy, in the general population.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We included 865 healthy adults and recorded their answers regarding their perceived mental and sleep states. We asked how they felt and if they had slept well lately. Clinically validated questionnaires measuring depression, anxiety, insomnia, and fatigue severity were also used. We developed a novel speech and machine learning pipeline involving voice activity detection, feature extraction, and model training. We automatically modeled speech with pretrained deep learning models that were pretrained on a large, open, and free database, and we selected the best one on the validation set. Based on the best speech modeling approach, clinical threshold detection, individual score prediction, model uncertainty estimation, and performance fairness across demographics (age, sex, and education) were evaluated. We used a train-validation-test split for all evaluations: to develop our models, select the best ones, and assess the generalizability of held-out data.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The best model was Whisper M with a max pooling and oversampling method. Our methods achieved good detection performance for all symptoms, depression (Patient Health Questionnaire-9: area under the curve [AUC]=0.76; <italic>F</italic><sub>1</sub>-score=0.49 and Beck Depression Inventory: AUC=0.78; <italic>F</italic><sub>1</sub>-score=0.65), anxiety (Generalized Anxiety Disorder 7-item scale: AUC=0.77; <italic>F</italic><sub>1</sub>-score=0.50), insomnia (Athens Insomnia Scale: AUC=0.73; <italic>F</italic><sub>1</sub>-score=0.62), and fatigue (Multidimensional Fatigue Inventory total score: AUC=0.68; <italic>F</italic><sub>1</sub>-score=0.88). The system performed well when it needed to abstain from making predictions, as demonstrated by low abstention rates in depression detection with the Beck Depression Inventory and fatigue, with risk-coverage AUCs below 0.4. Individual symptom scores were accurately predicted (correlations were all significant with Pearson strengths between 0.31 and 0.49). Fairness analysis revealed that models were consistent for sex (average disparity ratio [DR] 0.86, SD 0.13), to a lesser extent for education level (average DR 0.47, SD 0.30), and worse for age groups (average DR 0.33, SD 0.30).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This study demonstrates the potential of speech-based systems for multifaceted mental health assessment in the general population, not only for detecting clinical thresholds but also for estimating their severity. Addressing fairness and incorporating uncertainty estimation with selective classification are key contributions that can enhance the clinical utility and responsible implementation of such systems.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>speech analysis</kwd>
        <kwd>voice detection</kwd>
        <kwd>voice analysis</kwd>
        <kwd>speech biomarkers</kwd>
        <kwd>speech-based systems</kwd>
        <kwd>computer-aided diagnosis</kwd>
        <kwd>mental health symptom detection</kwd>
        <kwd>machine learning</kwd>
        <kwd>mental health</kwd>
        <kwd>fatigue</kwd>
        <kwd>anxiety</kwd>
        <kwd>depression</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Depression and anxiety disorders are recognized as the leading causes of disease burden [<xref ref-type="bibr" rid="ref1">1</xref>], and their prevalences are high during the entire life span across the sexes and all around the globe [<xref ref-type="bibr" rid="ref2">2</xref>]. This burden was aggravated by the COVID-19 pandemic [<xref ref-type="bibr" rid="ref3">3</xref>]. In these disorders, early identification and evaluation of the severity of the symptoms are of prime importance since the incidence of suicide is associated with a diagnosis of depression more than 50% of the time [<xref ref-type="bibr" rid="ref4">4</xref>]. Besides, measurement-based care, via the use of clinically valid scales, improves the follow-up and treatment of affected individuals with mental health disorders [<xref ref-type="bibr" rid="ref5">5</xref>]. Timely interventions lead to better outcomes in mental health. This proactive approach can ensure early access to treatment and prevent significant complications. Yet, measuring mental health remains a challenge, since manifestations of depression and anxiety are heterogeneous [<xref ref-type="bibr" rid="ref6">6</xref>] and co-occur with insomnia [<xref ref-type="bibr" rid="ref7">7</xref>] and fatigue [<xref ref-type="bibr" rid="ref8">8</xref>]. The exhaustive and objective assessment of these different mental health dimensions through validated assessment scales is long and fastidious for clinical staff and is particularly not adapted to primary care, which is at the forefront of handling mental health disorders [<xref ref-type="bibr" rid="ref9">9</xref>]. The development of objective biomarkers, which are easy to collect without the synchronization of clinicians and patients, has the potential to overcome these limitations. These quantifiable measures, encompassing biological, genetic, or behavioral assessments, could revolutionize early detection, enabling timely and targeted interventions that ultimately improve the patient’s outcomes and well-being. This is particularly significant for screening the general population across diverse mental health dimensions. Indeed, screening the first signs of mental health problems or symptoms could help to avoid escalation of symptoms, as different dimensions interact in time [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref10">10</xref>].</p>
      <p>The study of speech biomarkers in mental health holds great potential, offering a noninvasive and easily accessible avenue to capture significant motor, cognitive, and behavioral changes due to mental health disorders such as depression [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref14">14</xref>]. Clinical evidence and research studies have increasingly linked specific automated extracted speech features, such as prosody, articulation, and fluency, with various mental health conditions, such as depression [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref15">15</xref>], anxiety [<xref ref-type="bibr" rid="ref16">16</xref>], suicide-risk assessment [<xref ref-type="bibr" rid="ref17">17</xref>], fatigue [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>], or sleep deprivation [<xref ref-type="bibr" rid="ref20">20</xref>]. The complexity of human speech extends beyond the intricate motor coordination involved. The speech production system within the brain relies on the synchronization of diverse cognitive, social, and motor processes [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. This intricate interplay involves hundreds of muscles across the respiratory, phonatory, and supralaryngeal systems, working in concert with critical cognitive skills like attention, memory, and planning. Additionally, social skills such as theory of mind and emotional processing play a vital role. Importantly, disruptions in any of the aforementioned motor, cognitive, or social skills, as well as mental health states, can introduce perturbations in the resulting speech signal. Besides, beyond research evidence, clinical practitioners also use voice unconsciously when evaluating individuals, and these subjective evaluations could be complemented and refined with objective measures from automatic speech analysis.</p>
      <p>Speech biomarkers emerge also as a promising avenue for mental health assessment due to their unique characteristics: they are noninvasive, cost-effective, and convenient tools. Recent hardware and software advancements have significantly simplified and reduced the cost of acquiring acoustic data, making it a more accessible option compared to traditional biological, imaging, or cognitive markers. In addition, speech data collection requires minimal effort from both patients and clinicians and can even be conducted remotely, further enhancing its feasibility in various settings.</p>
      <p>However, despite its promises, the study of speech biomarkers remains largely fragmented, in laboratory settings, or not evaluated for deployment into clinical practice. This gap calls for more evidence to be integrated into clinical practice [<xref ref-type="bibr" rid="ref23">23</xref>]. Research on speech in mental health in the general population often focuses on 1 isolated mental health dimension, even though there are proofs supporting the existence of networks of symptoms and syndromes in mental health that influence each other [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. In addition, previous speech studies were limited to specific populations such as students [<xref ref-type="bibr" rid="ref26">26</xref>] or older people [<xref ref-type="bibr" rid="ref27">27</xref>]. However, before machine learning (ML) models can be used in clinical settings to make predictions for individuals, they must be “fair”—providing equally accurate predictions across all demographic groups [<xref ref-type="bibr" rid="ref28">28</xref>]. Finally, speech-based systems should not be tested only for the simple classification of binary labels (eg, depressed or not depressed) but rather for the estimation of the severity of symptoms [<xref ref-type="bibr" rid="ref29">29</xref>], and their ability to refrain from giving an output when uncertainty is too high, therefore deferring decisions to the health staff in practice [<xref ref-type="bibr" rid="ref30">30</xref>].</p>
      <p>In this study, the main objective was to assess the predictive potential of speech models in detecting and estimating the severity of depression, anxiety, fatigue, and insomnia within the general population using mobile-collected speech and mental health data. Besides, to prove that these models could be effectively implemented in diverse real-world settings, they are assessed for their fairness and uncertainty capabilities.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Participants</title>
        <p>We recruited French healthy adult participants without any known severe psychiatric or neurological disorder (self-declaration) or speech pathologies such as stuttering or clutter.</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>All participants signed an informed consent form to participate in the study, in line with the Declaration of Helsinki, current Good Clinical Practice guidelines, and local laws and regulations. All procedures were approved by the French National Institutional Review Board (identifier 23.00748.OOO2L7#I for the Committee for the Protection of Persons). All data were stored on secure health data servers, without any identifying information in the metadata. Participants received a €15 gift card as compensation for their time.</p>
      </sec>
      <sec>
        <title>Study Procedure</title>
        <sec>
          <title>Overview</title>
          <p>The participants completed the protocol on smartphones through the Callyope research mobile app in a home environment. The participants completed self-assessment scales for different mental health dimensions and recorded different speech tasks. In this work, we only focused on 1 spontaneous and semistructured speech task where participants had to answer “Describe how you are feeling at the moment and how your nights’ sleep have been lately” [<xref ref-type="bibr" rid="ref31">31</xref>]. The participants were included by speech pathologist interns and recruited through social media platforms. Finally, self-reported symptoms were examined with clinically validated questionnaires (<xref rid="figure1" ref-type="fig">Figure 1</xref>A). Participants followed the instructions displayed on the Callyope app, and their vocal answers were recorded with the smartphone’s microphone. The audio was sampled at 44.1 kHz with a 16-bit resolution. Each participant was asked to place his phone on a flat surface (eg, a table) with the microphone pointing toward the speaker, that is, himself. The session should take place in a quiet environment, whenever possible.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>Overview of this study. (A) Overview of our Callyope General-Population (Callyope-GP) dataset with 865 included participants. (B) Flowchart of the pretraining phases of our speech encoders with an illustration of the pretraining speaker embedding process. (C) Graphical illustration of our speech and machine learning (ML) pipeline for a single individual. The pretrained speaker embedding is frozen, and only ML models on top are trained on training data. (D) Evaluation and comparison between true and predicted assessment on held-out participants (test set n=131). Si represents the speech turn vector embeddings obtained from the speech encoder model, colors represent speaker identities for each speech turn and embedding, and wi represents the words spoken in the given audio.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e58572_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>We refer to the dataset collected in this study as the Callyope General-Population (Callyope-GP) dataset. We split randomly the Callyope-GP dataset into 3 sets: training, validation, and testing. Demographic data, such as sex, age, and education level, were collected. We compared groups with adequate tests for their demographics and self-assessments to ensure that groups were consistent.</p>
        </sec>
        <sec>
          <title>Measures of Depressive Symptoms</title>
          <p>To allow broader use of our solutions, the severity of depression was assessed through the Beck Depression Inventory (BDI) [<xref ref-type="bibr" rid="ref32">32</xref>] and Patient Health Questionnaire-9 (PHQ-9) [<xref ref-type="bibr" rid="ref33">33</xref>] self-report questionnaires. In current clinical practices, it is common that different professionals interacting with a patient use different metrics to monitor depression. While depression assessment through these 2 measures exhibits a robust correlation at the group level [<xref ref-type="bibr" rid="ref34">34</xref>], thus facilitating the development of an equational conversion for research uses, their limited efficacy at the individual level impedes their reliable conversions to predict individual depressive status [<xref ref-type="bibr" rid="ref35">35</xref>].</p>
          <p>The PHQ-9 is a short, self-administered questionnaire mainly used to screen and measure the severity of depression [<xref ref-type="bibr" rid="ref33">33</xref>] and is sensitive to potential changes [<xref ref-type="bibr" rid="ref34">34</xref>]. It includes the 2 cardinal signs of depression: anhedonia and depressed mood. We considered a risk of depression if the total score for the PHQ-9 was more than 10 (PHQ-9≥10).</p>
          <p>The BDI is a self-administered questionnaire with 21 items, each centered around a core theme [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Respondents are presented with statements for each item, and they are instructed to choose 1 statement, which is then associated with a score ranging from 0 to 3. The cumulative score for the scale can reach a maximum of 63 points. We considered the BDI threshold to be positive if the total score was more than 10 (BDI≥10), as it is above the normal range as defined by the authors of the BDI [<xref ref-type="bibr" rid="ref36">36</xref>].</p>
        </sec>
        <sec>
          <title>Measure of Anxiety</title>
          <p>The Generalized Anxiety Disorder 7-item scale (GAD-7) questionnaire is to measure or assess the severity of GAD [<xref ref-type="bibr" rid="ref37">37</xref>]. This is a self-administered questionnaire that takes less than 5 minutes to complete, and it was especially developed to be deployed efficiently in primary care. The optimal cutoff for the GAD-7 was found to be a cutoff for the total score of GAD-7≥10 [<xref ref-type="bibr" rid="ref37">37</xref>].</p>
        </sec>
        <sec>
          <title>Measure of Insomnia</title>
          <p>The Athens Insomnia Scale (AIS) is a self-administered questionnaire to assess the patient’s sleep difficulties according to the <italic>International Classification of Diseases, Tenth Revision</italic> (ICD-10) criteria [<xref ref-type="bibr" rid="ref38">38</xref>]. The AIS-8 comprises 8 items (5 minutes) and is a good tool for general sleep assessment and insomnia screening and to measure the intensity of sleep-related problems but also as a screening tool in reliably establishing the diagnosis of insomnia. The optimal cutoff for diagnosis to detect insomnia troubles, for the AIS scale, is 6 [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>].</p>
        </sec>
        <sec>
          <title>Measures of Fatigue</title>
          <p>We used the Multidimensional Fatigue Inventory (MFI) to assess the different dimensions of fatigue [<xref ref-type="bibr" rid="ref41">41</xref>-<xref ref-type="bibr" rid="ref43">43</xref>]. It is a short self-report questionnaire (5-10 minutes) based on 20 questions to determine 5 dimensions of fatigue: general fatigue, physical fatigue, reduced motivation, reduced activity, and mental fatigue. We also reported the total fatigue score as the sum of all subcomponents.</p>
          <p>We used the normative data from Schwarz et al [<xref ref-type="bibr" rid="ref42">42</xref>] and Hinz et al [<xref ref-type="bibr" rid="ref44">44</xref>] to choose thresholds for each subcomponent. Individual subcomponents of fatigue in the 75% quantile in the studied populations are all above 10. Therefore, we aimed to predict individuals’ scores, which are above or equal to 10, for each dimension. As mentioned also in Schwarz et al [<xref ref-type="bibr" rid="ref42">42</xref>], the total score has clinical significance and validity, as it was observed to have the highest correlations with anxiety, depression, and quality of life. There is no consensus cutoff for the total sum fatigue score; yet, based on the Colombian normative data [<xref ref-type="bibr" rid="ref44">44</xref>], we observed that the mean values for each studied subgroup were all above 40; therefore, we chose a clinical threshold of 40 for the total sum score.</p>
        </sec>
      </sec>
      <sec>
        <title>ML Analyses</title>
        <sec>
          <title>Overview</title>
          <p>Our ML analyses can be decomposed into three main steps: (1) the pretraining of the speech encoder model (<xref rid="figure1" ref-type="fig">Figure 1</xref>B) (2) the fine-tuning of ML models for each mental health aspect considered in this study (<xref rid="figure1" ref-type="fig">Figure 1</xref>C), and (3) extensive evaluations of the clinical threshold detection, selective detection, fairness assessments, and severity estimations for each clinical scale (<xref rid="figure1" ref-type="fig">Figures 1</xref>D and 2).</p>
        </sec>
        <sec>
          <title>Model Pretraining and Tuning</title>
          <p>Audio intensity is normalized per sample, and we compared the three main approaches to obtain representations for large-scale speech models: (1) Speaker recognition is performed using a ThinResNet model with 34 layers. The model takes speech samples as input, which are encoded as 40-Mel spectrograms, with a hop length of 10 ms and a Hamming window. This architecture is based on the ResNet design introduced by He et al [<xref ref-type="bibr" rid="ref45">45</xref>]. (2) We also considered a transformer [<xref ref-type="bibr" rid="ref46">46</xref>] architecture adapted for speech (HuBERT [<xref ref-type="bibr" rid="ref47">47</xref>]), trained in a self-supervision fashion, that is, to predict masked neighbor embeddings. (3) Finally, we evaluated a transformer architecture Whisper pretrained to tackle automatic speech recognition.</p>
          <p>Speaker recognition as a pretraining task has proven great results in mental health and neurology [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref48">48</xref>]. The ThinResNetis pretrained on the VoxCeleb2 dataset [<xref ref-type="bibr" rid="ref49">49</xref>], which is publicly available and contains over 1 million utterances from 6112 speakers, from 145 nationalities. The VoxCeleb2 dataset consists of almost only continuous speech. The pretraining learning forces the model to organize speech in terms of speaker characteristics, as we illustrated earlier in the right panel of <xref rid="figure1" ref-type="fig">Figure 1</xref>B. We used an additive-margin softmax loss for this speaker identification task [<xref ref-type="bibr" rid="ref50">50</xref>].</p>
          <p>We also compared these models to a self-supervised model, HuBERTXL, which exhibits great generalization for paralinguistic tasks such as emotion recognition [<xref ref-type="bibr" rid="ref51">51</xref>]. HuBERT was trained on 960 hours of Librispeech [<xref ref-type="bibr" rid="ref52">52</xref>]. Whisper is a recent robust automatic speech recognition system based on a transformer architecture trained on 680,000 hours of transcribed speech. Whisper training data are much bigger and more diverse with noisy labels than other speech models. We considered 3 versions of the model: small, medium, and large.</p>
          <p>In this work, we did not fine-tune any of the speech encoder models on the Callyope-GP dataset, which we represented in <xref rid="figure1" ref-type="fig">Figure 1</xref>C by a frozen speech encoder model. For each speaker <italic>i</italic>, we obtained a vector representation, a speech vector embedding denoted <italic>S<sub>i</sub></italic> We extracted segments of 20 seconds with 10 seconds overlap, and we compared different pooling of predictions with mean and max pooling. Besides, as our data are imbalanced, we also compared classic sampling of examples to train models with undersampling of majority class and oversampling of minority classes. The default values of undersampling and oversampling were used. We found no differences with and without voice activity detection, so we used windowing for simplicity. Extraction was performed using Python (version 3.9; Python Software Foundation), and the following packages were used to extract the acoustic features: <italic>pytorch 2.0.1</italic>, <italic>imbalanced learn</italic>, <italic>torchaudio 2.0.2</italic>, and <italic>voxceleb_trainer project</italic> [<xref ref-type="bibr" rid="ref53">53</xref>].</p>
          <p>For the fine-tuning of each task and each clinical score, different ML algorithms were compared on the validation set. For each task, once a model was selected, we retrained this final model on the concatenation of the training and validation sets and tested on the held-out test to avoid any inflated results. We used the scikit-learn implementation of each algorithm, splitting and evaluation [<xref ref-type="bibr" rid="ref54">54</xref>].</p>
          <p>For the speech collected in the Callyope-GP dataset, we applied the frozen speech encoder to each speech turn and propagated mental health assessment labels at the speech turn level to train and compare the final ML model. At inference, for final evaluation, we pool predictions at the speaker level, after we obtained varying speech turns from a specific speaker. We illustrated in <xref rid="figure2" ref-type="fig">Figure 2</xref> each clinical end point, translated as an ML task based on the aforementioned procedure.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Schematic representation of the clinical tasks that are being assessed in this study. Each task is different in terms of the set and types of outputs. The different tasks were illustrated with the BDI clinical scale and its given threshold of 10. BDI: Beck Depression Inventory; ML: machine learning.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e58572_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Clinical Threshold Detection (Classification)</title>
          <p>We first compared the predictive power of the speech encoder to discriminate between individuals who are below or above the threshold for each clinical scale. The distributions of positive and negative labels vary across clinical dimensions, and to take into account imbalance, we reported the performances of the macro <italic>F</italic><sub>1</sub>-scores along the area under the curve (AUC) of the receiver operating characteristic on the test set.</p>
          <p>We compared linear-based models (logistic regression for classification with L2 regularization and elastic net linear model for regression), tree-based models (random forests with 100 estimators), and gradient-boosting algorithms (histogram-based gradient boosting). Even though the pretraining phase captured information about the participants’ mental health, it is important to build a final model for each mental health dimension to be more specific and more sensitive. In ML terms, the mental health characteristics in speech are not necessarily linearly separable in the last vector space of the speech encoder.</p>
        </sec>
        <sec>
          <title>Estimation of Severity Through Predictions of Individual Scores (Regression)</title>
          <p>The conventional approach in mental health assessment through speech analysis typically focuses on the group’s statistical analyses or binary classifications of categorical outcomes, primarily discerning the presence or absence of a specific dimension. Yet, the risks of depression, anxiety, fatigue, and insomnia exist along a spectrum of severity levels that exert varying degrees of influence on an individual’s well-being.</p>
          <p>We go beyond traditional prediction and categorizations by integrating the estimation of severity through predictions of individual scores using regression ML models. This offers a more nuanced and comprehensive understanding of mental health dynamics, allowing for a more refined and personalized assessment. We evaluated our estimation of the severity of each total score with the mean absolute error (MAE) between actual and predicted scores and Pearson correlations. MAE score directly measures how close the predicted scores are to the actual scores without considering the direction of error. We also reported the Pearson correlation and the <italic>P</italic> value between the actual test set and the predicted values.</p>
        </sec>
        <sec>
          <title>Fairness Assessments: Quality of Services for Sex, Age, and Education Level Demographics</title>
          <p>ML systems can behave unfairly for different reasons and in multiple ways [<xref ref-type="bibr" rid="ref28">28</xref>]. In medicine, the use of ML and predictive models should be carefully evaluated, especially for potential quality-of-service harms, that is, it can occur when a system is not as performant for 1 specific group of people as it is for another group. We conducted a thorough analysis of the quality of service regarding potential harms for each clinical scale in the final predicted model across every dimension: sex, age, and education level. The disparity ratio (DR) was reported based on clinical threshold detection <italic>F</italic><sub>1</sub>-scores [<xref ref-type="bibr" rid="ref55">55</xref>] to consider both false positives and false negatives and to be more stringent than equality of opportunity. The DR is computed as the fraction of the minimum <italic>F</italic><sub>1</sub>-score across subgroups divided by the average <italic>F</italic><sub>1</sub>-score on the full test set:</p>
          <graphic xlink:href="jmir_v26i1e58572_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>We did the same for Pearson correlation for regression. The higher the DR, the better it is, as it means that the model performs equally well across groups with the perfect DR being 1, that is, each group has the same level of performance. The <italic>fairlearn</italic> toolkit was used to perform our fairness evaluations [<xref ref-type="bibr" rid="ref56">56</xref>].</p>
        </sec>
        <sec>
          <title>Selective Clinical Threshold Detection (Selective Prediction)</title>
          <p>ML approaches have made great strides in several domains; yet, apps to high-stakes settings remain challenging. In our case, in mental health assessments, communicating appropriately the uncertainty associated with the system predictions is critical [<xref ref-type="bibr" rid="ref57">57</xref>]. Yet, the communication of probabilities to human users is hard [<xref ref-type="bibr" rid="ref58">58</xref>], and a pragmatic approach is to determine if an artificial intelligence system is more likely to make erroneous predictions and defer these cases to clinicians. This approach can be viewed as a selective prediction task, where the ML system has the ability to withhold a prediction when it is too uncertain (essentially, the model saying “I don’t know”) [<xref ref-type="bibr" rid="ref59">59</xref>,<xref ref-type="bibr" rid="ref60">60</xref>]. In this work, we followed the method from Hendrycks and Gimpel [<xref ref-type="bibr" rid="ref61">61</xref>], and we used the maximum output probabilities of the ML classification system as a way to measure uncertainty.</p>
          <p>Based on a moving threshold, we can obtain a specific ML system to choose to abstain when its output probabilities are too low. This specific ML system is evaluated based on the predictions it chooses to make only; thus, there is a specific coverage and a specific accuracy or risk. There is a natural tradeoff between the coverage of the ML system and its accuracy. Therefore, the way to evaluate a selective prediction task is the AUC for the risk-coverage curve.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Data Overview and Demographics of Participants</title>
        <p>A total of 1150 participants were eligible and agreed to participate in our study. Among them, 865 completed the study, giving a recruitment yield of 75.2%. There was an equal split between Android (n=475, 54.9%) and iOS (n=390, 45.1%) devices used by participants. The reasons for which participants were not included in this analysis were the following: one missing speech task, missing demographic information, or one missing answer in the self-report questionnaires.</p>
        <p>For our analyses, 605 participants were in the training set, 129 were in the validation set, and 131 were in the test set, and these groups did not differ in terms of demographics and mental health evaluations (<xref ref-type="table" rid="table1">Table 1</xref>). This yields a dataset sufficient (n&#62;500) to evaluate error bars and predictive algorithms to avoid over-optimistic results [<xref ref-type="bibr" rid="ref62">62</xref>]. Among the 865 participants, 275 (31.8%) were above the BDI screening threshold, 146 (16.9%) were above the PHQ-9 threshold, 133 (15.3%) were above the GAD-7 threshold, 371 (42.5%) were above the AIS threshold, 489 (56.5%) were above the MFI general fatigue threshold, 325 (37.5%) were above the MFI physical fatigue threshold, 283 (32.7%) were above the MFI reduced activity threshold, 379 (43.8%) were above the MFI mental fatigue threshold, 209 (24.1%) were above the MFI reduced motivation threshold, and 557 (64.4%) were above the threshold of the total score of the MFI. We reported a co-occurrence matrix of people at risk for each dimension in Figure S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Demographic characteristics of the training, validation, and test groups<sup>a</sup>.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="30"/>
            <col width="270"/>
            <col width="0"/>
            <col width="160"/>
            <col width="160"/>
            <col width="0"/>
            <col width="170"/>
            <col width="0"/>
            <col width="110"/>
            <col width="0"/>
            <col width="0"/>
            <col width="70"/>
            <thead>
              <tr valign="bottom">
                <td colspan="4">Characteristics<sup>b</sup></td>
                <td>Training (n=605)</td>
                <td colspan="2">Validation (n=129)</td>
                <td>Test (n=131)</td>
                <td colspan="5">Group comparisons</td>
              </tr>
              <tr valign="top">
                <td colspan="4">
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td colspan="2">
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td colspan="3"><italic>F</italic> test (<italic>df</italic>)</td>
                <td colspan="2"><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="13">
                  <bold>Demographics</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>Sex, n (%)</bold>
                </td>
                <td colspan="3">9.0<sup>c</sup> (4)</td>
                <td>.06</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Female</td>
                <td colspan="2">386 (76.4)</td>
                <td>85 (65.9)</td>
                <td colspan="2">69 (52.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Male</td>
                <td colspan="2">214 (35.3)</td>
                <td>44 (34.1)</td>
                <td colspan="2">63 (47.7)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Other</td>
                <td colspan="2">5 (0.8)</td>
                <td>0 (0)</td>
                <td colspan="2">0 (0)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>Education level, n (%)</bold>
                </td>
                <td colspan="3">5.2<sup>c</sup> (6)</td>
                <td>.52</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>No diploma</td>
                <td colspan="2">19 (3.1)</td>
                <td>7 (5.4)</td>
                <td colspan="2">8 (6.1)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Secondary studies</td>
                <td colspan="2">107 (17.7)</td>
                <td>21 (16.2)</td>
                <td colspan="2">24 (18.3)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Short postbaccalaureate studies (Bac+2)</td>
                <td colspan="2">71 (11.7)</td>
                <td>20 (15.5)</td>
                <td colspan="2">16 (12.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Long postbaccalaureate studies (Bac+3 and above)</td>
                <td colspan="2">408 (67.4)</td>
                <td>81 (62.8)</td>
                <td colspan="2">83 (63.3)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>Age (years)</bold>
                </td>
                <td colspan="3">0.78 (2)</td>
                <td>.45</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">38.8 (18.2)</td>
                <td>40.2 (18.7)</td>
                <td colspan="2">40.1 (20.0)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">18-92</td>
                <td>18-86</td>
                <td colspan="2">18-89</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="13">
                  <bold>Clinical evaluation</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>BDI<sup>d</sup></bold>
                </td>
                <td colspan="3">0.07 (2)</td>
                <td>.92</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">7.7 (7.3)</td>
                <td>7.9 (7.2)</td>
                <td colspan="2">7.9 (7.7)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">0.0-32.0</td>
                <td>0.0-29.0</td>
                <td colspan="2">0.0-36.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">413 (68.3)</td>
                <td>88 (68.2)</td>
                <td colspan="2">89 (67.9)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">192 (31.7)</td>
                <td>41 (31.8)</td>
                <td colspan="2">42 (32.1)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>PHQ-9<sup>f</sup></bold>
                </td>
                <td colspan="3">0.85 (2)</td>
                <td>.42</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">5.6 (4.5)</td>
                <td>5.2 (4.7)</td>
                <td colspan="2">5.1 (4.4)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">0.0-23.0</td>
                <td>0.0-25.0</td>
                <td colspan="2">0.0-22.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">499 (82.5)</td>
                <td>108 (83.7)</td>
                <td colspan="2">112 (85.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">106 (17.5)</td>
                <td>21 (16.3)</td>
                <td colspan="2">19 (14.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>GAD-7<sup>g</sup></bold>
                </td>
                <td colspan="3">0.50 (2)</td>
                <td>.60</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">5.0 (4.5)</td>
                <td>4.8 (4.7)</td>
                <td colspan="2">5.4 (5.0)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">0.0-21.0</td>
                <td>0.0-20.0</td>
                <td colspan="2">0.0-21.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">511 (84.5)</td>
                <td>109 (84.5)</td>
                <td colspan="2">112 (85.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">94 (15.5)</td>
                <td>20 (15.5)</td>
                <td colspan="2">19 (14.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>AIS<sup>h</sup></bold>
                </td>
                <td colspan="3">1.6 (2)</td>
                <td>.19</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">5.6 (3.9)</td>
                <td>5.0 (3.1)</td>
                <td colspan="2">5.2 (4.1)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">0.0-24.0</td>
                <td>0.0-16.0</td>
                <td colspan="2">0.0-19.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">338 (55.9)</td>
                <td>75 (58.1)</td>
                <td colspan="2">81 (61.8)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">267 (44.1)</td>
                <td>54 (41.8)</td>
                <td colspan="2">50 (38.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI<sup>i</sup> general fatigue</bold>
                </td>
                <td colspan="3">1.8 (2)</td>
                <td>.17</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">10.4 (4.2)</td>
                <td>11.1 (4.2)</td>
                <td colspan="2">10.4 (4.1)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">4.0-20.0</td>
                <td>4.0-20.0</td>
                <td colspan="2">4.0-20.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">335 (55.4)</td>
                <td>80 (62)</td>
                <td colspan="2">74 (56.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">270 (44.6)</td>
                <td>49 (38)</td>
                <td colspan="2">57 (43.5)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI physical fatigue</bold>
                </td>
                <td colspan="3">0.55 (2)</td>
                <td>0.57</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">8.7 (3.9)</td>
                <td>8.8 (4.0)</td>
                <td colspan="2">8.4 (3.9)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">4.0-20.0</td>
                <td>4.0-20.0</td>
                <td colspan="2">4.0-20.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">376 (62.1)</td>
                <td>78 (60.5)</td>
                <td colspan="2">86 (65.6)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">229 (37.9)</td>
                <td>51 (39.5)</td>
                <td colspan="2">45 (34.4)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI reduced activity</bold>
                </td>
                <td colspan="3">0.92 (2)</td>
                <td>.40</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">8.3 (3.8)</td>
                <td>7.9 (3.5)</td>
                <td colspan="2">8.1 (3.6)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">3.0-20.0</td>
                <td>4.0-20.0</td>
                <td colspan="2">4.0-18.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">400 (66.1)</td>
                <td>92 (71.3)</td>
                <td colspan="2">90 (68.7)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">205 (33.9)</td>
                <td>37 (28.7)</td>
                <td colspan="2">41 (31.3)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI mental fatigue</bold>
                </td>
                <td colspan="3">0.26 (2)</td>
                <td>.77</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">9.2 (4.2)</td>
                <td>9.5 (4.4)</td>
                <td colspan="2">9.3 (4.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">4.0-20.0</td>
                <td>4.0-20.0</td>
                <td colspan="2">4.0-20.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">340 (56.2)</td>
                <td>69 (53.5)</td>
                <td colspan="2">77 (58.8)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">265 (43.8)</td>
                <td>60 (46.5)</td>
                <td colspan="2">54 (41.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI reduced motivation</bold>
                </td>
                <td colspan="3">0.74 (2)</td>
                <td>.47</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">7.5 (3.0)</td>
                <td>7.2 (3.2)</td>
                <td colspan="2">7.7 (3.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">4.0-20.0</td>
                <td>4.0-20.0</td>
                <td colspan="2">4.0-18.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">458 (75.7)</td>
                <td>102 (79.1)</td>
                <td colspan="2">96 (73.3)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">147 (24.3)</td>
                <td>27 (20.9)</td>
                <td colspan="2">35 (26.7)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="8">
                  <bold>MFI total score</bold>
                </td>
                <td colspan="3">0.09 (2)</td>
                <td>.91</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Mean (SD)</td>
                <td colspan="2">44.2 (15.1)</td>
                <td>44.6 (15.0)</td>
                <td colspan="2">43.9 (15.1)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Range</td>
                <td colspan="2">19.0-99.0</td>
                <td>20.0-90.0</td>
                <td colspan="2">20.0-83.0</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Negative, n (%)</td>
                <td colspan="2">338 (55.9)</td>
                <td>79 (61.2)</td>
                <td colspan="2">77 (58.8)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>Positive<sup>e</sup>, n (%)</td>
                <td colspan="2">267 (44.1)</td>
                <td>50 (38.8)</td>
                <td colspan="2">54 (41.2)</td>
                <td colspan="2">
                  <break/>
                </td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>Categorical variables are compared with Pearson chi-square test, and continuous variables are compared with 1-way ANOVA.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>Characteristics for the 3 splits of the dataset to ensure generalization.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>Chi-square test.</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup>BDI: Beck Depression Inventory.</p>
            </fn>
            <fn id="table1fn5">
              <p><sup>e</sup>In addition to statistics for each clinical scale score, we report the participants below and above the cutoff (see the <italic>Methods</italic> section for each threshold).</p>
            </fn>
            <fn id="table1fn6">
              <p><sup>f</sup>PHQ-9: Patient Health Questionnaire-9.</p>
            </fn>
            <fn id="table1fn7">
              <p><sup>g</sup>GAD-7: General Anxiety Disorder 7-item scale.</p>
            </fn>
            <fn id="table1fn8">
              <p><sup>h</sup>AIS: Athens Insomnia Scale.</p>
            </fn>
            <fn id="table1fn9">
              <p><sup>i</sup>MFI: Multidimensional Fatigue Inventory.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>ML Analyses</title>
        <sec>
          <title>Overview</title>
          <p>Models and pipelines were compared with the results on the validation set (Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). Overall, all the Whisper models outperformed other approaches (Whisper M being the best with mean <italic>F</italic><sub>1</sub>-score=0.56, SD 0.09), while the speaker model performed the worst, even though still performing well. We found out that the performance of the pooling with the maximum prediction always outperformed the mean pooling, and undersampling and oversampling were helping. It was found out that the linear-based models were outperforming random forest and gradient boosting on the validation set. Thus, we retrained linear algorithms with max pooling and the Whisper M frozen speech encoder on the combination of the training and validation sets and reported the final results on the held-out test set in <xref rid="figure3" ref-type="fig">Figures 3</xref> and <xref rid="figure4" ref-type="fig">4</xref> and <xref ref-type="table" rid="table2">Tables 2</xref>-<xref ref-type="table" rid="table4">4</xref>.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>ROC curves for the clinical threshold detection task on the held-out test set. (A) ROC curves to detect clinically relevant thresholds for depression (PHQ-9 and BDI), anxiety (GAD-7), and insomnia (AIS). (B) ROC curves to detect clinically relevant thresholds for fatigue components (MFI), general fatigue, physical fatigue, reduced activity, mental fatigue, reduced motivation, and total fatigue. AUC: area under the curve; AIS: Athens Insomnia Scale; BDI: Beck Depression Inventory; GAD-7: General Anxiety Disorder 7-item scale; MFI: Multidimensional Fatigue Inventory; PHQ-9: Patient Health Questionnaire-9; ROC: receiver operating characteristic curve.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e58572_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>Risk-coverage curves for the selective clinical threshold detection task on the held-out test set illustrate the models’ selective screening ability, that is, risk detection capabilities with the ability to abstain when too uncertain. Curves are smoothed for clarity with a Gaussian blur but not used to compute AUC. A lower AUC is better, and 0 is the perfect score. (A) Risk-coverage curves selectively detect clinically relevant thresholds for depression (PHQ-9 and BDI), anxiety (GAD-7), and insomnia (AIS). (B) Risk-coverage curves to detect clinically relevant thresholds for fatigue components (MFI), general fatigue, physical fatigue, reduced activity, mental fatigue, reduced motivation, and total fatigue. AIS: Athens Insomnia Scale; AUC: area under the curve; BDI: Beck Depression Inventory; GAD-7: General Anxiety Disorder 7-item scale, MFI: Multidimensional Fatigue Inventory; PHQ-9: Patient Health Questionnaire-9.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e58572_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <table-wrap position="float" id="table2">
            <label>Table 2</label>
            <caption>
              <p>Estimation of clinical threshold detection severity results on the held-out test set for the different considered dimensions of mental health (classification).</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="730"/>
              <col width="270"/>
              <thead>
                <tr valign="bottom">
                  <td>Dimensions of mental health</td>
                  <td><italic>F</italic><sub>1</sub>-score<sup>a</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>PHQ-9<sup>b</sup></td>
                  <td>0.49</td>
                </tr>
                <tr valign="top">
                  <td>GAD-7<sup>c</sup></td>
                  <td>0.50</td>
                </tr>
                <tr valign="top">
                  <td>BDI<sup>d</sup></td>
                  <td>0.65</td>
                </tr>
                <tr valign="top">
                  <td>AIS<sup>e</sup></td>
                  <td>0.62</td>
                </tr>
                <tr valign="top">
                  <td>MFI<sup>f</sup> general fatigue</td>
                  <td>0.66</td>
                </tr>
                <tr valign="top">
                  <td>MFI physical fatigue</td>
                  <td>0.54</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced activity</td>
                  <td>0.52</td>
                </tr>
                <tr valign="top">
                  <td>MFI mental fatigue</td>
                  <td>0.53</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced motivation</td>
                  <td>0.43</td>
                </tr>
                <tr valign="top">
                  <td>MFI total fatigue score</td>
                  <td>0.69</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table2fn1">
                <p><sup>a</sup>Higher <italic>F</italic><sub>1</sub>-score is better, and 1 is perfect.</p>
              </fn>
              <fn id="table2fn2">
                <p><sup>b</sup>PHQ-9: Patient Health Questionnaire-9.</p>
              </fn>
              <fn id="table2fn3">
                <p><sup>c</sup>GAD-7: General Anxiety Disorder 7-item scale.</p>
              </fn>
              <fn id="table2fn4">
                <p><sup>d</sup>BDI: Beck Depression Inventory.</p>
              </fn>
              <fn id="table2fn5">
                <p><sup>e</sup>AIS: Athens Insomnia Scale.</p>
              </fn>
              <fn id="table2fn6">
                <p><sup>f</sup>MFI: Multidimensional Fatigue Inventory.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <table-wrap position="float" id="table3">
            <label>Table 3</label>
            <caption>
              <p>Disparity ratios (DRs) based on the F1-scores for sex, age, and education levels to assess fairness on the held-out test set for the clinical threshold detection (classification) for the different considered dimensions of mental health<sup>a</sup>.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="350"/>
              <col width="140"/>
              <col width="140"/>
              <col width="210"/>
              <col width="0"/>
              <col width="160"/>
              <thead>
                <tr valign="bottom">
                  <td>Dimensions of mental health</td>
                  <td colspan="4">DR</td>
                  <td>Mean (SD)<sup>b</sup></td>
                </tr>
                <tr valign="bottom">
                  <td>
                    <break/>
                  </td>
                  <td>Sex</td>
                  <td>Age<sup>c</sup></td>
                  <td>Education level<sup>d</sup></td>
                  <td colspan="2">
                    <break/>
                  </td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>PHQ-9<sup>e</sup></td>
                  <td>0.61</td>
                  <td>0.00</td>
                  <td>0.00</td>
                  <td colspan="2">0.20 (0.35)</td>
                </tr>
                <tr valign="top">
                  <td>GAD-7<sup>f</sup></td>
                  <td>0.66</td>
                  <td>0.00</td>
                  <td>0.00</td>
                  <td colspan="2">0.22 (0.38)</td>
                </tr>
                <tr valign="top">
                  <td>BDI<sup>g</sup></td>
                  <td>0.85</td>
                  <td>0.25</td>
                  <td>0.57</td>
                  <td colspan="2">0.56 (0.30)</td>
                </tr>
                <tr valign="top">
                  <td>AIS<sup>h</sup></td>
                  <td>0.98</td>
                  <td>0.45</td>
                  <td>0.69</td>
                  <td colspan="2">0.71 (0.27)</td>
                </tr>
                <tr valign="top">
                  <td>MFI<sup>i</sup> general fatigue</td>
                  <td>0.77</td>
                  <td>0.73</td>
                  <td>0.75</td>
                  <td colspan="2">0.75 (0.02)</td>
                </tr>
                <tr valign="top">
                  <td>MFI physical fatigue</td>
                  <td>0.97</td>
                  <td>0.76</td>
                  <td>0.90</td>
                  <td colspan="2">0.88 (0.11)</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced activity</td>
                  <td>0.95</td>
                  <td>0.27</td>
                  <td>0.56</td>
                  <td colspan="2">0.59 (0.34)</td>
                </tr>
                <tr valign="top">
                  <td>MFI mental fatigue</td>
                  <td>0.91</td>
                  <td>0.00</td>
                  <td>0.58</td>
                  <td colspan="2">0.50 (0.46)</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced motivation</td>
                  <td>0.94</td>
                  <td>0.25</td>
                  <td>0.30</td>
                  <td colspan="2">0.50 (0.38)</td>
                </tr>
                <tr valign="top">
                  <td>MFI total fatigue score</td>
                  <td>0.91</td>
                  <td>0.62</td>
                  <td>0.35</td>
                  <td colspan="2">0.63 (0.28)</td>
                </tr>
                <tr valign="top">
                  <td>Mean (SD)<sup>j</sup></td>
                  <td>0.86 (0.13)</td>
                  <td>0.33 (0.30)</td>
                  <td>0.47 (0.30)</td>
                  <td colspan="2">0.55 (0.33)</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table3fn1">
                <p><sup>a</sup>Higher DR is better, and 1 is perfect DR.</p>
              </fn>
              <fn id="table3fn2">
                <p><sup>b</sup>Mean and SD values per score.</p>
              </fn>
              <fn id="table3fn3">
                <p><sup>c</sup>Participants are grouped into age categories to allow analysis: 18-30, 30-45, 45-65, and &#62;65 years.</p>
              </fn>
              <fn id="table3fn4">
                <p><sup>d</sup>Highest achieved study level: no diploma or secondary studies or short postbaccalaureate studies (Bac+2) or long postbaccalaureate studies (Bac+3 and above).</p>
              </fn>
              <fn id="table3fn5">
                <p><sup>e</sup>PHQ-9: Patient Health Questionnaire-9.</p>
              </fn>
              <fn id="table3fn6">
                <p><sup>f</sup>GAD-7: General Anxiety Disorder 7-item scale.</p>
              </fn>
              <fn id="table3fn7">
                <p><sup>g</sup>BDI: Beck Depression Inventory.</p>
              </fn>
              <fn id="table3fn8">
                <p><sup>h</sup>AIS: Athens Insomnia Scale.</p>
              </fn>
              <fn id="table3fn9">
                <p><sup>i</sup>MFI: Multidimensional Fatigue Inventory.</p>
              </fn>
              <fn id="table3fn10">
                <p><sup>j</sup>Mean and SD are reported per sensitive dimension.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <table-wrap position="float" id="table4">
            <label>Table 4</label>
            <caption>
              <p>Estimation of severity results on the held-out test set for the different considered dimensions of mental health (regression)<sup>a</sup>.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="420"/>
              <col width="290"/>
              <col width="290"/>
              <thead>
                <tr valign="bottom">
                  <td>Dimensions of mental health</td>
                  <td>Pearson <italic>r</italic></td>
                  <td>MAE<sup>b</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>PHQ-9<sup>c</sup></td>
                  <td>0.47</td>
                  <td>3.1</td>
                </tr>
                <tr valign="top">
                  <td>GAD-7<sup>d</sup></td>
                  <td>0.48</td>
                  <td>3.2</td>
                </tr>
                <tr valign="top">
                  <td>BDI<sup>e</sup></td>
                  <td>0.49</td>
                  <td>4.9</td>
                </tr>
                <tr valign="top">
                  <td>AIS<sup>f</sup></td>
                  <td>0.43</td>
                  <td>2.9</td>
                </tr>
                <tr valign="top">
                  <td>MFI<sup>g</sup> general fatigue</td>
                  <td>0.38</td>
                  <td>3.3</td>
                </tr>
                <tr valign="top">
                  <td>MFI physical fatigue</td>
                  <td>0.32</td>
                  <td>3.0</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced activity</td>
                  <td>0.31</td>
                  <td>2.9</td>
                </tr>
                <tr valign="top">
                  <td>MFI mental fatigue</td>
                  <td>0.34</td>
                  <td>3.1</td>
                </tr>
                <tr valign="top">
                  <td>MFI reduced motivation</td>
                  <td>0.32</td>
                  <td>2.5</td>
                </tr>
                <tr valign="top">
                  <td>MFI total fatigue score</td>
                  <td>0.44</td>
                  <td>11.3</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table4fn1">
                <p><sup>a</sup>We reported mean absolute errors and Pearson correlations between actual and predicted values. Lower mean absolute error is better, and 0 is perfect. A higher Pearson correlation is better, and +1 is perfect. All correlations were significant (<italic>P</italic>&#60;1×10<sup>–14</sup>).</p>
              </fn>
              <fn id="table4fn2">
                <p><sup>b</sup>MAE: mean absolute error.</p>
              </fn>
              <fn id="table4fn3">
                <p><sup>c</sup>PHQ-9: Patient Health Questionnaire-9.</p>
              </fn>
              <fn id="table4fn4">
                <p><sup>d</sup>GAD-7: General Anxiety Disorder 7-item scale.</p>
              </fn>
              <fn id="table4fn5">
                <p><sup>e</sup>BDI: Beck Depression Inventory.</p>
              </fn>
              <fn id="table4fn6">
                <p><sup>f</sup>AIS: Athens Insomnia Scale.</p>
              </fn>
              <fn id="table4fn7">
                <p><sup>g</sup>MFI: Multidimensional Fatigue Inventory.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Clinical Threshold Detection (Classification)</title>
          <p>The clinical threshold detection performed well based on the speech data and our developed system (<xref rid="figure3" ref-type="fig">Figure 3</xref> and <xref ref-type="table" rid="table2">Table 2</xref>). All systems outperformed the chance levels. Based on the AUC, the classification results were the highest for the BDI score (AUC=0.78; <italic>F</italic><sub>1</sub>-score=0.65). Based on the <italic>F</italic><sub>1</sub>-score, it was the total MFI score (AUC=0.68; <italic>F</italic><sub>1</sub>-score=0.69). The lowest for both metrics was the MFI reduced motivation (AUC=0.61; <italic>F</italic><sub>1</sub>-score=0.43).</p>
        </sec>
        <sec>
          <title>Fairness Assessments: Quality of Services for Sex, Age, and Education Level for Classification</title>
          <p>We computed the sensitive attribute DRs (sex, age, and education level) and assessed the differences in the quality of the service made by the speech-based system (<xref ref-type="table" rid="table3">Table 3</xref>) for classification. Overall, the classification of the speech-based system had a better quality of service for sex (mean 0.86, SD 0.13), and the worst was for age (mean 0.33, SD 0.30). We also identified that the detection of PHQ-9 had the worst quality-of-service disparity (mean of DRs 0.20, SD 0.35), and the best quality-of-service was obtained with the AIS (mean of DRs 0.71, SD 0.27), the MFI general fatigue (mean of DRs 0.75, SD 0.02), and the MFI physical fatigue (mean of DRs 0.88, SD 0.11). We also observed that only the MFI general fatigue and MFI physical fatigue obtained a good performance for age (MFI general fatigue: DR=0.73 and MFI physical fatigue: DR=0.76), and, except for PHQ-9 and GAD-7, all DRs were satisfactory or high for sex.</p>
        </sec>
        <sec>
          <title>Selective Clinical Threshold Detection (Selective Prediction)</title>
          <p>The capabilities of speech-based models were also evaluated to selectively predict the different clinical thresholds. Great performances were observed for BDI, AIS, and MFI general fatigue (<xref rid="figure4" ref-type="fig">Figure 4</xref>). The model that selectively predicts the risk of depression based on the BDI score achieved the best result (BDI risk-coverage: AUC=0.28). The other scores could not achieve such feats of important coverage with no risk.</p>
        </sec>
        <sec>
          <title>Estimation of Severity Through Predictions of Individual Scores (Regression)</title>
          <p>The regression results for the estimation of severity are reported in <xref ref-type="table" rid="table3">Table 3</xref>. Speech-based models obtained significant results for all clinical variables based on the evaluation with the Pearson correlations (all <italic>P</italic>&#60;1×10<sup>–14</sup>). The strongest correlations between the prediction and the actual scores on the held-out test were found for BDI score (<italic>r</italic>=0.49), GAD-7 (<italic>r</italic>=0.48), and PHQ-9 (<italic>r</italic>=0.47). The lowest correlation was found for the MFI reduced activity (<italic>r</italic>=0.31).</p>
          <p>Speech-based models also obtained great results in terms of absolute errors. We observed less than 3 points of MAE for AIS and MFI reduced activity. All other scores were predicted on average with less than 5 points, except for the MFI total fatigue score, since its range is 13-88.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>We aimed to explore the full capabilities and limitations of using speech data extracted from 865 participants in the general population to predict the presence or absence of different mental health self-reported symptoms: depression, anxiety, and insomnia and the different dimensions of fatigue. We built a fully automated speech-based ML system that takes as input the audio waveform collected from 1 simple speech task performed on our smartphone app. The models were trained and calibrated on training and validation sets of participants, and we demonstrated the system’s generalization on the held-out test set of 131 participants.</p>
        <p>The results indicated that ML-based systems using speech only as input could identify participants above clinical thresholds for depression, insomnia, total fatigue components, but to a lesser extent, anxiety and fatigue subcomponents. All classification results were above chance levels for each clinical threshold.</p>
        <p>This result was confirmed with an extensive fairness analysis of quality of service for age, sex, and education levels. Depression, insomnia, and different dimensions of fatigue clinical threshold detection results were particularly consistent for sex, slightly less for age, and to a lesser extent for education level. Anxiety risk identification fell behind in accuracy overall and was also unequal per group. The extension of our clinical threshold detection system to be able to abstain, with selective prediction, was conclusive, even for anxiety. Risk-coverage AUCs remained low for insomnia, total fatigue, and depression detection through BDI. Finally, we showed that speech-based models could also predict the severity, with the prediction of exact scores moving beyond binary interpretations of score thresholds. All correlations between the predicted scores and the actual scores given by participants were significant, exhibiting strengths ranging from 0.31 to 0.49.</p>
        <p>Our study builds upon existing mental health research on speech analysis and extends the insights for deployment into clinical practice. For risk and anxiety depression assessments from speech in the general population, we found similar strong performances such as in previous studies using speech analyses [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref63">63</xref>,<xref ref-type="bibr" rid="ref64">64</xref>]. Our recruitment and involvement of participants was in person. For medium-sized datasets with face-to-face recruitment (below 1000 participants, such as ours), it can be observed that crowdsourced recruitment of participants via web-based platforms needs more training data to yield the same level of model accuracy [<xref ref-type="bibr" rid="ref63">63</xref>,<xref ref-type="bibr" rid="ref65">65</xref>]. This was also observed in a large study, with over 6000 participants, with web-based data recruitment for risk detection in the general population in American English [<xref ref-type="bibr" rid="ref64">64</xref>]. This discrepancy could be explained by the fact that data quality is variable on web crowdsourcing platforms, especially for participants’ psychiatric evaluations [<xref ref-type="bibr" rid="ref66">66</xref>], and also voice recording through laptops.</p>
        <p>Our study revealed discrepancies in both clinical threshold detection and estimation of severity through self-reported depression scores between BDI and PHQ-9. This underlines the inherent limitations of score conversion and the crucial role of individual-level assessment in capturing the nuanced and different expressions of depression [<xref ref-type="bibr" rid="ref35">35</xref>]. This reinforces the necessity of developing assessment tools and interpreting results with meticulous attention to individual variability, particularly by scrutinizing model performance at the individual level, mirroring real-world clinical scenarios.</p>
        <p>Our study uniquely addresses the co-occurrence of perceived fatigue and reported insomnia, both prevalent mental health concerns, which could be detected simultaneously through speech analysis. While prior research, like the work by Thoret et al [<xref ref-type="bibr" rid="ref20">20</xref>], has explored how sleep deprivation impacts specific vocal features like prosody and voice quality, no previous study has delved into the combined influence of fatigue and insomnia on speech. Addressing this gap is crucial because these conditions often co-occur and significantly impact the symptom trajectory and potential development of other mental health issues [<xref ref-type="bibr" rid="ref67">67</xref>]. The prevention of recurrent sleep problems can prevent other mental health troubles or relapses. The observed co-occurrence of symptoms, particularly insomnia with other clinical dimensions, highlights the interconnected nature of these symptoms and syndromes.</p>
        <p>Finally, to the best of our knowledge, this is the first study to assess the fairness and selective prediction capabilities in speech-based mental health assessments. This is of prime importance since speech signals can be heavily influenced by a multitude of factors such as age, sex, weight, and height [<xref ref-type="bibr" rid="ref68">68</xref>]. Age factor was the least preserved in our grouped performances, this can be attributed to voice changes due to hormones [<xref ref-type="bibr" rid="ref69">69</xref>], and normal aging affects the different parts of the vocal production system: larynx, respiratory system, resonators, saliva system, and the individual’s global emotional status [<xref ref-type="bibr" rid="ref70">70</xref>]. In addition to lower performances compared to other mental health dimensions, the anxiety risk detection performance collapsed for certain groups of demographics. This could be explained by the heterogeneity and low positive examples in our Callyope-GP dataset. Even though there are limits concerning some groups of individuals, selective classification offers an option to potentially remediate these variable quality of services, ensuring a deployment in clinical settings and still bringing overall clinical utility.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>While valuable, this study has some limitations. The French monolingual, medium-sized dataset (300&#60;n&#60;1000) needs more diverse data to achieve better generalizability, and the nonlongitudinal design misses insights on symptom evolution. Besides, the use of self-assessment scales introduces a potential bias because they rely heavily on the insight of participants. A limitation of our study is the use of a fixed train-dev-test split. This split approach, while convenient for comparing results across different tasks, can introduce bias and limit the generalizability of our findings. Future research with larger samples, longitudinal designs, and the inclusion of pathological data is crucial for exploiting the full potential of voice biomarkers in mental health. Studying various and diverse speech tasks and prompts also holds the potential for multiple benefits related to user adherence. Indeed, tailored tasks can address specific mental health needs, cater to individual preferences, and boost engagement. Varying prompts can reduce user fatigue and sustain interest, leading to more consistent system use and richer data collection.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>This study demonstrates the potential of speech-based systems for detecting and predicting various mental health symptoms in the general population. While challenges remain regarding real-world apps and ensuring fairness across the population demographics, our findings pave the way for further development and responsible integration of such tools into clinical settings, advancing personalized mental health assessment and intervention. In future work, we will extend this study by including longitudinal data, adding more diverse linguistic and geographic data, and including more severely affected patients who are already followed by mental health practitioners. We will also look into fairness and uncertainty mitigation methods to improve the performance of our systems.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Co-occurrences of symptoms, and validation and comparison results for each speech modeling approach.</p>
        <media xlink:href="jmir_v26i1e58572_app1.docx" xlink:title="DOCX File , 436 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AIS</term>
          <def>
            <p>Athens Insomnia Scale</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AUC</term>
          <def>
            <p>area under the curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BDI</term>
          <def>
            <p>Beck Depression Inventory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">Callyope-GP</term>
          <def>
            <p>Callyope General-Population</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">DR</term>
          <def>
            <p>disparity ratio</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">GAD-7</term>
          <def>
            <p>General Anxiety Disorder 7-item scale</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">ICD-10</term>
          <def>
            <p>International Classification of Diseases, Tenth Revision</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">MAE</term>
          <def>
            <p>mean absolute error</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">MFI</term>
          <def>
            <p>Multidimensional Fatigue Inventory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">ML</term>
          <def>
            <p>machine learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">PHQ-9</term>
          <def>
            <p>Patient Health Questionnaire-9</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">RCT</term>
          <def>
            <p>randomized controlled trial</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>The authors are thankful to all the participants who volunteered for this research study. Without their active involvement, this study would not have been possible. The authors also would like to thank each of the speech pathology interns who helped with the participant recruitment and made sure that the protocol was completed successfully.</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>RR, XNC, AL, MDG, MD, and AB are shareholders of Callyope, and VO was a former employee of Callyope.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <collab>GBD 2019 Diseases Injuries Collaborators</collab>
          </person-group>
          <article-title>Global burden of 369 diseases and injuries in 204 countries and territories, 1990-2019: a systematic analysis for the Global Burden of Disease Study 2019</article-title>
          <source>Lancet</source>
          <year>2020</year>
          <volume>396</volume>
          <issue>10258</issue>
          <fpage>1204</fpage>
          <lpage>1222</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0140-6736(20)30925-9"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/S0140-6736(20)30925-9</pub-id>
          <pub-id pub-id-type="medline">33069326</pub-id>
          <pub-id pub-id-type="pii">S0140-6736(20)30925-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC7567026</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <collab>GBD 2019 Mental Disorders Collaborators</collab>
          </person-group>
          <article-title>Global, regional, and national burden of 12 mental disorders in 204 countries and territories, 1990-2019: a systematic analysis for the Global Burden of Disease Study 2019</article-title>
          <source>Lancet Psychiatry</source>
          <year>2022</year>
          <volume>9</volume>
          <issue>2</issue>
          <fpage>137</fpage>
          <lpage>150</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2215-0366(21)00395-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/S2215-0366(21)00395-3</pub-id>
          <pub-id pub-id-type="medline">35026139</pub-id>
          <pub-id pub-id-type="pii">S2215-0366(21)00395-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC8776563</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <collab>COVID-19 Mental Disorders Collaborators</collab>
          </person-group>
          <article-title>Global prevalence and burden of depressive and anxiety disorders in 204 countries and territories in 2020 due to the COVID-19 pandemic</article-title>
          <source>Lancet</source>
          <year>2021</year>
          <volume>398</volume>
          <issue>10312</issue>
          <fpage>1700</fpage>
          <lpage>1712</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0140-6736(21)02143-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/S0140-6736(21)02143-7</pub-id>
          <pub-id pub-id-type="medline">34634250</pub-id>
          <pub-id pub-id-type="pii">S0140-6736(21)02143-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC8500697</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Henriksson</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Aro</surname>
              <given-names>HM</given-names>
            </name>
            <name name-style="western">
              <surname>Marttunen</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Heikkinen</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Isometsä</surname>
              <given-names>ET</given-names>
            </name>
            <name name-style="western">
              <surname>Kuoppasalmi</surname>
              <given-names>KI</given-names>
            </name>
            <name name-style="western">
              <surname>Lönnqvist</surname>
              <given-names>JK</given-names>
            </name>
          </person-group>
          <article-title>Mental disorders and comorbidity in suicide</article-title>
          <source>Am J Psychiatry</source>
          <year>1993</year>
          <volume>150</volume>
          <issue>6</issue>
          <fpage>935</fpage>
          <lpage>940</lpage>
          <pub-id pub-id-type="doi">10.1176/ajp.150.6.935</pub-id>
          <pub-id pub-id-type="medline">8494072</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Xiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chiu</surname>
              <given-names>Hf</given-names>
            </name>
            <name name-style="western">
              <surname>Ungvari</surname>
              <given-names>Gs</given-names>
            </name>
            <name name-style="western">
              <surname>Correll</surname>
              <given-names>Cu</given-names>
            </name>
            <name name-style="western">
              <surname>Lai</surname>
              <given-names>Ky</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Geng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Measurement-based care versus standard care for major depression: a randomized controlled trial with blind raters</article-title>
          <source>AJP</source>
          <year>2015</year>
          <month>10</month>
          <day>01</day>
          <volume>172</volume>
          <issue>10</issue>
          <fpage>1004</fpage>
          <lpage>1013</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ajp.psychiatryonline.org/doi/full/10.1176/appi.ajp.2015.14050652"/>
          </comment>
          <pub-id pub-id-type="doi">10.1176/appi.ajp.2015.14050652</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zimmerman</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ellison</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Young</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Chelminski</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dalrymple</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>How many different ways do patients meet the diagnostic criteria for major depressive disorder?</article-title>
          <source>Compr Psychiatry</source>
          <year>2015</year>
          <volume>56</volume>
          <fpage>29</fpage>
          <lpage>34</lpage>
          <pub-id pub-id-type="doi">10.1016/j.comppsych.2014.09.007</pub-id>
          <pub-id pub-id-type="medline">25266848</pub-id>
          <pub-id pub-id-type="pii">S0010-440X(14)00258-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soehner</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Kaplan</surname>
              <given-names>KA</given-names>
            </name>
            <name name-style="western">
              <surname>Harvey</surname>
              <given-names>AG</given-names>
            </name>
          </person-group>
          <article-title>Prevalence and clinical correlates of co-occurring insomnia and hypersomnia symptoms in depression</article-title>
          <source>J Affect Disord</source>
          <year>2014</year>
          <volume>167</volume>
          <fpage>93</fpage>
          <lpage>97</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/24953480"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jad.2014.05.060</pub-id>
          <pub-id pub-id-type="medline">24953480</pub-id>
          <pub-id pub-id-type="pii">S0165-0327(14)00355-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC4291280</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Corfield</surname>
              <given-names>EC</given-names>
            </name>
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>NG</given-names>
            </name>
            <name name-style="western">
              <surname>Nyholt</surname>
              <given-names>DR</given-names>
            </name>
          </person-group>
          <article-title>Co-occurrence and symptomatology of fatigue and depression</article-title>
          <source>Compr Psychiatry</source>
          <year>2016</year>
          <volume>71</volume>
          <fpage>1</fpage>
          <lpage>10</lpage>
          <pub-id pub-id-type="doi">10.1016/j.comppsych.2016.08.004</pub-id>
          <pub-id pub-id-type="medline">27567301</pub-id>
          <pub-id pub-id-type="pii">S0010-440X(16)30029-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Park</surname>
              <given-names>LT</given-names>
            </name>
            <name name-style="western">
              <surname>Zarate</surname>
              <given-names>CA</given-names>
            </name>
          </person-group>
          <article-title>Depression in the primary care setting</article-title>
          <source>N Engl J Med</source>
          <year>2019</year>
          <volume>380</volume>
          <issue>6</issue>
          <fpage>559</fpage>
          <lpage>568</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/30726688"/>
          </comment>
          <pub-id pub-id-type="doi">10.1056/NEJMcp1712493</pub-id>
          <pub-id pub-id-type="medline">30726688</pub-id>
          <pub-id pub-id-type="pmcid">PMC6727965</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bard</surname>
              <given-names>HA</given-names>
            </name>
            <name name-style="western">
              <surname>O'Driscoll</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>CB</given-names>
            </name>
            <name name-style="western">
              <surname>Henry</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Cape</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Espie</surname>
              <given-names>CA</given-names>
            </name>
          </person-group>
          <article-title>Insomnia, depression, and anxiety symptoms interact and individually impact functioning: a network and relative importance analysis in the context of insomnia</article-title>
          <source>Sleep Med</source>
          <year>2023</year>
          <volume>101</volume>
          <fpage>505</fpage>
          <lpage>514</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1389-9457(22)01254-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.sleep.2022.12.005</pub-id>
          <pub-id pub-id-type="medline">36527942</pub-id>
          <pub-id pub-id-type="pii">S1389-9457(22)01254-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cummins</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Dineley</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Conde</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Matcham</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Siddi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lamers</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Carr</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lavelle</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Leightley</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>White</surname>
              <given-names>KM</given-names>
            </name>
            <name name-style="western">
              <surname>Oetzmann</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Campbell</surname>
              <given-names>EL</given-names>
            </name>
            <name name-style="western">
              <surname>Simblett</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bruce</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Haro</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Penninx</surname>
              <given-names>BWJH</given-names>
            </name>
            <name name-style="western">
              <surname>Ranjan</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Rashid</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Stewart</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Folarin</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Bailón</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Schuller</surname>
              <given-names>BW</given-names>
            </name>
            <name name-style="western">
              <surname>Wykes</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Vairavan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Dobson</surname>
              <given-names>RJB</given-names>
            </name>
            <name name-style="western">
              <surname>Narayan</surname>
              <given-names>VA</given-names>
            </name>
            <name name-style="western">
              <surname>Hotopf</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Multilingual markers of depression in remotely collected speech samples: a preliminary analysis</article-title>
          <source>J Affect Disord</source>
          <year>2023</year>
          <volume>341</volume>
          <fpage>128</fpage>
          <lpage>136</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0165-0327(23)01076-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jad.2023.08.097</pub-id>
          <pub-id pub-id-type="medline">37598722</pub-id>
          <pub-id pub-id-type="pii">S0165-0327(23)01076-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dumpala</surname>
              <given-names>SH</given-names>
            </name>
            <name name-style="western">
              <surname>Dikaios</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Langley</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rempel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Uher</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Oore</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Manifestation of depression in speech overlaps with characteristics used to represent and recognize speaker identity</article-title>
          <source>Sci Rep</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>1</issue>
          <fpage>11155</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-023-35184-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-023-35184-7</pub-id>
          <pub-id pub-id-type="medline">37429935</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-023-35184-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC10333314</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bone</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Gibson</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chaspari</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Can</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Speech and language processing for mental health research and care</article-title>
          <year>2016</year>
          <conf-name>2016 50th Asilomar Conference on Signals, Systems and Computers</conf-name>
          <conf-date>November 6-9, 2016</conf-date>
          <conf-loc>Pacific Grove, CA</conf-loc>
          <fpage>831</fpage>
          <lpage>835</lpage>
          <pub-id pub-id-type="doi">10.1109/acssc.2016.7869164</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cummins</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Scherer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Krajewski</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Schnieder</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Epps</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Quatieri</surname>
              <given-names>TF</given-names>
            </name>
          </person-group>
          <article-title>A review of depression and suicide risk assessment using speech analysis</article-title>
          <source>Speech Commun</source>
          <year>2015</year>
          <volume>71</volume>
          <fpage>10</fpage>
          <lpage>49</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.specom.2015.03.004"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.specom.2015.03.004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Williamson</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Quatieri</surname>
              <given-names>TF</given-names>
            </name>
            <name name-style="western">
              <surname>Helfer</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Vocal and facial biomarkers of depression based on motor incoordination and timing</article-title>
          <year>2014</year>
          <conf-name>Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge</conf-name>
          <conf-date>November 7, 2014</conf-date>
          <conf-loc>Orlando, FL</conf-loc>
          <fpage>65</fpage>
          <lpage>72</lpage>
          <pub-id pub-id-type="doi">10.1145/2661806.2661809</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Teferra</surname>
              <given-names>BG</given-names>
            </name>
            <name name-style="western">
              <surname>Borwein</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>DeSouza</surname>
              <given-names>DD</given-names>
            </name>
            <name name-style="western">
              <surname>Simpson</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Rheault</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Acoustic and linguistic features of impromptu speech and their association with anxiety: validation study</article-title>
          <source>JMIR Ment Health</source>
          <year>2022</year>
          <volume>9</volume>
          <issue>7</issue>
          <fpage>e36828</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mental.jmir.org/2022/7/e36828/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/36828</pub-id>
          <pub-id pub-id-type="medline">35802401</pub-id>
          <pub-id pub-id-type="pii">v9i7e36828</pub-id>
          <pub-id pub-id-type="pmcid">PMC9308078</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Stasak</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Epps</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Schatten</surname>
              <given-names>HT</given-names>
            </name>
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>IW</given-names>
            </name>
            <name name-style="western">
              <surname>Provost</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Armey</surname>
              <given-names>MF</given-names>
            </name>
          </person-group>
          <article-title>Read speech voice quality and disfluency in individuals with recent suicidal ideation or suicide attempt</article-title>
          <source>Speech Commun</source>
          <year>2021</year>
          <volume>132</volume>
          <fpage>10</fpage>
          <lpage>20</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.specom.2021.05.004"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.specom.2021.05.004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bayerl</surname>
              <given-names>SP</given-names>
            </name>
            <name name-style="western">
              <surname>Wagner</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Baumann</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Bocklet</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Riedhammer</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Detecting vocal fatigue with neural embeddings</article-title>
          <source>J Voice</source>
          <year>2023</year>
          <fpage>11</fpage>
          <lpage>15</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jvoice.2023.01.012</pub-id>
          <pub-id pub-id-type="medline">36774263</pub-id>
          <pub-id pub-id-type="pii">S0892-1997(23)00011-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Elbéji</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Higa</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Fischer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Despotovic</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Nazarov</surname>
              <given-names>PV</given-names>
            </name>
            <name name-style="western">
              <surname>Aguayo</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Fagherazzi</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Vocal biomarker predicts fatigue in people with COVID-19: results from the prospective Predi-COVID cohort study</article-title>
          <source>BMJ Open</source>
          <year>2022</year>
          <volume>12</volume>
          <issue>11</issue>
          <fpage>e062463</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmjopen.bmj.com/lookup/pmidlookup?view=long&#38;pmid=36414294"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmjopen-2022-062463</pub-id>
          <pub-id pub-id-type="medline">36414294</pub-id>
          <pub-id pub-id-type="pii">bmjopen-2022-062463</pub-id>
          <pub-id pub-id-type="pmcid">PMC9684280</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thoret</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Andrillon</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Gauriau</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Léger</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Pressnitzer</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Sleep deprivation detected by voice analysis</article-title>
          <source>PLoS Comput Biol</source>
          <year>2024</year>
          <volume>20</volume>
          <issue>2</issue>
          <fpage>e1011849</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dx.plos.org/10.1371/journal.pcbi.1011849"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pcbi.1011849</pub-id>
          <pub-id pub-id-type="medline">38315733</pub-id>
          <pub-id pub-id-type="pii">PCOMPBIOL-D-23-00407</pub-id>
          <pub-id pub-id-type="pmcid">PMC10890756</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Levelt</surname>
              <given-names>WJM</given-names>
            </name>
          </person-group>
          <source>Speaking: From Intention to Articulation</source>
          <year>1989</year>
          <publisher-loc>Cambridge, MA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guenther</surname>
              <given-names>FH</given-names>
            </name>
          </person-group>
          <source>Neural Control of Speech</source>
          <year>2016</year>
          <publisher-loc>Cambridge, MA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Low</surname>
              <given-names>DM</given-names>
            </name>
            <name name-style="western">
              <surname>Bentley</surname>
              <given-names>KH</given-names>
            </name>
            <name name-style="western">
              <surname>Ghosh</surname>
              <given-names>SS</given-names>
            </name>
          </person-group>
          <article-title>Automated assessment of psychiatric disorders using speech: a systematic review</article-title>
          <source>Laryngoscope Investig Otolaryngol</source>
          <year>2020</year>
          <volume>5</volume>
          <issue>1</issue>
          <fpage>96</fpage>
          <lpage>116</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/32128436"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/lio2.354</pub-id>
          <pub-id pub-id-type="medline">32128436</pub-id>
          <pub-id pub-id-type="pii">LIO2354</pub-id>
          <pub-id pub-id-type="pmcid">PMC7042657</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Berta</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ángel</surname>
              <given-names>CM</given-names>
            </name>
            <name name-style="western">
              <surname>Clara</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Rubén</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>A bibliometric analysis of 10 years of research on symptom networks in psychopathology and mental health</article-title>
          <source>Psychiatry Res</source>
          <year>2022</year>
          <volume>308</volume>
          <fpage>114380</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0165-1781(21)00674-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.psychres.2021.114380</pub-id>
          <pub-id pub-id-type="medline">34999293</pub-id>
          <pub-id pub-id-type="pii">S0165-1781(21)00674-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Borsboom</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>A network theory of mental disorders</article-title>
          <source>World Psychiatry</source>
          <year>2017</year>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>5</fpage>
          <lpage>13</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28127906"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/wps.20375</pub-id>
          <pub-id pub-id-type="medline">28127906</pub-id>
          <pub-id pub-id-type="pmcid">PMC5269502</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>König</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tröger</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mallick</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Mina</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Linz</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Wagnon</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Karbach</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kuhn</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Peter</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Detecting subtle signs of depression with automated speech analysis in a non-clinical sample</article-title>
          <source>BMC Psychiatry</source>
          <year>2022</year>
          <volume>22</volume>
          <issue>1</issue>
          <fpage>830</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcpsychiatry.biomedcentral.com/articles/10.1186/s12888-022-04475-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12888-022-04475-0</pub-id>
          <pub-id pub-id-type="medline">36575442</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12888-022-04475-0</pub-id>
          <pub-id pub-id-type="pmcid">PMC9793349</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Nazreen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Rutowski</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Harati</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shriberg</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Chlebek</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Aratow</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Feasibility of a machine learning-based smartphone application in detecting depression and anxiety in a generally senior population</article-title>
          <source>Front Psychol</source>
          <year>2022</year>
          <volume>13</volume>
          <fpage>811517</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/35478769"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpsyg.2022.811517</pub-id>
          <pub-id pub-id-type="medline">35478769</pub-id>
          <pub-id pub-id-type="pmcid">PMC9037748</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Barocas</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hardt</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Narayanan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <source>Fairness and Machine Learning: Limitations and Opportunities</source>
          <year>2023</year>
          <publisher-loc>Cambridge, MA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>VP</given-names>
            </name>
            <name name-style="western">
              <surname>Rouas</surname>
              <given-names>JL</given-names>
            </name>
          </person-group>
          <article-title>Estimating symptoms and clinical signs instead of disorders: the path toward the clinical use of voice and speech biomarkers in psychiatry</article-title>
          <year>2024</year>
          <conf-name>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</conf-name>
          <conf-date>April 14-19, 2024</conf-date>
          <conf-loc>Seoul, South Korea</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icassp48485.2024.10445888</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bondi</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Koster</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Sheahan</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Chadwick</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Bachrach</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cemgil</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Paquet</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Dvijotham</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Role of human-AI interaction in selective prediction</article-title>
          <source>Proc AAAI Conf Artif Intell</source>
          <year>2022</year>
          <volume>36</volume>
          <issue>5</issue>
          <fpage>5286</fpage>
          <lpage>5294</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1609/aaai.v36i5.20465"/>
          </comment>
          <pub-id pub-id-type="doi">10.1609/aaai.v36i5.20465</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boschi</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Catricalà</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Consonni</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chesi</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Moro</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cappa</surname>
              <given-names>SF</given-names>
            </name>
          </person-group>
          <article-title>Connected speech in neurodegenerative language disorders: a review</article-title>
          <source>Front Psychol</source>
          <year>2017</year>
          <volume>8</volume>
          <fpage>269</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28321196"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpsyg.2017.00269</pub-id>
          <pub-id pub-id-type="medline">28321196</pub-id>
          <pub-id pub-id-type="pmcid">PMC5337522</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beck</surname>
              <given-names>AT</given-names>
            </name>
            <name name-style="western">
              <surname>Ward</surname>
              <given-names>CH</given-names>
            </name>
            <name name-style="western">
              <surname>Mendelson</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mock</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>An inventory for measuring depression</article-title>
          <source>Arch Gen Psychiatry</source>
          <year>1961</year>
          <volume>4</volume>
          <fpage>561</fpage>
          <lpage>574</lpage>
          <pub-id pub-id-type="doi">10.1001/archpsyc.1961.01710120031004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kroenke</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Spitzer</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>JB</given-names>
            </name>
          </person-group>
          <article-title>The PHQ-9: validity of a brief depression severity measure</article-title>
          <source>J Gen Intern Med</source>
          <year>2001</year>
          <volume>16</volume>
          <issue>9</issue>
          <fpage>606</fpage>
          <lpage>613</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/11556941"/>
          </comment>
          <pub-id pub-id-type="doi">10.1046/j.1525-1497.2001.016009606.x</pub-id>
          <pub-id pub-id-type="medline">11556941</pub-id>
          <pub-id pub-id-type="pii">jgi01114</pub-id>
          <pub-id pub-id-type="pmcid">PMC1495268</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cameron</surname>
              <given-names>IM</given-names>
            </name>
            <name name-style="western">
              <surname>Crawford</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Lawton</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Reid</surname>
              <given-names>IC</given-names>
            </name>
          </person-group>
          <article-title>Psychometric comparison of PHQ-9 and HADS for measuring depression severity in primary care</article-title>
          <source>Br J Gen Pract</source>
          <year>2008</year>
          <volume>58</volume>
          <issue>546</issue>
          <fpage>32</fpage>
          <lpage>36</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bjgp.org/lookup/pmidlookup?view=long&#38;pmid=18186994"/>
          </comment>
          <pub-id pub-id-type="doi">10.3399/bjgp08X263794</pub-id>
          <pub-id pub-id-type="medline">18186994</pub-id>
          <pub-id pub-id-type="pmcid">PMC2148236</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hawley</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Gale</surname>
              <given-names>TM</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>PSJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jain</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Farag</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kondan</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Avent</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Graham</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Equations for converting scores between depression scales (MÅDRS, SRS, PHQ-9 and BDI-II): good statistical, but weak idiographic, validity</article-title>
          <source>Hum Psychopharmacol</source>
          <year>2013</year>
          <volume>28</volume>
          <issue>6</issue>
          <fpage>544</fpage>
          <lpage>551</lpage>
          <pub-id pub-id-type="doi">10.1002/hup.2341</pub-id>
          <pub-id pub-id-type="medline">24519690</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beck</surname>
              <given-names>AT</given-names>
            </name>
            <name name-style="western">
              <surname>Steer</surname>
              <given-names>RA</given-names>
            </name>
            <name name-style="western">
              <surname>Garbin</surname>
              <given-names>MG</given-names>
            </name>
          </person-group>
          <article-title>Psychometric properties of the beck depression inventory: twenty-five years of evaluation</article-title>
          <source>Clin Psychol Rev</source>
          <year>1988</year>
          <volume>8</volume>
          <fpage>77</fpage>
          <lpage>100</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/0272-7358(88)90050-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/0272-7358(88)90050-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Spitzer</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Kroenke</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>JBW</given-names>
            </name>
            <name name-style="western">
              <surname>Löwe</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>A brief measure for assessing generalized anxiety disorder: the GAD-7</article-title>
          <source>Arch Intern Med</source>
          <year>2006</year>
          <volume>166</volume>
          <issue>10</issue>
          <fpage>1092</fpage>
          <lpage>1097</lpage>
          <pub-id pub-id-type="doi">10.1001/archinte.166.10.1092</pub-id>
          <pub-id pub-id-type="medline">16717171</pub-id>
          <pub-id pub-id-type="pii">166/10/1092</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soldatos</surname>
              <given-names>CR</given-names>
            </name>
            <name name-style="western">
              <surname>Dikeos</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Paparrigopoulos</surname>
              <given-names>TJ</given-names>
            </name>
          </person-group>
          <article-title>Athens Insomnia Scale: validation of an instrument based on ICD-10 criteria</article-title>
          <source>J Psychosom Res</source>
          <year>2000</year>
          <volume>48</volume>
          <issue>6</issue>
          <fpage>555</fpage>
          <lpage>560</lpage>
          <pub-id pub-id-type="doi">10.1016/s0022-3999(00)00095-7</pub-id>
          <pub-id pub-id-type="medline">11033374</pub-id>
          <pub-id pub-id-type="pii">S0022399900000957</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soldatos</surname>
              <given-names>CR</given-names>
            </name>
            <name name-style="western">
              <surname>Dikeos</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Paparrigopoulos</surname>
              <given-names>TJ</given-names>
            </name>
          </person-group>
          <article-title>The diagnostic validity of the Athens Insomnia Scale</article-title>
          <source>J Psychosom Res</source>
          <year>2003</year>
          <volume>55</volume>
          <issue>3</issue>
          <fpage>263</fpage>
          <lpage>267</lpage>
          <pub-id pub-id-type="doi">10.1016/s0022-3999(02)00604-9</pub-id>
          <pub-id pub-id-type="medline">12932801</pub-id>
          <pub-id pub-id-type="pii">S0022399902006049</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Okajima</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Miyamoto</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Ubara</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Omichi</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Matsuda</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sumi</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Matsuo</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ito</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Kadotani</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of severity levels of the Athens Insomnia Scale based on the criterion of Insomnia Severity Index</article-title>
          <source>Int J Environ Res Public Health</source>
          <year>2020</year>
          <volume>17</volume>
          <issue>23</issue>
          <fpage>8789</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=ijerph17238789"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/ijerph17238789</pub-id>
          <pub-id pub-id-type="medline">33256097</pub-id>
          <pub-id pub-id-type="pii">ijerph17238789</pub-id>
          <pub-id pub-id-type="pmcid">PMC7730071</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Smets</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Garssen</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Bonke</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>de Haes</surname>
              <given-names>JC</given-names>
            </name>
          </person-group>
          <article-title>The Multidimensional Fatigue Inventory (MFI) psychometric qualities of an instrument to assess fatigue</article-title>
          <source>J Psychosom Res</source>
          <year>1995</year>
          <volume>39</volume>
          <issue>3</issue>
          <fpage>315</fpage>
          <lpage>325</lpage>
          <pub-id pub-id-type="doi">10.1016/0022-3999(94)00125-o</pub-id>
          <pub-id pub-id-type="medline">7636775</pub-id>
          <pub-id pub-id-type="pii">002239999400125O</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schwarz</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Krauss</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Hinz</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Fatigue in the general population</article-title>
          <source>Onkologie</source>
          <year>2003</year>
          <volume>26</volume>
          <issue>2</issue>
          <fpage>140</fpage>
          <lpage>144</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1159/000069834"/>
          </comment>
          <pub-id pub-id-type="doi">10.1159/000069834</pub-id>
          <pub-id pub-id-type="medline">12771522</pub-id>
          <pub-id pub-id-type="pii">69834</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gentile</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Delarozière</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Favre</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Sambuc</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>San Marco</surname>
              <given-names>JL</given-names>
            </name>
          </person-group>
          <article-title>Validation of the French 'Multidimensional Fatigue Inventory' (MFI 20)</article-title>
          <source>Eur J Cancer Care (Engl)</source>
          <year>2003</year>
          <volume>12</volume>
          <issue>1</issue>
          <fpage>58</fpage>
          <lpage>64</lpage>
          <pub-id pub-id-type="doi">10.1046/j.1365-2354.2003.00295.x</pub-id>
          <pub-id pub-id-type="medline">12641557</pub-id>
          <pub-id pub-id-type="pii">295</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hinz</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Barboza</surname>
              <given-names>CF</given-names>
            </name>
            <name name-style="western">
              <surname>Barradas</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Körner</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Beierlein</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Singer</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Fatigue in the general population of Colombia—normative values for the Multidimensional Fatigue Inventory MFI-20</article-title>
          <source>Onkologie</source>
          <year>2013</year>
          <volume>36</volume>
          <issue>7-8</issue>
          <fpage>403</fpage>
          <lpage>407</lpage>
          <pub-id pub-id-type="doi">10.1159/000353606</pub-id>
          <pub-id pub-id-type="medline">23921758</pub-id>
          <pub-id pub-id-type="pii">000353606</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Deep residual learning for image recognition</article-title>
          <year>2016</year>
          <conf-name>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>
          <conf-date>June 27-30, 2016</conf-date>
          <conf-loc>Las Vegas, NV</conf-loc>
          <fpage>770</fpage>
          <lpage>778</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.90</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaswani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Parmar</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Uszkoreit</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Attention is all you need</article-title>
          <year>2017</year>
          <conf-name>Advances in Neural Information Processing Systems</conf-name>
          <conf-date>December 4-9, 2017</conf-date>
          <conf-loc>Long Beach, CA</conf-loc>
          <fpage>6000</fpage>
          <lpage>6010</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/10.5555/3295222.3295349"/>
          </comment>
          <pub-id pub-id-type="doi">10.5555/3295222.3295349</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hsu</surname>
              <given-names>WN</given-names>
            </name>
            <name name-style="western">
              <surname>Bolte</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Tsai</surname>
              <given-names>YHH</given-names>
            </name>
            <name name-style="western">
              <surname>Lakhotia</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mohamed</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>HuBERT: self-supervised speech representation learning by masked prediction of hidden units</article-title>
          <source>IEEE/ACM Transactions on Audio, Speech, and Language Processing</source>
          <year>2021</year>
          <volume>29</volume>
          <fpage>3451</fpage>
          <lpage>3461</lpage>
          <pub-id pub-id-type="doi">10.1109/taslp.2021.3122291</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arias-Vergara</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Vásquez-Correa</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Orozco-Arroyave</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Nöth</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Speaker models for monitoring Parkinson’s disease progression considering different communication channels and acoustic conditions</article-title>
          <source>Speech Commun</source>
          <year>2018</year>
          <volume>101</volume>
          <fpage>11</fpage>
          <lpage>25</lpage>
          <pub-id pub-id-type="doi">10.1016/j.specom.2018.05.007</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chung</surname>
              <given-names>JS</given-names>
            </name>
            <name name-style="western">
              <surname>Nagrani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zisserman</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>VoxCeleb2: deep speaker recognition</article-title>
          <year>2018</year>
          <conf-name>Interspeech 2018</conf-name>
          <conf-date>September 2-6, 2018</conf-date>
          <conf-loc>Hyderabad, India</conf-loc>
          <fpage>1086</fpage>
          <lpage>1090</lpage>
          <pub-id pub-id-type="doi">10.21437/interspeech.2018-1929</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Additive margin softmax for face verification</article-title>
          <source>IEEE Signal Process Lett</source>
          <year>2018</year>
          <volume>25</volume>
          <issue>7</issue>
          <fpage>926</fpage>
          <lpage>930</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/LSP.2018.2822810"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/lsp.2018.2822810</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Morais</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Hoory</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Speech emotion recognition using self-supervised features</article-title>
          <year>2022</year>
          <conf-name>ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</conf-name>
          <conf-date>May 23-27, 2022</conf-date>
          <conf-loc>Singapore</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icassp43922.2022.9747870</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Panayotov</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Povey</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Librispeech: an ASR corpus based on public domain audio books</article-title>
          <year>2015</year>
          <conf-name>2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</conf-name>
          <conf-date>April 19-24, 2015</conf-date>
          <conf-loc>South Brisbane, Queensland, Australia</conf-loc>
          <fpage>5206</fpage>
          <lpage>5210</lpage>
          <pub-id pub-id-type="doi">10.1109/icassp.2015.7178964</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chung S</surname>
              <given-names>JS</given-names>
            </name>
            <name name-style="western">
              <surname>Huh</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mun</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Heo</surname>
              <given-names>HS</given-names>
            </name>
            <name name-style="western">
              <surname>Choe</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ham</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>In defence of metric learning for speaker recognition</article-title>
          <year>2024</year>
          <conf-name>Interspeech 2020</conf-name>
          <conf-date>October 25-29, 2020</conf-date>
          <conf-loc>Shanghai, China</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.isca-archive.org/interspeech_2020/chung20b_interspeech.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pedregosa</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Varoquaux</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Gramfort</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Scikit-learn: machine learning in Python</article-title>
          <source>J Mach Learn Res</source>
          <year>2011</year>
          <volume>12</volume>
          <fpage>2825</fpage>
          <lpage>2830</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmlr.org/papers/volume12/pedregosa11a/pedregosa11a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agarwal</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Beygelzimer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dudik</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Langford</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wallach</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>A reductions approach to fair classification</article-title>
          <year>2018</year>
          <conf-name>Proceedings of the 35th International Conference on Machine Learning</conf-name>
          <conf-date>March 6, 2018</conf-date>
          <conf-loc>Stockholm, Sweden</conf-loc>
          <fpage>66</fpage>
          <lpage>69</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://proceedings.mlr.press/v80/agarwal18a/agarwal18a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weerts</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Dudík</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Edgar</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Jalali</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lutz</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Fairlearn: assessing and improving fairness of AI systems</article-title>
          <source>J Mach Learn Res</source>
          <year>2023</year>
          <volume>24</volume>
          <fpage>1</fpage>
          <lpage>8</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmlr.org/papers/volume24/23-0389/23-0389.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kompa</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Snoek</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Beam</surname>
              <given-names>AL</given-names>
            </name>
          </person-group>
          <article-title>Second opinion needed: communicating uncertainty in medical machine learning</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>4</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-020-00367-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-020-00367-3</pub-id>
          <pub-id pub-id-type="medline">33402680</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-020-00367-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC7785732</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Maloney</surname>
              <given-names>LT</given-names>
            </name>
          </person-group>
          <article-title>Ubiquitous log odds: a common representation of probability and frequency distortion in perception, action, and cognition</article-title>
          <source>Front Neurosci</source>
          <year>2012</year>
          <volume>6</volume>
          <fpage>1</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22294978"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fnins.2012.00001</pub-id>
          <pub-id pub-id-type="medline">22294978</pub-id>
          <pub-id pub-id-type="pmcid">PMC3261445</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>The art of abstention: selective prediction and error regularization for natural language processing</article-title>
          <year>2021</year>
          <conf-name>Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing</conf-name>
          <conf-date>August 28, 2021</conf-date>
          <conf-loc>Online</conf-loc>
          <fpage>1040</fpage>
          <lpage>1051</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/2021.acl-long.84</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>El-Yaniv</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Wiener</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>On the foundations of noise-free selective classification</article-title>
          <source>J Mach Learn Res</source>
          <year>2010</year>
          <fpage>1605</fpage>
          <lpage>1641</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://jmlr2020.csail.mit.edu/papers/volume11/el-yaniv10a/el-yaniv10a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref61">
        <label>61</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hendrycks</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Gimpel</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>A baseline for detecting misclassified and out-of-distribution examples in neural networks</article-title>
          <source>ArXiv. Preprint posted online on October 3, 2018</source>
          <pub-id pub-id-type="doi">10.48550/arXiv.1610.02136</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref62">
        <label>62</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Varoquaux</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Cross-validation failure: small sample sizes lead to large error bars</article-title>
          <source>Neuroimage</source>
          <year>2018</year>
          <volume>180</volume>
          <issue>Pt A</issue>
          <fpage>68</fpage>
          <lpage>77</lpage>
          <pub-id pub-id-type="doi">10.1016/j.neuroimage.2017.06.061</pub-id>
          <pub-id pub-id-type="medline">28655633</pub-id>
          <pub-id pub-id-type="pii">S1053-8119(17)30531-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref63">
        <label>63</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tasnim</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ehghaghi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Diep</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Novikova</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>DEPAC: a corpus for depression and anxiety detection from speech</article-title>
          <year>2022</year>
          <conf-name>Proceedings of the Eighth Workshop on Computational Linguistics and Clinical Psychology</conf-name>
          <conf-date>July 17, 2022</conf-date>
          <conf-loc>Seattle, WA</conf-loc>
          <fpage>1</fpage>
          <lpage>16</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/2022.clpsych-1.1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref64">
        <label>64</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rutowski</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Harati</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shriberg</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Toward corpus size requirements for training and evaluating depression risk models using spoken language</article-title>
          <year>2022</year>
          <conf-name>Interspeech 2022</conf-name>
          <conf-date>September 18-22, 2022</conf-date>
          <conf-loc>Incheon, South Korea</conf-loc>
          <fpage>3343</fpage>
          <lpage>3347</lpage>
          <pub-id pub-id-type="doi">10.21437/interspeech.2022-10888</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref65">
        <label>65</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Teferra</surname>
              <given-names>BG</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Predicting generalized anxiety disorder from impromptu speech transcripts using context-aware transformer-based neural networks: model evaluation study</article-title>
          <source>JMIR Ment Health</source>
          <year>2023</year>
          <volume>10</volume>
          <fpage>e44325</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mental.jmir.org/2023//e44325/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/44325</pub-id>
          <pub-id pub-id-type="medline">36976636</pub-id>
          <pub-id pub-id-type="pii">v10i1e44325</pub-id>
          <pub-id pub-id-type="pmcid">PMC10131846</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref66">
        <label>66</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kolc</surname>
              <given-names>KL</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>YXK</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>AZY</given-names>
            </name>
            <name name-style="western">
              <surname>Shvetcov</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mitchell</surname>
              <given-names>PB</given-names>
            </name>
            <name name-style="western">
              <surname>Perkes</surname>
              <given-names>IE</given-names>
            </name>
          </person-group>
          <article-title>Measuring psychiatric symptoms online: a systematic review of the use of inventories on Amazon Mechanical Turk (mTurk)</article-title>
          <source>J Psychiatr Res</source>
          <year>2023</year>
          <volume>163</volume>
          <fpage>118</fpage>
          <lpage>126</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0022-3956(23)00213-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jpsychires.2023.05.027</pub-id>
          <pub-id pub-id-type="medline">37209617</pub-id>
          <pub-id pub-id-type="pii">S0022-3956(23)00213-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref67">
        <label>67</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>McCallum</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Batterham</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Calear</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Sunderland</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Carragher</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Kazan</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Associations of fatigue and sleep disturbance with nine common mental disorders</article-title>
          <source>J Psychosom Res</source>
          <year>2019</year>
          <volume>123</volume>
          <fpage>109727</fpage>
          <pub-id pub-id-type="doi">10.1016/j.jpsychores.2019.05.005</pub-id>
          <pub-id pub-id-type="medline">31376877</pub-id>
          <pub-id pub-id-type="pii">S0022-3999(19)30156-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref68">
        <label>68</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sorokowski</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Pisanski</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Frąckowiak</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Kobylarek</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Groyecka-Bernard</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Voice-based judgments of sex, height, weight, attractiveness, health, and psychological traits based on free speech versus scripted speech</article-title>
          <source>Psychon Bull Rev</source>
          <year>2024</year>
          <volume>31</volume>
          <issue>4</issue>
          <fpage>1680</fpage>
          <lpage>1689</lpage>
          <pub-id pub-id-type="doi">10.3758/s13423-023-02445-5</pub-id>
          <pub-id pub-id-type="medline">38238560</pub-id>
          <pub-id pub-id-type="pii">10.3758/s13423-023-02445-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref69">
        <label>69</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Amir</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Biron-Shental</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>The impact of hormonal fluctuations on female vocal folds</article-title>
          <source>Curr Opin Otolaryngol Head Neck Surg</source>
          <year>2004</year>
          <volume>12</volume>
          <issue>3</issue>
          <fpage>180</fpage>
          <lpage>184</lpage>
          <pub-id pub-id-type="doi">10.1097/01.moo.0000120304.58882.94</pub-id>
          <pub-id pub-id-type="medline">15167026</pub-id>
          <pub-id pub-id-type="pii">00020840-200406000-00006</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref70">
        <label>70</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beton</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yücel</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Başak</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Çiler Büyükatalay</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>The elderly voice: mechanisms, disorders and treatment methods</article-title>
          <source>Turk Arch Otorhinolaryngol</source>
          <year>2022</year>
          <volume>60</volume>
          <issue>4</issue>
          <fpage>220</fpage>
          <lpage>226</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/37456599"/>
          </comment>
          <pub-id pub-id-type="doi">10.4274/tao.2022.2022-8-1</pub-id>
          <pub-id pub-id-type="medline">37456599</pub-id>
          <pub-id pub-id-type="pii">60533</pub-id>
          <pub-id pub-id-type="pmcid">PMC10339270</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
