Review
Abstract
Background: Depression is highly recurrent and heterogeneous. The unobtrusive, continuous collection of mobile sensing data via smartphones and wearable devices offers a promising approach to monitor and predict individual depression trajectories, distinguish illness states, and anticipate changes in symptom severity.
Objective: This systematic review evaluates whether objective data from wearable devices and smartphones can (1) monitor and distinguish different states of depression, (2) predict changes in symptom severity, and (3) identify clinically relevant objective features for tracking and forecasting depression within diagnosed individuals.
Methods: We searched PubMed and Web of Science databases for English-language studies (published 2012-2022) that used smartphone or wearable device data, included participants aged ≥14 years with a depression diagnosis, and collected continuous data for at least 12 weeks.
Results: Out of 12,997 peer-reviewed articles, 9 original studies met the inclusion criteria, with sample sizes ranging from 45 to 2200 and durations of 12-52 weeks. Of the 9 studies, 3 used smartphone data, 1 used wearable device data, and 5 used both data types. Commonly collected variables were step count, distance moved, smartphone usage, call logs, sleep, heart rate, light exposure, and speech patterns. One study (11%) successfully differentiated between depressive states (worsening, relapse, or recovery). Six studies (67%) showed that mobile sensing data could predict depressive episodes or symptom severity. Four studies reported the predictive accuracy for depression using mobile sensing data from smartphones and wearable devices, ranging from 81% to 91%. Higher accuracy was achieved with personalized models or multimodal data.
Conclusions: Real-time passive monitoring via wearable devices and smartphones holds promise for personalized self-management, but key gaps remain, such as a lack of longitudinal and long-term studies with data collection for 1 year or longer, studies with confirmatory parameters on an individual level, and studies with a strong correlation between parameters in individual patients to support clinical decision-making. Improvements in reporting standards are highly recommended to provide better-informed insights for clinicians. Throughout this process, there is a clear need to address various other issues, such as limited types of collected data, reliability, user adherence, and privacy concerns.
Trial Registration: PROSPERO CRD42022355696; https://www.crd.york.ac.uk/PROSPERO/view/CRD42022355696
doi:10.2196/57418
Keywords
Introduction
Background
Depression (International Classification of Diseases, version 10 [ICD-10]: F32, F33) is a highly prevalent, usually recurrent or chronic, severe, and sometimes life-threatening disease []. The World Health Organization proclaimed depression as one of the leading causes of disease and disability in the world []. The duration of major depressive episodes has been found to vary widely, with median durations ranging from 3 to 6 months [-]. The Collaborative Depression Study reported that a primary episode usually recovers within 1 year, and the initial recurrence rate is 25%-40% after 2 years []. Broadly, individuals with depression might experience 5 to 9 episodes of depression in their lifetime, with each episode potentially presenting with worsened symptoms and having a progressively devastating impact on well-being []. In between the episodes, there are symptom-free intervals, during which well-managed strategies might prevent the risk of a worsening status [,]. If the period of remission and recovery could be guided by personalized self-management strategies, it would improve the experience and quality of life for patients diagnosed with depression []. However, until now, this has not been feasible, as most studies have identified group-based risk factors rather than within-individual risk factors [-]. For example, several studies identified sleep duration as an important predictor of depressive states and depressive episodes in group-based analysis [,]. Likewise, there was an association between longer sleep duration and worsening of depressive core symptoms among some individuals, but the causal direction varied between duration of sleep and depression in each individual []. Due to the heterogeneity of the course of depression in individuals, intraindividual effects may differ from those found in between-participant studies. This is where n-of-1 studies come into play, as they allow longitudinal testing of the hypotheses within individuals [].
During the last few decades, digital technologies have remarkably permeated and reshaped the daily lives of people. Globally, there are approximately 6.84 billion smartphone users []. To put that into perspective, the figure accounts for around 85% of the 8 billion global population.
Patients can be collectors and owners of large, diverse, long-term datasets through biosensors (wearable devices, including activity trackers), and “biosensor” (short form of biological sensor) data have been defined as physiological measures combined with biological components detected by an analytical device (eg, glucometer, pulse oximeter, and smartwatch) []. This raises the question of whether these patient-generated bio- and self-monitoring data can be used for the benefit of patients and systematically integrated into the care processes [-].
It has been almost 2 decades since the notion of digital interventions emerged, and efforts have been made to monitor the states of depression and predict them using sensor data. For instance, wearable devices (electronic devices designed to be worn on the body, with the capability to collect, process, and sometimes transmit data) monitor physiological parameters uninterruptedly with high precision []. These devices make continual passive sensing easier, that is, recording data of a person with only minimal effort required from their side (eg, to wear and charge the device regularly). Using smartphones and wearable devices may therefore provide an easy and effective way to monitor depressive symptoms, and parameters derived from these monitoring data could serve as digital markers of mood symptoms in depression []. Such data could allow further exploration of patterns that are related to the severity and changes in the severity of depressive symptoms.
Objectives
To our knowledge, numerous systematic and scoping reviews have examined the growing body of research on wearable devices and smartphone-based passive sensing [,-]. However, the included studies were cross-sectional or had short durations (ranging from 1 day to 8 weeks), limiting their ability to capture the long-term course of depressive disorders. Longer studies are better suited for identifying patterns related to symptom severity and its fluctuations over time within the individual. This systematic review aims to fill this gap in the psychiatric and biomedical informatics literature by being the first to focus on longitudinal studies with continuous passive sensor-based monitoring lasting more than 3 months.
Our aim was to systematically review longitudinal studies on patients with depressive disorders and answer the following research questions:
- Can objective data, automatically collected from wearable devices and smartphones, be used to effectively monitor and differentiate between varying states of depression within individuals with a clinical diagnosis of depression?
- Can objective data, automatically collected from wearable devices and smartphones, be used to predict changes in depressive symptom severity within individuals with a clinical diagnosis of depression?
- Which objective features derived from wearable devices and smartphones are clinically relevant for monitoring and predicting depression within individuals?
Methods
Design
This systematic review was conducted based on the guidelines outlined by the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) statement for transparent and comprehensive reporting of the methodology and results [].
To minimize the likelihood of researcher bias, the search strategy, inclusion criteria, and data extraction were specified and preregistered as an online protocol (PROSPERO registration number: CRD42022355696).
Search Strategy and Selection Criteria
A preliminary literature search in PubMed was performed using key terms related to depression and biosensors (performed in March 2022). The retrieved articles were used to identify further keywords and build an adequate search string.
We included studies assessing the contribution of available sensor technologies (eg, wearable devices and smartphones) in terms of monitoring and predicting depressive symptoms (depression progression and any fluctuations, ie, improvement or worsening, in symptoms) for a minimum period of 3 months among individuals.
Inclusion Criteria
The inclusion criteria were as follows: (1) patients belonging to any population with a primary or co-morbid diagnosis of clinical depression (eg, individuals with postpartum or perinatal depression; individuals from the LGBT [lesbian, gay, bisexual, transgender] community; patients with chronic obstructive pulmonary disease, cancer, parkinsonism, and diabetes; students; substance users; individuals with obesity; and smokers); (2) patients from any age group (eg, adolescents [age 14-18 years] and elderly individuals); (3) articles involving any study design (longitudinal, cross-sectional, retrospective, prospective, interventional, observational, etc); (4) studies involving longitudinal monitoring or prediction of depressive symptoms for a minimum period of 3 months; (5) studies published between 2012 and 2022; (6) studies using smartphones, wearable devices, or other sensory devices to assess or monitor clinical depression; (7) studies involving any context, such as geographical location and cultural factors; and (8) studies using any outcome measures for depression.
Exclusion Criteria
The exclusion criteria were as follows: (1) diary studies (only self-report, no biosensor data collected); (2) non-English articles; (3) opinion pieces, overview articles, reviews, notes, case reports, letters to the editor, extended abstracts, proceedings, patents, editorials, commentaries, conference proceedings, website discussions, blogs, and magazine and newspaper articles; and (4) articles not openly available for full-text reading.
Definitions
“Smartphone” was defined in line with its MeSH (Medical Subject Headings) term as any phone equipped with a mobile operating system (Android, Apple iOS, Symbian OS, or Windows Mobile) on which apps can be installed to capture data from the phone’s sensors, along with internet connectivity.
“Passive” data collection was defined as data collection without user input or without active participation (eg, collected without disrupting normal activities) []. “Active” data collection was defined as data collection through active or conscious participation (eg, journaling or diary technique).
Literature Search
To collect relevant publications, an electronic database search was conducted. We searched in 2 high-order databases, PubMed (primary database for this work) and Web of Science (also known as Web of Knowledge; secondary database), using Boolean search operators and combinations of keywords of conditions (major depressive disorder and depressive disorder) and technology (smartphone, sensor data, wearable device, mobile phone, smartphone application, sensor, mobile app [only used in PubMed], mHealth, ecological momentary intervention, and digital phenotype). The search was conducted initially on May 17, 2022, and updated on June 7, 2022 (PubMed) and June 9, 2022 (Web of Science). The search term “mobile app” identified a high number of articles that were not matching with the objectives of this review during our search in PubMed. Therefore, we only used “smartphone application” (but not “mobile app”) as a search term during the second search (Web of Science). Using the same search strings, the searches were repeated on February 1, 2023, to include articles from June to December 2022. To identify any further eligible studies and minimize selection bias, an additional search of the reference lists of retrieved articles (pearling) was conducted.
This search identified a total of 12,997 articles. After eliminating duplicate papers (n=112), the initial title and abstract screening took place, and the retrieved papers were screened against the inclusion and exclusion criteria (eg, nonhuman experiments; mobile phone addiction topics; focus on diary methods, which only involve subjective data; and nonmedical-related topics such as bipolar electricity). A total of 49 articles were retrieved for full-text review, and 9 articles met all the requirements for inclusion.
In order to make this search replicable in the future, more details are provided in and .
| Search string | Database | Search information (search date, search duration, and hits [N=12,997]) |
| (Depression*[Title/Abstract] OR Major Depressive Disorder*[Title/Abstract] OR Depressive disorder*[Title/Abstract] OR major depression*[Title/Abstract] OR unipolar depressive disorder*[Title/Abstract]) AND (smartphone*[Title/Abstract] OR sensor data[Title/Abstract] OR wearable device*[Title/Abstract] OR mobile phone*[Title/Abstract] OR smartphone application*[Title/Abstract] OR sensor*[Title/Abstract] OR mobile app*[Title/Abstract] OR mHealth[Title/Abstract] OR ecological momentary intervention[Title/Abstract] OR digital phenotype*[Title/Abstract]) | PubMed |
|
| (AB=Depression* OR AB=Major Depressive Disorder* OR AB=Depressive disorder* OR AB=major depression* OR AB=unipolar depressive disorder*) AND (AB=smartphone* OR AB=mHealth OR AB=sensor data OR AB=wearable device* OR AB=mobile phone* OR AB=smartphone application* OR AB=sensor* OR AB=ecological momentary intervention* OR AB=digital phenotype*) | Web of Science |
|

Study Selection
To control the risk of bias, the PRISMA 2020 recommendations for systematic literature analysis were followed. The PRISMA checklist has been prepared according to the PRISMA 2020 statement () []. Studies were independently selected by 3 different authors (RA, SS, and HO), who first analyzed the titles and abstracts and subsequently selected the full papers that met the inclusion criteria, resolving disagreements through consensus. The reasons for rejection were annotated. Subsequently, the authors retrieved the full-text copies of the remaining articles and selected those meeting the inclusion criteria. Disagreements were resolved through discussion and consensus.
Data Extraction and Synthesis
A systematic extraction form (Microsoft Excel data abstraction sheet) was used for each article to collect the following data: general data (authors, article title, type of publication, year of publication, and journal), study characteristics (aim of the study, study design, main variables of interest, type of electronic device, type of biosensor data, study length, and trial quality), participant data (number of participants, mean age of participants, type of control group, inclusion and exclusion criteria, and dropout rates), and outcome measures (primary outcomes, compliance rates, and missing data).
Data extraction was performed by 3 different authors (RA, SS, and HO), and discrepancies were resolved by internal discussion and majority agreement. Data synthesis was conducted for the extracted data. A narrative synthesis was applied due to the heterogeneity of the extracted data.
Critical Appraisal
Three independent reviewers (RA, SS, and HO) assessed the quality of all included studies, using Critical Appraisal Skills Program (CASP) checklists []. All articles were assessed following CASP questions, which determine if all the required steps for successful scientific reporting are completed and if the relevant information is presented clearly in the selected papers.
Results
Study Selection and Flow
The search yielded 12,997 records based on our search parameters across the PubMed and Web of Science databases. After screening the abstracts (n=116), 49 articles potentially eligible for inclusion were retrieved for full-text review. Many studies were excluded due to irrelevance, meaning that their abstracts did not mention the use of wearable technologies or smart devices for the detection of depression, or they did not use wearable devices to measure any physiological parameters. We conducted a further investigation to identify suitable articles through “pearling.” Ultimately, a total of 9 articles were included in the analysis. The most frequent reasons for exclusion were the duration of the study period and the length of continuous data collection [,-]. We collected information about study designs, patient populations, types of digital health interventions, and their intended uses. The PRISMA flowchart is presented in . Given the variability in study populations, study designs, and outcomes, we performed a descriptive analysis of the data. Comprehensive details of each study are presented in .

| Study | Digital technology | Data source | Duration of longitudinal data collection | ||||||
| Wearable | Smartphone app | Platform | Developer | PROMsa | Sensor data | ||||
| Tønning et al [], 2021 | No | Yes | Android and iOS | Monsenso | Mood and activity (daily) | Physical activity (steps and total distance), social activity, and smartphone usage (call logs and text messages) | PROMs: average of 130 days/participant; sensor data: average of 158 days/participant (data collection) | ||
| McIntyre et al [], 2021 | No | “mind.me” app | Android | Mind Mental Health Technologies Inc | PHQ-9b (monthly) | Daily call count, daily SMS text message count, location variance, normalized entropy, number of clusters, total distance (km), and mean absolute deviation in distance (km) | 12 weeks (study duration) | ||
| Sarda et al [], 2019 | No | Yes | Android | Touchkin | PHQ-9 (biweekly) | Accelerometer, GPS, and ambient light sensor | 20 weeks (study duration) | ||
| Fang et al [], 2021 | Fitbit Charge 2 | Yes | —c | Remedy Health Media LLC | Mood (daily), PHQ-9 (quarterly) | Sleep parameters | Sensor data: average of 115 (SD 111) days/participant (data collection) | ||
| Cho et al [], 2020 | Fitbit Charge HR 2 or 3 | “Circadian Rhythm for Mood (CRM)” app; daily “eMoodChart” | Android and iOS | — | Sleep (daily) | Light exposure (Android only), activity, and heart rate | 52 weeks (study duration) | ||
| Mullick et al [], 2022 | Fitbit Inspire HR, software version 1.84.5 | “AWARE” app | Android and iOS | — | PHQ-9 (weekly) | Heart rate, sleep, steps, calls, conversations, location, Wi-Fi, and screen use | 24 weeks (study duration) | ||
| Zhang et al [], 2021 | Fitbit Charge 2 or 3 | Several remote technology apps | — | — | PHQ-8d (biweekly) | Sleep, physical activity, stress, speech patterns, and cognitive function | Up to 2 years (study duration) | ||
| Lee et al [], 2022 | Fitbit Charge HR 2 or 3 | “eMoodChart” | Android and iOS | Study team | Mood (daily) | Light exposure (Android only), steps, heart rate, and sleep | Average of 279.7 days/participant (data collection) | ||
| Bai et al [], 2021 | Wristband Mi Band 2, Xiaomi Corporation | “Mood Mirror” | Android | — | Mood (daily), PHQ-9 (biweekly) | Sleep data, step count, heart rate, and phone usage (call logs, text messages, app usage, GPS, and screen time) | 12 weeks (study duration) | ||
aPROMs: patient-reported outcome measures.
bPHQ-9: Patient Health Questionnaire-9.
cNot applicable.
dPHQ-8: Patient Health Questionnaire-8.
Study Characteristics
We present a narrative synthesis of the major themes extracted from the 9 studies. This will be followed by a summary of the characteristics of the digital health technologies (DHTs) and a section addressing the findings related to our research questions. Finally, we will summarize key challenges of the approach.
Characteristics of the Included Studies
The characteristics of the studies reported in the selected articles are comprehensively presented in .
Among the 9 articles, only 1 discussed the results of a randomized study []. Eight studies were nonrandomized, single-arm studies [-]. Among the studies, 3 assessed the feasibility and validity of DHTs [,], while 3 others evaluated the efficacy or effectiveness of DHTs [-]. Across the chosen studies, participant recruitment ranged from 45 to 2200 individuals. However, justifications for the chosen sample sizes were seldom provided. The reporting of recruitment strategies was notably concise [-]. Study durations varied, ranging from 12 weeks [] to 1 year [,]. Either the study completion rate or adherence to the technology was reported. Geographically, 4 studies originated in Asia [,,,], 2 in Europe [,], and 3 in North America [,,]. Among these, 2 were multisite studies (different countries) [,]. The study by Sarda et al [] mentioned providing data recharge (1 GB) per month to cover usage costs without any additional incentives. In the remaining studies, information concerning economic compensation for participation was not provided.
Characteristics of Patient Populations
All studies included patients diagnosed with major depressive disorder and unipolar depressive disorder according to ICD-10. Female participants consistently outnumbered male participants in all studies [-]. All studies enrolled adults aged >18 years [-,-], except for 1 study that included adolescents aged 14-18 years []. One study recruited participants from residency training [], while the remaining studies included participants from diverse backgrounds.
Summary of Collected Data
The studies included in the analysis used the Patient Health Questionnaire-8/-9 (PHQ-8/-9) [-,] and Beck Depression Inventory-Second Edition [] self-report instruments as principal outcome measures. There were variations in the timing of outcome assessment (daily [,,,,], weekly [], biweekly (every 2 weeks) [,,], monthly [], and quarterly [] assessments).
In addition to self-ratings of depressive symptoms, the most commonly collected data included the number of steps per day [,] and the total distance moved per day [,], which were based on GPS [,], Wi-Fi signals [], and mobile cell towers. Smartphone usage metrics, such as total screen-on time per day and the number of times the screen was turned on per day, were also collected [].
Moreover, the studies gathered data on device analytics, encompassing call logs (daily call count and the duration of calls per day) [,] and SMS text message patterns (daily frequency). One study also incorporated accelerometer and ambient light sensor data into the data collection efforts [].
Characteristics of DHTs
Electronic Devices and Use of Sensors
The DHTs evaluated across the studies could be broadly classified into 2 categories: mobile or smartphone apps and wearable devices. In certain studies, smartphone apps were employed to collect patient data in real time, coupled with daily mood and symptom monitoring through self-reports [-,,,]. In 6 studies, wearable devices, such as Fitbit Charge, were used in conjunction with smartphone apps to passively accumulate activity data [-]. One study used daily push notifications at a specific time to assess mood []. Another study implemented SMS text message reminders if participants missed mood recording sessions [].
Operating Systems
Four studies used the Android operating system [,,,], while another study used both Apple iOS and Android []. The operating system used in the remaining studies could not be ascertained.
Data Processing and Use
All studies conducted time-series analysis. Among them, 3 studies conducted comparisons within individuals [,,], while the other studies conducted comparisons between individuals and between groups [-,]. Complex calculations, such as data classification and prediction modeling, were performed in all studies, but not directly on the smartphone. For some studies, it could not be determined whether a remote server was used [,,]. In 7 studies, data were processed and correlated with clinical self-rating scales for depression as validation measures to test the validity of interpretations made through passive sensing [-,,,]. The correlation analysis was typically performed after the study’s completion. Various families of algorithms were employed for data processing to interpret or predict participant status in different studies.
Usability and Effectiveness of Health Technologies
In 1 study, user satisfaction with the app was assessed using a qualitative questionnaire, and all respondents who completed the questionnaire indicated a high level of satisfaction with the app []. Overall, the selected studies indicated high feasibility and acceptability of technologies for mental health care.
Synthesis of Findings Related to the Research Questions
To summarize, of the 9 studies, 1 (11%) was able to discriminate among different states of depression (worsening, relapse, or recovery) and 6 (67%) reported prediction capabilities for depressive episodes and symptom severity. Detailed descriptions of the results regarding the 3 research questions of this systematic review are presented below.
Passively Collected Data for Monitoring Depressive Symptom Severity
The primary intended use of digital interventions in the selected studies was the monitoring of depressive symptoms to gain an understanding of disease progression and clinical assessment []. For monitoring purposes, different parameters were collected through sensors by different studies (eg, call logs [,,,,], step count [,,-], GPS data [-,,], smartphone usage [,], light sensor data [,,], sleep [-], and heart rate [,,,]).
Discriminating among different states of depression (worsening, relapse, or recovery) was found to be feasible in 1 study []. In the study by Zhang et al [], daily sleep records of participants were collected using the Charge 2 or Charge 3 device (Fitbit Inc), alongside the variability of each participant’s depressive symptom severity measured through the PHQ-8. The study demonstrated that 5 sleep features (sleep architecture, sleep stability, sleep quality, insomnia, and hypersomnia) exhibited a significant relationship (P<.001) with the worsening of depressive status over the following 2 weeks [].
Passively Collected Data for Predicting Changes in Depressive Symptom Severity
The capability of passively collected data for predicting depressive episodes and states was shown by 6 studies [,,,,,].
In 1 study, a model was trained using user PHQ-9 scores and phone usage data, including call logs, SMS data, and GPS information []. The study illustrated a predictive accuracy of 0.91 (SD 0.06) for identifying individuals with clinically significant depressive symptoms, accompanied by a sensitivity of 0.98 and specificity of 0.93. Notably, phone usage features were extracted from data collected 14 days before the assessment of the PHQ-9 [].
Sarda et al [] collected a total of 53 passive sensing variables derived from the activity, mobility, sleep, and communication data from smartphone sensors (aggregated daily) and PHQ-9 data biweekly. Subsequently, using these derived sensing variables, a classification predictive model using a gradient boosting machine learning classifier was constructed. This model demonstrated the best performance, achieving a cross-validation accuracy of 79.07% (95% CI 74%-84%) and a test accuracy of 81.05%. Daily predictions were conducted for 950 days (with missing features in one or more of the 53 derived sensing variables) out of 2694 instances [].
The study conducted by Cho et al [] employed a mood prediction feedback model using a smartphone app called Circadian Rhythm for Mood. This model used machine learning to analyze real-time acquired personal digital phenotypes, including factors such as heart rate, activity, sleep, and light exposure. The mood prediction was visually conveyed to users through a facial expression icon (emoticon), providing an intuitive indication of how their mood was predicted to change over the following 3 days. Users received this daily visual feedback when they activated the app. However, the study did not specify the predictive accuracy achieved [].
Mullick et al [] conducted a study aiming to predict depression scores and changes in depression levels among adolescents, with the intention of understanding how sensor features contribute to this prediction. Passively sensed data from mobile phones were collected through the AWARE app, alongside data from a wearable device (Fitbit Inspire HR). The researchers applied regression-based machine learning algorithms to forecast depression scores based on weekly PHQ-9 surveys. The personalized models, which demonstrated superior performance in predicting depression in comparison to universal models, yielded the best outcomes. These models were constructed from feature sets that incorporated data from Fitbit, location, calls, and screen usage. Remarkably, this personalized model achieved the prediction of depression scores and weekly changes in depression with root mean squared errors of 2.83 and 3.21, respectively [].
Lee et al [] used random forest (a supervised learning algorithm) to train the episode prediction model. By selecting features, a set of optimal parameters (circadian rhythm–related digital phenotypes: light exposure, steps, and sleep) was identified and used for predicting whether a mood episode would occur within the next 3 days. They showed that features that contributed to increased risk of forthcoming episodes were lower mean sleep efficiency, higher standard deviation of sleep length, higher mean step count during bedtime, and lower mean step count during afternoon and evening, with a predictive accuracy of 87% for model-predicted mood episodes compared with predictions made as a result of clinical interviews [].
Another study assessed the prediction of mood changes (assessed through biweekly self-rating via the PHQ-9) by selected features collected through the Mood Mirror app and a wristband (call logs, sleep, step count, and heart rate) [], and an accuracy rate of 84.27% was obtained.
Machine Learning Approaches Used for Prediction
The reviewed studies share a common goal of leveraging passive sensor data and machine learning to predict depressive symptoms and mood fluctuations. Despite differences in data sources, modeling strategies, and clinical objectives, they provide valuable insights into the role of digital phenotyping in mental health prediction. summarizes the analytical approaches that were applied in the selected studies.
| Study | Goala | Main findings | Data used | Algorithms used | Key results |
| Tønning et al [], 2021 | —b | Smartphone usage data and clinical depression measures are associated in unipolar depressive disorder. |
| Inferential statistical tests (eg, t tests and chi-square) for analyzing associations between automatically generated smartphone data and clinical depression measures. | Association analysis showed links between phone usage (calls, steps, and screen time) and depressive symptoms. Lower psychosocial functioning was associated with fewer daily steps (P=.04) and increased phone calls (P<.05). |
| McIntyre et al [], 2021 | P | Sensor data show strong convergence with PHQ-9c depression scores in adults with depressive symptoms. |
| Multimodal data from GPS, SMS text messages, and call logs were processed using statistical models. Machine learning models for convergence analysis between sensor data and PHQ-9 depression scores. | The predictive algorithm calculated sensitivity (0.98) and specificity (0.93) while achieving an overall predictive accuracy of 0.91 (SD 0.06) for detecting depressive symptoms. User satisfaction with the app was high, and convergence with the PHQ-9 was statistically significant. |
| Sarda et al [], 2019 | P | Smartphone sensing for activity, mobility, and social interaction predicts depressive symptoms in patients with diabetes. |
| Supervised learning approach applied to features derived from smartphone-sensed data. XGBoost classifier used for the prediction of depression symptoms. Model performance was measured using cross-validation and test set accuracy. | The XGBoost model achieved 81.05% accuracy in identifying depression symptoms in patients with diabetes. Cross-validation accuracy was 79.07% (95% CI 74%-84%). Significant predictors included average activity rate (P=.005) and call activity (P<.001). |
| Fang et al [], 2021 | — | Sleep variability significantly affects mood and depressive symptoms in medical interns. |
| Linear regression for analyzing the effects of sleep variability on mood and depressive symptoms. | Beta coefficients for total sleep time (b=–0.11; P<.001), bedtime (b=0.068; P=.02), and sleep variability (b=0.4; P=.001). |
| Cho et al [], 2020 | P | A smartphone app with wearable device tracking reduces mood disorder episodes and improves patient health behaviors. |
| Machine learning model embedded in the CRMd app, using wearable device data to predict and prevent mood episodes. The feedback system used a prediction model based on circadian rhythm features, which triggered alerts for behavioral correction. | The CRM group had 97.4% fewer mood episodes and 98.9% shorter episode durations than the control group. The number of mood episodes was significantly reduced (exp β=0.026; P=.008), and the duration of mood episodes (exp β=0.011; P<.001) also decreased significantly. |
| Mullick et al [], 2022 | P | Passive smartphone and wearable device sensors predict adolescent depression using personalized machine learning models, which outperform universal models. |
| Machine learning models using both linear (eg, linear regression) and nonlinear algorithms (eg, support vector regression) to predict PHQ-9 scores and weekly change in depressive symptoms. Personalized modeling strategies were applied using methods like “leave-one-out” and “accumulated weeks” approaches. | Personalized machine learning models achieved a root mean square error of 2.83 for PHQ-9 score prediction and 3.21 for weekly change in depression severity. Personalized models outperformed universal models. |
| Zhang et al [], 2021 | M | Sleep features extracted from consumer wearable devices are associated with depressive symptom severity. |
| Linear mixed-effects models for analyzing the relationship between wearable device–derived sleep features and the PHQ-8. Statistical coefficients (z-scores) were computed to measure the effect sizes of features such as insomnia, hypersomnia, and sleep stability. | Linear mixed regression identified 14 significant sleep features, including awake time percentage (z=5.45; P<.001), awakening times (z=5.53; P<.001), insomnia (z=4.55; P<.001), mean sleep offset (z=6.19; P<.001), and hypersomnia (z=5.30; P<.001). |
| Lee et al [], 2022 | P | Digital phenotypes related to circadian rhythms predict impending mood episodes in depressive and bipolar disorder with high accuracy. |
| Machine learning models (supervised classification algorithms) to predict mood episodes using circadian rhythm parameters. | Prediction accuracy for mood episodes: depressive (90.1%), manic (92.6%), and hypomanic (93%) episodes. AUCf values for the mood episode prediction models: 0.937, 0.957, and 0.963, respectively. |
| Bai et al [], 2021 | P | Machine learning models using passive data from smartphones and wearable devices predict mood stability in patients with major depressive disorder. |
| Multiple classification models were trained on the data, using various combinations of data to classify mood stability into steady or swing states. Feature selection models were applied to identify relevant features, and 6 machine learning models were compared. | Prediction of steady versus mood swing states achieved 76.67% accuracy and 90.44% recall. The classifier predicted mood swings using call, sleep, step count, and heart rate data, with 252 features extracted for model training. |
aStudies aiming to monitor depression are labeled with M, and studies aiming to predict depression are labeled with P.
bNot applicable.
cPHQ-9: Patient Health Questionnaire-9.
dCRM: Circadian Rhythm for Mood.
ePHQ-8: Patient Health Questionnaire-8.
fAUC: area under the curve.
All studies integrated passive sensing data (eg, sleep, heart rate, activity, and smartphone usage) with self-reported measures (eg, PHQ-9, PHQ-8, and mood scores obtained through ambulatory assessment). While Fang et al [], Zhang et al [], and Sarda et al [] focused on sleep and activity data from wearable devices, studies like those by Mullick et al [] and Bai et al [] enhanced predictive accuracy by incorporating phone usage patterns (eg, calls, messages, and app activity). The studies differ in predictive targets, with some estimating continuous depressive severity scores [,] and others classifying mood states [] or predicting episode recurrence [,], influencing model selection and applicability.
Linear regression was used in studies, such as those by Fang et al [] and Zhang et al [], to establish statistical associations between sleep metrics and depressive symptoms. In contrast, more advanced machine learning models, such as XGBoost [] and support vector regression [], achieved higher predictive performance but sacrificed interpretability. Personalized modeling approaches [,] outperformed generalized models [,] by leveraging individual-specific baselines, improving prediction accuracy but requiring additional training data per user. Cho et al [] took an intervention-driven approach, integrating real-time mood prediction and behavioral feedback loops to reduce depressive episode recurrence.
A key methodological divergence lies in predictive objectives. While some studies estimated continuous depression severity, others classified mood transitions or predicted episode recurrence. This variation impacts how success is measured: classification models [,] support binary decision-making, whereas regression models [,] offer a more granular understanding of mood trajectories. Furthermore, explainability remains a challenge; linear models provide interpretable relationships between predictors and outcomes, while complex models (eg, XGBoost) optimize accuracy at the cost of transparency, potentially limiting their clinical adoption. Unfortunately, none of the studies assessed the reliability of their predictions over time.
Relevance of Individual Biosensor Data
All studies showed the potential of passive sensing using smartphones. The findings encompassed significant correlations with validation measures and the successful construction of prediction models for some or all of the variables studied. However, certain sensor data, such as sleep features (total duration and wake-up time), and other data, such as social contact (call logs) and activity (step count and GPS), could potentially serve as significant parameters for clinical aspects related to depression, including relapse or depressive symptoms.
For example, Zhang et al [] identified 14 sleep features that exhibited a significant association (P<.05) with depression status. Within this set, 5 sleep features, namely sleep architecture, sleep stability, sleep quality, insomnia, and hypersomnia, demonstrated a significant (P<.001) relationship with the deterioration of depressive symptoms over the subsequent 2 weeks. Among these sleep features, the percentages of light sleep, nonrapid eye movement sleep, and sleep efficiency were significantly and negatively associated with the PHQ-8 score. Conversely, the remaining significant features displayed a positive association with the PHQ-8 score.
The study by Tønning et al [] revealed a significant association between patient-reported mood and activity and automatically generated smartphone data. For instance, lower reported mood was linked to a lower step count (P=.04) and an increased number of incoming (P=.03) and outgoing (P=.02) calls, missed calls (P=.007), and longer phone calls (P=.01) [].
Univariate analysis conducted in 1 study revealed a significant difference in the average activity rates (P<.001), the number of screen-on times at night (P<.001), and the average total number of calls (P<.001) among individuals exhibiting symptoms of major depression [].
Fang et al [] demonstrated that, within individuals, an increase in total sleep time (P<.001), a later wake time (P<.001), and an earlier bedtime (P<.001) were associated with improved mood on the following day.
The study conducted by Mullick et al [] demonstrated that individuals experiencing depressive symptoms exhibited a tendency to be less active in terms of movement, a behavior that could be captured through location (GPS) data. Furthermore, depression led participants to decrease their interactions with friends and family, as indicated by call-related features. Other studies similarly identified significantly lower levels of social contact (call logs) and reduced daytime activity among individuals reporting symptoms of depression (PHQ-9 score >9) [,].
Critical Appraisal of Study Quality
Quality appraisal was performed taking into consideration 9 domains of research paper quality, including research aims, research design, recruitment strategy, data collection and analysis methods, ethical issues, outcome measures, etc []. The results are summarized in . The quality appraisal showed overall consistent quality across studies for the domains of research aims, research design, recruitment strategy, data collection and analysis methods, and ethical issues. However, there were some quality considerations that must be particularly noted, for instance, the 5th domain addressing the considerations between the researcher-participant relationship was rather scarce. In fact, there was no sufficient description of such rapport across the retrieved studies. Moreover, the value of research assessed in the 9th domain did not consist of the precise parameters that can be followed to assign the value of particular research within the field, therefore relying on the subjective understanding of the author rather than a quantitative unit of measurement.
| Research question | Study | ||||||||
| Tønning et al [], 2021 | McIntyre et al [], 2021 | Sarda et al [], 2019 | Fang et al [], 2021 | Cho et al [], 2020 | Mullick et al [], 2022 | Zhang et al [], 2021 | Lee et al [], 2022 | Bai et al [], 2021 | |
| 1. Was there a clear statement of the aims of the research? | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
| 2. Was the research design appropriate to address the aims of the research? | Yes | Yes | Cannot tell | Yes | Yes | Cannot tell | Yes | Yes | Yes |
| 3. Was the recruitment strategy appropriate for the aims of the research? | Yes | —a | Yes | Yes | Yes | Yes | Yes | — | Yes |
| 4. Was the data collected in a way that addressed the research issue? | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
| 5. Has the relationship between the researcher and participants been adequately considered? | No | — | Cannot tell | No | — | Cannot tell | — | Yes | No |
| 6. Have ethical issues been taken into consideration? | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
| 7. Was the data analysis sufficiently rigorous? | Yes | No | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
| 8. Is there a clear statement of the findings? | Yes | No | Yes | Yes | Yes | Yes | Yes | No | Yes |
| 9. How valuable is the research? | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field | Valuable for the research field |
aNot applicable.
Key Challenges
Advantages of Passive Sensing
Only 1 study reported providing feedback to study participants []. Feedback messages were automatically generated and displayed on the app screen to assist patients in modifying their behavior. In addition, warning alerts were conveyed to patients through SMS text messages with a message reading, “Recently, your life rhythm is irregular.” This study computed data locally and delivered feedback directly on the phone. The analyses conducted in this study revealed significant positive behavioral changes after participants received warning alert feedback (P<.05) in relation to light exposure during the daytime and steps taken during the daytime. Most of the advice given in the warning alert pertained to increasing physical activity. However, concerning sleep behavior, it was observed that sleep efficiency and duration tended to increase following the point at which the warning alert was delivered.
The study conducted by Cho et al [] observed the prognosis over a period of 1 year, contingent upon whether the intervention group incorporated feedback alongside conventional treatment. This study identified a robust recurrence prevention effect within the intervention group. Additionally, the study noted a significantly shortened duration of mood episodes and a quicker recovery time subsequent to the recurrence of mood episodes among participants in the intervention group (univariate general linear analysis identified 60.7% fewer total depressive episodes [P=.03] and 48.5% shorter depressive episodes [P<.001], and multivariate general linear analysis identified 96.7% fewer total depressive episodes [P=.03] and 99.5% shorter depressive episodes [P<.001]) [].
Challenges of Passive Sensing
The apparent ease of deploying passive sensing was counter-balanced by several reported challenges. Although not systematically reported across studies, these challenges could be divided into 3 categories: technological, methodological, and privacy issues.
Regarding technological challenges, 1 retrieved study mentioned that battery life affected participants’ adherence to Fitbit []. Regarding methodological challenges, 1 study noted concerns about generalizability due to a low sample size, possible sample bias, and variability in the study sample data []. Regarding the protection of privacy, while all studies did not extensively discuss or explicitly mention privacy concerns, they generally described their methods for safeguarding data privacy. These methods included secure communication with external servers, data anonymization, and audio scrambling [-].
Dealing With Missing Data
One of the significant challenges to address is the issue of missing data. Adherence rates were reported in only 2 studies [,]. The occurrence of missing data and the strategies employed to handle it were rarely outlined. Furthermore, none of the studies referenced a previously published protocol. The included studies exhibited poor reporting of missing data, both at the sample level (eg, study noncompletion) and the individual level (eg, nonadherence).
In the study by Sarda et al [], participant-day instances with missing values in one or more of the 53 derived sensing variables were removed, resulting in 950 out of 2694 instances available for analysis. Another study reported that the passive sensing data and Fitbit data were on average 69.11% and 32.36% complete, respectively, while PHQ-9 surveys were completed 69.01% of the time []. Zhang et al [] only included datasets with a completion rate of over 85% []. In the study by Lee et al [], participants who wore wearable devices for at least 30 days were included in the analysis, which constituted 54.5% of the original cohort.
Missing data can arise from technological issues, such as device and system failures, or from user-related problems that might be associated with depressed mood []. Identifying the underlying reasons for missing data and addressing them in terms of either exclusion or analysis are crucial. Furthermore, thresholds for defining what constitutes missing data varied across the studies.
Discontinuation
Reasons for dropout were provided in 1 study and were attributed to the unwillingness to participate in a noncompensated study for a period of 12 weeks []. Other studies did not provide specific information regarding discontinuation.
Discussion
Principal Findings
Developing and implementing new tools for the self-management and prevention of depression are priorities due to its high prevalence and recurrent nature [,-]. However, a significant gap remains between research and clinical practice. This study included 9 original articles and systematically reviewed evidence on objectively measured data from smartphones and wearable devices used to monitor depression for at least 3 months. To our knowledge, this is the first review to assess the capabilities of these technologies for monitoring and predicting depressive symptoms over an extended period, introducing a lifetime perspective into the field.
Only 1 study in this review addressed whether different depression states (eg, worsening, relapse, or recovery) can be distinguished using sleep data from a wearable device and linear mixed-effects models []. Most other studies examined general associations between objective features and depression [,] or focused on symptom or episode prediction (see the discussion of the principal findings regarding research question 2 below), rather than real-time monitoring. This shows that behavioral patterns and physiological changes, which may precede conscious emotional awareness, remain underutilized for assessing depression severity. Because self-perception often lags observable symptoms or physiological shifts, individuals may recognize their need for help only when the condition has already worsened. Objective markers from smartphones and wearable devices could help bridge this gap and support the timely detection and monitoring of depressive disease states. While following contemporary diagnostic manuals [,], depression is primarily defined by affective and cognitive symptoms, such as low mood, anhedonia, fatigue, and hopelessness. These rely on self-report and are not directly measurable by passive sensing (or other biomarkers []). In contrast, behavioral and physiological domains, also characteristic of depression, offer more promise for objective, continuous monitoring and should be explored further in future research.
Six studies demonstrated that objectively measured data can predict depressive episodes or symptom severity. Four of these studies reported predictive accuracies for depression measures based on data collected via smartphones and wearable devices (predictive accuracy ranged from 81% to 91%) [,,,]. Three studies predicted depressive symptom severity assessed by the PHQ-9 (biweekly; predictive accuracies: 81% [], 84% [], and 91% []). Higher predictive accuracies were achieved in studies with a personalized modeling approach [] or those that were using multimodal data streams as model input []. Although these results are promising, the long-term prediction of depression using sensor data from smartphones and wearable devices faces several potential limitations. Depression is highly heterogeneous, and generic models may fail to capture person-specific symptom trajectories over time. However, only 1 study applied personalized modeling [], leaving space for exploring idiographic approaches in future studies. Our review focused on longitudinal studies investigating depression from a lifespan perspective. However, models trained on short-term datasets (eg, 12 weeks [,]) may not generalize well over longer periods. As health-related behaviors and associated symptoms evolve over time, past data may no longer reflect current patterns, and model accuracy would be reduced. Technical issues (eg, sensor malfunction and inconsistent data sampling) or a lack of user engagement and compliance (eg, not wearing a device and disabling tracking) can lead to data attrition. Missing data can compromise prediction quality, but only 1 of the included studies [] reported both adherence rates and handling of missing data during analysis.
The objective features most commonly examined in the reviewed studies and significantly associated with depression were primarily behavioral in nature, including activity [,,,], sleep [,,,], and smartphone usage [,,,]. Several studies reported significant associations between depression and physiological features obtained from wearable devices, including heart rate [,] and circadian rhythm [,]. This review could not determine whether single features are clinically relevant for monitoring and predicting depression within individuals, as the methodological heterogeneity across studies prevented reliable and generalizable conclusions. Even when feature importance was reported, it was specific to the analytical methods used, limiting the ability to draw conclusive comparisons across studies.
Comparison With Prior Work
Traditionally, semistructured interviews complemented by self-report measures have been used as diagnostic procedures, capturing a static moment in time. A few former valuable reviews in a similar field provided the groundwork for our review. For example, a systematic review by Köhnen et al [] included studies with durations ranging from 1 to 52 weeks and using offline computer programs, mobile texts, or videoconferencing tools as a medium of intervention []. Two other reviews only included studies with younger populations and adolescents [,]. Another review included only randomized controlled trials []. Owing to recent advancements in sensor technologies, physiological signals can now be recorded unobtrusively using devices like smartwatches and smartphones. This approach does not entail recording sensitive information (unlike cameras and audio signals), and the associated sensors do not disrupt users’ daily routines. Consequently, patient-specific models can be automatically generated to continuously estimate the patient’s affective state [,]. Certainly, we agree that self-reports play a pivotal role in monitoring depressive symptoms, given that objective observations might not always align with patients’ subjective experiences. Nonetheless, our study encompassed research that gathered self-reports along with data collected from wearable biosensors or smartphones, and applied machine learning techniques to the acquired data whenever feasible.
From the results, it is clear that monitoring and predicting depressive symptoms through digital phenotyping can be considered worthy measures for improving the quality of life of patients with depression. The existing evidence indicates that the use of passively collected data could potentially lead to enhanced management of depressive symptoms [-]. For instance, Tønning et al [] conducted a study that identified correlations between automatically generated data and daily self-reports. This information could subsequently be employed to implement just-in-time adaptive interventions, which involve providing smartphone-based treatments when machine learning models predict instances of depressed mood. While the studies under consideration did not uncover any notable associations between specific biosensor data and relapse, these findings can serve as a foundational stepping stone for further exploration in this domain. Three studies passively collected GPS and location data [-,], which can be considered very sensitive information. Therefore, it is crucial to address ethical concerns pertaining to privacy, consent, and self-awareness. While individuals might be open to sharing personal data for research purposes, they could exhibit greater caution when commercial interests come into play [,,]. Additionally, it is argued that the measurement of distance traveled by smartphone-based GPS receivers might differ [,]. Similarly, accelerometers are commonly considered reliable, but variations in their output and validity in measuring physical activity have been noted [,].
Digital interventions have faced criticism due to concerns about them potentially exacerbating anxiety through continuous monitoring and impeding the enjoyment of physical activities [,]. However, despite these critiques, the integration of wearable devices and smartphones with feedback mechanisms has shown potential in effectively alleviating symptoms and could contribute to relieving the burden on patients by managing their episodes and consequently mitigating the burden on clinicians [,]. For individuals experiencing milder levels of depression, smartphone apps could potentially serve as an excellent self-management tool. Doherty et al [] demonstrated that user engagement is also influenced by the severity of depression; however, the specific direction in which this influence operates remains to be clarified []. Furthermore, the studies examined in this review concentrated on a limited set of predictor variables. Consequently, further research involving the analysis of multiple variables over an extended duration will be essential to identify significant predictors for the relapse of depressive episodes. While assessing socialization, call logs were taken into account. However, features related to proximity using Bluetooth and microphone sensors appear to be more attuned to mood than mere counts of calls and messages. This presents a novel avenue for exploration and warrants further investigation [].
Strengths and Limitations
Our review was conducted in accordance with the PRISMA guidelines [], reflecting rigorous methodological standards. Our broad search strategy identified 12,997 peer-reviewed articles focused on approaches to support mental health. This substantial number revealed the rapid evolution of the literature in this domain, which has grown considerably larger compared to smartphone interventions for other health conditions []. Furthermore, we used a comprehensive search term to ensure the inclusion of all pertinent articles and minimize the risk of overlooking relevant information.
Nevertheless, we must acknowledge certain limitations. First, we may have overlooked eligible trials due to different reasons: (1) studies may have been published in languages other than English; (2) eligible studies exclusively published in technology-oriented journals could have been inadvertently omitted, as our literature search was limited to databases specializing in medical and psychological publications; and (3) the search terms and selected keywords did not cover all potential studies. Due to the nascent stage of the research field, MeSH terms have not yet been fully established, presenting challenges to the consolidation of knowledge across the research field. Second, due to the absence of standardized evaluation and reporting practices, comparing technologies across studies posed a significant challenge. This review included studies that used diverse methods for diagnosing depression, making it challenging to establish strict inclusion criteria concerning diagnostic assessments [-]. For instance, variations were observed in basic descriptors of sample characteristics, recruitment methods, data collection durations, depression assessment measures, the handling of missing data, the types and quantities of sensors employed, the timing and location of data processing, and the manner in which feedback was provided to users. Most importantly, retrieved studies were not uniform in terms of data collection (both self-reported and sensor data). The types of sensors used (ie, wearable devices and smartphone sensors) also varied. These differences make it difficult to reliably compare parameters. Third, the selected studies adhered to the criterion of selecting adults aged 14 to 65 years. In future research, it will be essential to assess functionality (in terms of monitoring and prediction) in a population that better reflects the diversity of the real world, thereby enhancing the generalizability of the findings. Among the chosen studies, only the study by Fang et al [] used education level as a selection criterion; unfortunately, no specific information was provided regarding the education levels of participants in the other studies. This factor could potentially influence the use of and openness to health technology []. Fourth, the duration of the studies included in this review varied, with the continuous data collection period typically being relatively short (averaging around 3 months). As a result, there was insufficient data available to assess the long-term usage of passive sensing DHTs for mental health. While the outcomes observed in the short term are encouraging, it remains uncertain whether the substantial volume of data derived from DHTs can truly revolutionize mental health care and enhance patient outcomes over their lifespan.
Future Directions
To enhance long-term self-management practices for depressive symptoms, the provision of personalized feedback based on passively sensed data would be advantageous. Patient adherence to any intervention for self-management plays a very crucial role, and this needs to be addressed in future studies. Upcoming studies should undertake comprehensive assessments of the acceptance, potential adverse outcomes, and longitudinal use of passive sensing systems, given their expanding technological reach.
All studies that used passive data for the prediction of depression collectively highlighted the promise of digital biomarkers and machine learning in mental health prediction, but differences in target populations, modeling approaches, and predictive outputs restricted direct comparability. Personalized models demonstrate strong predictive performance but require individualized calibration, whereas generalized models are more scalable but risk lower accuracy for specific users. The trade-off between model interpretability and predictive power remains a challenge, particularly for clinical integration. Moving forward, standardized benchmarks, clearer definitions of predictive targets, and transparent reporting on feature selection and performance trade-offs are essential to improve the comparability and real-world applicability of predictive models in mental health care.
This review highlighted significant heterogeneity in the configurations of data collection, reporting, and processing. These disparities emphasize the urgency of adopting a more consistent adherence to reporting guidelines that have already been established in closely related fields []. Based on our clinical and research experience, we make the following recommendations to enhance the reporting and generalizability of research concerning passive sensing in patients with depression. Regarding the study population, studies should (1) provide more precise reporting of recruitment strategies to increase the scope of reproducibility; (2) present basic demographic and clinical data, including age, gender, ethnicity, and comorbidities; and (3) report more clearly on user engagement and acceptability. Regarding data collection and analysis, studies should (1) use validated scales for assessing depression; (2) preregister the study protocol to increase reliability; and (3) offer a clear description of missing data management.
Conclusion
Our review demonstrated the potential of longitudinal passive data from smartphones and wearable devices for the prediction of depressive symptoms among individuals. However, few studies have addressed their potential for illness monitoring. Passive sensing via wearable devices and smartphone apps is in harmony with the principles of minimally disruptive medicine and may help to identify the early progression of symptoms. Predicting individual illness trajectories ahead of time facilitates early interventions like self-management strategies that could be offered by smartphone apps or adjusting traditional therapeutic interventions for depression to the intensity and frequency that is needed in a specific moment for each individual.
Acknowledgments
This work was funded as part of the MONDY project “Secure and open platform for AI-based healthcare apps” by the German Federal Ministry of Education and Research (BMBF; grant number: 13GW0576D; date awarded: November 1, 2021). The funding source had no involvement in study design; collection, analysis, and interpretation of data; writing of the report; and decision to submit the article for publication. The authors would like to express gratitude and acknowledge the contributions of all members of the MONDY consortium: Mohamed Alhaskir, Angela Carell, Tobias Dunker, Kismet Ekinci, Florian Fischer, Leonie Hipper, Sascha Ludwig, Elisabeth Schriewer, Yvonne Weber, Stefan Wolking, and Jil Zippelius. The authors attest that there was no use of generative artificial intelligence technology in the generation of text, figures, or other informational content in this manuscript.
Data Availability
Retrieved data are provided as a supplementary file, and all retrieved articles and generated data are available upon request from the corresponding author.
Authors' Contributions
RA contributed to data curation, formal analysis, investigation, methodology, visualization, and writing – original draft. SS contributed to formal analysis, investigation, project administration, validation, and writing – original draft. HO contributed to formal analysis, investigation, and validation. MP contributed to methodology and supervision. UH contributed to conceptualization, funding acquisition, and supervision. HR contributed to conceptualization, funding acquisition, project administration, methodology, supervision, and writing – original draft. All authors contributed to the review and editing of the manuscript. All authors read and approved the final manuscript.
Conflicts of Interest
None declared.
PRISMA checklist.
DOCX File , 32 KBCharacteristics of the included studies.
XLSX File (Microsoft Excel File), 18 KBReferences
- Hickey BA, Chalmers T, Newton P, Lin C, Sibbritt D, McLachlan CS, et al. Smart devices and wearable technologies to detect and monitor mental health conditions and stress: a systematic review. Sensors. May 16, 2021;21(10):3461. [CrossRef]
- Depression and Other Common Mental Disorders: Global Health Estimates. World Health Organization. 2017. URL: https://www.who.int/publications/i/item/depression-global-health-estimates [accessed 2025-07-27]
- Spijker J, De Graaf R, Bijl RV, Beekman ATF, Ormel J, Nolen WA. Duration of major depressive episodes in the general population: Results from the Netherlands Mental Health Survey and Incidence Study (NEMESIS). Br J Psychiatry. 2002;181(3):208-213. [CrossRef]
- Richards D. Prevalence and clinical course of depression: A review. Clinical Psychology Review. Nov 2011;31(7):1117-1125. [CrossRef] [Medline]
- Murray CJL, Lopez AD, World Health Organization, World Bank, Harvard School of Public Health. The Global burden of disease: a comprehensive assessment of mortality and disability from diseases, injuries, and risk factors in 1990 and projected to 2020. World Health Organization. 1996. URL: https://iris.who.int/handle/10665/41864 [accessed 2025-07-27]
- Eaton WW. Natural history of Diagnostic Interview Schedule/ DSM-IV Major Depression. Arch Gen Psychiatry. Nov 01, 1997;54(11):993. [CrossRef]
- Roca M, Armengol S, García-García M, Rodriguez-Bayón A, Ballesta I, Serrano MJ, et al. Clinical differences between first and recurrent episodes in depressive patients. Comprehensive Psychiatry. Jan 2011;52(1):26-32. [CrossRef]
- Tokgöz P, Hrynyschyn R, Hafner J, Schönfeld S, Dockweiler C. Digital health interventions in prevention, relapse, and therapy of mild and moderate depression: scoping review. JMIR Ment Health. Apr 16, 2021;8(4):e26268. [FREE Full text] [CrossRef] [Medline]
- Zwerenz R, Baumgarten C, Becker J, Tibubos A, Siepmann M, Knickenberg RJ, et al. Improving the course of depressive symptoms after inpatient psychotherapy using adjunct web-based self-help: follow-up results of a randomized controlled trial. J Med Internet Res. Oct 24, 2019;21(10):e13655. [CrossRef]
- Maj M, Stein DJ, Parker G, Zimmerman M, Fava GA, De Hert M, et al. The clinical characterization of the adult patient with depression aimed at personalization of management. World Psychiatry. Sep 15, 2020;19(3):269-293. [CrossRef]
- Cormack F, McCue M, Taptiklis N, Skirrow C, Glazer E, Panagopoulos E, et al. Wearable technology for high-frequency cognitive and mood assessment in major depressive disorder: longitudinal observational study. JMIR Ment Health. Nov 18, 2019;6(11):e12814. [CrossRef]
- Mehrotra A, Tsapeli F, Hendley R, Musolesi M. MyTraces: investigating correlation and causation between users? Emotional states and mobile phone interaction. Proc ACM Interact Mob Wearable Ubiquitous Technol. Sep 11, 2017;1(3):1-21. [FREE Full text] [CrossRef]
- Carney RM, Freedland KE. Depression and heart rate variability in patients with coronary heart disease. CCJM. Feb 01, 2009;76(4 suppl 2):S13-S17. [CrossRef]
- Burns MN, Begale M, Duffecy J, Gergle D, Karr CJ, Giangrande E, et al. Harnessing context sensing to develop a mobile intervention for depression. J Med Internet Res. 2011;13(3):e55. [FREE Full text] [CrossRef] [Medline]
- Rollo S, Antsygina O, Tremblay MS. The whole day matters: Understanding 24-hour movement guideline adherence and relationships with health indicators across the lifespan. Journal of Sport and Health Science. Dec 2020;9(6):493-510. [CrossRef]
- Tazawa Y, Liang K, Yoshimura M, Kitazawa M, Kaise Y, Takamiya A, et al. Evaluating depression with multimodal wristband-type wearable device: screening and assessing patient severity utilizing machine-learning. Heliyon. Feb 2020;6(2):e03274. [CrossRef]
- Moshe I, Terhorst Y, Philippi P, Domhardt M, Cuijpers P, Cristea I, et al. Digital interventions for the treatment of depression: A meta-analytic review. Psychological Bulletin. Aug 2021;147(8):749-786. [CrossRef] [Medline]
- Lorenz N, Sander C, Ivanova G, Hegerl U. Temporal associations of daily changes in sleep and depression core symptoms in patients suffering from major depressive disorder: idiographic time-series analysis. JMIR Ment Health. Apr 23, 2020;7(4):e17071. [CrossRef]
- McDonald S, Quinn F, Vieira R, O’Brien N, White M, Johnston DW, et al. The state of the art and future opportunities for using longitudinal n-of-1 methods in health behaviour research: a systematic literature overview. Health Psychology Review. Apr 25, 2017;11(4):307-323. [CrossRef] [Medline]
- Smith J. Mobile eCommerce Stats in 2023: What Percentage of eCommerce Sales Are on Mobile Devices? OuterBox Solutions. URL: https://www.outerboxdesign.com/web-design-articles/mobile-ecommerce-statistics [accessed 2023-04-17]
- Bhalla N, Jolly P, Formisano N, Estrela P. Introduction to biosensors. Essays Biochem. 2016;60(1):1-8. [FREE Full text] [CrossRef] [Medline]
- Moukaddam N, Truong A, Cao J, Shah A, Sabharwal A. Findings from a trial of the Smartphone and OnLine Usage-based eValuation for Depression (SOLVD) application: what do apps really tell us about patients with depression? concordance between app-generated data and standard psychiatric questionnaires for depression and anxiety. J Psychiatr Pract. 2019;25(5):365-373. [CrossRef] [Medline]
- Jacobson NC, Weingarden H, Wilhelm S. Using digital phenotyping to accurately detect depression severity. J Nerv Ment Dis. 2019;207(10):893-896. [CrossRef]
- Gould CE, Carlson C, Ma F, Forman-Hoffman V, Ranta K, Kuhn E. Effects of mobile app–based intervention for depression in middle-aged and older adults: mixed methods feasibility study. JMIR Form Res. Jun 29, 2021;5(6):e25808. [CrossRef]
- Pedrelli P, Fedor S, Ghandeharioun A, Howe E, Ionescu DF, Bhathena D, et al. Monitoring changes in depression severity using wearable and mobile sensors. Front Psychiatry. Dec 18, 2020;11:584711. [CrossRef]
- Marzano L, Bardill A, Fields B, Herd K, Veale D, Grey N, et al. The application of mHealth to mental health: opportunities and challenges. The Lancet Psychiatry. Oct 2015;2(10):942-948. [CrossRef]
- Mohr DC, Zhang M, Schueller SM. Personal sensing: understanding mental health using ubiquitous sensors and machine learning. Annu Rev Clin Psychol. May 08, 2017;13(1):23-47. [CrossRef] [Medline]
- Batra S, Baker R, Wang T, Forma F, DiBiasi F, Peters-Strickland T. Digital health technology for use in patients with serious mental illness: a systematic review of the literature. MDER. Oct 2017;Volume 10:237-251. [CrossRef]
- Rohani DA, Faurholt-Jepsen M, Kessing LV, Bardram JE. Correlations between objective behavioral features collected from mobile and wearable devices and depressive mood symptoms in patients with affective disorders: systematic review. JMIR Mhealth Uhealth. Aug 13, 2018;6(8):e165. [CrossRef]
- Cornet VP, Holden RJ. Systematic review of smartphone-based passive sensing for health and wellbeing. Journal of Biomedical Informatics. Jan 2018;77:120-132. [CrossRef] [Medline]
- Frank E, Pong J, Asher Y, Soares CN. Smart phone technologies and ecological momentary data: is this the way forward on depression management and research? Curr Opin Psychiatry. 2018;31(1):3-6. [CrossRef] [Medline]
- Köhnen M, Dreier M, Seeralan T, Kriston L, Härter M, Baumeister H, et al. Evidence on technology-based psychological interventions in diagnosed depression: systematic review. JMIR Ment Health. Feb 10, 2021;8(2):e21700. [CrossRef]
- Matcham F, Barattieri di San Pietro C, Bulgari V, de Girolamo G, Dobson R, Eriksson H, et al. Remote assessment of disease and relapse in major depressive disorder (RADAR-MDD): a multi-centre prospective cohort study protocol. BMC Psychiatry. Feb 18, 2019;19(1):72. [CrossRef]
- Melbye S, Kessing LV, Bardram JE, Faurholt-Jepsen M. Smartphone-based self-monitoring, treatment, and automatically generated data in children, adolescents, and young adults with psychiatric disorders: systematic review. JMIR Ment Health. Oct 29, 2020;7(10):e17453. [FREE Full text] [CrossRef] [Medline]
- Sequeira L, Perrotta S, LaGrassa J, Merikangas K, Kreindler D, Kundur D, et al. Mobile and wearable technology for monitoring depressive symptoms in children and adolescents: A scoping review. Journal of Affective Disorders. Mar 2020;265:314-324. [CrossRef]
- Wahle F, Bollhalder L, Kowatsch T, Fleisch E. Toward the design of evidence-based mental health information systems for people with depression: a systematic literature review and meta-analysis. J Med Internet Res. May 31, 2017;19(5):e191. [FREE Full text] [CrossRef] [Medline]
- Yim SJ, Lui LM, Lee Y, Rosenblat JD, Ragguett R, Park C, et al. The utility of smartphone-based, ecological momentary assessment for depressive symptoms. Journal of Affective Disorders. Sep 2020;274:602-609. [CrossRef]
- Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gotzsche PC, Ioannidis JPA, et al. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate healthcare interventions: explanation and elaboration. BMJ. Jul 21, 2009;339(jul21 1):b2700-b2700. [CrossRef]
- Patil P, Sekar A. Passive data collection: What it is and advantages. QuestionPro. URL: https://www.questionpro.com/blog/passive-data-collection/ [accessed 2025-02-13]
- Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. Mar 29, 2021;372:n71. [FREE Full text] [CrossRef] [Medline]
- Long HA, French DP, Brooks JM. Optimising the value of the critical appraisal skills programme (CASP) tool for quality appraisal in qualitative evidence synthesis. Research Methods in Medicine & Health Sciences. Aug 06, 2020;1(1):31-42. [CrossRef]
- Saeb S, Lattie EG, Schueller SM, Kording KP, Mohr DC. The relationship between mobile phone location sensor data and depressive symptom severity. PeerJ. 2016;4:e2537. [FREE Full text] [CrossRef] [Medline]
- Di Matteo D, Wang W, Fotinos K, Lokuge S, Yu J, Sternat T, et al. Smartphone-detected ambient speech and self-reported measures of anxiety and depression: exploratory observational study. JMIR Form Res. Jan 29, 2021;5(1):e22723. [FREE Full text] [CrossRef] [Medline]
- Di Matteo D, Fotinos K, Lokuge S, Mason G, Sternat T, Katzman MA, et al. Automated screening for social anxiety, generalized anxiety, and depression from objective smartphone-collected data: cross-sectional study. J Med Internet Res. Aug 13, 2021;23(8):e28918. [CrossRef]
- Di Matteo D, Fotinos K, Lokuge S, Yu J, Sternat T, Katzman MA, et al. The relationship between smartphone-recorded environmental audio and symptomatology of anxiety and depression: exploratory study. JMIR Form Res. Aug 13, 2020;4(8):e18751. [FREE Full text] [CrossRef] [Medline]
- Rykov Y, Thach T, Bojic I, Christopoulos G, Car J. Digital biomarkers for depression screening with wearable devices: cross-sectional study with machine learning modeling. JMIR Mhealth Uhealth. Oct 25, 2021;9(10):e24872. [CrossRef]
- Teepe GW, Da Fonseca A, Kleim B, Jacobson NC, Salamanca Sanabria A, Tudor Car L, et al. Just-in-time adaptive mechanisms of popular mobile apps for individuals with depression: systematic app search and literature review. J Med Internet Res. Sep 28, 2021;23(9):e29412. [CrossRef]
- Boonstra TW, Nicholas J, Wong QJ, Shaw F, Townsend S, Christensen H. Using mobile phone sensor technology for mental health research: integrated analysis to identify hidden challenges and potential solutions. J Med Internet Res. Jul 30, 2018;20(7):e10131. [CrossRef]
- Jacobson NC, Summers B, Wilhelm S. Digital biomarkers of social anxiety severity: digital phenotyping using passive smartphone sensors. J Med Internet Res. May 29, 2020;22(5):e16875. [CrossRef]
- Deady M, Johnston DA, Glozier N, Milne D, Choi I, Mackinnon A, et al. Smartphone application for preventing depression: study protocol for a workplace randomised controlled trial. BMJ Open. Jul 13, 2018;8(7):e020510. [CrossRef]
- Pellegrini AM, Huang EJ, Staples PC, Hart KL, Lorme JM, Brown HE, et al. Estimating longitudinal depressive symptoms from smartphone data in a transdiagnostic cohort. Brain and Behavior. Jan 25, 2022;12(2):2077. [CrossRef]
- Braund TA, Zin MT, Boonstra TW, Wong QJJ, Larsen ME, Christensen H, et al. Smartphone sensor data for identifying and monitoring symptoms of mood disorders: a longitudinal observational study. JMIR Ment Health. May 4, 2022;9(5):e35549. [CrossRef]
- Lee J, Solomonov N, Banerjee S, Alexopoulos GS, Sirey JA. Use of passive sensing in psychotherapy studies in late life: a pilot example, opportunities and challenges. Front Psychiatry. Oct 28, 2021;12:732773. [CrossRef]
- Zulueta J, Piscitello A, Rasic M, Easter R, Babu P, Langenecker SA, et al. Predicting mood disturbance severity with mobile phone keystroke metadata: a biaffect digital phenotyping study. J Med Internet Res. Dec 20, 2018;20(7):e241. [FREE Full text] [CrossRef] [Medline]
- Jacobson NC, Chung YJ. Passive sensing of prediction of moment-to-moment depressed mood among undergraduates with clinical levels of depression sample using smartphones. Sensors. Jun 24, 2020;20(12):3572. [FREE Full text] [CrossRef] [Medline]
- Choudhary S, Thomas N, Ellenberger J, Srinivasan G, Cohen R. A machine learning approach for detecting digital behavioral patterns of depression using nonintrusive smartphone data (complementary path to Patient Health Questionnaire-9 assessment): prospective observational study. JMIR Form Res. May 16, 2022;6(5):e37736. [CrossRef]
- Everitt N, Broadbent J, Richardson B, Smyth JM, Heron K, Teague S, et al. Exploring the features of an app-based just-in-time intervention for depression. Journal of Affective Disorders. Aug 2021;291:279-287. [CrossRef]
- Zhou H, Al-Ali F, Kang GE, Hamad AI, Ibrahim RA, Talal TK, et al. Application of wearables to facilitate virtually supervised intradialytic exercise for reducing depression symptoms. Sensors. Mar 12, 2020;20(6):1571. [FREE Full text] [CrossRef] [Medline]
- Opoku Asare K, Terhorst Y, Vega J, Peltonen E, Lagerspetz E, Ferreira D. Predicting depression from smartphone behavioral markers using machine learning methods, hyperparameter optimization, and feature importance analysis: exploratory study. JMIR Mhealth Uhealth. Jul 12, 2021;9(7):e26540. [CrossRef]
- Kathan A, Harrer M, Küster L, Triantafyllopoulos A, He X, Milling M, et al. Personalised depression forecasting using mobile sensor data and ecological momentary assessment. Front Digit Health. Nov 18, 2022;4:964582. [FREE Full text] [CrossRef] [Medline]
- Tønning ML, Faurholt-Jepsen M, Frost M, Bardram JE, Kessing LV. Mood and activity measured using smartphones in unipolar depressive disorder. Front Psychiatry. Jul 9, 2021;12:701360. [FREE Full text] [CrossRef] [Medline]
- McIntyre RS, Lee Y, Rong C, Rosenblat JD, Brietzke E, Pan Z, et al. Ecological momentary assessment of depressive symptoms using the mind.me application: convergence with the Patient Health Questionnaire-9 (PHQ-9). Journal of Psychiatric Research. Mar 2021;135:311-317. [CrossRef]
- Sarda A, Munuswamy S, Sarda S, Subramanian V. Using passive smartphone sensing for improved risk stratification of patients with depression and diabetes: cross-sectional observational study. JMIR Mhealth Uhealth. Jan 29, 2019;7(1):e11041. [CrossRef]
- Fang Y, Forger DB, Frank E, Sen S, Goldstein C. Day-to-day variability in sleep parameters and depression risk: a prospective cohort study of training physicians. NPJ Digit Med. Feb 18, 2021;4(1):28. [CrossRef]
- Cho C, Lee T, Lee J, Seo JY, Jee H, Son S, et al. Effectiveness of a smartphone app with a wearable activity tracker in preventing the recurrence of mood disorders: prospective case-control study. JMIR Ment Health. Aug 5, 2020;7(8):e21283. [FREE Full text] [CrossRef] [Medline]
- Mullick T, Radovic A, Shaaban S, Doryab A. Predicting depression in adolescents using mobile and wearable sensors: multimodal machine learning–based exploratory study. JMIR Form Res. Jun 24, 2022;6(6):e35807. [CrossRef]
- Zhang Y, Folarin AA, Sun S, Cummins N, Bendayan R, Ranjan Y, et al. Relationship between major depression symptom severity and sleep collected using a wristband wearable device: multicenter longitudinal observational study. JMIR Mhealth Uhealth. Apr 12, 2021;9(4):e24604. [CrossRef]
- Lee H, Cho C, Lee T, Jeong J, Yeom JW, Kim S, et al. Prediction of impending mood episode recurrence using real-time digital phenotypes in major depression and bipolar disorders in South Korea: a prospective nationwide cohort study. Psychol Med. Sep 23, 2022;53(12):5636-5644. [CrossRef]
- Bai R, Xiao L, Guo Y, Zhu X, Li N, Wang Y, et al. Tracking and monitoring mood stability of patients with major depressive disorder by machine learning models using passive digital data: prospective naturalistic multicenter study. JMIR Mhealth Uhealth. Mar 8, 2021;9(3):e24365. [CrossRef]
- Colombo D, Palacios AG, Alvarez JF, Patané A, Semonella M, Cipresso P, et al. Current state and future directions of technology-based ecological momentary assessments and interventions for major depressive disorder: protocol for a systematic review. Syst Rev. Dec 13, 2018;7(1):233. [CrossRef]
- Colombo D, Fernández-Álvarez J, Patané A, Semonella M, Kwiatkowska M, García-Palacios A, et al. Current state and future directions of technology-based ecological momentary assessment and intervention for major depressive disorder: a systematic review. JCM. Apr 05, 2019;8(4):465. [CrossRef]
- Firth J, Torous J, Nicholas J, Carney R, Pratap A, Rosenbaum S, et al. The efficacy of smartphone‐based mental health interventions for depressive symptoms: a meta‐analysis of randomized controlled trials. World Psychiatry. Sep 21, 2017;16(3):287-298. [FREE Full text] [CrossRef] [Medline]
- ICD-11: International Classification of Diseases 11th Revision. World Health Organization. URL: https://icd.who.int/en [accessed 2025-02-20]
- Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition. Washington, DC. American Psychiatric Association; 2013.
- Kraus C, Kadriu B, Lanzenberger R, Zarate CA, Kasper S. Prognosis and improved outcomes in major depression: a review. Transl Psychiatry. Apr 03, 2019;9(1):127. [CrossRef]
- Wahle F, Kowatsch T, Fleisch E, Rufer M, Weidt S. Mobile sensing and support for people with depression: a pilot trial in the wild. JMIR Mhealth Uhealth. Sep 21, 2016;4(3):e111. [FREE Full text] [CrossRef] [Medline]
- Lisetti CL, Nasoz F. Using noninvasive wearable computers to recognize human emotions from physiological signals. EURASIP J Adv Signal Process. Sep 18, 2004;2004(11):929414. [CrossRef]
- Naughton F, Johnston D. A starter kit for undertaking n-of-1 trials. The European Health Psychologist. 2014;16(5):196-205. [FREE Full text]
- Bietz MJ, Bloss CS, Calvert S, Godino JG, Gregory J, Claffey MP, et al. Opportunities and challenges in the use of personal health data for health research. J Am Med Inform Assoc. 2016;23(e1):e42-e48. [CrossRef] [Medline]
- De Angel V, Lewis S, White K, Oetzmann C, Leightley D, Oprea E, et al. Digital health tools for the passive monitoring of depression: a systematic review of methods. NPJ Digit Med. Jan 11, 2022;5(1):3. [CrossRef]
- Adamakis M. Comparing the validity of a GPS monitor and a smartphone application to measure physical activity. JournalMTM. Aug 2017;6(2):28-38. [CrossRef]
- Plasqui G, Bonomi AG, Westerterp KR. Daily physical activity assessment with accelerometers: new insights and validation studies. Obesity Reviews. Feb 07, 2013;14(6):451-462. [CrossRef] [Medline]
- Schukat M, McCaldin D, Wang K, Schreier G, Lovell NH, Marschollek M, et al. Unintended consequences of wearable sensor use in healthcare. Yearb Med Inform. Mar 06, 2018;25(01):73-86. [FREE Full text] [CrossRef] [Medline]
- Canali S, Schiaffonati V, Aliverti A. Challenges and recommendations for wearable devices in digital health: Data quality, interoperability, health equity, fairness. PLOS Digit Health. Oct 13, 2022;1(10):e0000104. [CrossRef]
- Doherty G, Coyle D, Sharry J. Engagement with online mental health interventions: an exploratory clinical study of a treatment for depression. In: CHI '12: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 2012. Presented at: SIGCHI Conference on Human Factors in Computing Systems; May 5-10, 2012; Austin, TX. [CrossRef]
- Riddell WC, Song X. The role of education in technology use and adoption: evidence from the Canadian workplace and employee survey. ILR Review. Jan 13, 2017;70(5):1219-1253. [CrossRef]
- Trull TJ, Ebner-Priemer UW. Ambulatory assessment in psychopathology research: a review of recommended reporting guidelines and current practices. Journal of Abnormal Psychology. Jan 2020;129(1):56-63. [CrossRef]
Abbreviations
| CASP: Critical Appraisal Skills Program |
| DHT: digital health technology |
| ICD-10: International Classification of Diseases, version 10 |
| MeSH: Medical Subject Headings |
| PHQ-8: Patient Health Questionnaire-8 |
| PHQ-9: Patient Health Questionnaire-9 |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
Edited by T de Azevedo Cardoso; submitted 19.02.24; peer-reviewed by MM Qirtas, B Kinast; comments to author 01.11.24; revised version received 14.02.25; accepted 21.07.25; published 21.08.25.
Copyright©Rebeka Amin, Simon Schreynemackers, Hannah Oppenheimer, Milica Petrovic, Ulrich Hegerl, Hanna Reich. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 21.08.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

