<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v21i4e12910</article-id>
      <article-id pub-id-type="pmid">31025942</article-id>
      <article-id pub-id-type="doi">10.2196/12910</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Using Machine Learning to Derive Just-In-Time and Personalized Predictors of Stress: Observational Study Bridging the Gap Between Nomothetic and Ideographic Approaches</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Hekler</surname>
            <given-names>Erik</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Fernández</surname>
            <given-names>César</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="author" id="contrib1" corresp="yes">
          <name name-style="western">
            <surname>Rozet</surname>
            <given-names>Alan</given-names>
          </name>
          <degrees>BA</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Center for Behavioral Cardiovascular Health</institution>
            <institution>Columbia University Irving Medical Center</institution>
            <addr-line>Presbyterian Hospital Building, 9th Floor</addr-line>
            <addr-line>622 W 168th Street</addr-line>
            <addr-line>New York, NY, 10032</addr-line>
            <country>United States</country>
            <phone>1 212 342 4493</phone>
            <email>ar3793@cumc.columbia.edu</email>
          </address>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-4526-1891</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib2">
          <name name-style="western">
            <surname>Kronish</surname>
            <given-names>Ian M</given-names>
          </name>
          <degrees>MPH, MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0003-0945-2380</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib3">
          <name name-style="western">
            <surname>Schwartz</surname>
            <given-names>Joseph E</given-names>
          </name>
          <degrees>MS, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-8944-3566</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib4">
          <name name-style="western">
            <surname>Davidson</surname>
            <given-names>Karina W</given-names>
          </name>
          <degrees>MASc, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-9162-477X</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
      <label>1</label>
      <institution>Center for Behavioral Cardiovascular Health</institution>
      <institution>Columbia University Irving Medical Center</institution>  
      <addr-line>New York, NY</addr-line>
      <country>United States</country></aff>
      <aff id="aff2">
      <label>2</label>
      <institution>Feinstein Institute for Medical Research</institution>
      <institution>Northwell Health</institution>  
      <addr-line>New York, NY</addr-line>
      <country>United States</country></aff>
      <author-notes>
        <corresp>Corresponding Author: Alan Rozet 
        <email>ar3793@cumc.columbia.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection"><month>04</month><year>2019</year></pub-date>
      <pub-date pub-type="epub">
        <day>26</day>
        <month>04</month>
        <year>2019</year>
      </pub-date>
      <volume>21</volume>
      <issue>4</issue>
      <elocation-id>e12910</elocation-id>
      <!--history from ojs - api-xml-->
      <history>
        <date date-type="received">
          <day>28</day>
          <month>11</month>
          <year>2018</year>
        </date>
        <date date-type="rev-request">
          <day>3</day>
          <month>1</month>
          <year>2019</year>
        </date>
        <date date-type="rev-recd">
          <day>24</day>
          <month>1</month>
          <year>2019</year>
        </date>
        <date date-type="accepted">
          <day>3</day>
          <month>3</month>
          <year>2019</year>
        </date>
      </history>
      <copyright-statement>©Alan Rozet, Ian M Kronish, Joseph E Schwartz, Karina W Davidson. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 26.04.2019.</copyright-statement>
      <copyright-year>2019</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="http://www.jmir.org/2019/4/e12910/" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Investigations into person-specific predictors of stress have typically taken either a population-level nomothetic approach or an individualized ideographic approach. Nomothetic approaches can quickly identify predictors but can be hindered by the heterogeneity of these predictors across individuals and time. Ideographic approaches may result in more predictive models at the individual level but require a longer period of data collection to identify robust predictors.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>Our objectives were to compare predictors of stress identified through nomothetic and ideographic models and to assess whether sequentially combining nomothetic and ideographic models could yield more accurate and actionable predictions of stress than relying on either model. At the same time, we sought to maintain the interpretability necessary to retrieve individual predictors of stress despite using nomothetic models.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>Data collected in a 1-year observational study of 79 participants performing low levels of exercise were used. Physical activity was continuously and objectively monitored by actigraphy. Perceived stress was recorded by participants via daily ecological momentary assessments on a mobile app. Environmental variables including daylight time, temperature, and precipitation were retrieved from the public archives. Using these environmental, actigraphy, and mobile assessment data, we built machine learning models to predict individual stress ratings using linear, decision tree, and neural network techniques employing nomothetic and ideographic approaches. The accuracy of the approaches for predicting individual stress ratings was compared based on classification errors.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Across the group of patients, an individual’s recent history of stress ratings was most heavily weighted in predicting a future stress rating in the nomothetic recurrent neural network model, whereas environmental factors such as temperature and daylight, as well as duration and frequency of bouts of exercise, were more heavily weighted in the ideographic models. The nomothetic recurrent neural network model was the highest performing nomothetic model and yielded 72% accuracy for an 80%/20% train/test split. Using the same 80/20 split, the ideographic models yielded 75% accuracy. However, restricting ideographic models to participants with more than 50 valid days in the training set, with the same 80/20 split, yielded 85% accuracy.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>We conclude that for some applications, nomothetic models may be useful for yielding higher initial performance while still surfacing personalized predictors of stress, before switching to ideographic models upon sufficient data collection.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>ecological momentary assessment</kwd>
        <kwd>machine learning</kwd>
        <kwd>stress-behavior pathway</kwd>
        <kwd>personal informatics</kwd>
        <kwd>self-quantification</kwd>
        <kwd>exercise</kwd>
        <kwd>weather</kwd>
        <kwd>just-in-time interventions</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Deeper knowledge of the day-to-day effects of both weather and physical activity on stress can be valuable for creating personalized stress-reduction interventions on a just-in-time basis. Previous investigations have often focused on a nomothetic approach, pooling data to identify influential features across individuals [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. However, this approach typically has a drawback: Insight into any particular individual is limited due to heterogeneous effects of factors on individual-level stress and may not be generalized due to biological variability or overfitting [<xref ref-type="bibr" rid="ref3">3</xref>]. For example, hot weather may reduce stress for some participants, but increase stress for others. To remedy this, some have taken an ideographic approach, developing personalized stress-prediction models for each person [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>].</p>
      <p>Although this ideographic strategy resolves the issue of achieving per-individual insights, it does so by discarding potentially useful data from other individuals. The ideographic approach also requires the acquisition of at least some data about a given individual before making predictions for that individual. Identifying predictors of potentially low-frequency events such as occurrences of high stress may require substantial data collection before the ideographic model becomes sufficiently robust to confidently identify predictors. Thus, prior to obtaining reliable predictors of stress, a given individual may have to complete a long period of self-tracking, which may not be acceptable to some individuals.</p>
      <p>Here, we first compared individual-level predictors of stress identified through nomothetic models to those identified in ideographic models. We next explored the accuracy of a model-switching paradigm that begins with a nomothetic model and progressively changes to an ideographic model for data for the individual accumulate. We hypothesized that beginning with a nomothetic model would maximize accuracy during the early phase of data collection (referred to as a “warm start”) and subsequently, switching to ideographic modeling for higher personalization and performance would be the most effective approach to maximizing accuracy throughout data collection.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Overview</title>
        <p>This paper used the data collected in an observational study of 79 participants who were followed for up to 1 year, starting between January and July 2014; the study continuously and objectively monitored the physical activity of the participants by actigraphy and the perceived stress rating through ecological momentary assessment (EMA) reports on a mobile app [<xref ref-type="bibr" rid="ref6">6</xref>]. Participants were healthy individuals, aged 18 years or older, who responded to fliers posted throughout the buildings of Columbia University Medical Center (New York City, NY) and who, on phone screening, reported only intermittent engagement in exercise and having access to a personal computer and iOS or Android smartphone. Individuals with significant medical comorbidities, occupational work demands requiring rigorous physical activity, or inability to read and speak English were excluded. During a baseline interview, demographic characteristics including age, sex, race, ethnicity, education, partner status, and living situation were collected.</p>
      </sec>
      <sec>
        <title>Measures</title>
        <p>Stress was measured using an end-of-day text message survey on the participant’s own iPhone or Android phone, with the question “Overall, how stressful was your day?” Each evening, the participant was also asked, “Overall, how stressful do you think tomorrow will be?” Each morning, the participant responded to the questions “How stressful do you expect today to be?” and “How likely are you to exercise today?” All responses were rated on a scale from 0 (not at all) to 10 (extremely). All surveys were administered using Qualtrics software (Qualtrics, Seattle, WA). Two participants’ data were excluded for almost no variance in the self-reported stress ratings, leading to a total of 77 participants for the analysis.</p>
        <p>Physical activity was measured using a wrist-worn Fitbit (Fitbit, Inc, San Francisco, CA) to track daily physical activity, including the steps taken, calories burned, and intensity of physical activity for each minute of the day. Participants were instructed to sync and charge the device every 5 to 7 days. In this analysis, a bout of “exercise” was defined as any consecutive 30-minute period within which 24 or more minutes of moderate- or vigorous-intensity activity was performed. We followed the recommendations of Ward et al [<xref ref-type="bibr" rid="ref7">7</xref>] regarding best practices for the use of accelerometer data in research on physical activity. Specifically, physical activity guidelines recommend exercising for at least 30 minutes a day while accommodating interruptions. Further, when analyzing accelerometer data, the conventional approach is to quantify exercise in bouts of 10 minutes with allowances for 2 minutes of interruption (ie, total exercise for 8 of 10 min). Extrapolating the definition based on 10 to 30 minutes of activity, this yielded 24 of 30 minutes of activity. Software was written to determine, for each day, whether there was any 30-minute period within which at least 24 minutes of moderate or vigorous activity was performed; this was our objectively assessed measure of a 30-minute period of exercise. Days in which the Fitbit device was worn for fewer than 10 hours were excluded from all analyses.</p>
        <p>External and environmental variables, including temperature (high, low, average, and range), hours of daylight, precipitation, and day of the week, were retrieved from the meteorological station in Central Park (New York City, NY); these data are made publicly available by the National Oceanic and Atmospheric Administration’s National Center for Environmental Information.</p>
      </sec>
      <sec>
        <title>Statistical Analysis</title>
        <p>We developed models for stress rating using neural network, decision tree, and linear approaches across all participants, as well as participant-specific decision trees. Each model predicted a person’s self-reported stress rating (range, 0-10), using the previous 3 days of weather, self-reported stress, and actigraphy data. Because the previous 3 days were used as inputs, periods missing a stress rating were excluded.</p>
        <p>The continuous prediction for stress was then converted into a binary classification as either above or below a participant’s median self-reported stress value in the training set. We chose this approach, because if the stress rating was left as a continuous value, it was not clear at what rating an intervention should be initiated. Further, using the participant’s median as a reference slightly adjusts for participants who did not utilize the full range of the 11-point scale. In this way, the high-stress rating was individualized for each participant. The Keras Python library [<xref ref-type="bibr" rid="ref8">8</xref>] was used to train neural network models, and the scikit-learn Python library [<xref ref-type="bibr" rid="ref9">9</xref>] was used to train linear and decision tree-type models. Dropout, a regularization method for neural networks, was also used in the neural network models during training to decrease overfitting.</p>
        <p>Model performance was compared based on regression (mean absolute error) and classification (area under the curve [AUC], F<sub>1</sub> score, accuracy) error in predicting stress self-report. Feature importances were also compared between models. The AUC of a classifier is equivalent to the probability that the classifier will rank a randomly chosen positive instance</p>
        <p>higher than a randomly chosen negative instance [<xref ref-type="bibr" rid="ref10">10</xref>]. A model with an AUC closer to 1 is generally better. An F<sub>1</sub> score is the harmonic mean of a model’s precision and recall, with values between 0 and 1, in which values closer to 1 are better. Model selection was performed through exhaustive grid search of the corresponding hyperparameters for each model using 3-fold cross-validation and optimizing on the mean absolute error.</p>
        <p>For the ideographic decision trees, the Gini importance, as implemented in scikit-learn, was used to derive the significance of each input variable for each participant. Layer-wise relevance propagation (LRP) [<xref ref-type="bibr" rid="ref11">11</xref>] was used to interpret our neural network models. LRP propagates the relevance of each input variable back through the network from a specific prediction. In other words, for this dataset, a participant with 30 valid days in the dataset would yield 30 sets of LRP relevance scores, each set having one score for each input variable. These scores can be positive or negative in a similar fashion as linear coefficients, and the greater the magnitude of the score, the more that particular variable increased or decreased that particular prediction for stress rating. LRP was used in favor of other methods such as Deep Learning Important Features (DeepLIFT) and Integrated Gradients, because it does not require a baseline or reference value. However, LRP can be viewed as an approximation of DeepLIFT when bias terms are 0 and the reference values are set to 0 [<xref ref-type="bibr" rid="ref12">12</xref>]. This, in turn, approximates Shapley values [<xref ref-type="bibr" rid="ref13">13</xref>], which is another method of interpreting neural network output. As such, to leverage the visualizations for Shapley values built in the Shapley Additive Explanations Python library and to add an additional method of visual interpretation, bias terms for the neural net were locked to 0; this did not impact model performance.</p>
        <p>The following variables were used for the ideographic models: the number of hours with ≥10 activities, total exercise duration in minutes during bouts of exercise with ≥24 minutes of activity out of 30 minutes, the number of exercise bouts with ≥24 minutes of activity out of 30 minutes, the binary presence of any exercise bout of ≥24 minutes of activity out of 30 minutes, the total number of exercise bouts, the binary presence of any exercise bout, total exercise duration, stress rating, minimum outdoor temperature, maximum outdoor temperature, average outdoor temperature, total daylight minutes, and total precipitation. Several person-level variables were included in the nomothetic models that were not included in the per-participant models, as they would have been static for a given participant and would not have contributed to performance. These included height, weight, age, and sex. The inclusion of these variables in the nomothetic models slightly improved their performance.</p>
        <p>Additionally, all models were trained with varying training set sizes to test the hypothesis that ideographic models may be more reliable at large-enough training set sizes for each participant. Training set sizes ranged from 30% to 80% of valid days for each participant. For example, the first 30% of days in which a participant documented his or her stress was used to train a model predicting the subsequent 70% of days. The results for an 80% train/20% test split are highlighted here, with a total of 4050 training set samples and 678 testing set samples available to all nomothetic models.</p>
        <p>All models were compared to a baseline model that simply predicted the median of the stress self-report values from a participant’s training set, for all test samples.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Overview</title>
        <p>The average age of the participants was 32 years (range, 20-58 years), with a height of 169 cm and weight of 75 kg. The study sample included 34 men and 43 women in the dataset. <xref ref-type="table" rid="table1">Table 1</xref> depicts the basic descriptive statistics on the age, height, and weight of our set of participants.</p>
        <p>The pattern of stress ratings differed significantly among participants. <xref ref-type="fig" rid="figure1">Figure 1</xref> shows plots of stress by day for the 10 participants with the greatest number of valid responses in the dataset. Days without responses are shown without points plotted.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Basic summary statistics for the participants (N=77).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <thead>
              <tr valign="bottom">
                <td>Statistic</td>
                <td>Mean (SD)</td>
                <td>Min</td>
                <td>25%<sup>a</sup></td>
                <td>50%<sup>a</sup></td>
                <td>75%<sup>a</sup></td>
                <td>Max</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Age (years)</td>
                <td>31.62 (9.42)</td>
                <td>20</td>
                <td>24</td>
                <td>27</td>
                <td>38</td>
                <td>58</td>
              </tr>
              <tr valign="top">
                <td>Height (cm)</td>
                <td>168.71 (8.49)</td>
                <td>149.4</td>
                <td>162</td>
                <td>168</td>
                <td>176</td>
                <td>187</td>
              </tr>
              <tr valign="top">
                <td>Weight (kg)</td>
                <td>75.68 (17.46)</td>
                <td>43.3</td>
                <td>62.5</td>
                <td>74.7</td>
                <td>85.5</td>
                <td>138.6</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>Quartile ranges.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Plots of stress ratings for the participants with the greatest amount of responses.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Summary statistics and count of stress ratings, grouped by occurrence of missing stress ratings for the 3 days before the date of the predicted stress rating.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="120"/>
            <col width="120"/>
            <col width="160"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <thead>
              <tr valign="bottom">
                <td>Response</td>
                <td>Count</td>
                <td>Mean stress (SD)<break/><break/></td>
                <td>Min</td>
                <td>25%<sup>a</sup></td>
                <td>50%<sup>a</sup></td>
                <td>75%<sup>a</sup></td>
                <td>Max</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>000</td>
                <td>1728</td>
                <td>3.60 (2.58)</td>
                <td>0</td>
                <td>2</td>
                <td>3</td>
                <td>6</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>001</td>
                <td>1077</td>
                <td>3.40 (2.57)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>5</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>010</td>
                <td>292</td>
                <td>3.36 (2.63)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>5</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>011</td>
                <td>973</td>
                <td>3.31 (2.51)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>5</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>100</td>
                <td>973</td>
                <td>3.16 (2.54)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>4</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>101</td>
                <td>239</td>
                <td>3.35 (2.63)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>5</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>110</td>
                <td>889</td>
                <td>3.17 (2.48)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>4</td>
                <td>10</td>
              </tr>
              <tr valign="top">
                <td>111</td>
                <td>4060</td>
                <td>2.93 (2.35)</td>
                <td>0</td>
                <td>1</td>
                <td>3</td>
                <td>4</td>
                <td>10</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>Quartile ranges.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Percentage of stress ratings grouped by the occurrence of missing stress ratings for the 3 days before the date of the predicted stress rating. Each column corresponds to the value of the stress rating, and each cell represents what percentage of stress ratings had that value and response pattern.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="90"/>
            <col width="90"/>
            <col width="90"/>
            <col width="90"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <col width="80"/>
            <thead>
              <tr valign="top">
                <td>Response</td>
                <td>0</td>
                <td>1</td>
                <td>2</td>
                <td>3</td>
                <td>4</td>
                <td>5</td>
                <td>6</td>
                <td>7</td>
                <td>8</td>
                <td>9</td>
                <td>10</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>000</td>
                <td>11.92</td>
                <td>12.21</td>
                <td>14.47</td>
                <td>14.47</td>
                <td>15.74</td>
                <td>5.73</td>
                <td>8.91</td>
                <td>8.39</td>
                <td>4.40</td>
                <td>1.50</td>
                <td>2.26</td>
              </tr>
              <tr valign="top">
                <td>001</td>
                <td>13.65</td>
                <td>12.26</td>
                <td>16.06</td>
                <td>16.34</td>
                <td>12.91</td>
                <td>6.41</td>
                <td>8.73</td>
                <td>6.41</td>
                <td>2.79</td>
                <td>1.58</td>
                <td>2.88</td>
              </tr>
              <tr valign="top">
                <td>010</td>
                <td>17.47</td>
                <td>9.93</td>
                <td>13.36</td>
                <td>17.12</td>
                <td>13.70</td>
                <td>6.16</td>
                <td>7.88</td>
                <td>7.19</td>
                <td>3.42</td>
                <td>0.00</td>
                <td>3.77</td>
              </tr>
              <tr valign="top">
                <td>011</td>
                <td>14.59</td>
                <td>11.51</td>
                <td>14.59</td>
                <td>20.45</td>
                <td>11.51</td>
                <td>7.09</td>
                <td>7.71</td>
                <td>5.96</td>
                <td>3.19</td>
                <td>0.41</td>
                <td>2.98</td>
              </tr>
              <tr valign="top">
                <td>100</td>
                <td>17.57</td>
                <td>13.05</td>
                <td>13.77</td>
                <td>14.80</td>
                <td>16.14</td>
                <td>6.06</td>
                <td>6.37</td>
                <td>6.37</td>
                <td>2.26</td>
                <td>0.72</td>
                <td>2.88</td>
              </tr>
              <tr valign="top">
                <td>101</td>
                <td>17.57</td>
                <td>10.88</td>
                <td>13.39</td>
                <td>15.90</td>
                <td>13.81</td>
                <td>5.02</td>
                <td>6.69</td>
                <td>9.21</td>
                <td>5.02</td>
                <td>0.42</td>
                <td>2.09</td>
              </tr>
              <tr valign="top">
                <td>110</td>
                <td>16.20</td>
                <td>11.25</td>
                <td>15.75</td>
                <td>18.67</td>
                <td>14.40</td>
                <td>5.29</td>
                <td>7.31</td>
                <td>5.29</td>
                <td>2.25</td>
                <td>0.79</td>
                <td>2.81</td>
              </tr>
              <tr valign="top">
                <td>111</td>
                <td>15.62</td>
                <td>15.12</td>
                <td>17.02</td>
                <td>18.92</td>
                <td>13.42</td>
                <td>5.91</td>
                <td>4.98</td>
                <td>3.89</td>
                <td>2.09</td>
                <td>0.64</td>
                <td>2.39</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p><xref ref-type="table" rid="table2">Table 2</xref> shows the mean stress rating grouped by the occurrence of missing stress ratings for the 3 days before the date of the predicted stress rating. The response column describes the occurrence of stress ratings: 000 represents no stress ratings in any of the 3 days, 001 represents presence of only one stress rating recorded just before the predicted day, and 111 represents presence of all three stress ratings. Periods missing a stress rating were more often followed by a higher stress rating than periods that were not missing any stress ratings.</p>
        <p><xref ref-type="table" rid="table3">Table 3</xref> shows the percentage of stress ratings grouped by the occurrence of missing stress ratings for the 3 days before the date of the predicted stress rating. The frequency of high-stress ratings is slightly higher for response patterns with missing stress ratings.</p>
      </sec>
      <sec>
        <title>Accuracy of Nomothetic Versus Ideographic Models</title>
        <p>With an 80% training set/20% testing set split, the nomothetic recurrent neural network model AUC was 74.20% and the F<sub>1</sub> score was 79.21%. In addition, the per-participant decision tree AUC was 0.67 and the F<sub>1</sub> score was 0.83. The full results can be found in <xref ref-type="app" rid="app1">Multimedia Appendix 1</xref>. Nomothetic models, on an average, outperformed ideographic models for shorter training set sizes. As the training set size increased (ie, allowing for longer duration of assessments), group-level model performance generally increased across all models, and the nomothetic and ideographic model performances converged. However, when separately applying models to individual participants, ideographic models generally outperformed the nomothetic ones once the days with valid data exceeded 50 days.</p>
        <p><xref ref-type="table" rid="table4">Table 4</xref> lists performance metrics for each model, grouped by training set proportion and model used. Precision, recall, F<sub>1</sub> score, AUC, and accuracy are performance metrics. Training set size ranges from 30% to 80%. “Ideographic decision tree over 100 days” represents the performance metrics, only for participants who had at least 100 valid days, which is the same for the 50-day model.</p>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Performance metrics across all models, grouped by training set size.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="180"/>
            <col width="80"/>
            <col width="100"/>
            <col width="120"/>
            <col width="80"/>
            <col width="140"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Average number of days for training set size</td>
                <td>Precision (%)</td>
                <td>F<sub>1</sub> score (%)</td>
                <td>Area under the curve</td>
                <td>Accuracy (%)</td>
                <td>Training set size (%)</td>
                <td>Model</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>16.7215</td>
                <td>74.22</td>
                <td>71.95</td>
                <td>53.39</td>
                <td>60.69</td>
                <td>30.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>75.76</td>
                <td>72.63</td>
                <td>54.40</td>
                <td>61.53</td>
                <td>30.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>76.61</td>
                <td>71.86</td>
                <td>52.73</td>
                <td>60.21</td>
                <td>30.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>77.80</td>
                <td>78.77</td>
                <td>62.88</td>
                <td>69.75</td>
                <td>30.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>78.83</td>
                <td>78.72</td>
                <td>64.25</td>
                <td>70.10</td>
                <td>30.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>79.90</td>
                <td>77.93</td>
                <td>65.32</td>
                <td>69.69</td>
                <td>30.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>81.09</td>
                <td>72.69</td>
                <td>64.71</td>
                <td>65.18</td>
                <td>30.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>81.20</td>
                <td>81.03</td>
                <td>68.21</td>
                <td>73.36</td>
                <td>30.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>16.7215</td>
                <td>82.41</td>
                <td>77.39</td>
                <td>68.00</td>
                <td>70.01</td>
                <td>30.00</td>
                <td>Neural Network</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>75.35</td>
                <td>72.40</td>
                <td>54.44</td>
                <td>61.37</td>
                <td>40.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>76.48</td>
                <td>73.06</td>
                <td>57.12</td>
                <td>62.79</td>
                <td>40.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>76.96</td>
                <td>72.96</td>
                <td>56.32</td>
                <td>62.37</td>
                <td>40.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>77.87</td>
                <td>77.09</td>
                <td>62.08</td>
                <td>67.96</td>
                <td>40.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>78.15</td>
                <td>79.23</td>
                <td>63.16</td>
                <td>70.25</td>
                <td>40.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>80.93</td>
                <td>71.11</td>
                <td>63.74</td>
                <td>63.61</td>
                <td>40.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>81.47</td>
                <td>76.25</td>
                <td>66.23</td>
                <td>68.47</td>
                <td>40.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>81.70</td>
                <td>80.15</td>
                <td>68.15</td>
                <td>72.48</td>
                <td>40.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>21.7595</td>
                <td>82.33</td>
                <td>79.40</td>
                <td>68.55</td>
                <td>71.90</td>
                <td>40.00</td>
                <td>Neural Network</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>75.76</td>
                <td>75.36</td>
                <td>52.41</td>
                <td>63.47</td>
                <td>50.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>77.09</td>
                <td>75.95</td>
                <td>52.52</td>
                <td>64.06</td>
                <td>50.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>77.27</td>
                <td>75.56</td>
                <td>50.10</td>
                <td>63.07</td>
                <td>50.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>82.41</td>
                <td>82.65</td>
                <td>67.54</td>
                <td>74.60</td>
                <td>50.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>82.45</td>
                <td>81.94</td>
                <td>67.29</td>
                <td>73.80</td>
                <td>50.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>84.02</td>
                <td>82.71</td>
                <td>69.79</td>
                <td>75.15</td>
                <td>50.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>84.81</td>
                <td>76.31</td>
                <td>67.89</td>
                <td>68.57</td>
                <td>50.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>85.84</td>
                <td>83.89</td>
                <td>72.73</td>
                <td>77.00</td>
                <td>50.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>27.3291</td>
                <td>88.76</td>
                <td>83.09</td>
                <td>75.69</td>
                <td>76.79</td>
                <td>50.00</td>
                <td>Neural Network</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>78.54</td>
                <td>76.99</td>
                <td>54.63</td>
                <td>65.59</td>
                <td>60.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>78.85</td>
                <td>77.45</td>
                <td>53.40</td>
                <td>65.77</td>
                <td>60.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>79.41</td>
                <td>77.69</td>
                <td>54.87</td>
                <td>66.34</td>
                <td>60.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>85.90</td>
                <td>83.73</td>
                <td>69.13</td>
                <td>75.75</td>
                <td>60.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>86.53</td>
                <td>76.05</td>
                <td>66.83</td>
                <td>67.36</td>
                <td>60.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>88.76</td>
                <td>83.49</td>
                <td>73.24</td>
                <td>76.19</td>
                <td>60.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>89.80</td>
                <td>85.19</td>
                <td>75.62</td>
                <td>78.47</td>
                <td>60.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>90.78</td>
                <td>82.20</td>
                <td>75.20</td>
                <td>75.15</td>
                <td>60.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>32.7215</td>
                <td>91.35</td>
                <td>80.73</td>
                <td>75.08</td>
                <td>73.62</td>
                <td>60.00</td>
                <td>Neural Network</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>80.69</td>
                <td>80.90</td>
                <td>62.87</td>
                <td>71.65</td>
                <td>70.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>84.52</td>
                <td>82.50</td>
                <td>67.17</td>
                <td>74.09</td>
                <td>70.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>84.64</td>
                <td>83.78</td>
                <td>69.06</td>
                <td>75.96</td>
                <td>70.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>87.35</td>
                <td>85.71</td>
                <td>74.69</td>
                <td>79.23</td>
                <td>70.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>87.83</td>
                <td>83.81</td>
                <td>74.23</td>
                <td>77.08</td>
                <td>70.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>88.57</td>
                <td>84.21</td>
                <td>75.36</td>
                <td>77.72</td>
                <td>70.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>88.60</td>
                <td>88.99</td>
                <td>76.76</td>
                <td>83.25</td>
                <td>70.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>89.16</td>
                <td>80.69</td>
                <td>74.07</td>
                <td>73.88</td>
                <td>70.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>38.1646</td>
                <td>89.57</td>
                <td>82.76</td>
                <td>75.69</td>
                <td>76.28</td>
                <td>70.00</td>
                <td>Neural Network</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>81.44</td>
                <td>81.79</td>
                <td>63.94</td>
                <td>72.81</td>
                <td>80.00</td>
                <td>Baseline</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>84.10</td>
                <td>83.00</td>
                <td>67.72</td>
                <td>74.82</td>
                <td>80.00</td>
                <td>Ideographic Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>88.31</td>
                <td>90.07</td>
                <td>76.80</td>
                <td>84.62</td>
                <td>80.00</td>
                <td>Ideographic Decision Tree Over 50 Days</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>89.00</td>
                <td>84.20</td>
                <td>75.63</td>
                <td>77.71</td>
                <td>80.00</td>
                <td>Elastic Net</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>89.00</td>
                <td>83.43</td>
                <td>75.18</td>
                <td>76.80</td>
                <td>80.00</td>
                <td>Gradient Boosted Decision Tree</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>89.25</td>
                <td>84.70</td>
                <td>76.22</td>
                <td>78.35</td>
                <td>80.00</td>
                <td>Random Forest</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>90.38</td>
                <td>78.91</td>
                <td>74.20</td>
                <td>72.16</td>
                <td>80.00</td>
                <td>Recurrent Neural Network</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>90.57</td>
                <td>90.78</td>
                <td>79.87</td>
                <td>85.82</td>
                <td>80.00</td>
                <td>Ideographic Decision Tree Over 100 Days</td>
              </tr>
              <tr valign="top">
                <td>43.2025</td>
                <td>90.89</td>
                <td>73.87</td>
                <td>72.06</td>
                <td>67.27</td>
                <td>80.00</td>
                <td>Neural Network</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>Predictors of Stress</title>
        <p>In our ideographic models, there was significant heterogeneity in the effect of features of weather and exercise. <xref ref-type="fig" rid="figure2">Figure 2</xref> shows a series of horizontal bar plots, visualizing the Gini importance values, or the factors that significantly predicted day-to-day variability in stress rating for the 9 individuals with at least 100 valid days of data.</p>
        <p>The most frequent important predictors for each participant included daylight minutes, temperature, and exercise behavior for the current or preceding 3 days. Often, variables that were assessed closer (temporally) to the stress rating being predicted exhibited greater importance.</p>
        <p>Viewing feature importance in aggregate for all the participants in the dataset, the Gini importance values from their corresponding decision tree model were sorted and then ranked. Thereafter, the number of times that each input variable ranked in the top 5 spots across participants’ rankings was retrieved to create <xref ref-type="fig" rid="figure3">Figure 3</xref>. Decision tree models appear to have most frequently and highly ranked exercise-related features, followed by environmental variables.</p>
        <p><xref ref-type="fig" rid="figure4">Figure 4</xref> depicts the relevance scores for the recurrent neural network model, which were derived using LRP. Like the decision tree models, the scores ranged widely, even among participants. To retrieve a metric more comparable with the strictly positive Gini importance values derived from the decision trees, the absolute value of the LRP scores was taken for each participant, followed by the median value for each input variable. The neural network model often ranked the highest, preceding stress self-reports, but there was also significant variety across features such as weather (eg, average or minimum temperature on the day of the stress rating), exercise (eg, number of minutes of exercise or number of exercise bouts the previous day), age, height, and weight.</p>
        <p>In <xref ref-type="fig" rid="figure5">Figure 5</xref>, as with the decision tree models, the relevance of each input variable was ranked and then counted across participants.</p>
        <p><xref ref-type="fig" rid="figure6">Figure 6</xref> depicts the F<sub>1</sub> score, AUC, and accuracy for the nomothetic recurrent neural network model and the ideographic decision trees across multiple training set sizes. For individuals with fewer than 50 valid days, the nomothetic models generally outperformed or performed comparably as the ideographic ones. Nomothetic model performance and ideographic model performance converged as training set size per participant increased. However, for the 16 individuals with more than 50 valid days and the 9 individuals with over 100 valid days, ideographic performance overtook nomothetic performance. Each dot corresponds to a training set size proportion, ranging from 30% to 80% in increments of 10%. The number of days included in the training set for each proportion across all participants, on an average, was as follows: 17 days for 30%, 22 days for 40%, 27 days for 50%, 33 days for 60%, 38 days for 70%, and 43 days for 80%.</p>
        <p>Using LRP for the nomothetic model, a per-sample feature importance method, allows us to gain insight on what influenced a participant’s predicted stress score for a given day. In <xref ref-type="fig" rid="figure7">Figure 7</xref>, a particular participant’s actual stress self-reports, predicted stress self-reports, outside temperature, and the neural network model’s LRP values are depicted across several consecutive days. On day 5, the neural network model assigned a high importance to the average temperature. As shown in the figure, as the average temperature increased (temperature here is standardized to fit on the same scale as the other values), predicted stress and the true stress value increased. Note that in the following graphs, the left-hand y-axis contains the scale for the input variable and the LRP value, and the right-hand y-axis contains the scale for both predicted and true stress rating. The x-axis simply denotes consecutive dates, marked as integers, rather than true dates.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Most important predictors for the ideographic decision tree models for the 9 individuals with at least 100 valid days of data.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>A horizontally stacked bar chart of occurrences of the most frequently appearing predictor variables, and how often they ranked in the top 5 spots across participants’ predictor variable importance rankings from the ideographic models.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>A horizontal bar chart of the 15 predictor variables with the highest median absolute layer-wise relevance propagation scores. LRP: layer-wise relevance propagation.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>A horizontally stacked bar chart of occurrences of the most frequently appearing predictor variables, and how often they ranked in the top 5 spots across participants’ predictor variable importance rankings from the nomothetic recurrent neural network model.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>A plot comparing the F1 score, area under the curve, and accuracy across the neural network model and the ideographic models. Training set size varies on the x-axis, and performance of participants with more than 50 valid days and more than 100 valid days is shown separately for the ideographic models. The y-axis is the same in all 3 figures.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>A plot of the actual stress rating, predicted stress rating, layer-wise relevance propagation value, and average temperature for an individual participant, with average temperature standardized to fit on the same graph. LRP: layer-wise relevance propagation.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure8" position="float">
          <label>Figure 8</label>
          <caption>
            <p>A plot of the actual stress rating, predicted stress rating, layer-wise relevance propagation value, and average temperature for an individual participant, with the average temperature standardized to fit on the same graph. LRP: layer-wise relevance propagation.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>For another participant represented in <xref ref-type="fig" rid="figure8">Figure 8</xref>, an increase in the maximum temperature on a particular day was associated with a decrease in predicted stress, demonstrating individual-specific response patterns to environmental variables.</p>
        <p>Exercise, as measured through total minutes or number of separate bouts, was often inversely related with stress rating, where less exercise increased the predicted stress rating (<xref ref-type="fig" rid="figure9">Figure 9</xref>).</p>
        <p>However, there is a risk when interpreting feature importance presented as a time series. Although a variable may carry a high positive influence for a given day, it may be outweighed by a high cumulative negative impact of other variables, causing an improper inference. As such, it is useful to visualize a prediction for a given day not as a time series, but as a force plot of contributions from different variables, allowing us to more easily disentangle their influences. Treating LRP values as approximations for Shapley values, we can retrieve the following force plot depicted in <xref ref-type="fig" rid="figure10">Figure 10</xref> of feature importance for a specific prediction.</p>
        <p>Here, the predicted stress self-report value is 1.07. The stress rating 2 days ago of 2 and the stress rating 3 days ago of 6 push the prediction downward. The minimum temperature of 35°F 3 days ago and the absence of any exercise bouts of greater than 24 minutes of moderate-to-vigorous physical activity out of a 30-minute period push the prediction upward.</p>
        <p>We can also aggregate these force plots across a set of predictions to understand how our nomothetic neural network model behaves more generally. In <xref ref-type="fig" rid="figure11">Figure 11</xref>, each horizontal line displays the LRP values for an input variable, across the entire testing set. The variables are sorted by the sum of the magnitudes of their values, in descending order. The figure shows that higher stress ratings preceding a predicted stress rating typically increase predicted stress, whereas lower minimum temperatures typically increase predicted stress.</p>
        <fig id="figure9" position="float">
          <label>Figure 9</label>
          <caption>
            <p>A plot of the actual stress rating, predicted stress rating, layer-wise relevance propagation value, and exercise minutes for a participant, with exercise minutes standardized to fit on the same graph. A lower value for exercise minutes on the previous day was associated with a high layer-wise relevance propagation value and a higher predicted stress rating. LRP: layer-wise relevance propagation.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure10" position="float">
          <label>Figure 10</label>
          <caption>
            <p>A force plot visualizing the contributions of different input variables to a predicted stress rating for a single participant.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure11" position="float">
          <label>Figure 11</label>
          <caption>
            <p>Layer-wise relevance propagation values for each input variable across the entire testing set. Each horizontal line corresponds to a single input variable. LRP: layer-wise relevance propagation.</p>
          </caption>
          <graphic xlink:href="jmir_v21i4e12910_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Overview</title>
        <p>Many studies have examined relationships among exercise, weather, and stress using both linear and nonlinear approaches, and a mix of self-report questionnaire responses and automatically gathered sensing data. Some have taken an exclusively nomothetic approach; Wang et al [<xref ref-type="bibr" rid="ref1">1</xref>] used both EMA data and automatically sensed activity and sociability data to explore correlates of stress, depression, and loneliness, but did not use a personalized machine learning approach or investigate individual predictors. Using a large set of meteorological data paired with responses to a self-report questionnaire and a mixed modeling approach, Beecher et al [<xref ref-type="bibr" rid="ref2">2</xref>] found that increased sun exposure decreased reported distress.</p>
        <p>In contrast, other studies have taken an ideographic approach. Tuarob et al [<xref ref-type="bibr" rid="ref4">4</xref>] used a variety of machine learning techniques and questionnaire data to train ideographic models predicting participant mental states but relied on self-reported input data and did not investigate predictors for each individual. Sharmin et al [<xref ref-type="bibr" rid="ref14">14</xref>] used sensor data and self-reported stress reports to create personalized visualizations that were then manually examined for temporal trends in stress. Plarre et al [<xref ref-type="bibr" rid="ref5">5</xref>] trained ideographic decision trees using electrocardiographic and respiration-related data to predict self-reported stress after performing activities in the laboratory setting, such as public speaking or mental arithmetic. Burg et al [<xref ref-type="bibr" rid="ref6">6</xref>] analyzed the same EMA and exercise data as those used here and estimated ideographic random coefficient mixed models; they found that the influence of exercise on self-reported stress was heterogeneous, as was the effect of self-reported stress on exercise.</p>
        <p>Finally, taking an integrative approach and using the same dataset as that used by Burg et al [<xref ref-type="bibr" rid="ref6">6</xref>] and us, Cheung et al [<xref ref-type="bibr" rid="ref15">15</xref>] compared the performance of ideographic and nomothetic methods to predict whether an individual will exercise on a given day, again showing that for some, but not all participants, self-reported stress was a predictor of exercise.</p>
        <p>In this paper, heterogeneity in the effects of predictors on stress was confirmed, highlighting the value of using an ideographic modeling approach. Further, it was demonstrated that the nomothetic model performs better (ie, is more accurate) than the ideographic model initially, but as data collected accumulates, the performance of the ideographic model equals and then surpasses that of the nomothetic model, providing a motivating example for a “warm start” strategy to leverage the advantages of each modeling approach. Put into practice, one might continuously monitor the performance of a nomothetic model and an ideographic model on a particular individual and adjust the weight of each model’s contribution to the predicted output accordingly.</p>
        <p>Retrieving individual predictors yields hypotheses that we can test for a particular participant and, perhaps, act on. For example, if a clinician repeatedly sees that low average temperatures are driving higher predicted stress values, they may be able to recommend a particular intervention for the patient. Of course, not all situations are so easily interpretable, and the relationship of a predictor to the outcome may not be linear. This is both an advantage and a disadvantage, allowing a model to potentially be flexible to multiple climates but limiting the ease of generalizability to other participants, and requires either manual inspection as mentioned above or a more rigorous and automated method of consistent linear predictor detection.</p>
        <p>Here, the LRP values from our nomothetic neural network model, and Gini importance values from our ideographic decision trees, suggested different predictors for each individual. The nomothetic neural network model gave preceding stress self-reports higher weight on an average (<xref ref-type="fig" rid="figure10">Figure 10</xref>). This may be a result of training the neural network using dropout, in which increasing the weight given to preceding stress reports was an optimal way to manage the heterogeneity of participant response patterns. Notably, Sarker et al [<xref ref-type="bibr" rid="ref16">16</xref>] reported similar results that stress episodes increase the likelihood of subsequent episodes, although these were within-day data.</p>
        <p>Although <xref ref-type="fig" rid="figure4">Figure 4</xref> presents a small sample of individuals and a relatively small dataset overall, it depicts the LRP values for the nomothetic neural network model and motivates discussion of real-world applications of such a model for just-in-time predictions. From this, it seems likely that it is effective to start with a nomothetic model to maximize predictive performance and gain initial insights into the possible correlates of stress before switching to (or increasing the weight of, with an ensemble-type implementation) an ideographic model once enough responses have been collected. This assumes that immediate interventions are necessary, valuable, and worth the cost of potentially introducing bias in the dataset for that individual. Alternatively, predictions could be withheld until sufficient data are collected for ideographic models to be used, but this might come at the expense of disengaging participants during the process of data collection. More generally, adopting a framework, as proposed by Nahum-Shani et al [<xref ref-type="bibr" rid="ref17">17</xref>], to specify specific proximal outcomes while managing participant engagement is prudent.</p>
      </sec>
      <sec>
        <title>Future Directions and Limitations</title>
        <p>There is no current industry or academic standard for retrieving feature importance values from neural networks. Layer-wise relevance propagation, the method used here, has drawbacks of not meeting certain axiomatic properties of ideal feature importance methods [<xref ref-type="bibr" rid="ref18">18</xref>]. In this case, however, LRP was used because it does not require a baseline, whereas other methods require some reference input for comparison. This instance is not one in which it is immediately clear what that input would be. Other model-agnostic, sensitivity-based approaches exist, such as Locally Interpretable Model Explanations [<xref ref-type="bibr" rid="ref19">19</xref>]. Regardless of the approach for retrieving feature importance values, we retrieved per-participant predictors from the nomothetic neural network model by taking the median of the absolute value of LRP scores across a subset of the highest-stress events in a participant’s test set. Other strategies may yield features that are more representative of an individual’s stress. Further research in the field may be required to elaborate on strengths and weaknesses of different approaches in terms of interpretability, generalizability, and suitability for inclusion in a clinical decision process.</p>
        <p>Next, in this setting, individuals who deviate significantly from the rest of the population may create large gradient updates to a neural network, potentially reducing performance for other individuals. Although this can be mitigated through techniques such as batch normalization and dropout, these strategies, in turn, reduce the degree to which the model can closely fit the data and obscure important differences among individuals. Similarly, individuals contributing a larger number of samples to the training set can bias the model. If an exhaustive hyperparameter search is performed without regularization strategies like L1/L2 penalties, max-norm constraints, dropout, or early stopping, the model may be further overfit to these individuals. As a result, some settings might still be best served by an ideographic approach for safety, but even N-of-1 decision trees may need to be constrained in their complexity to prevent overfit, especially while still acquiring data.</p>
        <p>With self-report values, similar to item ratings, variable distributions may be skewed for particular individuals based on their perception of the scale. For example, in this dataset, some individuals never reported their stress to be above a value of 7, despite the scale going up to 10. Examples of these differing response patterns can be found in <xref ref-type="fig" rid="figure1">Figure 1</xref>. It is also highly unlikely that responses are missing at random. In fact, these days may be more stressful than the ones reported. The same issue may exist with Fitbit nonwear data. Time series forecasting methods often rely on either excluding or interpolating time windows that have missing data. Here, we do not interpolate missing data and instead, choose to exclude them. As a result, in situations in which self-report data are used, individual response patterns should be taken into account, whether through techniques such as feature engineering or increased data collection.</p>
        <p>In our analyses, samples with missing stress ratings prior to the predicted stress rating were excluded. Based on the observed difference in distributions between stress ratings preceded by missing stress ratings, and ones that were not preceded by such ratings, it is possible that the ratings may not be missing at random. Instead, unobserved stress ratings may be indicative of higher stress in some patients. As a result, we primarily explored imputation using a dummy value, but models trained using these values severely underperformed, likely due to the relatively small size of the dataset; therefore, those results were not reported here. Alternatively, missing stress ratings could be imputed using a rule-based approach such as incrementing the most recently observed stress rating. However, for a practical implementation, we believe that with sufficient dataset size, missing stress ratings could be replaced by dummy values to avoid manually biasing predictions and improve generalization. Further, current Fitbit devices and other accelerometers provide continuous heart rate monitoring data, which may provide additional useful predictors and mitigate the effect of missing stress ratings.</p>
        <p>Finally, as confidence intervals were not retrieved, we lacked a measure of certainty per prediction. This could be alleviated by using Bayesian neural network or dropout-based methods.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>Through the combination of a nomothetic neural network model, recent advances in retrieving per-sample feature importance, and ideographic decision trees, we show that high predictive performance can be achieved while recognizing individual differences and surfacing personalized predictors of stress. Key predictors in the nomothetic models were typically related to recent stress experience and weather activity. In addition, key predictors in the ideographic models displayed significant heterogeneity but were often weather or exercise related for individuals from whom more data were collected. Environmental variables were also shown to affect stress differently in different participants; for example, high temperatures predicted high stress in one individual but low stress in another. These predictors can be used to provide individuals with insights into what may contribute to their stress, as indicated by Yoon et al [<xref ref-type="bibr" rid="ref20">20</xref>]. These models can also be operationalized to generate interventions or encouragements just before instances of high stress when the model predicts, with a sufficient degree of confidence, impending stress based on what is expected to be effective for that particular individual. Finally, ideographic models surpassed a nomothetic one after sufficient data collection, supporting the use of a “warm start” model-switching approach. Further work is needed to explore interpretable and repeatable ways to assess personalized predictors in nonlinear settings, as applied to disentangling correlates of stress.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <app id="app1">
        <title>Multimedia Appendix 1</title>
        <p>Performance metrics for each model.</p>
        <media xlink:href="jmir_v21i4e12910_app1.pdf" xlink:title="PDF File (Adobe PDF File), 90KB"/>
      </app>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AUC</term>
          <def>
            <p>area under the curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">EMA</term>
          <def>
            <p>ecological momentary assessment</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">LRP</term>
          <def>
            <p>layer-wise relevance propagation</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was supported by funding (R01-HL115941) from the National Heart, Lung, and Blood Institute at the National Institutes of Health. IK was also supported by the National Center for Advancing Translational Sciences (UL1TR001873).</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>F</given-names>
          </name>
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Z</given-names>
          </name>
          <name name-style="western">
            <surname>Li</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Harari</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Tignor</surname>
            <given-names>S</given-names>
          </name>
        </person-group>
        <article-title>StudentLife: assessing mental health, academic performance and behavioral trends of college students using smartphones</article-title>
        <year>2014</year>  
        <conf-name>UbiComp: ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
        <conf-date>Sep 13-17, 2014</conf-date>
        <conf-loc>Seattle, WA</conf-loc>
        <pub-id pub-id-type="doi">10.1145/2632048.2632054</pub-id></nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Beecher</surname>
            <given-names>ME</given-names>
          </name>
          <name name-style="western">
            <surname>Eggett</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Erekson</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Rees</surname>
            <given-names>LB</given-names>
          </name>
          <name name-style="western">
            <surname>Bingham</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Klundt</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Bailey</surname>
            <given-names>RJ</given-names>
          </name>
          <name name-style="western">
            <surname>Ripplinger</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Kirchhoefer</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Gibson</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Griner</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Cox</surname>
            <given-names>JC</given-names>
          </name>
          <name name-style="western">
            <surname>Boardman</surname>
            <given-names>RD</given-names>
          </name>
        </person-group>
        <article-title>Sunshine on my shoulders: Weather, pollution, and emotional distress</article-title>
        <source>J Affect Disord</source>  
        <year>2016</year>  
        <month>12</month>  
        <day>15</day>  
        <volume>205</volume>  
        <fpage>234</fpage>  
        <lpage>238</lpage>  
        <pub-id pub-id-type="doi">10.1016/j.jad.2016.07.021</pub-id>
        <pub-id pub-id-type="medline">27449556</pub-id>
        <pub-id pub-id-type="pii">S0165-0327(16)30655-3</pub-id></nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Steyerberg</surname>
            <given-names>E</given-names>
          </name>
        </person-group>
        <source>Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating</source>  
        <year>2009</year>  
        <publisher-loc>New York</publisher-loc>
        <publisher-name>Springer</publisher-name></nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Tuarob</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Tucker</surname>
            <given-names>CS</given-names>
          </name>
          <name name-style="western">
            <surname>Kumara</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Giles</surname>
            <given-names>CL</given-names>
          </name>
          <name name-style="western">
            <surname>Pincus</surname>
            <given-names>AL</given-names>
          </name>
          <name name-style="western">
            <surname>Conroy</surname>
            <given-names>DE</given-names>
          </name>
          <name name-style="western">
            <surname>Ram</surname>
            <given-names>N</given-names>
          </name>
        </person-group>
        <article-title>How are you feeling?: A personalized methodology for predicting mental states from temporally observable physical and behavioral information</article-title>
        <source>J Biomed Inform</source>  
        <year>2017</year>  
        <month>12</month>  
        <volume>68</volume>  
        <fpage>1</fpage>  
        <lpage>19</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(17)30035-7"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1016/j.jbi.2017.02.010</pub-id>
        <pub-id pub-id-type="medline">28213145</pub-id>
        <pub-id pub-id-type="pii">S1532-0464(17)30035-7</pub-id>
        <pub-id pub-id-type="pmcid">PMC5453908</pub-id></nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Plarre</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Raij</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Hossain</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Ali</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Nakajima</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Al-absi</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Ertin</surname>
            <given-names>E</given-names>
          </name>
          <name name-style="western">
            <surname>Kamarck</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Kumar</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Scott</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Siewiorek</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Smailagic</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Wittmers</surname>
            <given-names>L</given-names>
          </name>
        </person-group>
        <article-title>Continuous inference of psychological stress from sensory measurements collected in the natural environment</article-title>
        <year>2011</year>  
        <conf-name>Proceedings of the 10th ACM/IEEE International Conference on Information Processing in Sensor Networks</conf-name>
        <conf-date>April 12-14, 2011</conf-date>
        <conf-loc>Chicago, IL</conf-loc></nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Burg</surname>
            <given-names>MM</given-names>
          </name>
          <name name-style="western">
            <surname>Schwartz</surname>
            <given-names>JE</given-names>
          </name>
          <name name-style="western">
            <surname>Kronish</surname>
            <given-names>IM</given-names>
          </name>
          <name name-style="western">
            <surname>Diaz</surname>
            <given-names>KM</given-names>
          </name>
          <name name-style="western">
            <surname>Alcantara</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Duer-Hefele</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Davidson</surname>
            <given-names>KW</given-names>
          </name>
        </person-group>
        <article-title>Does stress result in you exercising less? Or does exercising result in you being less stressed? Or is it both? Testing the bi-directional stress-exercise association at the group and person (N of 1) level</article-title>
        <source>Ann Behav Med</source>  
        <year>2017</year>  
        <month>12</month>  
        <volume>51</volume>  
        <issue>6</issue>  
        <fpage>799</fpage>  
        <lpage>809</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/28290065"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1007/s12160-017-9902-4</pub-id>
        <pub-id pub-id-type="medline">28290065</pub-id>
        <pub-id pub-id-type="pii">10.1007/s12160-017-9902-4</pub-id>
        <pub-id pub-id-type="pmcid">PMC5597451</pub-id></nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Ward</surname>
            <given-names>DS</given-names>
          </name>
          <name name-style="western">
            <surname>Evenson</surname>
            <given-names>KR</given-names>
          </name>
          <name name-style="western">
            <surname>Vaughn</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Rodgers</surname>
            <given-names>AB</given-names>
          </name>
          <name name-style="western">
            <surname>Troiano</surname>
            <given-names>RP</given-names>
          </name>
        </person-group>
        <article-title>Accelerometer use in physical activity: best practices and research recommendations</article-title>
        <source>Med Sci Sports Exerc</source>  
        <year>2005</year>  
        <month>11</month>  
        <volume>37</volume>  
        <issue>11 Suppl</issue>  
        <fpage>S582</fpage>  
        <lpage>8</lpage>  
        <pub-id pub-id-type="medline">16294121</pub-id>
        <pub-id pub-id-type="pii">00005768-200511001-00011</pub-id></nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="web">
        <source>GitHub</source>  
        <year>2018</year>  
        <access-date>2018-11-27</access-date>
        <comment>Keras team 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/keras-team/keras">https://github.com/keras-team/keras</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="74ENlDUSZ"/></comment> </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Pedregosa</surname>
            <given-names>F</given-names>
          </name>
          <name name-style="western">
            <surname>Varoquaux</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Gramfort</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Michel</surname>
            <given-names>V</given-names>
          </name>
          <name name-style="western">
            <surname>Thirion</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Grisel</surname>
            <given-names>O</given-names>
          </name>
        </person-group>
        <article-title>Scikit-learn: Machine learning in Python</article-title>
        <source>JMLR</source>  
        <year>2011</year>  
        <volume>12</volume>  
        <fpage>2825</fpage>  
        <lpage>2830</lpage> </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Fawcett</surname>
            <given-names>T</given-names>
          </name>
        </person-group>
        <article-title>An introduction to ROC analysis</article-title>
        <source>Pattern Recognition Letters</source>  
        <year>2006</year>  
        <month>6</month>  
        <volume>27</volume>  
        <issue>8</issue>  
        <fpage>861</fpage>  
        <lpage>874</lpage>  
        <pub-id pub-id-type="doi">10.1016/j.patrec.2005.10.010</pub-id></nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Bach</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Binder</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Montavon</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Klauschen</surname>
            <given-names>F</given-names>
          </name>
          <name name-style="western">
            <surname>Müller</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Samek</surname>
            <given-names>W</given-names>
          </name>
        </person-group>
        <article-title>On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation</article-title>
        <source>PLoS One</source>  
        <year>2015</year>  
        <month>7</month>  
        <volume>10</volume>  
        <issue>7</issue>  
        <fpage>e0130140</fpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://dx.plos.org/10.1371/journal.pone.0130140"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1371/journal.pone.0130140</pub-id>
        <pub-id pub-id-type="medline">26161953</pub-id>
        <pub-id pub-id-type="pii">PONE-D-14-20115</pub-id>
        <pub-id pub-id-type="pmcid">PMC4498753</pub-id></nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Shrikumar</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Greenside</surname>
            <given-names>P</given-names>
          </name>
          <name name-style="western">
            <surname>Kundaje</surname>
            <given-names>A</given-names>
          </name>
        </person-group>
        <article-title>Learning Important Features Through Propagating Activation Differences</article-title>
        <year>2017</year>  
        <conf-name>Proceedings of the 34th International Conference on Machine Learning</conf-name>
        <conf-date>August 6-11, 2017</conf-date>
        <conf-loc>Sydney, Australia</conf-loc>
        <fpage>3145</fpage>  
        <lpage>3153</lpage> </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Lundberg</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>S</given-names>
          </name>
        </person-group>
        <article-title>A unified approach to interpreting model predictions</article-title>
        <year>2017</year>  
        <conf-name>Neural Information Processing Systems</conf-name>
        <conf-date>Dec 4-9, 2017</conf-date>
        <conf-loc>Long Beach, CA</conf-loc></nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Sharmin</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Raij</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Epstien</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Nahum-Shani</surname>
            <given-names>I</given-names>
          </name>
          <name name-style="western">
            <surname>Beck</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Vhaduri</surname>
            <given-names>S</given-names>
          </name>
        </person-group>
        <article-title>Visualization of time-series sensor data to inform the design of just-in-time adaptive stress interventions</article-title>
        <year>2015</year>  
        <conf-name>Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
        <conf-date>Sep 07-11, 2015</conf-date>
        <conf-loc>Osaka, Japan</conf-loc></nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
        <name name-style="western">
        <surname>Cheung</surname>  
        <given-names>Y</given-names></name>  
        <name name-style="western">
        <surname>Hsueh</surname>  
        <given-names>P</given-names></name>  
        <name name-style="western">
        <surname>Qian</surname>  
        <given-names>M</given-names></name>  
        <name name-style="western">
        <surname>Yoon</surname>  
        <given-names>Sunmoo</given-names></name>  
        <name name-style="western">
        <surname>Meli</surname>  
        <given-names>Laura</given-names></name>  
        <name name-style="western">
        <surname>Diaz</surname>  
        <given-names>Keith M</given-names></name>  
        <name name-style="western">
        <surname>Schwartz</surname>  
        <given-names>Joseph E</given-names></name>  
        <name name-style="western">
        <surname>Kronish</surname>  
        <given-names>Ian M</given-names></name>  
        <name name-style="western">
        <surname>Davidson</surname>  
        <given-names>Karina W</given-names></name> </person-group>
        <article-title>Are Nomothetic or Ideographic Approaches Superior in Predicting Daily Exercise Behaviors?</article-title>
        <source>Methods Inf Med</source>  
        <year>2017</year>  
        <volume>56</volume>  
        <issue>6</issue>  
        <fpage>452</fpage>  
        <lpage>460</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29582914"/>
        </comment>  
        <pub-id pub-id-type="doi">10.3414/ME16-02-0051</pub-id>
        <pub-id pub-id-type="medline">29582914</pub-id>
        <pub-id pub-id-type="pmcid">PMC5897129</pub-id></nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Sarker</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Hovsepian</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Chatterjee</surname>
            <given-names>S</given-names>
          </name>
        </person-group>
        <article-title>From markers to interventions: the case of just-in-time stress intervention</article-title>
        <source>Mobile Health: Sensors, Analytic Methods, and Applications</source>  
        <year>2017</year>  
        <publisher-loc>New York</publisher-loc>
        <publisher-name>Springer</publisher-name>
        <fpage>411</fpage>  
        <lpage>433</lpage> </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Nahum-Shani</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Smith</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Tewari</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Witkiewitz</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Collins</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Spring</surname>
            <given-names>B</given-names>
          </name>
        </person-group>
        <source>Just-in-Time Adaptive Interventions (JITAIs): An Organizing Framework for Ongoing Health Behavior Support (Technical Report No. 14-126)</source>  
        <year>2014</year>  
        <publisher-loc>University Park, PA</publisher-loc>
        <publisher-name>The Methodology Center</publisher-name></nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Sundararajan</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Taly</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Yan</surname>
            <given-names>Q</given-names>
          </name>
        </person-group>
        <article-title>Axiomatic Attribution for Deep Networks</article-title>
        <source>ICML</source>  
        <year>2017</year>  
        <conf-name>ICML</conf-name>
        <conf-date>June 10-15, 2017</conf-date>
        <conf-loc>Long Beach, CA</conf-loc></nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Ribeiro</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Singh</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Guestrin</surname>
            <given-names>C</given-names>
          </name>
        </person-group>
        <article-title>"Why Should I Trust You?": Explaining the predictions of any classifier</article-title>
        <year>2016</year>  
        <conf-name>ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>
        <conf-date>August 13-17, 2016</conf-date>
        <conf-loc>San Francisco, CA</conf-loc>
        <pub-id pub-id-type="doi">10.1145/2939672.2939778</pub-id></nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
        <name name-style="western">
        <surname>Yoon</surname>  
        <given-names>S</given-names></name>  
        <name name-style="western">
        <surname>Schwartz</surname>  
        <given-names>J</given-names></name>  
        <name name-style="western">
        <surname>Burg</surname>  
        <given-names>M</given-names></name>  
        <name name-style="western">
        <surname>Kronish</surname>  
        <given-names>Ian M</given-names></name>  
        <name name-style="western">
        <surname>Alcantara</surname>  
        <given-names>Carmela</given-names></name>  
        <name name-style="western">
        <surname>Julian</surname>  
        <given-names>Jacob</given-names></name>  
        <name name-style="western">
        <surname>Parsons</surname>  
        <given-names>Faith</given-names></name>  
        <name name-style="western">
        <surname>Davidson</surname>  
        <given-names>Karina W</given-names></name>  
        <name name-style="western">
        <surname>Diaz</surname>  
        <given-names>Keith M</given-names></name> </person-group>
        <article-title>Using Behavioral Analytics to Increase Exercise: A Randomized N-of-1 Study</article-title>
        <source>Am J Prev Med</source>  
        <year>2018</year>  
        <month>12</month>  
        <volume>54</volume>  
        <issue>4</issue>  
        <fpage>559</fpage>  
        <lpage>567</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29429607"/>
        </comment>  
        <pub-id pub-id-type="doi">10.1016/j.amepre.2017.12.011</pub-id>
        <pub-id pub-id-type="medline">29429607</pub-id>
        <pub-id pub-id-type="pii">S0749-3797(17)30743-2</pub-id>
        <pub-id pub-id-type="pmcid">PMC5860951</pub-id></nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
