<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v27i1e66100</article-id><article-id pub-id-type="doi">10.2196/66100</article-id><article-categories><subj-group subj-group-type="heading"><subject>Tutorial</subject></subj-group></article-categories><title-group><article-title>AI and Machine Learning Terminology in Medicine, Psychology, and Social Sciences: Tutorial and Practical Recommendations</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Cao</surname><given-names>Bo</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Greiner</surname><given-names>Russell</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Greenshaw</surname><given-names>Andrew</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Sui</surname><given-names>Jie</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff5">5</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Psychiatry, University of Alberta</institution><addr-line>4-142A Katz Group Centre for Research, 11315 - 87 Ave NW</addr-line><addr-line>Edmonton</addr-line><addr-line>AB</addr-line><country>Canada</country></aff><aff id="aff2"><institution>Department of Computing Science, Faculty of Science, University of Alberta</institution><addr-line>Edmonton</addr-line><addr-line>AB</addr-line><country>Canada</country></aff><aff id="aff3"><institution>School of Public Health, University of Alberta</institution><addr-line>Edmonton</addr-line><addr-line>AB</addr-line><country>Canada</country></aff><aff id="aff4"><institution>Alberta Machine Intelligence Institute (Amii)</institution><addr-line>Edmonton</addr-line><addr-line>AB</addr-line><country>Canada</country></aff><aff id="aff5"><institution>School of Psychology, University of Aberdeen</institution><addr-line>Aberdeen</addr-line><country>United Kingdom</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mavragani</surname><given-names>Amaryllis</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Raymond Guo</surname><given-names>L</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Wei Lam</surname><given-names>Sean Shao</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to  Bo Cao, PhD, Department of Psychiatry, University of Alberta, 4-142A Katz Group Centre for Research, 11315 - 87 Ave NW, Edmonton, AB, T6G 2B7, Canada, 1 7804929576; <email>cloudbocao@gmail.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>18</day><month>8</month><year>2025</year></pub-date><volume>27</volume><elocation-id>e66100</elocation-id><history><date date-type="received"><day>03</day><month>09</month><year>2024</year></date><date date-type="rev-recd"><day>16</day><month>05</month><year>2025</year></date><date date-type="accepted"><day>23</day><month>05</month><year>2025</year></date></history><copyright-statement>&#x00A9; Bo Cao, Russell Greiner, Andrew Greenshaw, Jie Sui. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 18.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2025/1/e66100"/><abstract><p>Recent applications of artificial intelligence (AI) and machine learning in medicine, psychology, and social sciences have led to common terminological confusions. In this paper, we review emerging evidence from systematic reviews documenting widespread misuse of key terms, particularly &#x201C;prediction&#x201D; being applied to studies merely demonstrating association or retrospective analysis. We clarify when &#x201C;prediction&#x201D; should be used and recommend using &#x201C;prospective prediction&#x201D; for future prediction; explain validation procedures essential for model generalizability; discuss overfitting and generalization in machine learning and traditional regression methods; clarify relationships between features, independent variables, predictors, risk factors, and causal factors; and clarify the hierarchical relationship between AI, machine learning, deep learning, large language models, and generative AI. We provide evidence-based recommendations for terminology use that can facilitate clearer communication among researchers from different disciplines and between the research community and the public, ultimately advancing the rigorous application of AI in medicine, psychology, and social sciences.</p></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>machine learning</kwd><kwd>terminology</kwd><kwd>medicine</kwd><kwd>psychology</kwd><kwd>social sciences</kwd><kwd>prediction</kwd><kwd>regression</kwd><kwd>deep learning</kwd><kwd>tutorial</kwd><kwd>prospective prediction</kwd><kwd>validation</kwd></kwd-group></article-meta></front><body><sec id="s1"><title>Challenges</title><p>The rapid growth of artificial intelligence (AI) and machine learning (ML) in medicine, psychology, and social sciences has led to a proliferation of terminology that is often inconsistently and imprecisely applied [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. This inconsistency creates confusion for researchers, clinicians, and policy makers who need to interpret and apply these technologies. Perhaps the most notable example of terminological confusion in the field involves the concept of &#x201C;prediction,&#x201D; which is frequently misused in the literature [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. However, similar confusion extends to other key terms such as &#x201C;validation,&#x201D; &#x201C;features,&#x201D; and even the distinctions between AI, ML, and deep learning (DL) [<xref ref-type="bibr" rid="ref1">1</xref>].</p><p>The confusion about &#x201C;prediction&#x201D; is particularly widespread and merits special attention. Yarkoni and Westfall [<xref ref-type="bibr" rid="ref5">5</xref>] have argued that psychology&#x2019;s emphasis on explaining the causes of behavior rather than predicting future behavior has led to research programs that provide intricate theories but have little ability to predict future behaviors with appreciable accuracy. McCall et al [<xref ref-type="bibr" rid="ref4">4</xref>] highlight this issue in the context of sports science and medicine, noting that much of the confusion stems from a mismatch between statistical modeling and subsequent interpretation of findings. After examining the literature in sports science that claimed to predict performance, talent, or injury, they found that the vast majority of studies actually analyzed association rather than prediction. This confusion can have serious consequences for practitioners who rely on these studies to make decisions about athlete training, selection, or rehabilitation. Varga et al [<xref ref-type="bibr" rid="ref3">3</xref>] conducted a systematic review in the field of diabetes and found that the term &#x201C;prediction&#x201D; is often incorrectly used to refer to association statistics. The confusion between association and prediction can lead to inflated expectations about the capabilities of models, and even biomarkers with strong statistical associations, causal biomarkers, or biomarkers with clear roles in disease pathophysiology can be poor predictors of future disease in individuals [<xref ref-type="bibr" rid="ref3">3</xref>].</p><p>In addition to misinterpreting association as predictions, the term &#x201C;prediction&#x201D; has been frequently misused as a marketing hot term for retrospective observational studies that lack external or future (prospective) validation in broader medical literature. In a review of 152 ML studies, Navarro et al [<xref ref-type="bibr" rid="ref2">2</xref>] found that most ML prediction model studies suffered from poor methodology and reporting. They noted that the majority of these studies were retrospective in nature and lacked proper external validation (87% with only internal validation). Ramspek et al [<xref ref-type="bibr" rid="ref6">6</xref>] noted that external validation remains rare in prediction model studies, with few models being validated in independent datasets. This lack of validation raises serious concerns about the generalizability of these models to new populations or settings and true predictions for future outcomes. Similarly, Nagendran et al [<xref ref-type="bibr" rid="ref7">7</xref>] found that of 81 nonrandomized clinical trials identified using DL, only 9 were prospective and 6 were tested in a real-world clinical settin, and Abdulazeem et al [<xref ref-type="bibr" rid="ref8">8</xref>] found only 8 out of 106 (7.6%) ML studies in primary health care were prospective in desgin.</p><p>According to all the aforementioned studies, there have been at least four distinct categories of studies that are commonly conflated under the umbrella of &#x201C;prediction.&#x201D; (1) Association studies (correlation or risk factor analysis mislabeled as &#x201C;prediction&#x201D;). (2) Retrospective modeling without external validation (models aiming to &#x201C;predict&#x201D; unknown outcomes and developed on an existing dataset without external or future validation), including studies showing prediction metrics without mentioning validation. (3) Retrospective modeling with external validation (models aiming to &#x201C;predict&#x201D; unknown outcomes and developed on existing data with external validation on an independent dataset but without future prediction). (4) Prospective prediction studies (models predicting future outcomes on new data or through a prospective design).</p><p><xref ref-type="table" rid="table1">Table 1</xref> summarizes the prevalence of these categories across domains based on systematic reviews and perspective papers from a sample of papers from medicine and psychology literature reporting potential issues of &#x201C;prediction&#x201D; studies. This table reveals several concerning trends.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Prevalence of different types of studies labeled as &#x201C;prediction.&#x201D;</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Domain</td><td align="left" valign="bottom">Association studies</td><td align="left" valign="bottom">Retrospective without external validation</td><td align="left" valign="bottom">Retrospective with external validation</td><td align="left" valign="bottom">Prospective prediction</td><td align="left" valign="bottom">Reference</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="6">Review and perspective studies on prediction</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Diabetes</td><td align="left" valign="top">61%</td><td align="left" valign="top">39%</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td><td align="left" valign="top">N/A</td><td align="left" valign="top">Varga et al [<xref ref-type="bibr" rid="ref3">3</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Psychology</td><td align="left" valign="top">Widespread</td><td align="left" valign="top">Widespread</td><td align="left" valign="top">Rare</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Yarkoni and Westfall [<xref ref-type="bibr" rid="ref5">5</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Sports science</td><td align="left" valign="top">77% (performance), 90% (talent), and 65% (injury)</td><td align="left" valign="top">10%&#x2010;35%</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">McCall et al [<xref ref-type="bibr" rid="ref4">4</xref>]</td></tr><tr><td align="left" valign="top" colspan="6">Review studies on machine learning</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Clinical prediction</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Most</td><td align="left" valign="top">0.5% in 1990<break/>7% by 2019</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Ramspek et al [<xref ref-type="bibr" rid="ref6">6</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Health care</td><td align="left" valign="top">N/A</td><td align="left" valign="top">87.00%</td><td align="left" valign="top">13.00%</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Navarro et al [<xref ref-type="bibr" rid="ref2">2</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Medical imaging</td><td align="left" valign="top">N/A</td><td align="left" valign="top">45.70%</td><td align="left" valign="top">43.00%</td><td align="left" valign="top">11.00%</td><td align="left" valign="top">Nagendran et al [<xref ref-type="bibr" rid="ref7">7</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Primary health care</td><td align="left" valign="top">N/A</td><td align="left" valign="top">76.50%</td><td align="left" valign="top">24%</td><td align="left" valign="top">7.60%</td><td align="left" valign="top">Abdulazeem et al [<xref ref-type="bibr" rid="ref8">8</xref>]</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>Not applicable.</p></fn></table-wrap-foot></table-wrap><p>First, the majority of studies labeled as &#x201C;prediction&#x201D; are actually association studies, characterized by the absence of model validation with new or unseen data. Second, these studies discuss the confusion surrounding other key terminological confusions that are relevant to but beyond &#x201C;prediction&#x201D;; for example, there are often misunderstandings of validation techniques, the assessment of generalizability [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref5">5</xref>], and the role of features, and independent variables. Finally, there is considerable confusion about the distinctions between traditional statistical approaches such as linear regression and contemporary ML techniques. It is important to note that such categories may not reflect the crucial temporal essence of prediction (eg, predicting future outcomes) or the predictability or generalizability of a model (eg, the ability to predict the outcome on validation data collected in the future). Additionally, there is also confusion about the basic terminology of AI methodologies. Terms such as &#x201C;artificial intelligence,&#x201D; &#x201C;machine learning,&#x201D; and &#x201C;deep learning&#x201D; are often used without appropriate definitional boundaries or used interchangeably. This terminological confusion has significant implications for research, practice, and policy in medicine, psychology, and social sciences. It can lead to misleading claims about model performance, inappropriate application of models in clinical settings, and inefficient use of research resources. As AI and ML continue to permeate these fields, there is an urgent need for greater clarity and consistency in the use of terminology.</p><p>In summary, most &#x201C;prediction&#x201D; studies predominantly explain the relationships between variables and outcomes within existing data, rather than establishing generalizable prediction through rigorous validation with new data. Even among studies using appropriate ML prediction models with both training and validation components, most implement only basic predictive methodologies, focusing on prediction with retrospective modeling and lacking external validation. True prospective prediction studies remain rare, with comprehensive explanations of prediction terminologies absent (see section &#x201C;Prediction&#x201D;).</p><p>The following sections will examine these terminological challenges and provide practical recommendations. We begin with the concept of prediction, with a focus on the temporal perspective of prospective prediction and offering operational implementation tutorial (see section &#x201C;Prediction&#x201D;), followed by testing and validation methodologies in the context of evaluating the generalizability of prediction (see section &#x201C;Testing, Validation, Cross-Validation, and External Validation&#x201D;), addressing overfitting phenomena, and comparing regular and ML regression models (see section &#x201C;Overfitting, Linear Regression, Regularization, and ML&#x201D;), distinctions between features, independent variables, predictors, risk factors, and causal factors (see section &#x201C;Feature, Independent Variable, Predictor, Risk Factor, and Causal Factor&#x201D;), and finally, clarification of different AI methodologies (see section &#x201C;AI, ML, DL, Large Language Models, and Generative AI&#x201D;). Through this terminological precision, we aim to provide a foundation for more rigorous and transparent interdisciplinary research that offers practical implementation pathways for AI applications in medicine, psychology, and social sciences.</p></sec><sec id="s2"><title>Prediction</title><sec id="s2-1"><title>Overview</title><p>The term &#x201C;prediction&#x201D; has been widely used in different disciplines, while no real prediction of future outcomes is present [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. Historically, a regression or correlation model may be considered in medicine, psychology, and social sciences as a prediction model [<xref ref-type="bibr" rid="ref9">9</xref>] simply because it reflects the fit or mathematical modeling between independent variables (eg, X, a set of clinical or cognitive assessments, or brain structure or activity) and dependent variables (eg, Y, the risk for a specific mental disorder, or social behavior or norms). This is reasonable, as the model can inform us about the potential value of Y when we have an observation of X. However, such models and analyses usually cannot be generalized to individual-level prediction on new data [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref10">10</xref>] (see section &#x201C;Overfitting, Linear Regression, Regularization, and ML&#x201D;).</p><p>Unlike traditional regression and correlation methods that primarily focus on identifying relationships between variables and outcomes within a single dataset, ML algorithms&#x2014;both classification (predicting binary or categorical outcomes) and regression (predicting continuous outcomes; see section &#x201C;Overfitting, Linear Regression, Regularization, and ML&#x201D;)&#x2014;are designed explicitly with generalization in mind. They typically involve validation procedures (eg, training models on a subset of data and validating its performance on an unseen subset; see section &#x201C;Testing, Validation, Cross-Validation, and External Validation&#x201D;), enabling robust prediction of new or future outcomes across individuals or within individuals over time [<xref ref-type="bibr" rid="ref11">11</xref>]. In this paper, we will focus on the prediction of an explicitly measurable outcome that can be used to verify the prediction (eg, a diagnosis or treatment outcome), as commonly considered as supervised learning in ML (compared with unsupervised learning, eg, clustering) [<xref ref-type="bibr" rid="ref12">12</xref>].</p><p>With increasing adoption of ML in research and practice in medicine, psychology, and social sciences, using the term &#x201C;prediction&#x201D; for the fit of a regression or correlation model is often no longer appropriate, and prediction, as its literal meaning, should have a component that is new to the existing observations or assessments. A summarized fitting of a mathematical model between variables in an existing dataset should be referred to as a fitting or explanatory model, rather than a prediction model, until it is validated on new data, and the corresponding study should be referred to as an association study.</p><p>In practice, we should refer to an outcome generated by a model as a &#x201C;prediction&#x201D; only when the outcome has not been observed yet and can only be obtained in the future. Ideally, this should be the case. However, future observations with sufficient validity may be challenging and take time to obtain and, often, not feasible for secondary data analysis on existing data from a research team that is not the original data owner. Given these practical constraints, many models and validation techniques can provide reasonable estimation about predictions to a different population (eg, models developed with a younger adult dataset generalized to older populations [<xref ref-type="bibr" rid="ref13">13</xref>]) or the same sample over time [<xref ref-type="bibr" rid="ref14">14</xref>], and we recommend that &#x201C;prediction&#x201D; should include scenarios when we can show evidence of how a model will perform on new data relative to the data used to develop the model. The new data need not be strictly &#x201C;future&#x201D; compared with the existing data but must remain previously unseen by the model during model training. With this expanded operational definition, we can use the term &#x201C;prediction&#x201D; for outcomes generated by models trained with existing data and then applied to new data, while we recommend using <italic>Prospective Prediction (or Forecasting</italic>) to specifically refer to the predictions aiming for future outcomes to emphasize the temporal aspect of the prediction [<xref ref-type="bibr" rid="ref14">14</xref>]. The &#x201C;future&#x201D; can be relative to the time of model training or the time of data collection of predicting variables in a retrospective dataset, where the validation data are held out or collected already at a later time, or relative to the current time, when the validation data are not yet collected. Thus, it is important for researchers to explicitly report the specific characteristics of the &#x201C;prediction&#x201D; as summarized in <xref ref-type="fig" rid="figure1">Figure 1</xref>, which distinguishes 4 types of prediction based on temporal relationships between training and validation. We focus on internal validation here, but external validation can be applied to replace the holdout internal validation data or be used as fully independent validation data (Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Prediction of current and future outcomes with current and future validation.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66100_fig01.png"/></fig><list list-type="order"><list-item><p>The ML model itself, with internal validation, is trained only on current outcomes observed within the same time frame of the variables used to predict the unknown outcomes (input to the ML model, ie, features; see section &#x201C;Feature, Independent Variable, Predictor, Risk Factor, and Causal Factor&#x201D;). Its predictability is estimated through the validation of the predicted outcomes using current features on a holdout dataset from the same dataset within the same time frame. Such a model and procedure can be used to develop prediction tools (eg, screening) for a current disease using assessments performed at the same time frame, but it needs further longitudinal validation on future data to confirm its generalizability over time or further external validation to confirm its generalizability across populations (Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p></list-item><list-item><p>The ML model itself is trained only on current outcomes observed within the same time window of features, and its predictability is estimated through the validation of the predicted outcomes using features on a future collection of data (eg, time 2). Such a future collection can be from the same people (eg, a whole population) or an independent cohort. Such a model and procedure can be used to develop prediction tools for a current disease using assessments collected at the same time frame, and its predictability is validated by future data, which greatly strengthens its potential application. However, such a model is not designed to predict future outcomes as their training design remains anchored to the relationships between variables and outcomes collected in the same time frame.</p></list-item><list-item><p>The ML model is trained on outcomes observed in a later time frame of the features. Its predictability is estimated through the validation of the predicted future outcomes (eg, time 2) using current predictors (eg, time 1) on a holdout dataset from the same dataset. Such a model and procedure can be used to develop prediction tools for a future disease (eg, onset of major depressive or anxiety disorder [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]) using assessments collected at an earlier time point. While providing valuable predictive potential, further longitudinal validation of future data is needed.</p></list-item><list-item><p>The ML model is trained on outcomes observed in a later time frame of the features, and its predictability is estimated through the validation of the predicted outcomes using features collected in future (eg, time 2) and outcomes observed in an even later time frame of the features (eg, time 3). Such a model and procedure can be used to develop prediction tools for a future disease or disease progression using earlier assessments, and its predictability is validated by future data, which greatly strengthens its potential application for real-world forecasting. This approach can also be applied to data collections that do not have discrete time points (eg, streaming data [<xref ref-type="bibr" rid="ref17">17</xref>]), as long as there are clear time stamps for features and outcomes.</p></list-item></list><p>One particular strength of ML approaches over traditional statistical analyses is that they provide predictions at the individual level rather than merely establishing group-level statistical significance. For example, in a typical statistical <italic>t</italic> test, the difference between the means of 2 distributions of an assessment from 2 populations (eg, patients and healthy subjects) is tested against the null hypothesis that the means are not different. Typically, as we have more observations, we will be more confident in estimating the means, thus reaching a higher significance level. However, this increased significance usually does not provide us with a better way to identify a new individual case randomly picked from the 2 populations (eg, patients vs healthy subjects) based on this assessment, because such identification depends only on the discrimination capability between the distributions of the original assessment metrics between the 2 populations and not the distributions of their means. Thus, a higher significance level of a variable at the group level does not automatically lead to better distinction of individual cases between 2 groups or make the variable a better predictor for one of the groups [<xref ref-type="bibr" rid="ref18">18</xref>]. On the other hand, ML can identify individual cases rather than group characteristic differences [<xref ref-type="bibr" rid="ref19">19</xref>-<xref ref-type="bibr" rid="ref23">23</xref>]. In many practical scenarios, identifying individual cases based on high-dimensional potential predicting variables is a much more difficult task than finding significant differences in these variables at the group level [<xref ref-type="bibr" rid="ref18">18</xref>]. However, finding statistically significant differences in variables could unveil the mechanisms associated with an outcome (eg, disease) and could be a complementary preliminary step for further individualized prediction and risk factor identification. Therefore, we recommend using the term Individualized Prediction for predictions at the individual level in general, as opposed to predictions of outcomes at the group level or population level.</p><p>Key terms related to &#x201C;Individualized Prediction&#x201D; include &#x201C;Personalized Medicine&#x201D; and &#x201C;Precision Medicine,&#x201D; which refer to treatment optimizations that consider all personal information. Personalized medicine and precision medicine usually need to engage multiple individualized prediction models for treatment outcome predictions of multiple or all candidate treatment options using multimodal data [<xref ref-type="bibr" rid="ref24">24</xref>] and thus should not be used casually if the model provides only 1 individual-level prediction for 1 condition or treatment. Currently, precision medicine is sometimes used to refer to the identification of individuals fitting to a highly specific treatment that can be further personalized (eg, gene or stem cell therapy); however, in general, this process usually involves previous attempts of or implicit comparisons with other treatment options.</p></sec><sec id="s2-2"><title>Recommendations for the Use of &#x201C;Prediction&#x201D;</title><p>The use of the term &#x201C;Prediction&#x201D; should consider the following recommendations. First, do not use the term &#x201C;prediction&#x201D; for an association study that does not predict any unknown outcomes (eg, correlation) or has no validation procedure to estimate the predictability of the estimated unknown outcome (eg, regression without proper validations). Second, if the model used in the study is trained to predict outcomes in the future of the input features, use the term &#x201C;Prospective Prediction.<italic>&#x201D;</italic> &#x201C;Prediction&#x201D; can be used if the model developed in the study is trained to predict unknown outcomes in the same time frame as the input features, with proper validation (<xref ref-type="fig" rid="figure2">Figure 2</xref>). Specify whether the validation is from a future collection (eg, &#x201C;prospective validation&#x201D;), a holdout of the same dataset (eg, &#x201C;internal validation&#x201D; and be specific about the validation methods), or an independent dataset (eg, &#x201C;external validation&#x201D;). Specify whether the &#x201C;future&#x201D; collection is a true prospective design, where the validation dataset is not yet collected, or is from retrospective data that were collected in the future of model training. External and prospective validation will greatly strengthen confidence in the generalizability and predictability of the model&#x2019;s performance. Third, the term &#x201C;Individualized Prediction&#x201D; can be used to differentiate predictions at the individual level from those focusing on outcomes at the group or population level. However, &#x201C;Personalized Medicine&#x201D; and &#x2019;Precision Medicine&#x201D; should not be used casually if the model provides only 1 individual-level prediction for a single condition or treatment.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Recommended decision process for the term &#x201C;prediction&#x201D; and relevant validations.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66100_fig02.png"/></fig></sec></sec><sec id="s3"><title>Testing, Validation, Cross-Validation, and External Validation</title><sec id="s3-1"><title>Overview</title><p>The ability of ML models to predict outcomes in new data must be demonstrated through a validation process. However, there is common confusion about the rationale for and implementation of testing and validation and the difference between the 2 terms. In ML, it is typical to test the model on a separate dataset. We usually train (or develop) the ML model on a dataset that we call the training set and then test it on another dataset that we call the testing set. Usually, the testing set is an independent data collection similar to the training set. In the case of 1 dataset, we can still try to mimic the training and testing procedure by separating the data into the training data and validation data, with the former used for training the model and the latter for internal testing of the trained model. Sometimes, we can perform the validation in a way that we randomly separate the data into certain portions and use 1 portion as the validation data each time until we exhaust the data, a common procedure that we refer to as cross-validation [<xref ref-type="bibr" rid="ref25">25</xref>]. Cross-validation is also used for model and variable selection (or, in more common terms, &#x201C;feature selection&#x201D;; see section &#x201C;Feature, Independent Variable, Predictor, Risk Factor, and Causal Factor&#x201D; about &#x201C;feature&#x201D;) [<xref ref-type="bibr" rid="ref26">26</xref>]. The most important aspect of validation and testing is that the model being trained should not encounter any of the validation data per se (comprising the validation) or any information specifically derived from validation data, including feature selection (eg, potential predicting variables), hyperparameter tuning, and model fitting or optimization. Testing and validation are sometimes used interchangeably. However, in the current literature of medicine, psychology, and social sciences, where prospective studies are still scarce while retrospective studies are dominant, it is usually more appropriate to report the specific validation procedures rather than use the term &#x201C;testing&#x201D; generally.</p><p>Note that the procedure of testing and validation is important to estimate how the model would perform on new, unseen data, but it is not and should not be a procedure exclusive to ML. For any scientific analysis, we want &#x201C;predictive results&#x201D; to be generalizable beyond our own study samples. The procedure of validation, including cross-validation, may not be perfect, especially when implemented improperly, as overestimation and bias still occur depending on how representative the testing and validation data are, but it is a crucial gateway for further real-world implementation of any model [<xref ref-type="bibr" rid="ref26">26</xref>].</p></sec><sec id="s3-2"><title>Recommendations for the Use of &#x201C;Validation&#x201D;</title><p>It is important to explicitly report whether models have undergone proper validation on independent datasets or cross-validation or holdout validation methods that strictly separate training and validation data. We recommend using specific terms about how validation is performed instead of using the term &#x201C;testing&#x201D; or &#x201C;validation&#x201D; in general. For example, if the model is developed on 1 dataset and validated on the same data using cross-validation, the term &#x201C;testing&#x201D; should be avoided, and the specific cross-validation should be clearly mentioned, for example, 5-fold cross-validation or leave-one-out cross-validation. If the model is developed on 1 dataset and validated on another independent dataset, usually collected from a separate cohort or another geographical region or facility, the term &#x201C;External Validation&#x201D; should be used. If the model is developed on 1 population dataset and validated in the future, where prospective prediction and longitudinal collection of the same cohort are usually involved, the term &#x201C;Prospective Validation&#x201D; should be used ( <xref ref-type="fig" rid="figure1">Figure 1</xref> and Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). Note that if the validation is performed on independent data collected in the future of training data, the validation is both prospective and external. In all the validation scenarios, the term &#x201C;testing&#x201D; is less relevant in medicine, psychology, and social sciences, especially for retrospective studies, compared to in computer science, and specific &#x201C;validation&#x201D; is always preferred.</p></sec></sec><sec id="s4"><title>Overfitting, Linear Regression, Regularization, and ML</title><sec id="s4-1"><title>Overview</title><p>Validating ML models on new data is essential to demonstrate their generalizability beyond the training data. Without validation, models risk overfitting [<xref ref-type="bibr" rid="ref10">10</xref>]&#x2014;a common problem where algorithms learn to fit both the meaningful patterns (&#x201C;signal&#x201D;) and random variations (&#x201C;noise&#x201D;) in the training data, limiting their applicability to new data. This issue is particularly relevant for linear regression models commonly used in medicine, psychology, and social sciences [<xref ref-type="bibr" rid="ref27">27</xref>], especially when the number of features is large while the number of observations is limited. Regularized linear regression models (eg, LASSO [Least Absolute Shrinkage and Selection Operator] and elastic net [<xref ref-type="bibr" rid="ref28">28</xref>]) address this challenge by controlling coefficient magnitudes to prevent overfitting, a principle that extends to other ML approaches [<xref ref-type="bibr" rid="ref29">29</xref>].</p><p>When we compare regularized linear regression algorithms with traditional linear regression in medical, psychological, and social applications, their primary advantage lies in their enhanced generalizability. These ML techniques use regularization to constrain coefficients and optimize them in ways that reduce overfitting [<xref ref-type="bibr" rid="ref30">30</xref>]. Unlike traditional linear regression, regularization parameters in the regularized linear regression can be automatically learned from training data (and thus considered as ML). While traditional approaches can implement similar constraints through techniques such as the Bayesian information criterion and Akaike&#x2019;s information criterion [<xref ref-type="bibr" rid="ref31">31</xref>], regularized methods typically offer superior generalization, although exceptions exist and should be verified through proper validation [<xref ref-type="bibr" rid="ref32">32</xref>]. The distinction between traditional statistics and ML continues to blur as classical regression models increasingly incorporate regularization techniques, yet meaningful differences persist [<xref ref-type="bibr" rid="ref33">33</xref>].</p><p>Beyond linear approaches, ML offers powerful techniques for capturing nonlinear and complex relationships between features. Methods such as support vector machines with nonlinear kernels [<xref ref-type="bibr" rid="ref34">34</xref>], random forests [<xref ref-type="bibr" rid="ref35">35</xref>], and neural networks [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref36">36</xref>] can model intricate patterns and interactions that linear models cannot accommodate. These algorithms automatically discover complex feature combinations and transformations, enabling them to represent relationships that would otherwise require extensive manual feature engineering when using traditional statistical approaches [<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref38">38</xref>].</p></sec><sec id="s4-2"><title>Recommendations for Reporting Prediction Performance and Model Choices to Prevent Overfitting</title><p>With proper validation, overfitting can be recognized during the validation stage. For example, the model may achieve a high accuracy of prediction on the training data but almost a random performance (eg, 50% in a binary classification task) on the validation data. Thus, it is important to report appropriate prediction performance metrics on both the training and especially validation data, including at least sensitivity, specificity, accuracy, precision, and the area under the receiver operating characteristic curve [<xref ref-type="bibr" rid="ref39">39</xref>]. When possible, external validation is always preferred.</p><p>Maintaining transparency about model choices during training and validation is crucial. Linear models are often preferred for their simplicity and explainability in clinical applications [<xref ref-type="bibr" rid="ref40">40</xref>,<xref ref-type="bibr" rid="ref41">41</xref>]. However, when more complex processes such as model selection, parameter tuning, and feature engineering are used, researchers should clearly document and justify these steps using flowcharts or concise textual descriptions [<xref ref-type="bibr" rid="ref42">42</xref>,<xref ref-type="bibr" rid="ref43">43</xref>]. This kind of rationale is increasingly neglected, partly due to unrealistic expectations that DL models can produce impressive results simply by feeding in complex data. Nonetheless, it is critical to justify every decision in model development&#x2014;even those involving seemingly opaque parameters&#x2014;because added complexity can compromise both reproducibility and explainability. Importantly, all selection processes must be performed exclusively within the training stage, using nested cross-validation approaches when necessary, to prevent overfitting and information leakage that could artificially inflate performance metrics [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref44">44</xref>].</p></sec></sec><sec id="s5"><title>Feature, Independent Variable, Predictor, Risk Factor, and Causal Factor</title><sec id="s5-1"><title>Overview</title><p>Terminology confusion persists regarding concepts, such as features, independent variables, predictors, risk factors, and causal factors. The term &#x201C;feature&#x201D; predominates in ML literature, while &#x201C;independent variable&#x201D; is more common in traditional statistics [<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref45">45</xref>]. Independent variables are designed to be independent of other variables and serve to explain or predict the dependent variables in statistical models [<xref ref-type="bibr" rid="ref46">46</xref>]. When these variables are established as influencing an outcome, they may be termed &#x201C;predictors.&#x201D;</p><p>ML experts typically use &#x201C;feature&#x201D; to refer to all input variables in a dataset (excluding the target variable being predicted), with the collection of all possible feature values constituting the &#x201C;feature space&#x201D; [<xref ref-type="bibr" rid="ref47">47</xref>]. In ML prediction models, features function as potential predictors for the target outcome variable, analogous to how independent variables relate to dependent variables in statistical modeling [<xref ref-type="bibr" rid="ref48">48</xref>], although the features are not necessarily independent of each other.</p></sec><sec id="s5-2"><title>Recommendations for the Use of &#x201C;Feature,&#x201D; &#x201C;Independent Variable,&#x201D; &#x201C;Predictor,&#x201D; &#x201C;Risk Factor,&#x201D; and &#x201C;Causal Factor&#x201D;</title><p>We recommend using &#x201C;Feature&#x201D; specifically for input variables in ML models, &#x201C;Independent Variables&#x201D; in statistical modeling contexts, and reserving &#x201C;Predictor&#x201D; for features or independent variables confirmed through rigorous validation to have predictive value for the outcome [<xref ref-type="bibr" rid="ref49">49</xref>]. Importantly, predictors should not be conflated with risk factors (or protective factors; referred to as risk factors in general) or causal factors [<xref ref-type="bibr" rid="ref3">3</xref>]. Risk factors typically emerge from association studies or require additional post hoc analysis to confirm their contribution to outcomes [<xref ref-type="bibr" rid="ref50">50</xref>]. Causal factors, by contrast, demand controlled experimental designs that can establish causative relationships [<xref ref-type="bibr" rid="ref51">51</xref>]. In general, features include all potential predictors, risk factors, and causal factors, and some predictors could be potential risk factors and causal factors and vice versa (<xref ref-type="fig" rid="figure3">Figure 3</xref>). Risk factors or causal factors are not naturally predictors and, similarly, predictors are not always risk factors or causal factors, especially when interaction between features is present.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Relationships between features, predictors, risk factors, and causal factors.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66100_fig03.png"/></fig></sec></sec><sec id="s6"><title>AI, ML, DL, Large Language Models, and Generative AI</title><sec id="s6-1"><title>Overview</title><p>The terms AI, ML, and DL are often used interchangeably in medicine, psychology, and social science literature, particularly since the emergence of generative AI, and could be potentially confusing for different audiences (<xref ref-type="fig" rid="figure4">Figure 4</xref>).</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Simplified relationships between artificial intelligence (AI), machine learning (ML), deep learning (DL), large language models (LLMs), and generative artificial intelligence. Modern LLMs and generative AI are still evolving and expanding the capabilities of both DL and ML, potentially surpassing traditional ML.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66100_fig04.png"/></fig><p>AI encompasses any man-made algorithm or agent (ie, computation) exhibiting aspects of intelligence, such as perceiving, reasoning, and acting [<xref ref-type="bibr" rid="ref52">52</xref>]. As a field, AI is broad and multifaceted. When referring to an AI agent, most experts expect it to mimic human cognition and behaviors, solve complex problems, and learn from experience [<xref ref-type="bibr" rid="ref53">53</xref>]. However, AI includes many subfields beyond just learning systems, such as logic systems, symbolic reasoning, or philosophical components.</p><p>ML represents a specific subset of AI focusing on algorithms that learn patterns from data to make predictions or decisions. For example, ML has successfully identified or predicted attention deficit, depression, bipolar disorder, suicidality, and substance use disorders and overdose using behavioral, cognitive, and health record data [<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref55">55</xref>]. With the proliferation of big data, ML has become the most visible component of AI, but it does not encompass all aspects of AI.</p><p>The defining characteristic of ML is that the model structure and parameters are learned automatically from data, usually high-dimensional, rather than being explicitly programmed. These models can range from relatively simple linear equations and support vector machines to complex deep neural networks.</p><p>DL represents a specialized form of ML using multilayer artificial neural networks inspired by the structure and function of the human brain [<xref ref-type="bibr" rid="ref36">36</xref>]. While this brain-inspired architecture may contribute to confusion between DL and broader AI, DL remains fundamentally a subset of ML focusing on making predictions based on data. The complexity of DL increases when researchers develop advanced architectures such as those used in generative AI applications or federated learning [<xref ref-type="bibr" rid="ref56">56</xref>], or when applying these models in reinforcement learning scenarios [<xref ref-type="bibr" rid="ref57">57</xref>]. Nevertheless, these advanced applications still fundamentally operate within the ML paradigm.</p><p>ML encompasses various pattern recognition tasks that identify meaningful data structures such as faces, objects, words, or sentiments [<xref ref-type="bibr" rid="ref37">37</xref>]. This aspect has gained particular prominence with the emergence of large language models (LLMs), which are primarily trained to predict the next term according to a given context. The definition of &#x201C;large&#x201D; may change depending on historical considerations, and previous language models could be preprogrammed and might not use DL paradigms [<xref ref-type="bibr" rid="ref58">58</xref>]. The current versions of LLMs are usually deep neural networks with billions or even trillions of parameters [<xref ref-type="bibr" rid="ref59">59</xref>]. While generative AI conceptually encompasses any system that creates new content (including rule-based and symbolic approaches), modern generative AI systems predominantly use DL techniques. Historical examples such as ELIZA or procedural content generation in games demonstrate non-ML generative approaches, but today&#x2019;s state-of-the-art generative AI relies fundamentally on DL architectures. Modern LLMs and generative AI are still evolving and expanding the capabilities of both DL and ML, potentially surpassing traditional ML, especially with the emerging agentic AI and world models [<xref ref-type="bibr" rid="ref60">60</xref>,<xref ref-type="bibr" rid="ref61">61</xref>].</p><p>The growing popularity of DL has sparked discussions about &#x201C;explainable AI&#x201D; [<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref63">63</xref>], reflecting concerns about the interpretability of complex models. While many ML algorithms provide reasonable transparency regarding influential variables, DL approaches often function as &#x201C;black boxes&#x201D; with limited interpretability. However, interpretability exists on a spectrum&#x2014;linear ML methods, LASSO regression, and support vector machines typically offer greater insight into the importance of specific variables for prediction than deep neural networks.</p><p>It is important to note that coefficients or importance scores from ML models should not be directly interpreted as statistical indicators (eg, odds ratios). Separate hypothesis-driven analyses dedicated to interpretability are usually necessary to complement ML predictions [<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref64">64</xref>,<xref ref-type="bibr" rid="ref65">65</xref>]. These hybrid approaches can be crucial for understanding AI models and their predictions in clinical and social contexts.</p></sec><sec id="s6-2"><title>Recommendations for the Use of AI, ML, DL, and LLMs</title><p>Although the recent advancement of LLMs and generative AI [<xref ref-type="bibr" rid="ref66">66</xref>,<xref ref-type="bibr" rid="ref67">67</xref>] has further blurred boundaries between them and AI, ML, and DL, we recommend using the most specific term possible when describing research applications. Studies using LLMs, generative AI, or other DL variants should be labeled accordingly rather than using the broader term &#x201C;ML.&#x201D; Similarly, studies using various ML algorithms to predict clinical outcomes should be described as &#x201C;ML&#x201D; rather than the more general term &#x201C;AI.&#x201D; Although AI is a buzzword, its general nature can be misleading without specific methodological details.</p><p>Due to the generative nature of LLMs and generative AI models that extends beyond traditional prediction tasks, it is important to specify their precise application, such as generating synthetic data (text, images, and videos), simulating human interactions (structured interviews, psychological assessments, diagnostics, cognitive-behavioral and group therapies, patient education, virtual patients, or health service providers), synthesizing knowledge from scientific literature, providing decision support with treatment alternatives, designing research protocols and experiments, improving accessibility through translation and simplification, and assisting with other specialized tasks. This specificity helps maintain terminological clarity and prevents confusion about the model&#x2019;s purpose, capabilities, and appropriate evaluation metrics.</p></sec></sec><sec id="s7" sec-type="conclusions"><title>Conclusions</title><p>In this paper, we have systematically addressed common terminological confusions surrounding AI applications in medicine, psychology, and social sciences. We have provided clarification on key terms and concepts that are frequently misused or misunderstood in the literature, with a particular focus on &#x201C;prediction&#x201D;&#x2014;a term that systematic reviews have shown is often incorrectly applied to association studies. We have established practical recommendations for when and how the term &#x201C;prediction&#x201D; should be used, emphasizing the importance of specifying the nature of prediction in a study, especially the use of &#x201C;prospective prediction&#x201D; for future prediction. We have also elucidated validation procedures essential for ensuring model generalizability and predictability and provided recommendations for using prospective, internal, and external validation.</p><p>Our discussion of overfitting and regularization has highlighted the advantages of ML approaches over traditional regression methods for developing generalizable models. By clarifying the relationships between features, independent variables, predictors, risk factors, and causal factors, we have provided researchers with a recommendation for more precise communication about model inputs and their interpretations.</p><p>Finally, we have delineated the hierarchical relationships between AI, ML, DL, LLMs, and generative AI, establishing a clear approach for discussing these increasingly prevalent technologies.</p><p>Maintaining terminological precision has become increasingly important and challenging as AI technologies continue to evolve rapidly. We hope that the recommendations provided in this tutorial will facilitate clearer communication among researchers and clinicians across different disciplines and between the research community and the public, advancing the rigorous and effective application of AI and ML in medicine, psychology, and social sciences, leading to improved research quality and, ultimately, better health and social outcomes (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>).</p></sec></body><back><ack><p>This research was undertaken, in part, thanks to funding from the Canada Research Chairs program, Alberta Innovates, the Institute for Advancements in Mental Health, Mental Health Foundation, Mental Health Research Canada, MITACS Accelerate program, Simon &#x0026; Martina Sochatsky Fund for Mental Health, Howard Berger Memorial Schizophrenia Research Fund, the Abraham &#x0026; Freda Berger Memorial Endowment Fund, the Alberta Synergies in Alzheimer&#x2019;s and Related Disorders (SynAD) program, University Hospital Foundation and University of Alberta, and Royal Society Leverhulme Trust Senior Fellowship. The funding sources had no impact on the design and conduct of the study; preparation, review, or approval of the manuscript; and decision to submit the manuscript for publication.</p></ack><fn-group><fn fn-type="con"><p>BC and JS developed the idea and the first draft of the paper. RG, AG, and JS provided critical feedback and revised the paper. BC critically revised the paper from its original version.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb3">LASSO</term><def><p>Least Absolute Shrinkage and Selection Operator</p></def></def-item><def-item><term id="abb4">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb5">ML</term><def><p>machine learning</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Cao</surname><given-names>B</given-names> </name><name name-style="western"><surname>Reilly</surname><given-names>J</given-names> </name></person-group><source>Major Challenges and Limitations of Big Data Analytics</source><year>2019</year><publisher-name>Springer International Publishing</publisher-name><pub-id pub-id-type="doi">10.1007/978-3-030-03553-2_2</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Andaur Navarro</surname><given-names>CL</given-names> </name><name name-style="western"><surname>Damen</surname><given-names>JAA</given-names> </name><name name-style="western"><surname>Takada</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Risk of bias in studies on prediction models developed using supervised machine learning techniques: systematic review</article-title><source>BMJ</source><year>2021</year><month>10</month><day>20</day><volume>375</volume><fpage>n2281</fpage><pub-id pub-id-type="doi">10.1136/bmj.n2281</pub-id><pub-id pub-id-type="medline">34670780</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Varga</surname><given-names>TV</given-names> </name><name name-style="western"><surname>Niss</surname><given-names>K</given-names> </name><name name-style="western"><surname>Estampador</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Collin</surname><given-names>CB</given-names> </name><name name-style="western"><surname>Moseley</surname><given-names>PL</given-names> </name></person-group><article-title>Association is not prediction: a landscape of confused reporting in diabetes&#x2014;a systematic review</article-title><source>Diabetes Res Clin Pract</source><year>2020</year><month>12</month><volume>170</volume><fpage>108497</fpage><pub-id pub-id-type="doi">10.1016/j.diabres.2020.108497</pub-id><pub-id pub-id-type="medline">33068662</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McCall</surname><given-names>A</given-names> </name><name name-style="western"><surname>Fanchini</surname><given-names>M</given-names> </name><name name-style="western"><surname>Coutts</surname><given-names>AJ</given-names> </name></person-group><article-title>Prediction: the modern-day sport-science and sports-medicine &#x201C;Quest for the Holy Grail&#x201D;</article-title><source>Int J Sports Physiol Perform</source><year>2017</year><month>05</month><volume>12</volume><issue>5</issue><fpage>704</fpage><lpage>706</lpage><pub-id pub-id-type="doi">10.1123/ijspp.2017-0137</pub-id><pub-id pub-id-type="medline">28488907</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yarkoni</surname><given-names>T</given-names> </name><name name-style="western"><surname>Westfall</surname><given-names>J</given-names> </name></person-group><article-title>Choosing prediction over explanation in psychology: lessons from machine learning</article-title><source>Perspect Psychol Sci</source><year>2017</year><month>11</month><volume>12</volume><issue>6</issue><fpage>1100</fpage><lpage>1122</lpage><pub-id pub-id-type="doi">10.1177/1745691617693393</pub-id><pub-id pub-id-type="medline">28841086</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ramspek</surname><given-names>CL</given-names> </name><name name-style="western"><surname>Jager</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>Dekker</surname><given-names>FW</given-names> </name><name name-style="western"><surname>Zoccali</surname><given-names>C</given-names> </name><name name-style="western"><surname>van Diepen</surname><given-names>M</given-names> </name></person-group><article-title>External validation of prognostic models: what, why, how, when and where?</article-title><source>Clin Kidney J</source><year>2021</year><month>01</month><volume>14</volume><issue>1</issue><fpage>49</fpage><lpage>58</lpage><pub-id pub-id-type="doi">10.1093/ckj/sfaa188</pub-id><pub-id pub-id-type="medline">33564405</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nagendran</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lovejoy</surname><given-names>CA</given-names> </name><etal/></person-group><article-title>Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies</article-title><source>BMJ</source><year>2020</year><month>03</month><day>25</day><volume>368</volume><fpage>m689</fpage><pub-id pub-id-type="doi">10.1136/bmj.m689</pub-id><pub-id pub-id-type="medline">32213531</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abdulazeem</surname><given-names>H</given-names> </name><name name-style="western"><surname>Whitelaw</surname><given-names>S</given-names> </name><name name-style="western"><surname>Schauberger</surname><given-names>G</given-names> </name><name name-style="western"><surname>Klug</surname><given-names>SJ</given-names> </name></person-group><article-title>A systematic review of clinical health conditions predicted by machine learning diagnostic and prognostic models trained or validated using real-world primary health care data</article-title><source>PLoS One</source><year>2023</year><volume>18</volume><issue>9</issue><fpage>e0274276</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0274276</pub-id><pub-id pub-id-type="medline">37682909</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Harrell</surname><given-names>FE</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>KL</given-names> </name><name name-style="western"><surname>Matchar</surname><given-names>DB</given-names> </name><name name-style="western"><surname>Reichert</surname><given-names>TA</given-names> </name></person-group><article-title>Regression models for prognostic prediction: advantages, problems, and suggested solutions</article-title><source>Cancer Treat Rep</source><year>1985</year><month>10</month><volume>69</volume><issue>10</issue><fpage>1071</fpage><lpage>1077</lpage><pub-id pub-id-type="medline">4042087</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Babyak</surname><given-names>MA</given-names> </name></person-group><article-title>What you see may not be what you get: a brief, nontechnical introduction to overfitting in regression-type models</article-title><source>Psychosom Med</source><year>2004</year><volume>66</volume><issue>3</issue><fpage>411</fpage><lpage>421</lpage><pub-id pub-id-type="doi">10.1097/01.psy.0000127692.23278.a9</pub-id><pub-id pub-id-type="medline">15184705</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Obermeyer</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Emanuel</surname><given-names>EJ</given-names> </name></person-group><article-title>Predicting the future&#x2014;big data, machine learning, and clinical medicine</article-title><source>N Engl J Med</source><year>2016</year><month>09</month><day>29</day><volume>375</volume><issue>13</issue><fpage>1216</fpage><lpage>1219</lpage><pub-id pub-id-type="doi">10.1056/NEJMp1606181</pub-id><pub-id pub-id-type="medline">27682033</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Michael</surname><given-names>W</given-names> </name><name name-style="western"><surname>Berry</surname><given-names>AM</given-names> </name></person-group><article-title>Biomedical signal processing and artificial intelligence in healthcare</article-title><source>Supervised and Unsupervised Learning for Data Science</source><year>2020</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1007/978-3-030-22475-2</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>F</given-names> </name><name name-style="western"><surname>Sui</surname><given-names>J</given-names> </name></person-group><article-title>Decoding individual differences in self-prioritization from the resting-state functional connectome</article-title><source>Neuroimage</source><year>2023</year><month>08</month><volume>276</volume><fpage>120205</fpage><pub-id pub-id-type="doi">10.1016/j.neuroimage.2023.120205</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Pierce</surname><given-names>DV</given-names> </name><name name-style="western"><surname>Metes</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Population-level individualized prospective prediction of opioid overdose using machine learning</article-title><source>Mol Psychiatry</source><year>2025</year><month>04</month><day>14</day><volume>5&#x2013;10</volume><pub-id pub-id-type="doi">10.1038/s41380-025-02992-4</pub-id><pub-id pub-id-type="medline">40229401</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Song</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Qian</surname><given-names>L</given-names> </name><name name-style="western"><surname>Sui</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Prediction of depression onset risk among middle-aged and elderly adults using machine learning and Canadian Longitudinal Study on Aging cohort</article-title><source>J Affect Disord</source><year>2023</year><month>10</month><day>15</day><volume>339</volume><fpage>52</fpage><lpage>57</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2023.06.031</pub-id><pub-id pub-id-type="medline">37380110</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Song</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Sui</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Prospective prediction of anxiety onset in the Canadian longitudinal study on aging (CLSA): a machine learning study</article-title><source>J Affect Disord</source><year>2024</year><month>07</month><volume>357</volume><fpage>148</fpage><lpage>155</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2024.04.098</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mavrogiorgou</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kiourtis</surname><given-names>A</given-names> </name><name name-style="western"><surname>Manias</surname><given-names>G</given-names> </name><name name-style="western"><surname>Symvoulidis</surname><given-names>C</given-names> </name><name name-style="western"><surname>Kyriazis</surname><given-names>D</given-names> </name></person-group><article-title>Batch and streaming data ingestion towards creating holistic health records</article-title><source>Emerg Sci J</source><year>2023</year><volume>7</volume><issue>2</issue><fpage>339</fpage><lpage>353</lpage><pub-id pub-id-type="doi">10.28991/ESJ-2023-07-02-03</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lo</surname><given-names>A</given-names> </name><name name-style="western"><surname>Chernoff</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>T</given-names> </name><name name-style="western"><surname>Lo</surname><given-names>SH</given-names> </name></person-group><article-title>Why significant variables aren&#x2019;t automatically good predictors</article-title><source>Proc Natl Acad Sci U S A</source><year>2015</year><month>11</month><day>10</day><volume>112</volume><issue>45</issue><fpage>13892</fpage><lpage>13897</lpage><pub-id pub-id-type="doi">10.1073/pnas.1518285112</pub-id><pub-id pub-id-type="medline">26504198</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Janssen</surname><given-names>RJ</given-names> </name><name name-style="western"><surname>Mour&#x00E3;o-Miranda</surname><given-names>J</given-names> </name><name name-style="western"><surname>Schnack</surname><given-names>HG</given-names> </name></person-group><article-title>Making individual prognoses in psychiatry using neuroimaging and machine learning</article-title><source>Biol Psychiatry Cogn Neurosci Neuroimaging</source><year>2018</year><month>09</month><volume>3</volume><issue>9</issue><fpage>798</fpage><lpage>808</lpage><pub-id pub-id-type="doi">10.1016/j.bpsc.2018.04.004</pub-id><pub-id pub-id-type="medline">29789268</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cao</surname><given-names>B</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Predicting individual responses to the electroconvulsive therapy with hippocampal subfield volumes in major depression disorder</article-title><source>Sci Rep</source><year>2018</year><month>04</month><day>3</day><volume>8</volume><issue>1</issue><fpage>5434</fpage><pub-id pub-id-type="doi">10.1038/s41598-018-23685-9</pub-id><pub-id pub-id-type="medline">29615675</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Librenza-Garcia</surname><given-names>D</given-names> </name><name name-style="western"><surname>Kotzian</surname><given-names>BJ</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><etal/></person-group><article-title>The impact of machine learning techniques in the study of bipolar disorder: a systematic review</article-title><source>Neurosci Biobehav Rev</source><year>2017</year><month>09</month><volume>80</volume><fpage>538</fpage><lpage>554</lpage><pub-id pub-id-type="doi">10.1016/j.neubiorev.2017.07.004</pub-id><pub-id pub-id-type="medline">28728937</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sawalha</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Individualized identification of first-episode bipolar disorder using machine learning and cognitive tests</article-title><source>J Affect Disord</source><year>2021</year><month>03</month><day>1</day><volume>282</volume><fpage>662</fpage><lpage>668</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2020.12.046</pub-id><pub-id pub-id-type="medline">33445089</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Song</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>NA</given-names> </name><etal/></person-group><article-title>Depression screening using a non-verbal self-association task: a machine-learning based pilot study</article-title><source>J Affect Disord</source><year>2022</year><month>08</month><day>1</day><volume>310</volume><fpage>87</fpage><lpage>95</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2022.04.122</pub-id><pub-id pub-id-type="medline">35472473</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Marks Sultan</surname><given-names>G</given-names> </name><name name-style="western"><surname>Dosch</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sprumont</surname><given-names>D</given-names> </name></person-group><article-title>Sant&#x00E9; personnalis&#x00E9;e: d&#x00E9;finition, caract&#x00E9;ristiques et perspectives pour le futur</article-title><source>Rev Med Suisse</source><year>2021</year><volume>17</volume><issue>732</issue><fpage>654</fpage><lpage>657</lpage><pub-id pub-id-type="doi">10.53738/REVMED.2021.17.732.0654</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Berrar</surname><given-names>D</given-names> </name></person-group><article-title>Cross-validation</article-title><source>Encyclopedia of Bioinformatics and Computational Biology: ABC of Bioinformatics</source><year>2018</year><access-date>2025-07-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.sciencedirect.com/referencework/9780128114322/encyclopedia-of-bioinformatics-and-computational-biology#book-info">https://www.sciencedirect.com/referencework/9780128114322/encyclopedia-of-bioinformatics-and-computational-biology#book-info</ext-link></comment><pub-id pub-id-type="doi">10.1016/B978-0-12-809633-8.20349-X</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Varma</surname><given-names>S</given-names> </name><name name-style="western"><surname>Simon</surname><given-names>R</given-names> </name></person-group><article-title>Bias in error estimation when using cross-validation for model selection</article-title><source>BMC Bioinformatics</source><year>2006</year><month>02</month><day>23</day><volume>7</volume><fpage>91</fpage><pub-id pub-id-type="doi">10.1186/1471-2105-7-91</pub-id><pub-id pub-id-type="medline">16504092</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Vach</surname><given-names>W</given-names> </name></person-group><source>Regression Models as a Tool in Medical Research</source><year>2012</year><publisher-name>Routledge</publisher-name><pub-id pub-id-type="other">9781032477510</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Hastie</surname><given-names>T</given-names> </name></person-group><article-title>Regularization and variable selection via the elastic net</article-title><source>J R Stat Soc Ser B</source><year>2005</year><month>04</month><day>1</day><volume>67</volume><issue>2</issue><fpage>301</fpage><lpage>320</lpage><pub-id pub-id-type="doi">10.1111/j.1467-9868.2005.00503.x</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schmidhuber</surname><given-names>J</given-names> </name></person-group><article-title>Deep learning in neural networks: an overview</article-title><source>Neural Netw</source><year>2015</year><month>01</month><volume>61</volume><fpage>85</fpage><lpage>117</lpage><pub-id pub-id-type="doi">10.1016/j.neunet.2014.09.003</pub-id><pub-id pub-id-type="medline">25462637</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Friedman</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hastie</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tibshirani</surname><given-names>R</given-names> </name></person-group><article-title>Regularization paths for generalized linear models via coordinate descent</article-title><source>J Stat Softw</source><year>2010</year><volume>33</volume><issue>1</issue><fpage>1</fpage><lpage>22</lpage><pub-id pub-id-type="medline">20808728</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kuha</surname><given-names>J</given-names> </name></person-group><article-title>AIC and BIC: comparisons of assumptions and performance</article-title><source>Sociol Methods Res</source><year>2004</year><volume>33</volume><issue>2</issue><fpage>188</fpage><lpage>229</lpage><pub-id pub-id-type="doi">10.1177/0049124103262065</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Talarico</surname><given-names>F</given-names> </name><name name-style="western"><surname>Metes</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Early identification of children with attention-deficit/hyperactivity disorder (ADHD)</article-title><source>PLOS Digit Health</source><year>2024</year><month>11</month><volume>3</volume><issue>11</issue><fpage>e0000620</fpage><pub-id pub-id-type="doi">10.1371/journal.pdig.0000620</pub-id><pub-id pub-id-type="medline">39509384</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sidey-Gibbons</surname><given-names>JAM</given-names> </name><name name-style="western"><surname>Sidey-Gibbons</surname><given-names>CJ</given-names> </name></person-group><article-title>Machine learning in medicine: a practical introduction</article-title><source>BMC Med Res Methodol</source><year>2019</year><month>03</month><day>19</day><volume>19</volume><issue>1</issue><fpage>64</fpage><pub-id pub-id-type="doi">10.1186/s12874-019-0681-4</pub-id><pub-id pub-id-type="medline">30890124</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ben-Hur</surname><given-names>A</given-names> </name><name name-style="western"><surname>Weston</surname><given-names>J</given-names> </name></person-group><article-title>A user&#x2019;s guide to support vector machines</article-title><source>Methods Mol Biol</source><year>2010</year><volume>609</volume><fpage>223</fpage><lpage>239</lpage><pub-id pub-id-type="doi">10.1007/978-1-60327-241-4_13</pub-id><pub-id pub-id-type="medline">20221922</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Breiman</surname><given-names>L</given-names> </name></person-group><article-title>Random forests</article-title><source>Mach Learn</source><year>2001</year><volume>45</volume><issue>1</issue><fpage>5</fpage><lpage>32</lpage><pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>LeCun</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Bengio</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hinton</surname><given-names>G</given-names> </name></person-group><article-title>Deep learning</article-title><source>Nature New Biol</source><year>2015</year><month>05</month><day>28</day><volume>521</volume><issue>7553</issue><fpage>436</fpage><lpage>444</lpage><pub-id pub-id-type="doi">10.1038/nature14539</pub-id><pub-id pub-id-type="medline">26017442</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Bishop</surname><given-names>CM</given-names> </name></person-group><source>Pattern Recognition And Machine Learning</source><year>2006</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1117/1.2819119</pub-id><pub-id pub-id-type="other">978-1-4939-3843-8</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Friedman</surname><given-names>JH</given-names> </name></person-group><article-title>Greedy function approximation: a gradient boosting machine</article-title><source>Ann Stat</source><year>2001</year><volume>29</volume><issue>5</issue><fpage>1189</fpage><lpage>1232</lpage><pub-id pub-id-type="doi">10.1214/aos/1013203451</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sokolova</surname><given-names>M</given-names> </name><name name-style="western"><surname>Lapalme</surname><given-names>G</given-names> </name></person-group><article-title>A systematic analysis of performance measures for classification tasks</article-title><source>Inf Process Manag</source><year>2009</year><month>07</month><volume>45</volume><issue>4</issue><fpage>427</fpage><lpage>437</lpage><pub-id pub-id-type="doi">10.1016/j.ipm.2009.03.002</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Christodoulou</surname><given-names>E</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>J</given-names> </name><name name-style="western"><surname>Collins</surname><given-names>GS</given-names> </name><name name-style="western"><surname>Steyerberg</surname><given-names>EW</given-names> </name><name name-style="western"><surname>Verbakel</surname><given-names>JY</given-names> </name><name name-style="western"><surname>Van Calster</surname><given-names>B</given-names> </name></person-group><article-title>A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models</article-title><source>J Clin Epidemiol</source><year>2019</year><month>06</month><volume>110</volume><fpage>12</fpage><lpage>22</lpage><pub-id pub-id-type="doi">10.1016/j.jclinepi.2019.02.004</pub-id><pub-id pub-id-type="medline">30763612</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rudin</surname><given-names>C</given-names> </name></person-group><article-title>Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead</article-title><source>Nat Mach Intell</source><year>2019</year><month>05</month><volume>1</volume><issue>5</issue><fpage>206</fpage><lpage>215</lpage><pub-id pub-id-type="doi">10.1038/s42256-019-0048-x</pub-id><pub-id pub-id-type="medline">35603010</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Collins</surname><given-names>GS</given-names> </name><name name-style="western"><surname>Reitsma</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Altman</surname><given-names>DG</given-names> </name><name name-style="western"><surname>Moons</surname><given-names>KGM</given-names> </name><collab>members of the TRIPOD group</collab></person-group><article-title>Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement</article-title><source>Eur Urol</source><year>2015</year><month>06</month><volume>67</volume><issue>6</issue><fpage>1142</fpage><lpage>1151</lpage><pub-id pub-id-type="doi">10.1016/j.eururo.2014.11.025</pub-id><pub-id pub-id-type="medline">25572824</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Luo</surname><given-names>W</given-names> </name><name name-style="western"><surname>Phung</surname><given-names>D</given-names> </name><name name-style="western"><surname>Tran</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Guidelines for developing and reporting machine learning predictive models in biomedical research: a multidisciplinary view</article-title><source>J Med Internet Res</source><year>2016</year><month>12</month><day>16</day><volume>18</volume><issue>12</issue><fpage>e323</fpage><pub-id pub-id-type="doi">10.2196/jmir.5870</pub-id><pub-id pub-id-type="medline">27986644</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cawley</surname><given-names>GC</given-names> </name><name name-style="western"><surname>Talbot</surname><given-names>NLC</given-names> </name></person-group><article-title>On over-fitting in model selection and subsequent selection bias in performance evaluation</article-title><source>J Mach Learn Res</source><year>2010</year><volume>11</volume><issue>70</issue><fpage>2079</fpage><lpage>2107</lpage></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hastie</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tibshirani</surname><given-names>R</given-names> </name><name name-style="western"><surname>Friedman</surname><given-names>J</given-names> </name></person-group><source>The Elements of Statistical Learning</source><year>2009</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1007/978-0-387-84858-7</pub-id><pub-id pub-id-type="other">978-0-387-84858-7</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Pearl</surname><given-names>J</given-names> </name></person-group><source>Causality: Models, Reasoning, and Inference</source><year>2011</year><edition>2</edition><publisher-name>Cambridge University Press</publisher-name><pub-id pub-id-type="doi">10.1017/CBO9780511803161</pub-id><pub-id pub-id-type="other">978-0521895606</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Kuhn</surname><given-names>M</given-names> </name><name name-style="western"><surname>Johnson</surname><given-names>K</given-names> </name></person-group><source>Applied Predictive Modeling</source><year>2013</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1007/978-1-4614-6849-3_2</pub-id><pub-id pub-id-type="other">978-1-4614-6849-3</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hastie</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tibshirani</surname><given-names>R</given-names> </name><name name-style="western"><surname>James</surname><given-names>G</given-names> </name><name name-style="western"><surname>Witten</surname><given-names>D</given-names> </name></person-group><source>An Introduction to Statistical Learning</source><year>2021</year><edition>2</edition><publisher-name>Springer</publisher-name><pub-id pub-id-type="other">978-1-0716-1418-1</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Breiman</surname><given-names>L</given-names> </name></person-group><article-title>Statistical modeling: the two cultures</article-title><source>Stat Sci</source><year>2001</year><volume>16</volume><issue>3</issue><fpage>199</fpage><lpage>231</lpage><pub-id pub-id-type="doi">10.1214/ss/1009213726</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shmueli</surname><given-names>G</given-names> </name></person-group><article-title>To explain or to predict?</article-title><source>Stat Sci</source><year>2010</year><volume>25</volume><issue>3</issue><fpage>289</fpage><lpage>310</lpage><pub-id pub-id-type="doi">10.1214/10-STS330</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>HILL</surname><given-names>AB</given-names> </name></person-group><article-title>The environment and disease: association or causation?</article-title><source>Proc R Soc Med</source><year>1965</year><month>05</month><volume>58</volume><issue>5</issue><fpage>295</fpage><lpage>300</lpage><pub-id pub-id-type="doi">10.1177/003591576505800503</pub-id><pub-id pub-id-type="medline">14283879</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Winston</surname><given-names>PH</given-names> </name></person-group><source>Artificial Intelligence</source><year>1992</year><publisher-name>Addison-Wesley Longman Publishing Co, Inc</publisher-name><pub-id pub-id-type="other">978-0201533774</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Ferber</surname><given-names>J</given-names> </name><name name-style="western"><surname>Weiss</surname><given-names>G</given-names> </name></person-group><source>Multi-Agent Systems: An Introduction to Distributed Artificial Intelligence</source><year>1999</year><publisher-name>Addison Wesley</publisher-name><pub-id pub-id-type="other">978-0-201-36048-6</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Kiyang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Hayward</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Individualized prospective prediction of opioid use disorder</article-title><source>Can J Psychiatry</source><year>2023</year><month>01</month><volume>68</volume><issue>1</issue><fpage>54</fpage><lpage>63</lpage><pub-id pub-id-type="doi">10.1177/07067437221114094</pub-id><pub-id pub-id-type="medline">35892186</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>S</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Identification of suicidality in adolescent major depressive disorder patients using sMRI: a machine learning approach</article-title><source>J Affect Disord</source><year>2021</year><month>02</month><day>1</day><volume>280</volume><issue>Pt A</issue><fpage>72</fpage><lpage>76</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2020.10.077</pub-id><pub-id pub-id-type="medline">33202340</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rieke</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hancox</surname><given-names>J</given-names> </name><name name-style="western"><surname>Li</surname><given-names>W</given-names> </name><etal/></person-group><article-title>The future of digital health with federated learning</article-title><source>NPJ Digit Med</source><year>2020</year><volume>3</volume><fpage>119</fpage><pub-id pub-id-type="doi">10.1038/s41746-020-00323-1</pub-id><pub-id pub-id-type="medline">33015372</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mahmud</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kaiser</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Hussain</surname><given-names>A</given-names> </name><name name-style="western"><surname>Vassanelli</surname><given-names>S</given-names> </name></person-group><article-title>Applications of deep learning and reinforcement learning to biological data</article-title><source>IEEE Trans Neural Netw Learning Syst</source><year>2018</year><volume>29</volume><issue>6</issue><fpage>2063</fpage><lpage>2079</lpage><pub-id pub-id-type="doi">10.1109/TNNLS.2018.2790388</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Weizenbaum</surname><given-names>J</given-names> </name></person-group><article-title>ELIZA&#x2014;a computer program for the study of natural language communication between man and machine</article-title><source>Commun ACM</source><year>1966</year><month>01</month><volume>9</volume><issue>1</issue><fpage>36</fpage><lpage>45</lpage><pub-id pub-id-type="doi">10.1145/365153.365168</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thirunavukarasu</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSJ</given-names> </name><name name-style="western"><surname>Elangovan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gutierrez</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>TF</given-names> </name><name name-style="western"><surname>Ting</surname><given-names>DSW</given-names> </name></person-group><article-title>Large language models in medicine</article-title><source>Nat Med</source><year>2023</year><month>08</month><volume>29</volume><issue>8</issue><fpage>1930</fpage><lpage>1940</lpage><pub-id pub-id-type="doi">10.1038/s41591-023-02448-8</pub-id><pub-id pub-id-type="medline">37460753</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Acharya</surname><given-names>DB</given-names> </name><name name-style="western"><surname>Kuppan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Divya</surname><given-names>B</given-names> </name></person-group><article-title>Agentic AI: autonomous intelligence for complex goals&#x2014;a comprehensive survey</article-title><source>IEEE Access</source><year>2025</year><volume>13</volume><fpage>18912</fpage><lpage>18936</lpage><pub-id pub-id-type="doi">10.1109/ACCESS.2025.3532853</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Ha</surname><given-names>D</given-names> </name><name name-style="western"><surname>Schmidhuber</surname><given-names>J</given-names> </name></person-group><article-title>World models</article-title><source>arXiv:1803.10122v4</source><comment>Preprint posted online on 2018</comment><pub-id pub-id-type="doi">10.48550/arXiv.1803.10122</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Holzinger</surname><given-names>A</given-names> </name></person-group><article-title>Explainable AI and multi-modal causability in medicine</article-title><source>I Com (Berl)</source><year>2021</year><month>01</month><day>26</day><volume>19</volume><issue>3</issue><fpage>171</fpage><lpage>179</lpage><pub-id pub-id-type="doi">10.1515/icom-2020-0024</pub-id><pub-id pub-id-type="medline">37014363</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Goebel</surname><given-names>R</given-names> </name><name name-style="western"><surname>Chander</surname><given-names>A</given-names> </name><name name-style="western"><surname>Holzinger</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Explainable AI: the new 42</article-title><conf-name>International Cross-Domain Conference for Machine Learning and Knowledge Extraction</conf-name><conf-date>Aug 27-30, 2018</conf-date><conf-loc>Hamburg, Germany</conf-loc><pub-id pub-id-type="doi">10.1007/978-3-319-99740-7_21</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cao</surname><given-names>B</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Selvitella</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Differential power of placebo across major psychiatric disorders: a preliminary meta-analysis and machine learning study</article-title><source>Sci Rep</source><year>2021</year><month>10</month><day>29</day><volume>11</volume><issue>1</issue><fpage>21301</fpage><pub-id pub-id-type="doi">10.1038/s41598-021-99534-z</pub-id><pub-id pub-id-type="medline">34716400</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cao</surname><given-names>B</given-names> </name><name name-style="western"><surname>Cho</surname><given-names>RY</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Treatment response prediction and individualized identification of first-episode drug-na&#x00EF;ve schizophrenia using brain functional connectivity</article-title><source>Mol Psychiatry</source><year>2020</year><month>04</month><volume>25</volume><issue>4</issue><fpage>906</fpage><lpage>913</lpage><pub-id pub-id-type="doi">10.1038/s41380-018-0106-5</pub-id><pub-id pub-id-type="medline">29921920</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mesk&#x00F3;</surname><given-names>B</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>The imperative for regulatory oversight of large language models (or generative AI) in healthcare</article-title><source>NPJ Digit Med</source><year>2023</year><month>07</month><day>6</day><volume>6</volume><issue>1</issue><fpage>120</fpage><pub-id pub-id-type="doi">10.1038/s41746-023-00873-0</pub-id><pub-id pub-id-type="medline">37414860</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Rueda</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Understanding LLM scientific reasoning through promptings and model&#x2019;s explanation on the answers</article-title><source>arXiv</source><comment>Preprint posted online on 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2505.01482</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Prospective internal or external validation.</p><media xlink:href="jmir_v27i1e66100_app1.docx" xlink:title="DOCX File, 43 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Glossary of terms.</p><media xlink:href="jmir_v27i1e66100_app2.docx" xlink:title="DOCX File, 18 KB"/></supplementary-material></app-group></back></article>