Review
Abstract
Background: People with stroke face a high mortality risk, and an accurate prediction model is essential to the guidance of clinical decision-making in this population. Recently, with growing attention paid to machine learning (ML) in stroke care, some researchers have investigated the effectiveness of ML in predicting the mortality risk in stroke. However, systematic evidence is still lacking for its effectiveness.
Objective: This systematic review aims to evaluate the value of ML in predicting the stroke mortality risk. The findings are expected to offer an evidence-based basis for developing and assessing clinical risk prediction tools.
Methods: A search was made in Cochrane Library, PubMed, Embase, and Web of Science up to June 23, 2025, and studies that reported a complete performance of ML in predicting stroke mortality were included. Studies with only risk factors analyzed were excluded. The risk of bias of the included studies was assessed using PROBAST (Prediction model Risk of Bias Assessment Tool). Pooled risk ratios with 95% CIs and prediction intervals (PIs) were derived using the Hartung-Knapp-Sidik-Jonkman method under a random-effects model. Subgroup analyses were also conducted by model type, stroke type, patient source, and treatment background. Moreover, a metaregression was conducted on the C-index for out-of-hospital mortality at different time points to explore the influence of time factors on the model’s predictive performance.
Results: Sixty-eight studies were included (23 predicting in-hospital mortality and 45 predicting out-of-hospital mortality), describing the development of 75 prediction models and 43 external validations. The follow-up period was 1 month to 15 years. For predicting in-hospital mortality, the external validation set had a pooled C-index of 0.727 (95% CI 0.677-0.781, 95% PI 0.521-1.000), with sensitivity and specificity of 0.64 (95% CI 0.57-0.70) and 0.74 (95% CI 0.70-0.77), respectively. For predicting out-of-hospital mortality, the pooled C-index was 0.847 (95% CI 0.808-0.887, 95% PI 0.750-0.956) in the external validation set, with sensitivity and specificity of 0.71 (95% CI 0.55-0.82) and 0.76 (95% CI 0.74-0.78), respectively. Comparatively, the overall pooled C-indexes were 0.788 (95% CI 0.766-0.810, 95% PI 0.621-0.999) and 0.812 (95% CI 0.798-0.826, 95% PI 0.693-0.952), respectively. The metaregression revealed a gradual decline in the predictive performance of the overall model and logistic regression model alone, whereas a random forest model maintained sustained performance. Age, National Institutes of Health Stroke Scale score, and stroke-related complications were the most frequently used variables for modeling.
Conclusions: This is the first meta-analysis to demonstrate that ML-based prediction of stroke mortality is feasible. The performance of ML supports its role as an auxiliary tool for identifying high-risk populations, thereby optimizing clinical monitoring and resource allocation. However, due to substantial heterogeneity and a relatively high risk of bias in available studies, caution is warranted in real-world application. The effectiveness of ML may vary across settings, and external validation is recommended before broader implementation.
Trial Registration: PROSPERO CRD420251086321; https://www.crd.york.ac.uk/PROSPERO/view/CRD420251086321
doi:10.2196/83821
Keywords
Introduction
Rationale
Stroke, an acute focal neurological deficit with cerebrovascular causes, is primarily classified into ischemic stroke and hemorrhagic stroke (intracerebral hemorrhage and subarachnoid hemorrhage) according to its pathogenesis. Acute ischemic stroke (AIS) is the most frequent type, accounting for 60%-80% of the total []. As a leading cause of mortality and long-term disability globally, stroke greatly reduces patients’ quality of life and brings a heavy socioeconomic burden [,]. Notably, stroke mortality is projected to rise by 50% from 2020 to 2050 []. Therefore, stroke has become a serious threat to human health.
The functional prognosis of some patients with ischemic stroke (especially cases within the time window) can benefit from aggressive treatments (eg, intravenous and mechanical thrombolysis). However, a large number of patients are still at risk of adverse outcomes, which deserves adequate attention. The global stroke mortality is high, with approximately 7.5 million deaths in 2021 []. In countries with high population density (eg, China), people with AIS have an in-hospital mortality or discharge against medical advice of about 6% [], a 3-month mortality of 1.5%-3.2%, and a 1-year mortality of 3.4%-6% [,]. The 90-day mortality of AIS shows no significant difference between patients undergoing endovascular thrombectomy and those not undergoing endovascular thrombectomy []. Moreover, people with intracerebral hemorrhage have an in-hospital mortality or discharge against medical advice of approximately 21.8%, and a 1-year mortality of 17.9% []. Notably, unlike highly lethal acute-phase hemorrhagic stroke, ischemic stroke presents a unique pattern of “low acute-phase mortality and high medium- to long-term risks” [,], and its mortality is mostly attributed to complications, recurrence, and long-term dysfunction. As a result, time-dynamic risk prediction tools are urgently needed for patients with stroke. Developing stage-specific prediction protocols for high-risk stroke is of great value in early intervention, medical counseling, and resource optimization.
Although numerous studies have sought to predict poststroke functional outcome [,], relevant models apply primarily to those patients with rehabilitation potential. However, critically ill intensive care unit patients are faced with a unique and urgent requirement during clinical decision-making, that is, to accurately assess mortality risk to guide physician-patient communication, treatment strategies, and allocation of health care resources []. Nowadays, validated prediction tools are still lacking for the stroke mortality risk, and clinical assessment relies on traditional tools (eg, Acute Physiology and Chronic Health Evaluation [APACHE] II score). However, such a single-mode assessment is static and subjective, ignoring the inherent complexity of conditions []. Some studies have also been conducted on specific biomarkers. For example, the triglyceride-glucose index (an important biochemical indicator for insulin resistance) can predict AIS mortality [], and adiponectin is independently associated with post-AIS 5-year mortality []. However, early prediction of poststroke death by these prediction tools is still lacking in clinical practice. In addition, the Delphi method can also be applied to the prediction of mortality risk, but the resulting bias is a serious clinical challenge. Therefore, developing early prediction tools for the mortality risk is urgently needed.
Objectives
As big data and artificial intelligence (AI) emerge, modeling by the growing volume of patient-related data has become a hot spot, which enables individualized outcome prediction []. For example, Bonkhoff and Grefkes [] argued that machine learning (ML) plus clinical, electrophysiological, and neuroimaging data exhibit a high potential for poststroke functional recovery; Akay et al [] found that AI achieves individualized outcome prediction and assists in treatment decision-making by analyzing multimodal data. The advantage of ML lies in its capability to integrate multisource heterogeneous data from traditional clinical practice, such as medical histories and medical images, and transform them into efficient predictive systems []. Unlike traditional methods reliant on rigorous statistical hypotheses, ML presets no data distributions and possesses the capability to process high-dimensional, large-scale data. Therefore, ML demonstrates higher accuracy and efficacy in disease classification and prognosis evaluation [,]. Given that patients with stroke often present with complex conditions and multiple complications, and predictions rely heavily on subjective neurological examinations nowadays, it is significant to integrate multimodal clinical data by ML for enhancing the predictive power for stroke mortality []. Under this context, some studies have also further investigated ML for predicting short- [,] or long-term [,] poststroke mortality based on individual data. However, the predictive performance of these models lacks systematic evidence due to the diversity of ML methods and variables, which is a challenge against the subsequent development of simpler clinical prediction tools. No studies have yet quantitatively assessed the effects of these ML models, so the clinical usefulness of the available predictive systems in routine practice remains unclear. Therefore, we conducted this meta-analysis with the following objectives: (1) to measure the predictive performance of ML for stroke mortality across scenarios and summarize the application status of different ML models in stroke mortality prediction, and (2) to evaluate the methodological quality of available studies and determine whether ML can serve as a novel tool for predicting in-hospital and out-of-hospital stroke mortality. The findings are expected to offer an evidence-based basis for facilitating AI development in this field and further developing simple and clinically useful ML-based prediction tools.
Methods
Study Registration
This meta-analysis followed PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 Guidelines [] in . The protocol was prospectively registered with PROSPERO (CRD420251086321). To incorporate constructive refinements suggested during the peer-review process, several deviations from the original PROSPERO protocol were implemented to enhance the study’s rigor. These updates primarily include (1) restructuring the inclusion or exclusion criteria based on the PICOS framework; (2) expanding the analytical scope, including the addition of subgroup and sensitivity (SE) analyses; and (3) refining the statistical methodology by adopting the Hartung-Knapp-Sidik-Jonkman (HKSJ) approach. Accordingly, the protocol has been updated to version 2.0 on the PROSPERO platform to ensure full alignment with the final manuscript.
Eligibility Criteria
The eligibility criteria are as follows ():
Inclusion criteria
- Populations: patients with stroke, including acute ischemic stroke and unspecified stroke types.
- Model construction: studies using artificial intelligence techniques such as machine learning for automated prediction.
- Outcomes: a complete prediction model for mortality risk.
- Study type: case-control, cohort, nested case-control, or case-cohort studies.
- Language: studies reported in English.
Exclusion criteria
- Populations: only patients with hemorrhagic stroke were included.
- Model construction: only risk factors were analyzed, without constructing a mortality prediction model; validation studies solely assessing established scales.
- Outcomes: any of the following evaluation metrics were lacking: receiver operating characteristic curve, C-index, sensitivity, specificity, accuracy, recall, precision, confusion matrix, F1-score, or calibration curves.
- Study type: conference abstracts published without peer review, meta-analysis/review/guidelines/letter/case report/protocol.
- Publication year: earlier than 2005.
- Language: non-English language original studies.
Information Sources and Search Strategy
Literature search was carried out in adherence to PRISMA-S (Preferred Reporting Items for Systematic Reviews and Meta-Analyses literature search extension) guidelines []. Cochrane Library (CENTRAL), PubMed (NCBI), Embase (Embase.com), and Web of Science (Clarivate) were thoroughly searched as of June 21, 2025. In addition, manual searches were conducted on references cited in systematic review reports on the same or similar topics. Developed based on the database requirements, the search strategy combined both medical subject headings (“Stroke,” “Machine learning,” “Deep learning,” and “Mortality”) and free-text keywords, as well as Boolean operators. To ensure a comprehensive retrieval, free-text keywords and relevant subordinate terms were identified via PubMed and Embase in addition to primary MeSH or Emtree terms and were further supplemented based on our clinical expertise. Details of the search strategy are available in . Duplicate records were removed with EndNote 20 software (Clarivate Analytics). Search parameters were set to include only peer-reviewed papers, with no restrictions on geographical location or publication year, and no published search filters were used. No peer review of the search strategy was conducted.
Selection Process
All identified studies were imported into EndNote 20 software. All identified studies were imported into EndNote 20 software. After duplicate publications were removed, we used EndNote’s filtering functionality to exclude records based on publication type (eg, reviews, case reports, letters, protocols, animal studies, and meeting abstracts), language (non-English), and publication date (pre-2005). Subsequently, 2 independent investigators (YJC and ZJO) screened the titles and abstracts of the remaining records based on predefined criteria regarding study population (stroke), intervention (machine learning), and outcomes (mortality). Thereafter, a rigorous full-text review was performed to exclude studies lacking model evaluation parameters or those that did not involve the development of original models. The screen was followed by cross-checking. In case of any disagreements, a third investigator (MKZ) was consulted for resolution.
Data Collection Process and Data Items
A spreadsheet was created to extract data, including title, first author, type of study, year of publication, patient source, dataset source, stroke type, treatment background (whether to undergo acute-phase vascular recanalization), follow-up time, total number of death cases in the training set, overfitting methods, number of cases in the validation set, handling methods for missing value, variable selection, model types, and modeling variables. During data extraction, studies with missing outcome data were excluded directly from the meta-analysis. Three independent investigators (YJC, YTD, and ALL) implemented the data extraction, followed by cross-checking. In case of any disagreements, a third investigator (XTW) was consulted for resolution.
Risk of Bias Assessment
The included studies were assessed for risk of bias (RoB) using the PROBAST (Prediction model Risk of Bias Assessment Tool) [] by structured questions across 4 domains: participants (2 questions), predictors (3 questions), outcomes (6 questions), and analysis (9 questions). Each question was answered as “Yes,” “Probably yes,” “No,” “Probably no,” or “No information.” Each domain was rated as high RoB if 1 or more questions were answered as “No” or “Probably no,” as low RoB if all answers were “Yes” or “Probably yes,” and as unclear RoB if answered “No information.” The overall risk was deemed low only when all domains were rated as low RoB, whereas any high RoB was considered overall high risk. In the absence of high-risk domains, the overall risk was unclear when 1 or more domains had unclear RoB. Two investigators (YJC and YJY) independently performed RoB assessment and cross-checking. Discrepancies were settled by third-party adjudication.
Synthesis Methods
The C-index, a measure for the overall predictive accuracy of ML models, underwent a meta-analysis. If the C-index was reported without its 95% CI or standard error, the SE was estimated using the method described by Debray et al []. Similarly, SE and specificity (SP) were synthesized by a bivariate mixed-effects model. Meta-analyses of these metrics typically rely on diagnostic 4-fold tables, but most primary studies included did not report such tables. In this case, we sought to reconstruct the diagnostic 4-fold table by leveraging other available performance metrics, such as SE, SP, precision, accuracy, and the total number of events (deaths). Some studies were based on the same database, but significant heterogeneity was present across these studies in the data extraction period, variable selection, model construction, and parameter adjustment rule. Therefore, all eligible studies were included, and given this potential heterogeneity, a random-effects model was adopted for meta-analysis []. Multiple entries from the same study represented distinct prediction models or validation methods, which were treated as independent units of evidence []. We performed random-effects meta-analyses using the HKSJ method to estimate pooled effect sizes with 95% CIs. The HKSJ method was chosen as it provides more robust inference, particularly when the number of studies is small. In addition, we calculated 95% prediction intervals (PIs) to present the expected range of true effects in similar studies, across clinical settings, or in future studies, and to better express the extent of heterogeneity [,]. The width of PIs was directly influenced by the number of included studies and the extent of heterogeneity []. Model performance was evaluated separately in training and validation sets. Furthermore, the performance of models for in-hospital and out-of-hospital mortality was separately synthesized.
To explore potential sources of heterogeneity, subgroup analyses were conducted, stratified by stroke type (AIS vs not specific), patient source (public database vs local hospital), treatment background (reperfusion therapy vs not specific), and model type. Furthermore, to evaluate the generalizability of the models and minimize optimistic bias associated with internal validation [], we conducted a subgroup analysis in the validation set by validation method (internal validation vs external validation). When reporting the results, we prioritized the performance of the external validation subgroup based on the primary importance of external validation results for assessing clinical generalization []. The overall pooled performance (combining internal and external validation) was subsequently reported as a reference during development setting. Given the limited number of external validation studies, subgroup analyses were based on the overall performance. We used a random-effects model to calculate pooled estimates within each subgroup and a fixed-effects (plural) model to compare effect sizes across subgroups. A random-effects metaregression was conducted to investigate the association between the C-index of model performance and the follow-up time after discharge in studies reporting out-of-hospital mortality. The small-study effect in the validation sets of included studies and different models (≥10 studies) in the validation sets was qualitatively and quantitatively assessed with funnel plots and Egger regression tests, respectively. Funnel plot asymmetry was visualized to detect the small-study effect.
SE analyses were performed in 2 stages. In the first stage, to assess the potential impact of patient cohort overlap from using the same public databases (Medical Information Mart for Intensive Care [MIMIC-III/IV], eICU Collaborative Research Database [eICU-CRD], and Sentinel Stroke National Audit Programme [SSNAP]) on the pooled results, only 1 study with a large sample size, high methodological quality (based on PROBAST), and complete analysis (using external validation) was retained per public database. All studies from single-center or independent regional registries were also retained. Using independent datasets filtered in the 2-stage ways, a meta-analysis of the C-index for in-hospital and out-of-hospital mortality prediction models was conducted.
To mitigate potential bias arising from studies reporting multiple models, we established a priority rule for the second stage: selecting only the best-performing model from each study. Specifically, for studies with multiple models, we prioritized the optimal model among those with external validation; if no external validation was performed, the model with the highest internal performance was selected. In this way, we could obtain the evidence of the model with the greatest potential for clinical generalization or optimal performance from each study of independent sets, thereby assessing the robustness of the main findings under this stricter criterion. Additionally, following this criterion (the principle of selecting the study with the largest sample size among overlapping validation sets and subsequently selecting the optimal model within each independent validation set), a separate meta-analysis of the C-index was conducted in the external validation set. It should be noted that the SE analyses were performed post hoc, and thus they were not prespecified in the preregistered study protocol or initial data analysis plan. Statistical analyses and visualizations were performed using Stata (version 15.0; StataCorp, LLC) and R (version 4.5.1; R Foundation for Statistical Computing), using metafor, forestplot, and ggplot2 packages.
Protocol Deviations and Updates
To incorporate constructive refinements from the peer-review process, several deviations from the original PROSPERO protocol were implemented to enhance the study’s rigor. These updates primarily include (1) restructuring the inclusion or exclusion criteria according to the PICOS framework, (2) refining the statistical methodology by adopting the HKSJ method and reporting PIs, and (3) expanding the scope of analysis by adding SE and subgroup analyses (eg, internal vs external validation). Accordingly, the study protocol has been updated to version 2.0 on the PROSPERO platform to ensure full alignment with the final paper.
Results
Study Selection
We initially retrieved 21,177 records, of which 6491 duplicate publications were eliminated. After title and abstract review, 14,544 studies were further excluded. Then the full text of the remainder was examined, of which 38 were excluded due to no modeling, 31 due to no definite outcome metrics, and 5 due to a focus only on validation of previous scores. Finally, 68 studies were included [-,-] ().

Study Characteristics
We included 68 studies published between 2005 and 2025, mainly from 2007 to 2025. They came from 19 countries and involved 1,759,472 participants, of whom 225,598 (12.8%) died. Twenty-three studies predicted the in-hospital mortality risk, and 45 studies predicted the out-of-hospital mortality risk. Three studies reported both in-hospital and out-of-hospital mortality [,,]. Out-of-hospital mortality was followed up for 1-12 months, as well as 2 years [], 3 years [,], 5 years [], 10 years [,], and 15 years [] in a few studies. Public databases, including MIMIC, eICU-CRD, the Minimum Basic Data Set, and SSNAP, were used in 28 studies, and there were 26 single-center studies and 14 multicenter studies. Forty-two studies focused on AIS, and the remaining 26 did not specify the stroke type. Seven studies predicted the mortality risk following acute-phase vascular recanalization. Logistic regression (LR) was the most frequently used model (51/68, 75%), followed by extreme gradient boosting (XGBoost) (21/68, 31%), random forest (RF) (20/68, 29%), support vector machine (11/68, 16%), and deep learning (DL) (2/68, 3%). For modeling variables, radiomics features were adopted in only 2 studies, and clinical features were adopted in the remainder. The generation method for the validation set was clearly described in 64 studies. Thirteen studies performed independent external validation, and 11 studies adopted both internal and external validation. For internal validation, random sampling was used in 44 studies, Bootstrap in 9 studies, k-fold cross-validation in 6 studies, holdout validation in 2 studies, and temporal validation in 1 study (Table S1 in ).
RoB in Studies
For participant selection, PROBAST considers retrospective cohort or case-control studies that do not use public databases with high RoB. Since 8 case-control studies did not use publicly available databases, they were rated as high RoB. All included studies developed reasonable eligibility criteria, so no RoB was introduced in this dimension.
The predictors had the same definition and assessment among participants in all studies, and they were all valid, so all studies were rated as low RoB. The outcome event was death, and the possibility of assessment with knowledge of the outcome could not be ruled out for case-control studies, so we rated 8 studies as high RoB.
The outcome in this meta-analysis was death or nondeath, and it was reasonably categorized and defined. The definition of the outcome eliminated the influence of predictors and was consistent among participants, and the outcome was assessed objectively with or without knowledge of the predictor. Therefore, all studies were rated as low RoB in this dimension.
For statistical analysis, 11 studies were assessed as high RoB because of inappropriate sample sizes; 11 as high RoB because subjects with missing data were directly excluded, without appropriate handling, and 3 as unclear RoB due to no description of the handling method for missing data. The RoB in 13 studies was high since only univariate analysis was conducted. The RoB in 10 studies was high for failing to account for data complexity, including high dimensionality, multicollinearity, or unbalanced data handling. Three studies were rated as high RoB for inappropriate internal validation. All studies were rated as low RoB since they all appropriately handled and incorporated continuous or categorical independent variables and assessed model performance.
To sum up, the overall RoB was low in 33 studies and high in 35 studies (Table S2 in [-,-] and ).
Results of Synthesis
In-hospital Mortality
Twenty-five ML models for predicting in-hospital mortality were extracted and included 1,345,411 patients, of whom 159,409 (11.7%) died. The pooled C-index was 0.815 (95% CI 0.783-0.848, 95% PI 0.666-0.997) ( and ). The C-index was 0.807 (95% CI 0.770-0.846, 95% PI 0.669-1.000) in studies on AIS, 0.811 (95% CI 0.771-0.853, 95% PI 0.652-1.000) in studies based on public databases, 0.825 (95% CI 0.759-0.897, 95% PI 0.657-1.000) in studies based on hospital databases, and 0.774 (95% CI 0.745-0.803, 95% PI 0.739-0.810) in studies explicitly reporting acute-phase vascular recanalization. LR, RF, and XGBoost models were dominated, and the LR model had a C-index of 0.803 (95% CI 0.742-0.868, 95% PI 0.638-1.000).
In addition, diagnostic metrics could be calculated in 24 studies in the training set, with a pooled SE of 0.81 (95% CI 0.70-0.88) and SP of 0.82 (95% CI 0.76-0.87). The SE and SP of the LR model were 0.71 (95% CI 0.61-0.79) and 0.78 (95% CI 0.69-0.85) ( and Figures S1 and S2 in ).
| Subgroup analysis | Training set | Validation set | ||||||||||||||||||||||||
| n | Sample size | C-index (95% CI) | 95% PIa | Tau | I2 (%) | n | Sample size | C-index (95% CI) | 95% PI | Tau | I2 (%) | |||||||||||||||
| Model | ||||||||||||||||||||||||||
| ANNb | 1 | 14,8891 | 0.651 (0.646-0.656) | N/Ac | N/A | N/A | 8 | 6837 | 0.794 (0.696-0.905) | 0.544-1.000 | 0.151 | 98.40 | ||||||||||||||
| DTd | —e | — | — | — | — | — | 5 | 5233 | 0.720 (0.648-0.800) | 0.581-0.892 | 0.068 | 73.00 | ||||||||||||||
| ELf | — | — | — | — | — | — | 2 | 845 | 0.864 (0.373-1.000) | N/A | N/A | N/A | ||||||||||||||
| GBMg | 2 | 2633 | 0.842 (0.400-1.000) | N/A | N/A | N/A | 2 | 3813 | 0.813 (0.248-1.000) | N/A | N/A | N/A | ||||||||||||||
| KNNh | — | — | — | — | — | — | 9 | 10,555 | 0.756 (0.707-0.805) | 0.590-0.967 | 0.101 | 92.80 | ||||||||||||||
| LASSOi | 1 | 77,653 | 0.879 (0.871-0.886) | N/A | N/A | N/A | 1 | 9141 | 0.884 (0.863-0.905) | N/A | N/A | N/A | ||||||||||||||
| LightGBMj | — | — | — | — | — | — | 4 | 2223 | 0.828 (0.685-1.000) | 0.549-1.000 | 0.114 | 92.80 | ||||||||||||||
| LRk | 8 | 17,2088 | 0.803 (0.742-0.868) | 0.638-1.000 | 0.091 | 99.10 | 26 | 419,933 | 0.802 (0.762-0.844) | 0.631-1.000 | 0.114 | 99.80 | ||||||||||||||
| NBl | 2 | 1300 | 0.838 (0.720-0.976) | N/A | N/A | N/A | 5 | 2464 | 0.621 (0.339-1.139) | 0.146-1.000 | 0.475 | 95.40 | ||||||||||||||
| RFm | 5 | 15,8453 | 0.789 (0.727-0.856) | 0.647-0.963 | 0.065 | 98.30 | 14 | 13,721 | 0.811 (0.773-0.851) | 0.682-0.963 | 0.077 | 93.20 | ||||||||||||||
| SVMn | 1 | 2031 | 0.924 (0.909-0.939) | N/A | N/A | N/A | 7 | 7056 | 0.763 (0.715-0.814) | 0.643-0.906 | 0.064 | 90.50 | ||||||||||||||
| XGBoosto | 5 | 12,128 | 0.844 (0.738-0.965) | 0.609-1.000 | 0.107 | 99.30 | 16 | 269,990 | 0.803 (0.761-0.847) | 0.650-0.991 | 0.096 | 97.60 | ||||||||||||||
| Stroke type | ||||||||||||||||||||||||||
| AISp | 20 | 56,6742 | 0.807 (0.770-0.846) | 0.669-1.000 | 0.099 | 99.80 | 65 | 108,059 | 0.788 (0.766-0.810) | 0.642-0.965 | 0.101 | 99.60 | ||||||||||||||
| Not specific | 5 | 8435 | 0.848 (0.769-0.936) | 0.669-1.000 | 0.078 | 97.00 | 34 | 643,752 | 0.784 (0.733-0.840) | 0.551-1.000 | 0.171 | 97.20 | ||||||||||||||
| Data resource | ||||||||||||||||||||||||||
| Public database | 18 | 561,389 | 0.811 (0.771-0.853) | 0.652-1.000 | 0.101 | 99.80 | 87 | 723,433 | 0.775 (0.753-0.799) | 0.613-0.980 | 0.117 | 99.50 | ||||||||||||||
| Local hospital | 7 | 13,788 | 0.825 (0.759-0.897) | 0.657-1.000 | 0.087 | 93.50 | 12 | 28,378 | 0.879 (0.840-0.921) | 0.749-1.000 | 0.070 | 95.70 | ||||||||||||||
| Treatment background | ||||||||||||||||||||||||||
| EVTq/IVTr | 4 | 572,769 | 0.774 (0.745-0.803) | 0.739-0.810 | 0.000 | 0.00 | — | — | — | — | — | — | ||||||||||||||
| Not specific | 21 | 2408 | 0.823 (0.786-0.862) | 0.663-1.000 | 0.101 | 99.80 | — | — | — | — | — | — | ||||||||||||||
| Validation method | ||||||||||||||||||||||||||
| Internal validation | — | — | — | — | — | — | 67 | 480,774 | 0.815 (0.796-0.835) | 0.681-0.976 | 0.089 | 99.60 | ||||||||||||||
| External validation | — | — | — | — | — | — | 32 | 271 037 | 0.727 (0.677-0.781) | 0.521-1.000 | 0.170 | 96.20 | ||||||||||||||
| Overall | 25 | 575,177 | 0.815 (0.783-0.848) | 0.666-0.997 | 0.096 | 99.70 | 99 | 751,811 | 0.788 (0.766-0.810) | 0.621-0.999 | 0.119 | 99.50 | ||||||||||||||
aPI: prediction interval.
bANN: artificial neural network.
cN/A: not applicable.
dDT: decision tree.
eNot available.
fEL: ensemble learning.
gGBM: gradient boosting machine.
hKNN: k-nearest neighbor.
iLASSO: least absolute shrinkage and selection operator.
jLightGBM: light gradient boosting machine.
kLR: logistic regression.
lNB: Naive Bayes.
mRF: random forest.
nSVM: support vector machine.
oXGBoost: extreme gradient boosting.
pAIS: acute ischemic stroke.
qEVT: endovascular thrombectomy.
rIVT: intravenous thrombolysis.

| Subgroup analysis | Training set | Validation set | ||||||||||||||||
| n | Sample size | SEa (95% CI) | SPb (95% CI) | n | Sample size | SE (95% CI) | SP (95% CI) | |||||||||||
| Model | ||||||||||||||||||
| ANNc | 2 | 14,9231 | 0.92 (0.86-0.98) | 0.81 (0.64-0.97) | 9 | 7177 | 0.74 (0.58-0.85) | 0.79 (0.72-0.85) | ||||||||||
| DTd | —e | — | — | — | 5 | 5233 | 0.65 (0.44-0.81) | 0.78 (0.52-0.92) | ||||||||||
| ELf | — | — | — | — | 2 | 845 | 0.88 (0.82-0.94) | 0.89 (0.86-0.91) | ||||||||||
| GBMg | 2 | 2633 | 0.88 (0.72-0.79) | 0.79 (0.75-0.83) | 2 | 3813 | 0.58 (0.33-0.83) | 0.78 (0.58-0.98) | ||||||||||
| KNNh | — | — | — | — | 9 | 10,555 | 0.52 (0.34-0.69) | 0.83 (0.70-0.91) | ||||||||||
| LASSOi | 1 | 77,653 | 0.81 | 0.81 | 1 | 9141 | 0.79 | 0.84 | ||||||||||
| LightGBMj | — | — | — | — | 4 | 2223 | 0.73 (0.58-0.83) | 0.82 (0.71-0.90) | ||||||||||
| LRk | 7 | 166,373 | 0.71 (0.61-0.79) | 0.78 (0.69-0.85) | 19 | 189,436 | 0.72 (0.66-0.78) | 0.78 (0.70-0.84) | ||||||||||
| NBl | 2 | 1300 | 0.82 (0.70-0.93) | 0.74 (0.69-0.77) | 5 | 2464 | 0.64 (0.36-0.85) | 0.70 (0.61-0.77) | ||||||||||
| RFm | 4 | 152,738 | 0.66 (0.62-0.70) | 0.78 (0.68,0.86) | 13 | 12,292 | 0.74 (0.63-0.82) | 0.81 (0.69-0.89) | ||||||||||
| SVMn | 1 | 2031 | 0.90 | 0.85 | 7 | 7056 | 0.61 (0.51-0.70) | 0.77 (0.71-0.82) | ||||||||||
| XGBoosto | 5 | 6753 | 0.78 (0.62-0.88) | 0.90 (0.77-0.96) | 13 | 138,201 | 0.69 (0.62-0.75) | 0.73 (0.68-0.78) | ||||||||||
| Stroke type | ||||||||||||||||||
| AISp | 17 | 549,597 | 0.73 (0.66-0.79) | 0.82 (0.75-0.88) | 58 | 102,373 | 0.70 (0.66-0.73) | 0.77 (0.74-0.80) | ||||||||||
| Not specific | 7 | 9115 | 0.92 (0.67-0.98) | 0.83 (0.71-0.90) | 31 | 286,063 | 0.67 (0.57-0.75) | 0.82 (0.74-0.88) | ||||||||||
| Data resource | ||||||||||||||||||
| Public database | 15 | 544,244 | 0.74 (0.67-0.80) | 0.82 (0.75-0.88) | 76 | 359,666 | 0.70 (0.66-0.73) | 0.75 (0.72-0.77) | ||||||||||
| Local hospital | 9 | 14,468 | 0.89 (0.67-0.97) | 0.82 (0.72-0.89) | 13 | 28,770 | 0.63 (0.43-0.80) | 0.93 (0.87-0.97) | ||||||||||
| Treatment background | ||||||||||||||||||
| EVTq/IVTr | 4 | 2408 | 0.68 (0.64-0.72) | 0.76 (0.69-0.82) | N/As | N/A | N/A | N/A | ||||||||||
| Not specific | 20 | 556,304 | 0.83 (0.71-0.91) | 0.83 (0.77-0.88) | N/A | N/A | N/A | N/A | ||||||||||
| Validation method | ||||||||||||||||||
| Internal validation | — | — | — | — | 56 | 111,5921 | 0.71 (0.66-0.76) | 0.81 (0.77-0.85) | ||||||||||
| External validation | — | — | — | — | 33 | 272,515 | 0.64 (0.57-0.70) | 0.74 (0.70-0.77) | ||||||||||
| Overall | 24 | 558,712 | 0.81 (0.70-0.88) | 0.82 (0.76-0.87) | 89 | 388,436 | 0.69 (0.65-0.73) | 0.79 (0.76-0.82) | ||||||||||
aSE: sensitivity.
bSP: specificity.
cANN: artificial neural network.
dDT: decision tree
eNot available.
fEL: ensemble learning.
gGBM: gradient boosting machine.
hKNN: k-nearest neighbor.
iLASSO: least absolute shrinkage and selection operator.
jLightGBM: light gradient boosting machine.
kLR: logistic regression.
lNB: Naive Bayes.
mRF: random forest.
nSVM: support vector machine.
oXGBoost: extreme gradient boosting.
pAIS: acute ischemic stroke.
qEVT: endovascular thrombectomy.
rIVT: intravenous thrombolysis.
sN/A: not applicable.
Based on externally validated models, the analysis indicated a moderate level of generalizability, with a pooled C-index of 0.727 (95% CI 0.677-0.781, 95% PI 0.521-1.000), with an SE of 0.64 (95% CI 0.57-0.70) and an SP of 0.74 (95% CI 0.70-0.77). These performance estimates were lower than the overall pooled results which yielded a C-index of 0.788 (95% CI 0.766-0.810, 95% PI 0.621-0.999), an overall SE of 0.69 (95% CI 0.65-0.73), and an SP of 0.79 (95% CI 0.76-0.82). The models using hospital databases outperformed those using public databases; the former had a C-index of 0.879 (95% CI 0.840-0.921, 95% PI 0.621-0.999), with SP and SE of 0.93 (95% CI 0.87-0.97) and 0.63 (95% CI 0.43-0.80), and the latter had a C-index of 0.775 (95% CI 0.753-0.799, 95% PI 0.613-0.980), with SP and SE of 0.75 (95% CI 0.72-0.77) and 0.70 (95% CI 0.66-0.73). The studies on AIS had a C-index of 0.788 (95% CI 0.766-0.810, 95% PI 0.642-0.965), with SP and SE of 0.70 (95% CI 0.66-0.73) and 0.77 (95% CI 0.74-0.80), and the LR model had a C-index of 0.802 (95% CI 0.762-0.844, 95% PI 0.631-1.000), with SE and SP of 0.72 (95% CI 0.66-0.78) and 0.78 (95% CI 0.70-0.84), consistent with the overall results. The internal validation subgroup aligned with the overall trend but yielded more optimistic estimates, with a pooled C-index of 0.815 (95% CI 0.796-0.835, 95% PI 0.681-0.976), an SE of 0.71 (95% CI 0.66-0.76), and an SP of 0.81 (95% CI 0.77-0.85) ( and , , and Figures S3 and S4 in ).

Out-of-Hospital Mortality
Fifty ML models for predicting out-of-hospital mortality were extracted and included 414,061 patients, of whom 66,189 (16%) died. The pooled C-index was 0.823 (95% CI 0.798-0.848, 95% PI 0.672-1.000) ( and ), and the diagnostic SE and SP were 0.66 (95% CI 0.60-0.72) and 0.88 (95% CI 0.83-0.91) (Figures S5 and S6 in ).
| Subgroup analysis | Training set | Validation set | |||||||||||||
| n | Sample size | C-index (95% CI) | 95% PIa | Tau | I2 (%) | n | Sample size | C-index (95% CI) | 95% PI | Tau | I2 (%) | ||||
| Model | |||||||||||||||
| ANNb | 3 | 4330 | 0.813 (0.768-0.860) | 0.779-0.847 | 0.001 | 46.80 | 4 | 2615 | 0.822 (0.653-1.000) | 0.498-1.000 | 0.140 | 95.3 | |||
| DTc | 1 | 474 | 0.793 (0.697-0.889) | N/Ad | N/A | N/A | 3 | 1906 | 0.69 5 (0.585-0.825) | 0.556-0.868 | 0.025 | 0.0 | |||
| ELe | —f | — | — | — | — | — | 3 | 17,053 | 0.857 (0.708-1.000) | 0.588-1.000 | 0.075 | 97.4 | |||
| GBMg | 3 | 3333 | 0.863 (0.660-1.000) | 0.515-1.000 | 0.103 | 92.20 | 5 | 2076 | 0.788 (0.717-0.866) | 0.652-0.953 | 0.058 | 57.6 | |||
| KNNh | 2 | 2859 | 0.669 (0.0356-1.000) | N/A | N/A | N/A | 4 | 5205 | 0.737 (0.637-0.852) | 0.578-0.939 | 0.065 | 80.1 | |||
| LASSOi | — | — | — | — | — | — | — | — | — | — | — | — | |||
| LightGBMj | 1 | 1388 | 0.840 (0.790-0.890) | N/A | N/A | N/A | 2 | 2116 | 0.926 (0.665-1.000) | N/A | N/A | N/A | |||
| LRk | 24 | 151,284 | 0.822 (0.795-0.850) | 0.698-0.968 | 0.077 | 98.10 | 27 | 149,308 | 0.817 (0.793-0.841) | 0.708-0.942 | 0.068 | 96.8 | |||
| NBl | 3 | 3333 | 0.832 (0.628-1.000) | 0.490-1.000 | 0.105 | 86.60 | 6 | 6687 | 0.809 (0.741-0.883) | 0.651-1.000 | 0.076 | 86.3 | |||
| RFm | 5 | 9502 | 0.835 (0.699-0.996) | 0.549-1.000 | 0.137 | 99.50 | 16 | 17,729 | 0.833 (0.801-0.865) | 0.724-0.957 | 0.063 | 83.1 | |||
| RSFn | — | — | — | — | — | — | 2 | 383 | 0.789 (0.630- 0.986) | N/A | N/A | N/A | |||
| SVMo | 2 | 2859 | 0.762 (0.249-1.000) | N/A | N/A | N/A | 7 | 7225 | 0.789 (0.721-0.863) | 0.620-1.000 | 0.091 | 92.1 | |||
| XGBoostp | 3 | 7660 | 0.857 (0.685-1.000) | 0.552-1.000 | 0.088 | 97.70 | 10 | 11,327 | 0.800 (0.747-0.856) | 0.643-0.994 | 0.091 | 96.8 | |||
| DLq | 1 | 1546 | 0.955 (0.925-0.984) | N/A | N/A | N/A | 2 | 1161 | 0.873 (0.799-0.955) | N/A | N/A | N/A | |||
| COXr | 1 | 425 | 0.822 (0.804-0.839) | N/A | N/A | N/A | 5 | 4991 | 0.792 (0.720-0.871) | 0.632-0.993 | 0.074 | 92.2 | |||
| CatBoosts | 1 | 1471 | 0.895 (0.878-0.912) | N/A | N/A | N/A | 5 | 3902 | 0.818 (0.715-0.937) | 0.599-1.000 | 0.102 | 94.1 | |||
| AdaBoostt | — | — | — | — | — | — | 1 | 1968 | 0.850 (0.811-0.889) | N/A | N/A | N/A | |||
| Stroke type | |||||||||||||||
| AISu | 43 | 176,770 | 0.827 (0.801-0.855) | 0.678-1.000 | 0.097 | 98.90 | 66 | 163,990 | 0.796 (0.778-0.813) | 0.677-0.935 | 0.080 | 91.4 | |||
| Not specific | 7 | 13,694 | 0.796 (0.719-0.882) | 0.597-1.000 | 0.110 | 99.20 | 36 | 71,662 | 0.841 (0.819-0.862) | 0.733-0.964 | 0.066 | 97.8 | |||
| Data resource | |||||||||||||||
| Public database | 24 | 171,811 | 0.826 (0.798-0.854) | 0.696-0.979 | 0.081 | 99.50 | 30 | 161,344 | 0.789 (0.763-0.815) | 0.661-0.941 | 0.085 | 99.4 | |||
| Local hospital | 26 | 18,653 | 0.818 (0.776-0.862) | 0.633-1.000 | 0.015 | 91.70 | 72 | 74,308 | 0.824 (0.808-0.841) | 0.712-0.955 | 0.073 | 87.1 | |||
| Treatment background | |||||||||||||||
| EVTv/IVTw | 3 | 920 | 0.853 (0.796-0.914) | 0.783-0.929 | 0.008 | 0.00 | 7 | 1338 | 0.817 (0.760-0.879) | 0.681-0.981 | 0.068 | 77.5 | |||
| Not specific | 47 | 189,544 | 0.821 (0.795-0.848) | 0.666-1.000 | 0.103 | 99.10 | 95 | 234,314 | 0.811 (0.797-0.826) | 0.691-0.953 | 0.081 | 98.4 | |||
| Validation method | |||||||||||||||
| Internal validation | — | — | — | — | — | — | 92 | 230,828 | 0.808 (0.793-0.823) | 0.687-0.950 | 0.081 | 98.5 | |||
| External validation | — | — | — | — | — | — | 10 | 4824 | 0.847 (0.808-0.887) | 0.750-0.956 | 0.050 | 71.3 | |||
| Overall | 50 | 190,464 | 0.823 (0.798-0.848) | 0.672-1.000 | 0.099 | 99.10 | 102 | 235,652 | 0.812 (0.798-0.826) | 0.693-0.952 | 0.080 | 98.3 | |||
aPI: prediction interval.
bANN: artificial neural network.
cDT: decision tree.
dNot applicable.
eEL: ensemble learning.
fNot available.
gGBM: gradient boosting machine.
hKNN: k-nearest neighbors.
iLASSO: least absolute shrinkage and selection operator.
jLightGBM: light gradient boosting machine.
kLR: logistic regression.
lNB: Naive Bayes.
mRF: random forest.
nRSF: random survival forest.
oSVM: support vector machine.
pXGBoost: extreme gradient boosting.
qDL: deep learning.
rCOX: Cox proportional hazards.
sCatBoost: categorical boosting.
tAdaBoost: adaptive boosting.
uAIS: acute ischemic stroke.
vEVT: endovascular thrombectomy.
wIVT: intravenous thrombolysis.

The pooled C-index was 0.827 (95% CI 0.801-0.855, 95% PI 0.678-1.000) in studies on AIS, 0.826 (95% CI 0.798-0.854, 95% PI 0.696-0.979) in studies based on public databases, 0.818 (95% CI 0.776-0.862, 95% PI 0.633-1.000) in studies based on hospital databases, and 0.853 (95% CI 0.796-0.914, 95% PI 0.783-0.929) in studies reporting acute-phase vascular recanalization. LR remained the most used model (24/50, 48%), with a C-index of 0.824 (95% CI 0.798-0.850) ( and ).
For out-of-hospital mortality, the pooled performance in external validation yielded a C-index of 0.847 (95% CI 0.808-0.887, 95% PI 0.750-0.956), with an SE of 0.71 (95% CI 0.55-0.82) and an SP of 0.76 (95% CI 0.74-0.78). The overall pooled analysis yielded a C-index of 0.812 (95% CI 0.798-0.826, 95% PI 0.693-0.952), with an SE of 0.68 (95% CI 0.63-0.72) and an SP of 0.82 (95% CI 0.78-0.85). Studies on AIS had a pooled C-index of 0.796 (95% CI 0.778-0.813, 95% PI 0.677-0.935), with SE and SP of 0.65 (95% CI 0.59-0.71) and 0.81 (95% CI 0.75-0.85). Studies based on public databases had a pooled C-index of 0.789 (95% CI 0.763-0.815, 95% PI 0.661-0.941), with SE and SP of 0.69 (95% CI 0.62-0.76) and 0.77 (95% CI 0.74-0.80). Studies based on hospital databases had a pooled C-index of 0.824 (95% CI 0.808-0.841, 95% PI 0.712-0.955), with SE and SP of 0.68 (95% CI 0.62-0.74) and 0.84 (95% CI 0.79-0.88). Studies reporting acute-phase vascular recanalization had a pooled C-index of 0.817 (95% CI 0.760-0.879, 95% PI 0.681-0.981), with SE and SP of 0.77 (95% CI 0.69-0.84) and 0.79 (95% CI 0.72-0.85). Internal validation subgroup yielded a C-index of 0.808 (95% CI 0.793-0.823, 95% PI 0.687-0.950), with SE and SP of 0.67 (95% CI 0.62-0.72) and 0.82 (95% CI 0.78-0.85), respectively. The subgroup results were all consistent with the overall trend. In the validation process for out-of-hospital mortality prediction models, those constructed with data sourced from private hospitals show superior overall performance when contrasted with models that rely on data from public datasets.
LR also demonstrated better performance, with a C-index, SE, and SP of 0.817 (95% CI 0.793-0.841, 95% PI 0.708-0.942), 0.74 ( 95% CI 0.69-0.79), and 0.79 (95% CI 0.73-0.84), respectively. DL models for out-of-hospital mortality were created in 2 studies, and they exhibited good discriminatory power (C-index: 0.873, 95% CI 0.799-0.955) but unsatisfactory diagnostic performance (SE 0.48, SP 0.76). ( and , , and Figures S7 and S8 in ).
| Subgroup analysis | Training set | Validation set | ||||||||
| n | Sample size | SEa (95% CI) | SPb (95% CI) | n | Sample size | SE (95% CI) | SP (95% CI) | |||
| Model | ||||||||||
| ANNc | 3 | 4330 | 0.58 (0.55-0.61) | 0.80 (0.73-0.87) | 5 | 2860 | 0.61 (0.38-0.80) | 0.85 (0.74-0.92) | ||
| DTd | —e | — | — | 4 | 2217 | 0.62 (0.24-0.89) | 0.72 (0.45-0.89) | |||
| ELf | — | — | — | — | 3 | 17,053 | 0.76 (0.71-0.81) | 0.80 (0.78-0.81) | ||
| GBMg | 2 | 2859 | 0.55 (0.39-0.71) | 0.84 (0.75-0.93) | 5 | 2076 | 0.56 (0.38-0.72) | 0.80 (0.52-0.94) | ||
| KNNh | 2 | 2859 | 0.45 (0.45-0.46) | 0.92 (0.92-0.93) | 4 | 5205 | 0.53 (0.33-0.72) | 0.80 (0.69-0.88) | ||
| AdaBoostii | — | — | — | — | 1 | 1968 | 0.80 | 0.81 | ||
| LightGBMj | 1 | 1388 | 0.53 | 0.83 | 2 | 2116 | 0.77 (0.67-0.86) | 0.90 (0.82-0.97) | ||
| LRk | 14 | 44,677 | 0.76 (0.69-0.81) | 0.80 (0.76-0.83) | 21 | 29,892 | 0.74 (0.69-0.79) | 0.79 (0.73-0.84) | ||
| NBl | 2 | 2859 | 0.51 (0.50-0.52) | 0.87 (0.82-0.92) | 6 | 6687 | 0.68 (0.55-0.79) | 0.81 (0.67-0.89) | ||
| RFm | 3 | 3313 | 0.68 (0.52-0.83) | 0.81 (0.68-0.93) | 11 | 13,978 | 0.64 (0.46-0.79) | 0.88 (0.72-0.95) | ||
| RSFn | — | — | — | — | 2 | 383 | 0.45 (0.21-0.69) | 0.59 (0.42-0.76) | ||
| SVMo | 2 | 2859 | 0.49 (0.41-0.57) | 0.81 (0.75-0.86) | 7 | 7225 | 0.67 (0.53-0.78) | 0.87 (0.73-0.94) | ||
| XGBoostp | 1 | 1471 | 0.33 | 0.98 | 8 | 9179 | 0.70 (0.48-0.86) | 0.84 (0.70-0.91) | ||
| DLq | 2 | 3821 | 0.92 (0.83-1.00) | 0.99 (0.99-1) | 2 | 1161 | 0.47 (0.47-0.48) | 0.76 | ||
| COXr | 1 | 425 | 0.62 | 0.87 | 3 | 1101 | 0.70 (0.61-0.78) | 0.74 (0.70-0.78) | ||
| CatBoosts | 1 | 1471 | 0.68 | 0.90 | 3 | 2741 | 0.71 (0.52-0.89) | 0.78 (0.77-0.79) | ||
| Stroke type | ||||||||||
| AISt | 29 | 62,174 | 0.64 (0.56-0.70) | 0.89 (0.84-0.93) | 49 | 33,411 | 0.65 (0.59-0.71) | 0.81 (0.75-0.85) | ||
| Not specific | 5 | 10,158 | 0.78 (0.74-0.80) | 0.75 (0.74-0.76) | 38 | 72,431 | 0.70 (0.64-0.76) | 0.83 (0.78-0.87) | ||
| Data resource | ||||||||||
| Public database | 17 | 55,453 | 0.64 (0.57-0.71) | 0.89 (0.84-0.93) | 30 | 54,088 | 0.69 (0.62-0.76) | 0.77 (0.74-0.80) | ||
| Local hospital | 17 | 16,879 | 0.69 (0.58-0.78) | 0.87 (0.77-0.93) | 57 | 51,754 | 0.68 (0.62-0.74) | 0.84 (0.79-0.88) | ||
| Treatment background | ||||||||||
| EVTu/IVTv | — | — | — | — | 5 | 1044 | 0.77 (0.69-0.84) | 0.79 (0.72-0.85) | ||
| Not specific | — | — | — | — | 82 | 104,798 | 0.67 (0.62-0.72) | 0.82 (0.78-0.85) | ||
| Validation method | ||||||||||
| Internal validation | — | — | — | — | 82 | 103,116 | 0.67 (0.62-0.72) | 0.82 (0.78-0.85) | ||
| External validation | — | — | — | — | 5 | 2726 | 0.71 (0.55-0.82) | 0.76 (0.74-0.78) | ||
| Overall | 34 | 72,332 | 0.66 (0.60-0.72) | 0.88 (0.83-0.91) | 87 | 105,842 | 0.68 (0.63- 0.72) | 0.82 (0.78-0.85) | ||
aSE: sensitivity.
bSP: specificity.
cANN: artificial neural network.
dDT: decision tree
eNot available.
fEL: ensemble learning.
gGBM: gradient boosting machine.
hKNN: k-nearest neighbor.
iAdaBoost: adaptive boosting.
jLightGBM: light gradient boosting machine.
kLR: logistic regression.
lNB: Naive Bayes.
mRF: random forest.
nRSF: random survival forest.
oSVM: support vector machine.
pXGBoost: extreme gradient boosting.
qDL: deep learning.
rCOX: Cox proportional hazards.
sCatBoost: categorical boosting.
tAIS: acute ischemic stroke.
uEVT: endovascular thrombectomy.
vIVT: intravenous thrombolysis.

Surprisingly, only a few studies adopted survival analysis modeling (only 1 COX [Cox Proportional Hazards] model in the training set, and only 5 COX models and 2 random survival forest models in the validation set). Furthermore, metaregression revealed a gradual decline in the predictive performance of the LR models and a sustained performance of the RF and XGBoost models (A and B and Figures S9-S11 in ).

Modeling Variables
Of the 155 clinical variables included, high-frequency variables (>5 occurrences) are shown in , with age (n=36), National Institutes of Health Stroke Scale (NIHSS) score (n=16), and hypertension (n=16) as the top 3. These variables were distributed across comorbidities (smoking, previous stroke history, cancer, and in-hospital infections), acute severity of illness (NIHSS, Glasgow Coma Scale, and premorbid modified Rankin Scale [mRS] scores), and admission characteristics (time of admission). Few studies involved biomarkers such as platelet-to-neutrophil ratio (PNR) (n=1) and neutrophil-percentage-to-albumin ratio (NPAR) (n=1).

Small-Study Effect Assessment
As revealed by the funnel plot asymmetry (with fewer studies skewing toward lower performance values) and significant results of Egger tests (P<.001), significant small study effects were present in the included studies on out-of-hospital mortality (Figure S13 in ). Meanwhile, no small study effects were present in the included studies on in-hospital mortality and LR and RF models (Figures S12-S15 in ).
SE Analysis
SE analyses were performed after controlling for cohort overlap, yielding pooled C-indexes based on selected studies: For in-hospital mortality, Ouyang et al [] (MIMIC-III), Cummins et al [] (MIMIC-IV), Huang et al [] (eICU-CRD), and Wang et al [] (SSNAP) were selected; for out-of-hospital mortality, Li et al [] (MIMIC-III) and Zhu et al [] (MIMIC-IV) were selected as representatives. Thus, the pooled C-indexes for in-hospital and out-of-hospital mortality prediction models were 0.801 (95% CI 0.777-0.828, 95% PI 0.682-0.992) and 0.825 (95% CI 0.811-0.840, 95% PI 0.711-0.957), respectively. Similar to the primary analysis, these results indicated that the main findings of this meta-analysis were robust to cohort overlap. From further single original studies, 16 and 32 optimal models for predicting in-hospital and out-of-hospital mortality were retained, respectively, with pooled C-indexes of 0.852 (95% CI 0.811-0.895, 95% PI 0.699-1.000) and 0.842 (95% CI 0.817-0.868, 95% PI 0.720-0.986), respectively (Figures S16 [,,,,,,,,,,,,,-] and S17 [, , , , , , -, , , , , , , , , , -, , , , , , , ] in ). Besides, 5 and 6 external validation studies were included for in-hospital and out-of-hospital mortality, respectively. They yielded pooled C-indexes of 0.785 (95% CI 0.698-0.884, 95% PI 0.594-1.000) and 0.853 (95% CI 0.781-0.931, 95% PI 0.708-1.000), respectively (Figures S18 [,,,,] and S19 [,,,,,] in ).
Discussion
Main Findings
To our knowledge, this is the first meta-analysis to evaluate ML as a promising tool for predicting stroke mortality. Different from previous reviews, we conducted subgroup analyses across multiple dimensions, including the model type, patient source, and time. Sixty-eight studies were included in this study, involving numerous ML models for predicting poststroke mortality. Despite methodological concerns of these studies, the most rigorous evidence on generalizability derived from the externally validated subset. The results demonstrated favorable discriminatory power, with pooled C-indices of approximately 0.73 for in-hospital and 0.85 for out-of-hospital mortality, alongside maintaining balanced SE and SP for both in-hospital and out-of-hospital mortality.
The predictive performance of LR models for long-term out-of-hospital mortality might gradually decline, while RF and XGBoost achieved a robust and sustained performance. Subgroup analyses suggested that the models for AIS and the presence or absence of acute-phase vascular recanalization had no great difference from the overall model. Accounting for approximately 50% of the total, LR had slightly inferior performance to other ML methods. Besides, the PI for the C-index for in-hospital or out-of-hospital mortality did not cover the reference line, suggesting that the LR model consistently outperforms random prediction across clinical settings and possesses translational potential. Notably, the PIs were broader in some subgroups due to the limited number of included studies, so further validation is needed in the future. To sum up, ML holds significant value in predicting both in-hospital and out-of-hospital stroke mortality. These findings offer an evidence basis for developing precise, dynamic clinical risk prediction tools in the future.
Comparison With Previous Reviews
Previous reviews have also investigated ML for early judgment of stroke prognosis. Yang et al [] conducted a meta-analysis of AI models for predicting ischemic stroke outcomes by binary prediction of 90-day mRS, with 7 studies included. They found a pooled area under the curve (AUC) up to 0.872 of the fixed-effects model, confirming the value of AI in predicting a good prognosis of stroke (mRS ≤2). However, they ignored the assessment of mortality (mRS=6). Schwartz et al [] and Wang et al [] included 25 and 7 original studies, respectively, and argued that ML achieves better predictive effects on short- and long-term stroke mortality. However, they performed only a qualitative assessment, thus quantitative evidence was lacking. In addition, available studies on ML for predicting stroke mortality do not classify prediction models and consider the influence of time factors on the model accuracy, generating concerns about heterogeneity in population and methodology. Therefore, we expanded the model performance beyond previous systematic reviews by quantitatively synthesizing evidence from external validations to evaluate model generalizability, in addition to assessing the goodness-of-fit in training sets.
Modeling Variables
Modeling variables (clinical features, radiomics features, genetic information, and microbiomics) also constituted a challenge to this study. Due to the critical conditions of severe stroke, some examinations such as magnetic resonance imaging are difficult and consume vast human and material resources. Only 2 studies used radiomics [,], which relied primarily on interpretable clinical features. We synthesized interpretable clinical features and confirmed that age, NIHSS score, and hypertension were the most used modeling variables, consistent with previous reviews [,]. Meanwhile, some studies also focused on specific laboratory indicators. NPAR combines inflammation with nutrient metabolic status, which is more reflective of the body’s acute stress response and overall reserve status []. Therefore, NPAR possesses greater predictive power. Chunjuan et al [] created a nomogram using NPAR and the systematic inflammatory response index for predicting stroke mortality, with an AUC of 0.637. Chung et al [] proposed that NPAR in combination with simple clinical indicators demonstrates excellent predictive performance for severe illness, with an AUC (0.929) comparable with the Sequential Organ Failure Assessment score, which enables better clinical judgment and interventions. Geng et al [] adopted PNR to reflect thrombotic and inflammatory status, in which AUC increases from 0.736 (PNR alone) to 0.888 (PNR incorporating age and NIHSS scores). Therefore, the association of changes in some variables with the mortality risk can be observed in the future.
For intensive care unit patients, simple biomarkers and clinical parameters can be combined with laboratory tests to predict the disease severity. In addition, radiomics can be further introduced to reflect individual neuropathologic features and predict prognosis []. Besides, models from individual hospitals generally yielded higher C-indexes. The possible reason is as follows: electronic health records at medical centers often provide more detailed, abundant, and clinically relevant predictor variables than public databases covering broad populations, enabling models to capture more complex prediction patterns [,].
Model Selection
Besides modeling variables, the model types (well-interpretable LR, COX, competing risk, and DT) should also be taken into account. These models can quantify the association of variables with risk and thus assist in determining the specific scoring tools, but their accuracy is unsatisfactory. In addition, traditional ML displays high accuracy [], but its interpretability is challenging. Although ML models interpret “black-box models” using various Shapley additive explanations (SHAP) diagrams at the local and global levels in some studies, the SHAP value under its computing paradigm provides the mean contribution of features in the global scope rather than local logical explanations for specific predictions. Due to this essential attribute, SHAP is difficult to meet the requirements for causality and fidelity in high-reliability scenarios such as clinical diagnosis. Consequently, SHAP values often fail to accurately capture the actual impact of features and may generate misleading feature importance rankings, which restricts their credibility and applicability in high-risk decision-making [,]. In the subgroup analysis, LR was dominant in the included studies, with favorable SE and C-index, suggesting that routine interpretable models can early predict stroke mortality. Because of its theoretical simplicity and ease of implementation, LR is widely favored in clinical prediction models []. In particular, datasets may be relatively simple and exhibit linear trends in the distribution when data predicting stroke mortality are primarily from clinical history and testing. Therefore, the performance of LR is often comparable with that of complex ML models [,]. Meanwhile, LR can quantify the association between various influencing factors and mortality risk and further create a nomogram to better visualize results and thus develop convenient online prediction tools []. Thus, despite growing attention to ML, LR remains dominant in data analysis.
Besides predictive accuracy, model interpretability must also be considered in current clinical applications. Generally, highly interpretable models, such as LR, COX, and DT, exhibit limitations in predictive accuracy. In contrast, less interpretable models, such as RF, support vector machine, neural networks, and DL, possess superior accuracy but low interpretability due to their “black-box” property [,]. Consequently, less interpretable models face challenges in clinical practice, including limited use by health care providers and difficulty in patient management []. Therefore, highly interpretable models should be prioritized. If their accuracy fails to meet clinical requirements, model interpretation techniques (eg, Grad-CAM and LIME) should be introduced to enhance understanding and trust in complex model outputs while maintaining predictive performance []. Such interpretable methods are crucial for advancing AI integration and application in clinical settings [,]. Besides, DL has high efficiency for image processing and excellent time effect capture ability when processing time series data [,], so it has the potential for judging stroke site and time. However, DL is less applied to the early prediction of poststroke mortality, possibly attributed to the difficulty in its interpretability and high requirements for techniques []. In the future, DL using multimodal data may be more accurate in prognostic prediction of stroke.
Our meta-analysis revealed a gradual decline in the predictive performance of the overall model and LR model for long-term mortality, consistent with previous reviews [,]. The reason may be related to variations in clinical background, alterations in body status, and introduction of time-varying confounders. Notably, RF achieved a robust performance in long-term mortality prediction, with high accuracy, suggesting that RF may be an effective ML model for predicting stroke mortality. Given rehabilitation progresses postdischarge, condition and complications of patients with stroke often present dynamic evolution, so the reliability of AI and ML models relying solely on information at a single time point during hospitalization is restricted in prediction and assessment. Therefore, real-time collection and analysis of patient data are crucial for accurately identifying individual phenotypic changes []. Establishing a collaborative real-time data network has become a key foundation for long-term, dynamic clinical decision-making []. Moreover, for predicting out-of-hospital mortality, survival analysis modeling was adopted in very few studies. The clinical significance greatly varied between survival analysis modeling and nonsurvival analysis modeling, while the 2 had a minor difference in short-term follow-up; dynamic mortality risk prediction was often a focus in clinical practice, which could be better achieved by survival analysis modeling, with a statistical basis for decision-making []. In contrast, nonsurvival analysis modeling was difficult to achieve dynamic mortality risk prediction. Although LR dominates in the analysis currently, dynamic prediction models should be more considered for long-term mortality in the future to develop predictive strategies in real time.
Prospects and Outlook
The potential of ML for predicting stroke mortality was confirmed in this study, but several technical challenges remain. The population is unbalanced data and the dataset balance is a serious challenge []. This meta-analysis found a mortality of about 13% from public databases and 11% from hospital databases. However, the impact of data imbalance was considered in few studies. We synthesized SE and SP of models of unbalanced data and found that despite good performance in predicting negative events (ie, survival), these models were less accurate in predicting positive events (ie, mortality) and were difficult to produce clinical benefits []. Therefore, the impact of data imbalance should be taken into account in future modeling.
Another challenge arises from the limited availability and divergent performance patterns of external validation evidence. For in-hospital mortality, the performance in external validation was more conservative than the overall performance and performance in internal validation, probably due to optimistic bias in high-dimensional prognostic settings and simple internal validations [,], However, the external validation set remained clinically meaningful and represented an appropriate basis for assessing model generalizability. Conversely, regarding out-of-hospital mortality, the external validation set had more satisfactory C-index than internal validation. This phenomenon was observed in all the 4 included studies [,,,], which might be related to the fact that patients had more severe conditions, higher mortality, and greater case-mix variation in external validation, thus enhancing the model’s discriminatory power [,]. However, due to the limited number and potential representativeness concerns [], the generalizability of this finding requires cautious interpretation.
This meta-analysis revealed significant small study effects in out-of-hospital mortality, with small sample studies tending to report lower C-index values. This may also be attributed to limited standardized data, introducing data noise that attenuated the model’s apparent performance []. Furthermore, the optimal performance still generally fell below the average of large sample studies. Small sample studies also possessed a narrower case-mix (ie, more homogeneous population), and established evidence suggests that the C-index is naturally lower in homogeneous populations []. The lower pooled C-index observed in small sample studies likely reflected clinical heterogeneity or implementation quality, rather than publication bias or outcome reporting bias []. Consequently, the inclusion of these studies probably yielded a conservative underestimate of the true performance of robust models developed with sufficient resources. Also, our findings of subgroup analyses should be interpreted with caution. Since algorithm types were not randomly assigned across studies, these differences were observed and might be confounded by other study-level factors rather than the algorithms themselves. Therefore, these results represented associations rather than definitive causal evidence [].
Furthermore, the translation of these models into clinical practice requires rigorous validation. The vast majority of included studies did not distinguish independent test sets and lacked valid external validation. As a result, the system suitability of the AI-based clinical decision support system was limited in clinical workflows []. Therefore, future research should conduct external validation of algorithms by program standardization in larger prospective cohorts and compare multiple existing models by external validation to determine the most useful one []. Moreover, the models can be compared with relevant clinicians’ judgment, such as providing clinicians (including intensivists and neurologists) with routine clinical information. Additional metrics can also be assessed, such as family acceptance of AI results, resource utilization and planning, and adjustments to subsequent medical plans []. Additionally, these algorithms still need improvement to identify mortality in all at-risk populations (low-, medium-, and high-risk populations), and they should be trained by a large number of outlier samples or oversampling techniques, such as random or boundary oversampling or adaptive synthetic sampling. These approaches will facilitate the critical deployment of AI in stroke mortality prediction.
Advantages and Limitations
Advantages
This meta-analysis first quantitatively reviewed the value of ML in predicting in-hospital and out-of-hospital mortality in stroke. The findings are expected to offer an accurate framework for poststroke mortality risk assessment and clinical decision-making.
Limitations
Some limitations are worth noting. First, this meta-analysis exhibited great heterogeneity, which can be attributed to the variations in algorithms and study populations. Few studies reporting acute-phase vascular recanalization were included, and most of the studies did not classify the treatment background in detail, which may affect the overall performance assessment. Also, data mostly came from available cohort studies or conventional registries, and some were single-center studies, introducing some bias and restricting the accuracy. Therefore, prospective multicenter studies with detailed subgroup analyses (including treatment backgrounds) are needed. Second, outcome metrics were missing in a small number of studies, so reasonable estimation was implemented, but the model’s predictive accuracy estimates were still affected. Third, internal validation was used in the majority of the studies and external validation in only 20% of the studies, weakening the model’s extrapolation capability. Fourth, multiple studies included were from a small number of large public databases, potentially leading to cohort overlap. Although the limited impact of potential cohort overlap on the pooled effect was confirmed by the SE analysis, we recognized that this method could not fully eliminate residual overfitting bias, given that such bias may arise from systematic similarities in patient populations and clinical protocols across these databases []. This concealed overlap could still inflate the summary effect estimates and complicate the interpretation of heterogeneity. Therefore, the extrapolation of our conclusions to more heterogeneous clinical settings warrants caution.
Conclusions
ML demonstrates promising performance in stroke mortality prediction, and some ML models are more valuable in predicting long-term mortality. However, more efforts are needed before application in clinical practice. For example, we should focus on population SP, search for more specific biomarkers, introduce more multimodal data, optimize model balance, and increase external validations. Furthermore, ML models should be used to develop prediction tools for stroke mortality risk applicable in clinical settings to advance their translation into clinical practice. In this way, data-based doctor-patient shared decision-making can be facilitated, and more efficient resource allocation can be achieved in health care institutions, ultimately ameliorating patient outcomes.
Acknowledgments
The authors attest that there was no use of generative artificial intelligence technology in the generation of text, figures, or other informational content of this manuscript.
Funding
This study was funded by the Regional Innovation Cooperation Project of the Science and Technology Department of Sichuan Province (2024YFHZ0142), Sichuan Science and Technology Program (no. 2023YFQ0109), and Health Commission of Sichuan Province Medical Science and Technology Program (2023-109 and 24LCYJPT08), as well as the National Key R&D Program of China (no. 2022YFC2503805). The funder had no involvement in the study design, data collection, analysis, interpretation, or the writing of the manuscript.
Data Availability
All data generated or analyzed during this study are included in this published paper and its supplementary information files.
Authors' Contributions
All authors contributed to the study conception and design. YC and ZO participated in writing – original draft preparation, and writing – review and editing. YD contributed to conceptualization. YC and YY contributed to methodology. YC, ZO, YD, AL, XL, YY, and QY conducted the formal analysis and investigation. MZ and XW participated in supervision. All authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Conflicts of Interest
None declared.
PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist.
PDF File (Adobe PDF File), 97 KBDetails of the search strategy.
DOCX File , 3381 KBResults of Prediction model Risk of Bias Assessment Tool assessment of risk of bias in included studies.
PNG File , 57 KBReferences
- Hilkens NA, Casolla B, Leung TW, de Leeuw FE. Stroke. Lancet. 2024;403(10446):2820-2836. [CrossRef] [Medline]
- Feigin VL, Brainin M, Norrving B, Martins SO, Pandian J, Lindsay P, et al. World stroke organization: Global stroke fact sheet 2025. Int J Stroke. 2025;20(2):132-144. [FREE Full text] [CrossRef] [Medline]
- Pendlebury ST, Rothwell PM. Prevalence, incidence, and factors associated with pre-stroke and post-stroke dementia: a systematic review and meta-analysis. Lancet Neurol. 2009;8(11):1006-1018. [CrossRef] [Medline]
- Feigin VL, Owolabi MO. Pragmatic solutions to reduce the global burden of stroke: a world stroke organization-lancet neurology commission. Lancet Neurol. 2023;22(12):1160-1206. [CrossRef] [Medline]
- Global, regional, and national burden of stroke and its risk factors, 1990-2021. A systematic analysis for the global burden of disease study 2021. Lancet Neurol. 2024;23(10):973-1003. [FREE Full text] [CrossRef] [Medline]
- Gu HQ, Yang X, Wang CJ, Zhao XQ, Wang YL, Liu LP, et al. Clinical characteristics, management, and in-hospital outcomes in patients with stroke or transient ischemic attack in china. JAMA Netw Open. 2021;4(8):e2120745. [FREE Full text] [CrossRef] [Medline]
- Wang M, Wang CJ, Gu HQ, Meng X, Jiang Y, Yang X, et al. Sex differences in short-term and long-term outcomes among patients with acute ischemic stroke in china. Stroke. 2022;53(7):2268-2275. [FREE Full text] [CrossRef] [Medline]
- Tu WJ, Chao BH, Ma L, Yan F, Cao L, Qiu H, et al. Case-fatality, disability and recurrence rates after first-ever stroke: a study from bigdata observatory platform for stroke of China. Brain Res Bull. 2021;175:130-135. [CrossRef] [Medline]
- Goyal M, Menon BK, van Zwam WH, Dippel DWJ, Mitchell PJ, Demchuk AM, et al. HERMES Collaborators. Endovascular thrombectomy after large-vessel ischaemic stroke: a meta-analysis of individual patient data from five randomised trials. Lancet. 2016;387(10029):1723-1731. [CrossRef] [Medline]
- Tu WJ, Wang LD. China stroke surveillance report 2021. Mil Med Res. 2023;10(1):33. [FREE Full text] [CrossRef] [Medline]
- van Asch CJ, Luitse MJ, Rinkel GJ, van der Tweel I, Algra A, Klijn CJ. Incidence, case fatality, and functional outcome of intracerebral haemorrhage over time, according to age, sex, and ethnic origin: a systematic review and meta-analysis. Lancet Neurol. 2010;9(2):167-176. [CrossRef] [Medline]
- Wu S, Wu B, Liu M, Chen Z, Wang W, Anderson CS, et al. Stroke in China: advances and challenges in epidemiology, prevention, and management. Lancet Neurol. 2019;18(4):394-405. [CrossRef] [Medline]
- Montellano FA, Ungethüm K, Ramiro L, Nacu A, Hellwig S, Fluri F, et al. Role of blood-based biomarkers in ischemic stroke prognosis. Stroke. 2021;52(2):543-551. [CrossRef]
- Kremers F, Venema E, Duvekot M, Yo L, Bokkers R, Lycklama À Nijeholt G, et al. Outcome prediction models for endovascular treatment of ischemic stroke: Systematic review and external validation. Stroke. 2022;53(3):825-836. [FREE Full text] [CrossRef] [Medline]
- Xie P, Hu Y, Li J, Ma Y, Xiao J. Unlocking the potential of real-time ICU mortality prediction: redefining risk assessment with continuous data recovery. NPJ Digit Med. 2025;8(1):733. [FREE Full text] [CrossRef] [Medline]
- Solomon OM, Gu Y, Wei Y. Risk factors for enteral nutrition feeding intolerance in stroke patients: a scoping review. Nurs Crit Care. 2025;30(3):e13183. [CrossRef] [Medline]
- Yang Y, Huang X, Wang Y, Leng L, Xu J, Feng L, et al. The impact of triglyceride-glucose index on ischemic stroke: a systematic review and meta-analysis. Cardiovasc Diabetol. 2023;22(1):2. [FREE Full text] [CrossRef] [Medline]
- Efstathiou SP, Tsioulos DI, Tsiakou AG, Gratsias YE, Pefanis AV, Mountokalakis TD. Plasma adiponectin levels and five-year survival after first-ever ischemic stroke. Stroke. 2005;36(9):1915-1919. [CrossRef] [Medline]
- Cecconi M, Greco M, Shickel B, Vincent JL, Bihorac A. Artificial intelligence in acute medicine: a call to action. Crit Care. 2024;28(1):258. [CrossRef] [Medline]
- Bonkhoff AK, Grefkes C. Precision medicine in stroke: towards personalized outcome predictions using artificial intelligence. Brain. 2022;145(2):457-475. [FREE Full text] [CrossRef] [Medline]
- Akay EMZ, Hilbert A, Carlisle BG, Madai VI, Mutke MA, Frey D. Artificial intelligence for clinical decision support in acute ischemic stroke: a systematic review. Stroke. 2023;54(6):1505-1516. [FREE Full text] [CrossRef] [Medline]
- Zhang J, Wu J, Zhou XS, Shi F, Shen D. Recent advancements in artificial intelligence for breast cancer: image augmentation, segmentation, diagnosis, and prognosis approaches. Semin Cancer Biol. 2023;96:11-25. [CrossRef] [Medline]
- Borchert RJ, Azevedo T, Badhwar A, Bernal J, Betts M, Bruffaerts R, et al. Artificial intelligence for diagnostic and prognostic neuroimaging in dementia: a systematic review. Alzheimers Dement. 2023;19(12):5885-5904. [CrossRef] [Medline]
- Alowais SA, Alghamdi SS, Alsuhebany N, Alqahtani T, Alshaya AI, Almohareb SN, et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med Educ. 2023;23(1):689. [FREE Full text] [CrossRef] [Medline]
- Radley NM, Soh I, Saad AM, Wickramarachchi M, Dawson A, Hin JNC, et al. Risk of bias assessment of post-stroke mortality machine learning predictive models: systematic review. J Stroke Cerebrovasc Dis. 2025;34(6):108291. [FREE Full text] [CrossRef] [Medline]
- Liu W, Ma W, Bai N, Li C, Liu K, Yang J, et al. Identification of key predictors of hospital mortality in critically ill patients with embolic stroke using machine learning. Biosci Rep. 2022;42(9):BSR20220995. [FREE Full text] [CrossRef] [Medline]
- Kurtz P, Peres IT, Soares M, Salluh JIF, Bozza FA. Hospital length of stay and 30-day mortality prediction in stroke: a machine learning analysis of 17,000 ICU admissions in Brazil. Neurocrit Care. 2022;37(Suppl 2):313-321. [CrossRef] [Medline]
- Kniep HC, Elsayed S, Nawabi J, Broocks G, Meyer L, Bechstein M, et al. Imaging-based outcome prediction in posterior circulation stroke. J Neurol. 2022;269(7):3800-3809. [FREE Full text] [CrossRef] [Medline]
- Bray BD, Campbell J, Cloud GC, Hoffman A, James M, Tyrrell PJ, et al. Derivation and external validation of a case mix model for the standardized reporting of 30-day stroke mortality rates. Stroke. 2014;45(11):3374-3380. [CrossRef] [Medline]
- Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71. [FREE Full text] [CrossRef] [Medline]
- Rethlefsen ML, Kirtley S, Waffenschmidt S, Ayala AP, Moher D, Page MJ, et al. PRISMA-S Group. PRISMA-S: an extension to the PRISMA statement for reporting literature searches in systematic reviews. Syst Rev. 2021;10(1):39. [FREE Full text] [CrossRef] [Medline]
- Moons KGM, Wolff RF, Riley RD, Whiting PF, Westwood M, Collins GS, et al. PROBAST: a tool to assess risk of bias and applicability of prediction model studies: explanation and elaboration. Ann Intern Med. 2019;170(1):W1-W33. [FREE Full text] [CrossRef] [Medline]
- Debray TP, Damen JA, Riley RD, Snell K, Reitsma JB, Hooft L, et al. A framework for meta-analysis of prediction model studies with binary and time-to-event outcomes. Stat Methods Med Res. 2019;28(9):2768-2786. [FREE Full text] [CrossRef] [Medline]
- Borenstein M, Hedges LV, Higgins JPT, Rothstein HR. A basic introduction to fixed-effect and random-effects models for meta-analysis. Res Synth Methods. 2010;1(2):97-111. [FREE Full text] [CrossRef] [Medline]
- Debray TPA, Damen JAAG, Snell KIE, Ensor J, Hooft L, Reitsma JB, et al. A guide to systematic review and meta-analysis of prediction model performance. BMJ. 2017;356:i6460. [FREE Full text] [CrossRef] [Medline]
- Nagashima K, Noma H, Furukawa TA. Prediction intervals for random-effects meta-analysis: a confidence distribution approach. Stat Methods Med Res. 2019;28(6):1689-1702. [CrossRef] [Medline]
- Röver C, Knapp G, Friede T. Hartung-Knapp-Sidik-Jonkman approach and its modification for random-effects meta-analysis with few studies. BMC Med Res Methodol. 2015;15:99. [FREE Full text] [CrossRef] [Medline]
- Partlett C, Riley RD. Random effects meta-analysis: coverage performance of 95% confidence and prediction intervals following REML estimation. Stat Med. 2017;36(2):301-317. [FREE Full text] [CrossRef] [Medline]
- Damen JAA, Moons KGM, van Smeden M, Hooft L. How to conduct a systematic review and meta-analysis of prognostic model studies. Clin Microbiol Infect. 2023;29(4):434-440. [FREE Full text] [CrossRef] [Medline]
- Norouzi S, Ahmadi S, Alinia S, Farzipoor F, Shahsavari A, Hajizadeh E, et al. Machine learning predictive models for survival in patients with brain stroke. Health Promot Perspect. 2025;15(1):63-72. [CrossRef] [Medline]
- Zhou Z, Chen B, Mei Z, Chen W, Cao W, Xu E, et al. Machine learning-based scoring model for predicting mortality in ICU-admitted ischemic stroke patients with moderate to severe consciousness disorders. Front Neurol. 2025;16:1534961. [FREE Full text] [CrossRef] [Medline]
- Xu L, Li C, Zhang J, Guan C, Zhao L, Shen X, et al. Personalized prediction of mortality in patients with acute ischemic stroke using explainable artificial intelligence. Eur J Med Res. 2024;29(1):341. [FREE Full text] [CrossRef] [Medline]
- Wang W, Rudd AG, Wang Y, Curcin V, Wolfe CD, Peek N, et al. Risk prediction of 30-day mortality after stroke using machine learning: a nationwide registry-based cohort study. BMC Neurol. 2022;22(1):195. [FREE Full text] [CrossRef] [Medline]
- Wang K, Gu L, Liu W, Xu C, Yin C, Liu H, et al. The predictors of death within 1 year in acute ischemic stroke patients based on machine learning. Front Neurol. 2023;14:1092534. [FREE Full text] [CrossRef] [Medline]
- Someeh N, Mirfeizi M, Asghari-Jafarabadi M, Alinia S, Farzipoor F, Shamshirgaran SM. Predicting mortality in brain stroke patients using neural networks: outcomes analysis in a longitudinal study. Sci Rep. 2023;13(1):18530. [FREE Full text] [CrossRef] [Medline]
- Scrutinio D, Ricciardi C, Donisi L, Losavio E, Battista P, Guida P, et al. Machine learning to predict mortality after rehabilitation among patients with severe stroke. Sci Rep. 2020;10(1):20127. [FREE Full text] [CrossRef] [Medline]
- Petrović I, Broggi S, Killer-Oberpfalzer M, Pfaff JAR, Griessenauer CJ, Milosavljević I, et al. Predictors of in-hospital mortality after thrombectomy in anterior circulation large vessel occlusion: a retrospective, machine learning study. Diagnostics (Basel). 2024;14(14):1531. [FREE Full text] [CrossRef] [Medline]
- Cummins JA, Gerber BS, Fukunaga MI, Henninger N, Kiefe CI, Liu F. In-hospital mortality prediction among intensive care unit patients with acute ischemic stroke: a machine learning approach. Health Data Sci. 2025;5:0179. [FREE Full text] [CrossRef] [Medline]
- Park E, Chang HJ, Nam HS. A bayesian network model for predicting post-stroke outcomes with available risk factors. Front Neurol. 2018;9:699. [FREE Full text] [CrossRef] [Medline]
- Ouyang Y, Cheng M, He B, Zhang F, Ouyang W, Zhao J, et al. Interpretable machine learning models for predicting in-hospital death in patients in the intensive care unit with cerebral infarction. Comput Methods Programs Biomed. 2023;231:107431. [CrossRef] [Medline]
- Luo X, Li B, Zhu R, Tai Y, Wang Z, He Q, et al. Development and validation of an interpretable machine learning model for predicting in-hospital mortality for ischemic stroke patients in ICU. Int J Med Inform. 2025;198:105874. [CrossRef] [Medline]
- Mai TD, Nguyen DT, Tran CC, Duong HQ, Nguyen HN, Dang DP, et al. Factors associated with 90-day mortality in Vietnamese stroke patients: prospective findings compared with explainable machine learning, multicenter study. PLoS One. 2024;19(9):e0310522. [FREE Full text] [CrossRef] [Medline]
- Lehmann ALCF, Alfieri DF, de Araújo MCM, Trevisani ER, Nagao MR, Pesente FS, et al. Immune-inflammatory, coagulation, adhesion, and imaging biomarkers combined in machine learning models improve the prediction of death 1 year after ischemic stroke. Clin Exp Med. 2022;22(1):111-123. [CrossRef] [Medline]
- Kim C, Kwon JM, Lee J, Jo H, Gwon D, Jang JH, et al. Deep learning model integrating radiologic and clinical data to predict mortality after ischemic stroke. Heliyon. 2024;10(10):e31000. [FREE Full text] [CrossRef] [Medline]
- Huang R, Liu J, Wan TK, Siriwanna D, Woo YMP, Vodencarevic A, et al. Stroke mortality prediction based on ensemble learning and the combination of structured and textual data. Comput Biol Med. 2023;155:106176. [CrossRef] [Medline]
- Huang J, Liu X, Jin W. Clinical decision support systems for 3-month mortality in elderly patients admitted to ICU with ischemic stroke using interpretable machine learning. Digit Health. 2024;10:20552076241280126. [FREE Full text] [CrossRef] [Medline]
- Huang J, Jin W, Duan X, Liu X, Shu T, Fu L, et al. Twenty-eight-day in-hospital mortality prediction for elderly patients with ischemic stroke in the intensive care unit: interpretable machine learning models. Front Public Health. 2022;10:1086339. [FREE Full text] [CrossRef] [Medline]
- Huang J, Chen H, Deng J, Liu X, Shu T, Yin C, et al. Interpretable machine learning for predicting 28-day all-cause in-hospital mortality for hypertensive ischemic or hemorrhagic stroke patients in the ICU: a multi-center retrospective cohort study with internal and external cross-validation. Front Neurol. 2023;14:1185447. [FREE Full text] [CrossRef] [Medline]
- Hoffman H, Wood J, Cote JR, Jalal MS, Otite FO, Masoud HE, et al. Development and internal validation of machine learning models to predict mortality and disability after mechanical thrombectomy for acute anterior circulation large vessel occlusion. World Neurosurg. 2024;182:e137-e154. [CrossRef] [Medline]
- Fernandez-Lozano C, Hervella P, Mato-Abad V, Rodríguez-Yáñez M, Suárez-Garaboa S, López-Dequidt I, et al. Random forest-based prediction of stroke outcome. Sci Rep. 2021;11(1):10071. [FREE Full text] [CrossRef] [Medline]
- Fang J, Song B, Li L, Tong L, Jiang M, Yan J. RGX ensemble model for advanced prediction of mortality outcomes in stroke patients. BME Front. 2024;5:0077. [FREE Full text] [CrossRef] [Medline]
- Chunjuan Z, Yulong W, Xicheng Z, Xiaodong M. Machine learning consensus clustering for inflammatory subtype analysis in stroke and its impact on mortality risk: a study based on NHANES (1999-2018). Front Neurol. 2025;16:1562247. [FREE Full text] [CrossRef] [Medline]
- Chen Y, Yang Z, Liu Y, Li Y, Zhong Z, McDowell G, et al. Exploring the prognostic impact of triglyceride-glucose index in critically ill patients with first-ever stroke: insights from traditional methods and machine learning-based mortality prediction. Cardiovasc Diabetol. 2024;23(1):443. [FREE Full text] [CrossRef] [Medline]
- Chen H, Yang F, Duan Y, Yang L, Li J. A novel higher performance nomogram based on explainable machine learning for predicting mortality risk in stroke patients within 30 days based on clinical features on the first day ICU admission. BMC Med Inform Decis Mak. 2024;24(1):161. [FREE Full text] [CrossRef] [Medline]
- Che Nawi CMNH, Mohd Hairon S, Wan Yahya WNN, Wan Zaidi WA, Musa KI. Machine learning models for predicting stroke mortality in malaysia: an application and comparative analysis. Cureus. 2023;15(12):e50426. [FREE Full text] [CrossRef] [Medline]
- Abujaber A, Yaseen S, Imam Y, Nashwan A, Akhtar N. Machine learning-based prediction of one-year mortality in ischemic stroke patients. Oxf Open Neurosci. 2024;3:kvae011. [CrossRef] [Medline]
- Abedi V, Avula V, Razavi SM, Bavishi S, Chaudhary D, Shahjouei S, et al. Predicting short and long-term mortality after acute ischemic stroke using EHR. J Neurol Sci. 2021;427:117560. [FREE Full text] [CrossRef] [Medline]
- Zhou L, Wu Y, Wang J, Wu H, Tan Y, Chen X, et al. Development of a predictive nomogram for intra-hospital mortality in acute ischemic stroke patients using LASSO regression. Clin Interv Aging. 2024;19:1423-1436. [FREE Full text] [CrossRef] [Medline]
- Zhang X, Yuan K, Wang H, Gong P, Jiang T, Xie Y, et al. Nomogram to predict mortality of endovascular thrombectomy for ischemic stroke despite successful recanalization. J Am Heart Assoc. 2020;9(3):e014899. [FREE Full text] [CrossRef] [Medline]
- Zhang RX, Zhang WW, Luo YT, Liu GW. An mNUTRIC-based nomogram for predicting the in-hospital death risk in patients with acute stroke. Eur J Clin Nutr. 2022;76(10):1464-1469. [CrossRef] [Medline]
- Zhang J, Zhuang S, Zhang Y, An L, Li D, Liu Z, et al. Ultrasonic optic nerve sheath diameter as a new predictor for the mortality of patients with large hemispheric infarction. Sci Rep. 2025;15(1):460. [FREE Full text] [CrossRef] [Medline]
- Zeng M, Smith L, Bird A, Trinh VQ, Bacchi S, Harvey J, et al. Predictions for functional outcome and mortality in acute ischaemic stroke following successful endovascular thrombectomy. BMJ Neurol Open. 2024;6(1):e000707. [FREE Full text] [CrossRef] [Medline]
- Çelik G, Baykan Ö, Kara Y, Tireli H. Predicting 10-day mortality in patients with strokes using neural networks and multivariate statistical methods. J Stroke Cerebrovasc Dis. 2014;23(6):1506-1512. [CrossRef] [Medline]
- Wen R, Wang M, Bian W, Zhu H, Xiao Y, He Q, et al. Nomogram to predict 6-month mortality in acute ischemic stroke patients treated with endovascular treatment. Front Neurol. 2023;14:1330959. [FREE Full text] [CrossRef] [Medline]
- Szlachetka WA, Pana TA, Mamas MA, Bettencourt-Silva JH, Metcalf AK, Potter JF, et al. Predicting 10-year stroke mortality: development and validation of a nomogram. Acta Neurol Belg. 2022;122(3):685-693. [FREE Full text] [CrossRef] [Medline]
- Sun W, Li G, Liu Z, Miao J, Yang Z, Zhou Q, et al. A nomogram for predicting the in-hospital mortality after large hemispheric infarction. BMC Neurol. 2019;19(1):347. [FREE Full text] [CrossRef] [Medline]
- Solberg OG, Dahl M, Mowinckel P, Stavem K. Derivation and validation of a simple risk score for predicting 1-year mortality in stroke. J Neurol. 2007;254(10):1376-1383. [CrossRef] [Medline]
- Sha L, Xu T, Ge X, Shi L, Zhang J, Guo H. Predictors of death within 6 months of stroke onset: a model with barthel index, platelet/lymphocyte ratio and serum albumin. Nurs Open. 2021;8(3):1380-1392. [FREE Full text] [CrossRef] [Medline]
- Ni H, Zhao L, Ma X, Lv P, Ding Y, Liu Z, et al. Technical risk stratification nomogram model for 90-day mortality prediction in patients with acute basilar artery occlusion undergoing endovascular thrombectomy: a multicenter cohort study. J Am Heart Assoc. 2024;13(6):e032107. [FREE Full text] [CrossRef] [Medline]
- Liljehult J, Christensen T, Christensen KB. Early prediction of one-year mortality in ischemic and haemorrhagic stroke. J Stroke Cerebrovasc Dis. 2020;29(4):104667. [CrossRef] [Medline]
- Li XD, Li MM. A novel nomogram to predict mortality in patients with stroke: a survival analysis based on the MIMIC-III clinical database. BMC Med Inform Decis Mak. 2022;22(1):92. [FREE Full text] [CrossRef] [Medline]
- Kim TJ, Lee JS, Oh MS, Kim JW, Park SH, Yu KH, et al. Stroke measures analysis of pRognostic testing-mortality nomogram predicts long-term mortality after ischemic stroke. Int J Stroke. 2025;20(1):75-84. [CrossRef] [Medline]
- Lee J, Morishima T, Kunisawa S, Sasaki N, Otsubo T, Ikai H, et al. Derivation and validation of in-hospital mortality prediction models in ischaemic stroke patients using administrative data. Cerebrovasc Dis. 2013;35(1):73-80. [CrossRef] [Medline]
- Jin G, Hu W, Zeng L, Ma B, Zhou M. Prediction of long-term mortality in patients with ischemic stroke based on clinical characteristics on the first day of ICU admission: an easy-to-use nomogram. Front Neurol. 2023;14:1148185. [FREE Full text] [CrossRef] [Medline]
- Jin G, Hu W, Zeng L, Diao M, Chen H, Chen J, et al. Development and verification of a nomogram for predicting short-term mortality in elderly ischemic stroke populations. Sci Rep. 2023;13(1):12580. [FREE Full text] [CrossRef] [Medline]
- Gao H, Sun X, Li W, Gao Q, Zhang J, Zhang Y, et al. Development and validation of a risk score to predict 30-day mortality in patients with atrial fibrillation-related stroke: GPS-GF score. Neurol Res. 2018;40(7):532-540. [CrossRef] [Medline]
- Fang L, Zhou M, Mao F, Diao M, Hu W, Jin G. Development and validation of a nomogram for predicting 28-day mortality in patients with ischemic stroke. PLoS One. 2024;19(4):e0302227. [FREE Full text] [CrossRef] [Medline]
- Cappellari M, Zivelonghi C, Turcato G, Forlivesi S, Micheletti N, Tomelleri G, et al. A nomogram to predict the probability of mortality after first-ever acute manifestations of cerebral small vessel disease. J Neurol Sci. 2018;385:92-95. [CrossRef] [Medline]
- Schwartz J, Wang Y, Qin L, Schwamm LH, Fonarow GC, Cormier N, et al. Incorporating stroke severity into hospital measures of 30-day mortality after ischemic stroke hospitalization. Stroke. 2017;48(11):3101-3107. [CrossRef] [Medline]
- Reitzle L, Rohmann JL, Kurth T, Audebert HJ, Piccininni M. External validation of risk prediction models for post-stroke mortality in Berlin. BMJ Open. 2025;15(6):e089320. [FREE Full text] [CrossRef] [Medline]
- Hwangbo L, Kang YJ, Kwon H, Lee JI, Cho H, Ko J, et al. Stacking ensemble learning model to predict 6-month mortality in ischemic stroke patients. Sci Rep. 2022;12(1):17389. [FREE Full text] [CrossRef] [Medline]
- Huang Y, Douiri A, Fahey M. A dynamic model for predicting survival up to 1 year after ischemic stroke. J Stroke Cerebrovasc Dis. 2020;29(10):105133. [FREE Full text] [CrossRef] [Medline]
- Geng L, Zuo L, Li H, Wang Y, Zhang Q, Ran J, et al. Association of platelet-to-neutrophil ratios with 1-year outcome and mortality in patients with acute ischemic stroke. Neurosci Lett. 2023;798:137016. [CrossRef] [Medline]
- Zhang S, Wang J, Pei L, Liu K, Gao Y, Fang H, et al. Interpretability analysis of one-year mortality prediction for stroke patients based on deep neural network. IEEE J Biomed Health Inform. 2022;26(4):1903-1910. [CrossRef] [Medline]
- Abujaber AA, Albalkhi I, Imam Y, Nashwan A, Akhtar N, Alkhawaldeh IM. Machine learning-based prognostication of mortality in stroke patients. Heliyon. 2024;10(7):e28869. [FREE Full text] [CrossRef] [Medline]
- Zhu E, Chen Z, Ai P, Wang J, Zhu M, Xu Z, et al. Analyzing and predicting the risk of death in stroke patients using machine learning. Front Neurol. 2023;14:1096153. [FREE Full text] [CrossRef] [Medline]
- Chang Y, Zhang L, Li Y, Wang D, Fang Q, Tang X. Derivation and validation of a new visceral adiposity index for predicting short-term mortality of patients with acute ischemic stroke in a chinese population. Brain Sci. 2023;13(2):297. [FREE Full text] [CrossRef] [Medline]
- Yan X, Xia P, Tong H, Lan C, Wang Q, Zhou Y, et al. Development and validation of a dynamic nomogram for predicting 3-month mortality in acute ischemic stroke patients with atrial fibrillation. Risk Manag Healthc Policy. 2024;17:145-158. [FREE Full text] [CrossRef] [Medline]
- Wang W, Otieno JA, Eriksson M, Wolfe CD, Curcin V, Bray BD. Developing and externally validating a machine learning risk prediction model for 30-day mortality after stroke using national stroke registers in the UK and Sweden. BMJ Open. 2023;13(11):e069811. [FREE Full text] [CrossRef] [Medline]
- Easton JF, Stephens CR, Angelova M. Risk factors and prediction of very short term versus short/intermediate term post-stroke mortality: a data mining approach. Comput Biol Med. 2014;54:199-210. [CrossRef] [Medline]
- Lea-Pereira MC, Amaya-Pascasio L, Martínez-Sánchez P, Rodríguez Salvador MDM, Galván-Espinosa J, Téllez-Ramírez L, et al. Predictive model and mortality risk score during admission for ischaemic stroke with conservative treatment. Int J Environ Res Public Health. 2022;19(6):3182. [FREE Full text] [CrossRef] [Medline]
- Süt N, Celik Y. Prediction of mortality in stroke patients using multilayer perceptron neural networks. Turk J Med Sci. 2011:886-893. [CrossRef]
- Gattringer T, Posekany A, Niederkorn K, Knoflach M, Poltrum B, Mutzenbach S, et al. Predicting early mortality of acute ischemic stroke. Stroke. 2019;50(2):349-356. [CrossRef]
- Yang Y, Tang L, Deng Y, Li X, Luo A, Zhang Z, et al. The predictive performance of artificial intelligence on the outcome of stroke: a systematic review and meta-analysis. Front Neurosci. 2023;17:1256592. [FREE Full text] [CrossRef] [Medline]
- Schwartz L, Anteby R, Klang E, Soffer S. Stroke mortality prediction using machine learning: systematic review. J Neurol Sci. 2023;444:120529. [CrossRef] [Medline]
- Wang W, Kiik M, Peek N, Curcin V, Marshall IJ, Rudd AG, et al. A systematic review of machine learning models for predicting outcomes of stroke with structured data. PLoS One. 2020;15(6):e0234722. [FREE Full text] [CrossRef] [Medline]
- Yu Y, Zhong Z, Yang W, Yu J, Li J, Guo X, et al. Neutrophil percentage-to-albumin ratio and risk of mortality in patients on peritoneal dialysis. J Inflamm Res. 2023;16:6271-6281. [FREE Full text] [CrossRef] [Medline]
- Chung J, Ahn J, Ryu JA. Beyond SOFA and APACHE II, novel risk stratification models using readily available biomarkers in critical care. Diagnostics (Basel). 2025;15(9):1122. [FREE Full text] [CrossRef] [Medline]
- Lin J, Frontera JA. Decompressive hemicraniectomy for large hemispheric strokes. Stroke. 2021;52(4):1500-1510. [FREE Full text] [CrossRef] [Medline]
- Maddox TM, Rumsfeld JS, Payne PRO. Questions for artificial intelligence in health care. JAMA. 2019;321(1):31-32. [CrossRef] [Medline]
- Lewis AE, Weiskopf N, Abrams ZB, Foraker R, Lai AM, Payne PRO, et al. Electronic health record data quality assessment and tools: a systematic review. J Am Med Inform Assoc. 2023;30(10):1730-1740. [FREE Full text] [CrossRef] [Medline]
- Christodoulou E, Ma J, Collins GS, Steyerberg EW, Verbakel JY, Van Calster B. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol. 2019;110:12-22. [CrossRef] [Medline]
- Oka S, Yamazaki T, Takefuji Y. Addressing bias in machine learning feature importance for food quality assessment. Food Chem. 2025;494:146171. [FREE Full text] [CrossRef] [Medline]
- Allgaier J, Mulansky L, Draelos RL, Pryss R. How does the model make predictions? A systematic literature review on the explainability power of machine learning in healthcare. Artif Intell Med. 2023;143:102616. [FREE Full text] [CrossRef] [Medline]
- Zabor EC, Reddy CA, Tendulkar RD, Patil S. Logistic regression in clinical studies. Int J Radiat Oncol Biol Phys. 2022;112(2):271-277. [CrossRef] [Medline]
- Yang J, Zeng S, Cui S, Zheng J, Wang H. Predictive modeling of acute respiratory distress syndrome using machine learning: systematic review and meta-analysis. J Med Internet Res. 2025;27:e66615. [FREE Full text] [CrossRef] [Medline]
- Bellazzi R, Zupan B. Predictive data mining in clinical medicine: current issues and guidelines. Int J Med Inform. 2008;77(2):81-97. [FREE Full text] [CrossRef] [Medline]
- La Cava WG, Lee PC, Ajmal I, Ding X, Solanki P, Cohen JB, et al. A flexible symbolic regression method for constructing interpretable clinical prediction models. NPJ Digit Med. 2023;6(1):107. [FREE Full text] [CrossRef] [Medline]
- Petch J, Di S, Nelson W. Opening the black box: the promise and limitations of explainable machine learning in cardiology. Can J Cardiol. 2022;38(2):204-213. [FREE Full text] [CrossRef] [Medline]
- Bergomi L, Nicora G, Orlowska MA, Podrecca C, Bellazzi R, Fregosi C, et al. Which explanations do clinicians prefer? A comparative evaluation of XAI understandability and actionability in predicting the need for hospitalization. BMC Med Inform Decis Mak. 2025;25(1):269. [FREE Full text] [CrossRef] [Medline]
- Nazim S, Alam MM, Rizvi SS, Mustapha JC, Hussain SS, Suud MM. Advancing malware imagery classification with explainable deep learning: a state-of-the-art approach using SHAP, LIME and grad-cam. PLoS One. 2025;20(5):e0318542. [FREE Full text] [CrossRef] [Medline]
- Kuo RYL, Harrison C, Curran TA, Jones B, Freethy A, Cussons D, et al. Artificial intelligence in fracture detection: a systematic review and meta-analysis. Radiology. 2022;304(1):50-62. [FREE Full text] [CrossRef] [Medline]
- Khokhar PB, Gravino C, Palomba F. Advances in artificial intelligence for diabetes prediction: insights from a systematic literature review. Artif Intell Med. 2025;164:103132. [FREE Full text] [CrossRef] [Medline]
- Ruan X, Fu S, Storlie CB, Mathis KL, Larson DW, Liu H. Real-time risk prediction of colorectal surgery-related post-surgical complications using GRU-D model. J Biomed Inform. 2022;135:104202. [FREE Full text] [CrossRef] [Medline]
- Bednarski BP, Singh AD, Zhang W, Jones WM, Naeim A, Ramezani R. Temporal convolutional networks and data rebalancing for clinical length of stay and mortality prediction. Sci Rep. 2022;12(1):21247. [FREE Full text] [CrossRef] [Medline]
- Shurrab S, Guerra-Manzanares A, Magid A, Piechowski-Jozwiak B, Atashzar SF, Shamout FE. Multimodal machine learning for stroke prognosis and diagnosis: a systematic review. IEEE J Biomed Health Inform. 2024;28(11):6958-6973. [CrossRef] [Medline]
- Armoundas AA, Narayan SM, Arnett DK, Spector-Bagdady K, Bennett DA, Celi LA, et al. Use of artificial intelligence in improving outcomes in heart disease: a scientific statement from the American Heart Association. Circulation. 2024;149(14):e1028-e1050. [FREE Full text] [CrossRef] [Medline]
- Mansournia MA, Etminan M, Danaei G, Kaufman JS, Collins G. Handling time varying confounding in observational research. BMJ. 2017;359:j4587. [CrossRef] [Medline]
- Braun RG, Edwardson MA, Aldridge CM, Cramer SC, Cruchaga C, Engelter ST, et al. Blood biospecimen recommendations for research on stroke outcomes and recovery. Stroke. 2025;56(10):e279-e290. [CrossRef] [Medline]
- Fisher LD, Lin DY. Time-dependent covariates in the cox proportional-hazards regression model. Annu Rev Public Health. 1999;20:145-157. [CrossRef] [Medline]
- Chen X, Wang L, W YYYEDCL. Mitigating data imbalance and noise: a divergence-based approach with enhanced sample selection. Neurocomputing. 2024;605:1-10. [CrossRef]
- Nagendran M, Chen Y, Lovejoy CA, Gordon AC, Komorowski M, Harvey H, et al. Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies. BMJ. 2020;368:m689. [FREE Full text] [CrossRef] [Medline]
- Steyerberg EW, Bleeker SE, Moll HA, Grobbee DE, Moons KG. Internal and external validation of predictive models: a simulation study of bias and precision in small samples. J Clin Epidemiol. 2003;56(5):441-447. [CrossRef] [Medline]
- Dubray-Vautrin A, Gravrand V, Marret G, Lamy C, Klijanienko J, Vacher S, et al. Internal validation strategy for high dimensional prognosis model: a simulation study and application to transcriptomic in head and neck tumors. Comput Struct Biotechnol J. 2025;27:3792-3802. [FREE Full text] [CrossRef] [Medline]
- Teale EA, Forster A, Munyombwe T, Young JB. A systematic review of case-mix adjustment models for stroke. Clin Rehabil. 2012;26(9):771-786. [CrossRef] [Medline]
- Peters U, Sherling HR, Chin-Yee B. Hasty generalizations and generics in medical research: a systematic review. PLoS One. 2024;19(7):e0306749. [FREE Full text] [CrossRef] [Medline]
- An D, Lim M, Lee S. Challenges for data quality in the clinical data life cycle: systematic review. J Med Internet Res. 2025;27:e60709. [FREE Full text] [CrossRef] [Medline]
- Sterne JAC, Sutton AJ, Ioannidis JPA, Terrin N, Jones DR, Lau J, et al. Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials. BMJ. 2011;343:d4002. [CrossRef] [Medline]
- Borenstein M, Higgins JPT. Meta-analysis and subgroups. Prev Sci. 2013;14(2):134-143. [CrossRef] [Medline]
- Labarère J, Renaud B, Fine MJ. How to derive and validate clinical prediction models for use in intensive care medicine. Intensive Care Med. 2014;40(4):513-527. [CrossRef] [Medline]
- Lulin Z, Yiranbon E, Asante Antwi H. Complementarity of clinician judgment and evidence based models in medical decision making: antecedents, prospects, and challenges. Biomed Res Int. 2016;2016:1425693. [FREE Full text] [CrossRef] [Medline]
- Jeong S, Shivakumar M, Jung S, Won H, Nho K, Huang H, et al. Addressing overfitting bias due to sample overlap in polygenic risk scoring. Alzheimers Dement. 2025;21(4):e70109. [CrossRef] [Medline]
Abbreviations
| AI: artificial intelligence |
| AIS: acute ischemic stroke |
| AUC: area under the curve |
| DL: deep learning |
| eICU-CRD: eICU Collaborative Research Database |
| HKSJ: Hartung-Knapp-Sidik-Jonkman |
| LR: logistic regression |
| MIMIC: Medical Information Mart for Intensive Care |
| ML: machine learning |
| mRS: modified Rankin Scale |
| NIHSS: National Institutes of Health Stroke Scale |
| NPAR: neutrophil-percentage-to-albumin ratio |
| PI: prediction interval |
| PNR: platelet-to-neutrophil ratio |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
| PRISMA-S: Preferred Reporting Items for Systematic Reviews and Meta-Analyses literature search extension |
| PROBAST: Prediction model Risk of Bias Assessment Tool |
| RF: random forest |
| RoB: risk of bias |
| SE: sensitivity |
| SHAP: Shapley additive explanations |
| SP: specificity |
| SSNAP: Sentinel Stroke National Audit Programme |
| XGBoost: extreme gradient boosting |
Edited by S Brini; submitted 09.Sep.2025; peer-reviewed by GI Han, JMI Arockiasamy; comments to author 31.Oct.2025; accepted 19.Feb.2026; published 02.Apr.2026.
Copyright©Yujie Chen, Zhujing Ou, Yiting Deng, Anling Luo, Xuzi Li, Yujia Yang, Qin Yang, Xintong Wu, Muke Zhou. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 02.Apr.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

