Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/68872, first published .
The Machine Learning Models in Major Cardiovascular Adverse Events Prediction Based on Coronary Computed Tomography Angiography: Systematic Review

The Machine Learning Models in Major Cardiovascular Adverse Events Prediction Based on Coronary Computed Tomography Angiography: Systematic Review

The Machine Learning Models in Major Cardiovascular Adverse Events Prediction Based on Coronary Computed Tomography Angiography: Systematic Review

Authors of this article:

Yuchen Ma1 Author Orcid Image ;   Mohan Li1 Author Orcid Image ;   Huiqun Wu1 Author Orcid Image

Review

Department of Medical Informatics, Medical School of Nantong University, Nantong, China

Corresponding Author:

Huiqun Wu, PhD

Department of Medical Informatics

Medical School of Nantong University

Chongchuan District

19 Qixiu Road

Nantong, 226001

China

Phone: 86 513 85051891

Fax:86 513 85051820

Email: wuhuiqun@ntu.edu.cn


Background: Coronary computed tomography angiography (CCTA) has emerged as the first-line noninvasive imaging test for patients at high risk of coronary artery disease (CAD). When combined with machine learning (ML), it provides more valid evidence in diagnosing major adverse cardiovascular events (MACEs). Radiomics provides informative multidimensional features that can help identify high-risk populations and can improve the diagnostic performance of CCTA. However, its role in predicting MACEs remains highly debated.

Objective: We evaluated the diagnostic value of ML models constructed using radiomic features extracted from CCTA in predicting MACEs, and compared the performance of different learning algorithms and models, thereby providing clinical recommendations for the diagnosis, treatment, and prognosis of MACEs.

Methods: We comprehensively searched 5 online databases, Cochrane Library, Web of Science, Elsevier, CNKI, and PubMed, up to September 10, 2024, for original studies that used ML models among patients who underwent CCTA to predict MACEs and reported clinical outcomes and endpoints related to it. Risk of bias in the ML models was assessed by the Prediction Model Risk of Bias Assessment Tool, while the radiomics quality score (RQS) was used to evaluate the methodological quality of the radiomics prediction model development and validation. We also followed the TRIPOD (Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis) guidelines to ensure transparency of ML models included. Meta-analysis was performed using Meta-DiSc software (version 1.4), which included the score and Cochran Q test, along with StataMP 17 (StataCorp) to assess heterogeneity and publication bias. Due to the high heterogeneity observed, subgroup analysis was conducted based on different model groups.

Results: Ten studies were included in the analysis, 5 (50%) of which differentiated between training and testing groups, where the training set collected 17 kinds of models and the testing set gathered 26 models. The pooled area under the receiver operating characteristic (AUROC) curve for ML models predicting MACEs was 0.7879 in the training set and 0.7981 in the testing set. Logistic regression (LR), the most commonly used algorithm, achieved an AUROC of 0.8229 in the testing group and 0.7983 in the training group. Non-LR models yielded AUROCs of 0.7390 in the testing set and 0.7648 in the training set, while the random forest (RF) models reached an AUROC of 0.8444 in the training group.

Conclusions: Study limitations included a limited number of studies, high heterogeneity, and the types of included studies. The performance of ML models for predicting MACEs was found to be superior to that of general models based on basic feature extraction and integration from CCTA. Specifically, LR-based ML diagnostic models demonstrated significant clinical potential, particularly when combined with clinical features, and are worth further validation through more clinical trials.

Trial Registration: PROSPERO CRD42024596364; https://www.crd.york.ac.uk/PROSPERO/view/CRD42024596364

J Med Internet Res 2025;27:e68872

doi:10.2196/68872

Keywords



Coronary computed tomography angiography (CCTA) has emerged as the first-line noninvasive imaging test for patients presenting with symptoms potentially attributable to coronary artery disease (CAD), owing to its high sensitivity and specificity in diagnosing this condition [1]. Supported by a growing body of evidence and advancements in cardiac imaging technology, CCTA has solidified its role as a primary investigative tool for individuals with suspected CAD. CCTA provides comprehensive information regarding the overall coronary circulation and luminal stenosis, while also enabling the assessment of the composition, morphology, and vulnerability of atherosclerotic plaques.

Radiomics uses advanced data characterization algorithms to extract extensive quantitative information that is often imperceptible to the naked eye [2]. These radiomic features can potentially unveil tumoral patterns and provide informative multidimensional characteristics. By using radiomics to identify individuals at risk, clinicians may better select patients who could benefit from more aggressive preventive measures aimed at improving clinical outcomes in those at higher risk of CAD [3]. Radiologic images contain far more information than can be interpreted visually or quantified through simple manual measurements. This process involves extracting thousands of imaging markers from radiologic images, which describe the heterogeneity and spatial complexity of lesions. Although quantitative radiomics analysis is feasible and has the potential to enhance the diagnostic performance of CCTA, its use in predicting major adverse cardiovascular events (MACEs) remains a subject of ongoing debate [4].

A large number of radiomic features have been extracted and processed by machine learning (ML), but their stability and robustness require further validation. ML goes through the process of data preprocessing, feature extraction, analysis and selection, and predictive modeling to ultimately predict clinical outcomes. The big difference between ML and deep learning (DL) is that the latter replaces the manual feature extraction of the former with the deep extraction of specific features using pretrained models automatically, which improves the performance of the model but cannot replace the former’s interpretation of clinical practice [5]. ML models based on traditional computed tomography (CT) have been shown to enhance the long-term prediction of MACEs. ML approaches leveraging radiomics help to integrate large datasets and optimize the use of CCTA data while minimizing interpretation bias due to interobserver variability. Furthermore, radiomics-based ML analysis improves the discriminatory power of CCTA in detecting advanced atherosclerotic lesions [6].

Traditionally, MACEs encompass cardiac death (including fatal myocardial infarction [MI]), nonfatal MI (both ST-segment elevation MI and non–ST-segment elevation MI), and unstable angina requiring coronary revascularization [7]. With an increasing number of studies investigating the adverse effects of coronary artery ischemia and stenosis [8], this meta-analysis chose to include ischemia and stenosis as key indicators for the diagnostic models based on ML. Although atherosclerosis and vascular morphology significantly influenced these factors [9], they were excluded from the analysis due to their indirect association with adverse cardiac outcomes. Here, we conducted a meta-analysis and systematic review to evaluate the prognostic and diagnostic value of ML and artificial intelligence models in conjunction with CCTA-derived radiomic features. Our objective is to compare the performance of various learning algorithms and ML models, thereby providing guiding recommendations for the clinical diagnosis, treatment, and prognosis of MACEs.


Database Search

This meta-analysis was conducted by the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines [10]. The checklist adhered to this study, followed by PRISMA, as shown in Table S1 in Multimedia Appendix 1. The protocol for this meta-analysis was registered with the International Prospective Register of Systematic Reviews (registration number CRD42024596364). The first submission was made on October 2, 2024, and the second revision was made on November 7, with no changes other than the title. The reason for the title change was to move the order of “ML model” forward and to emphasize one of the intervention methods, CCTA, to make the title more understandable and relevant to the full text. According to the PICO (population, intervention, control, and outcomes) framework, which are the key components of a clinical question outlined in Table S2 in Multimedia Appendix 2 [11-20] (P: Patients who underwent CCTA and were diagnosed with or predicted to develop a MACE; I: ML diagnostic or predictive models or learning algorithms; C: gold standards; O: MACE plus ischemia and stenosis), 2 independent researchers searched the Cochrane Library, Web of Science, Elsevier, CNKI, and PubMed for all available papers, with all databases searched until September 10, 2024. These researchers also recorded inclusion and exclusion reasons for these relevant papers according to the inclusion and exclusion criteria represented below, to avoid incorrect inclusion and exclusion. A third researcher resolved the dissent that occurred between the 2 independent researchers when selecting the included studies. Through discussion, keywords or phrases such as (“radiomic feature” OR “coronary plaque characteristics”) AND “CCTA[MeSH]” AND (“machine learning” OR “artificial intelligence”) AND “MACE[MeSH]” were used and searched. The full search strategies were presented in Multimedia Appendix 3. The inclusion and exclusion criteria of papers for this study can be seen in Textbox 1.

Textbox 1. Inclusion and exclusion criteria.

Inclusion criteria

  • Original research or studies containing original data.
  • Prospective observational trials, case-control studies, or cohort studies.
  • Research on extracting radiological features and evaluating the characteristics of coronary artery plaques using coronary computed tomography angiography (CCTA).
  • Studies reporting clinical outcomes and endpoints related to major adverse cardiovascular events (MACEs) during the follow-up period.
  • Studies involving registered populations that underwent ethical review or were granted an exemption from ethical review.

Exclusion criteria

  • Studies focusing on noncoronary arteries or those that did not provide data specifically related to coronary arteries.
  • Studies classified as nonoriginal research, including protocols, reviews, meta-analyses, case reports, editorials, and expert opinions.
  • Studies lacking sufficient statistical information regarding effect sizes.
  • Studies conducted on nonhuman participants.
  • Studies published in languages other than English.
  • Duplicated studies or those presenting redundant data.

Data Extraction

The titles and abstracts were screened, and information such as the names of the first (or cofirst) authors, year of publication, type of study, type of machine algorithm, sample size of patients receiving CCTA, MACE outcomes, and radiomic features was summarized by 2 independent researchers. It was necessary to extract information to help assess the risk of bias in the study design. Any disagreements in data extraction were resolved through consultation and referral to a third researcher.

Quality and Bias Assessments

The radiomics quality score (RQS) [21] was used to evaluate the radiomic quality of each included study to reduce bias and enhance model use in the context of radiomics. We used the Prediction Model Risk of Bias Assessment Tool (PROBAST) to assess the risk of bias and applicability concerns of each included study, using a total of 20 signaling questions in four domains—participants, predictors, outcome, and analysis [22]. TRIPOD (Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis) [23], a checklist of 22 items to validate prediction models, was used to provide more transparent details of the ML models included. Publication bias was minimized through manual searching and according to the PICO restriction. We used StataMP 17 to perform the publication bias assessment, and 2 independent researchers were involved in the quality and bias assessments. Disagreements and different judgments between reviewers were resolved by a third researcher.

Statistical Analysis

In this meta-analysis, the pooled area under the curve (AUC) of each ML model and the forest plots were generated. Receiver operating characteristics (ROC) is the common index to draw the plots of AUC, which means the area under the receiver operating characteristic (AUROC). The standard for its diagnostic or prognostic discrimination ability was set as none (AUROC≤0.6), poor (AUROC 0.6-0.7), fair (AUROC 0.7-0.8), good (AUROC 0.8-0.9), or optimum (AUROC 0.9-1) [24]. Model calibration is a metric of goodness of fit that assesses the agreement between observed and predicted outcomes and reflects the stability of the models via calibration plots. Heterogeneity tests are categorized into drawing forest plots and hypothesis testing. The common effect measures of the former are sensitivity, specificity, etc, and the latter is the main way to achieve better testing efficacy to reduce the false negative results compared with the G2 test. The Q-test calculated the I2 statistic, which was calculated by the formula presented in

Q is the chi-square value of the heterogeneity test (χ2), and (k-1) is the degree of freedom. Usually, I2≤25% indicates low heterogeneity, moderate heterogeneity means 25%<I2<50%, and I2≥50% suggests high heterogeneity, which means the random effects model would be of higher efficiency. With a cutoff of 0.05 or nontraditional 0.1, a P value less than .05 or .10 is considered to be of high heterogeneity, and greater than .05 or .10 is considered to be of low heterogeneity, which is usually judged in conjunction with the I2 value. The sample size can also have a corresponding effect on the results. For instance, if the P value is not significant (P>.05 or >.10) while the I2 value is higher than 50%, this may be due to the lack of test validity caused by the small sample size, but it does not seem to apply to our study. The high heterogeneity that occurred in our study was first chosen to be a random effects model and would be retrieved by subgroup analysis, sensitivity analysis, and meta-regression for searching heterogeneous sources and addressing potential high heterogeneity. The Meta-DiSc software program (version 1.4) [25] was used to calculate the pooled estimation of AUROC, sensitivity, specificity, positive likelihood ratio (PLR), negative likelihood ratio (NLR), and diagnostic odds ratio (DOR).

Ethical Considerations

Since our systematic review used publicly accessible documents as evidence and therefore is no need to seek institutional ethics approval.


Study Selection

A total of 48 studies were initially identified from the Cochrane Library, Web of Science, CNKI, Elsevier, and PubMed, supplemented by 13 additional studies identified through citation searching. After removing duplicates and records that failed to meet the inclusion criteria, 33 studies underwent full screening. About 5 studies were excluded due to the lack of relevant outcomes, 9 were considered irrelevant to coronary artery plaques, and 11 were removed for lacking radiomic features. Following the consideration of 2 additional eligible citations, 10 studies [11-20] were ultimately included in this meta-analysis. The selection process, including database outputs and the deduplication process, was illustrated in the PRISMA flow diagram (Figure 1).

Figure 1. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram for study screening and selection. CNKI: China National Knowledge Infrastructure.

Study Characteristics

The studies included in this meta-analysis were published between 2021 and 2024, with a total of 16 kinds of models identified in testing or validation groups, and 17 kinds of models found in training groups. Among the 10 studies, 5 out of 10 (50%) studies differentiated between training and testing groups with study flowcharts of inclusion and exclusion; one of these even included an additional validation group [14]. However, 1 study [18] lacked descriptions and explanations of validation or testing procedures, and thus was considered of high risk of bias [26], while 1 study [15] had 2 inner validation groups and the other had 2 external cohorts added with 1 internal cohort [19]. Most studies developed ML models at the patient level, with two exceptions [11,12]: one study [11] analyzed data at the segment level, and the second focused on the lesion level [12]. The names of the first (or cofirst) authors, year of publication, type of study, type of machine algorithm, sample size of patients included in the training group, the testing group or the validation group, MACE outcomes, and radiomic features were collected and presented in Table S3 in Multimedia Appendix 2 [11-20].

Various learning algorithms and their integrated versions were included in the analysis. Of the 10 studies, 5 used logistic regression (LR), 5 used a random forest (RF) algorithm, and 1 tested various ML algorithms to find the best one. Notably, Militello et al [18] developed L1-, tree-, and mutual information–based models, each representing different feature selection methods in 2023. Qin et al [11] implemented three algorithms: RF, LR, and artificial neural network (ANN). Overall, ML models accounted for the majority of those used, DL was applied to construct radiomic models from myocardial fibrosis features, and ANN had the best construction effect compared with RF and LR.

In the 2 studies that analyzed data at the nonpatient level, Li et al [12] in 2021 integrated 14 radiomic features and compared the RF model with a conventional LR model, achieving a sensitivity of 89.02%, a specificity of 64.91%, and an AUROC of 0.82 at the plaque level in the training set, with no statistically significant difference observed between the training set and the validation set (P=.58). Qin et al [11] achieved good diagnostic accuracy through the R-model (developed via multivariable LR extracting radiomic signatures) in both the training cohort (107 patients and 1712 segments) and the testing cohort (54 patients and 864 segments), achieving a sensitivity of 0.85, a specificity of 0.73, and an AUROC of 0.86.

Quality Assessment

We used the PROBAST checklist (Table S4 in Multimedia Appendix 2 [11-20]) to assess the risk of bias and applicability of the prognostic prediction model studies. According to the criteria, the bias of participants in 2 out of 10 (20%) studies was identified as having a high risk of bias, primarily because the models in the former study lacked any external validation and the latter one did not present the concrete process or outcomes of the training group even diagnostic value. Among the 4 domains, once 1 domain was associated with a high risk of bias, the study was classified as high risk due to these significant concerns. About 5 out of 10 (50%) studies were rated as unclear, as there was at least 1 domain rated as unclear risk of bias and without high risk. The main reason for most unclear risks of bias was the lack of explicit mention of terms such as the type of study or “cohort.” Overall, 4 (40%) studies raised unclear or high concerns regarding applicability, primarily due to excessive features in the models, complicating practical data collection. The remaining studies were rated as low risk in this respect.

Publication bias could be found in Figure S1 in Multimedia Appendix 4 [11-20]. The total of 17 models in the training set (P=.80) and 26 models in the testing or validation set (P=.63) showed there was no significant bias in this meta-analysis.

The scores of each item, along with the total score for evaluating radiomic quality, were presented in Figure S2 in Multimedia Appendix 4 [11-20]. The mean RQS across studies was 17.2 (47.78%), ranging from 12 (33.33%) to 23 (63.89%). The overall performance of the included studies was fair to good; however, there were notable deficiencies, such as a lack of phantom studies across all scanners, imaging conducted at multiple time points, failure to detect and discuss biological correlates, absence of prospective study registered in a trial database, and lack of cost-effectiveness analysis. These shortcomings were not unique to this meta-analysis, and 2 independent scores from the same rater also indicated above-average reliability [27]. Moreover, moderate to high adherence was observed for most TRIPOD checklist items (Figure S3 in Multimedia Appendix 4 [11-20]), except for seven domains with poor adherence: blind assessment of outcome, blind assessment of predictors, missing data, risk groups, validation, unadjusted association between each candidate predictor and outcome, and full prediction models.

DOR is typically used as the basis for weighting each diagnostic experiment when conducting combined tests. When the DOR value is greater than 1, it indicates a positive result in diagnostic performance. Conversely, a DOR value less than 1 indicates a negative diagnostic performance. In this meta-analysis, negative diagnosis results were not considered; thus, studies with a DOR value less than 1 were excluded based on publication bias (Figure S4 in Multimedia Appendix 4 [11-20]) in both the testing group and the training group.

Performance of ML Models for MACE Prediction

The overall pooled AUROC for ML models predicting MACEs was 0.7981 in the testing set (Figure 2) and 0.7879 in the training set (Figure 3). The original model evaluation metrics for each study were shown in Table S5 in Multimedia Appendix 2 [11-20]. The additional metrics were as follows.

In the training set, sensitivity was 0.71 (95% CI 0.68-0.74; P<.001; I2=68.3%); specificity was 0.73 (95% CI 0.70-0.75; P<.001; I2=88.1%); PLR was 2.57 (95% CI 2.10-3.14; P.<001; I2=76.1%); NLR was 0.39 (95% CI 0.33-0.47; P<.001; I2=71.1%); and DOR was 7.28 (95% CI 4.91-10.80; P<.001; I2=77.1%; Figure S5-S9 in Multimedia Appendix 4 [11-20]).

In the testing and validation set, sensitivity was 0.69 (95% CI 0.65-0.73; P<.001; I2=52.9%); specificity was 0.79 (95% CI 0.77-0.82; P<.001; I2=80.3%); PLR was 3.11 (95% CI 2.50-3.88; P<.001; I2=65.3%); NLR was 0.42 (95% CI 0.36-0.49; P=.02; I2=41.0%); and DOR was 8.71 (95% CI 6.22-12.20; P<.001; I2=47.5%; Figure S10-S14 in Multimedia Appendix 4 [11-20]).

Figure 2. The overall pooled area under the receiver operating characteristic (AUROC) curve of machine learning models for major adverse cardiovascular events (MACE) prediction in the testing set. AUC: area under the curve; Q*: the sensitivity at the intersection of the SROC curve and the straight line (sensitivity = specificity); SROC: summary receiver operating characteristic.
Figure 3. The overall pooled area under the receiver operating characteristic (AUROC) curve of machine learning models for major adverse cardiovascular events (MACE) prediction in the training set. AUC: area under the curve; Q*: the sensitivity at the intersection of the SROC curve and the straight line (sensitivity = specificity); SROC: summary receiver operating characteristic.

Sensitivity Analysis

Excluding the epicardial adipose tissue model (MEAT; AUROC 0.538, 95% CI 0.414-0.661) [13], which exhibited the lowest sensitivity and specificity with a DOR<1, the random effects meta-analysis in the training group yielded a better estimated pooled AUROC. To address the higher risk of heterogeneity, the MEAT model was also excluded from the testing group. Consequently, the pooled estimates were found to be insensitive to the exclusion of outliers.

Subgroup Analysis

Performance metrics for different ML models and MACE outcomes were examined in the subgroup analysis, in which pooled sensitivity, specificity, AUROC, DOR, PLR, and NLR were presented in Table 1. In this study, 5 included studies predicting MACEs were established using the algorithm of RF, 5 used the LR algorithm, while 3 of the 10 studies focused on ischemia during the follow-up, and 3 widely investigated MACE risks. For the different types of ML models, the LR subgroups achieved the AUROC of 0.8229 in the testing group and 0.7983 in the training group (Figure S15-S26 in Multimedia Appendix 4 [11-20]). The RF models reached an AUROC of 0.8444 in the training group (Figure S27-S32 in Multimedia Appendix 4 [11-20]), while non-LR models yielded AUROCs of 0.7390 in the testing set and 0.7648 in the training set (Figure S33-S44 in Multimedia Appendix 4 [11-20]). Moreover, the subgroup for specific MACE endpoints in the testing set produced an AUROC of 0.7343 (Figure S45-S50 in Multimedia Appendix 4 [11-20]). However, due to the limited data available from the included RF models, it was not possible to draw a direct conclusion about the predictive performance between LR and RF models in this analysis. Additionally, the tree-based model using radiomic features achieved the highest AUROC of 0.819 (SD 0.070), but when combined with clinical features using mutual information, the performance improved further to an AUROC of 0.820 (SD 0.076) [18].

Table 1. The pooled sensitivities, specificities, and area under the receiver operating characteristic for machine learning models predicting major adverse cardiovascular events in different subgroups.
Group and subgroupModel, NAUROCaSen (95% CI)bSpe (95% CI)cDOR (95% CI)dPLR (95% CI)eNLR (95% CI)f
Overall

Training160.78790.71 (0.68-0.74)0.73 (0.70-075)7.28 (4.91-10.80)2.57 (2.10-3.14)0.39 (0.33-0.47)

Testing or validation250.79810.69 (0.65-0.73)0.79 (0.77-0.82)8.71 (6.22-12.20)3.11 (2.50-3.88)0.42 (0.36-0.49)
LRg

Training70.79830.71 (0.67-0.75)0.81 (0.77-0.84)10.85 (4.58-25.74)3.54 (2.14-5.88)0.36 (0.25-0.52)

Testing or validation180.82290.68 (0.64-0.72)0.85 (0.82-0.87)11.09 (7.27-16.92)3.89 (2.92-5.20)0.41 (0.34-0.50)
NLRh

Training90.76480.71 (0.68-0.74)0.67 (0.64-0.70)5.45 (3.86-7.68)2.18 (1.86-2.55)0.42 (0.35-0.51)

Testing or validation60.73900.71 (0.63-0.78)0.65 (0.59-0.71)4.70 (2.96-7.44)2.03 (1.68-2.45)0.47 (0.36-0.61)
RFi

Training30.84440.77 (0.70-0.84)0.74 (0.67-0.79)10.92 (5.17-23.09)3.10 (1.92-5.01)0.30 (0.22-0.41)
MACEj

Testing or validation110.73430.61 (0.56-0.67)0.85 (0.82-0.88)8.34 (4.58-15.16)3.81 (2.36-6.14) 0.51 (0.43-0.60)

aAUROC: area under the receiver operating characteristic.

bSen: sensitivity.

cSpe: specificity.

dDOR: diagnostic odds ratio.

ePLR: positive likelihood ratio.

fNLR: negative likelihood ratio.

gLR: logistic regression.

hNLR: nonlogistic regression.

iRF: random forest.

jMACE: major adverse cardiovascular event.

Meta-Regression and Diagnostic Threshold

We performed meta-regression to address the high heterogeneity observed in subgroups. Diagnostic threshold was analyzed to address high heterogeneity and improve diagnostic accuracy, which showed no threshold effect in both the training set (Spearman correlation coefficient=–0.095; P=.73) and testing set (Spearman correlation coefficient=0.339; P=.10). The Spearman correlation coefficient was used to evaluate the correlation between certain factors occurred in this study and the calculated outcome. Generally, a coefficient greater than 0 indicated a positive correlation, while a coefficient less than 0 indicated a negative correlation. It was shown that learning algorithm, publication year, sample size, and type of ML models did not affect diagnostic accuracy (P>.05). The relative DOR and 95% CI were recorded in Table S6 in Multimedia Appendix 2 [11-20].

Key Results

  1. In this systematic review and meta-analysis of 10 studies, the pooled performance of ML models to diagnose MACEs had an AUROC of 0.7879 in the training set and 0.7981 in the testing set.
  2. The best diagnostic value was achieved by LR in the testing or validation set, with a pooled AUROC of 0.8229, and by RF in the training set, with an AUROC of 0.8444.
  3. The overall RQS assessment of the 10 included studies was moderate to good.

Principal Findings

Integrating ML models based on traditional CT imaging significantly enhances the long-term prediction of MACEs and improves risk stratification [28]. In many regions, the combination of CCTA and ML has made a substantial contribution to the diagnosis of MACEs. Overall, radiomics-based ML models have shown enhanced diagnostic capabilities over traditional CCTA features in assessing coronary plaque vulnerability [29] and in differentiating between culprit and nonculprit lesions [30]. The key advantages of these approaches include their noninvasive nature and high execution rates [31]. DL methods exhibited better predictive performance in the prediction of future culprit lesions compared with basic angiographic parameters [32]. Unfortunately, very few DL models were included in the discussion of our manuscript, mainly focused on the exploration of the performance of ML models to predict MACEs. It is hoped that in the future, more DL-related model studies that meet the inclusion criteria will be included to make up for this deficiency.

According to our meta-analysis, a total of 16 models from 7 studies were included in the training group. The LR method exhibited better diagnostic performance than non-LR (AUROC 0.7983 vs 0.7648), with both categories classified as fair to good. Among non-LR methods, the RF method achieved higher sensitivity and specificity, benefiting from the diversity of radiomic features [17], with the RF algorithm constructing a radiomic signature by combining multiple features into a single value.

While the LR model demonstrates robust diagnostic performance in our study, its inherent assumptions warrant critical discussion, particularly in the biomedical context. LR assumes linear separability of classes in the feature space and a monotonic relationship between predictors and log-odds, potentially oversimplifying complex biological systems. Highly correlated inputs (making LR sensitive to multicollinearity) can distort coefficient interpretation and reduce model stability [33]. Modern artificial intelligence approaches such as attention-based neural networks or kernelized models could potentially address these limitations by learning adaptive feature representations while maintaining interpretability through techniques such as saliency mapping or Shapley value analysis [34].

The RF model, despite its second ensemble advantages in our manuscript, faces specific risks in medical applications involving high-dimensional data. When processing multiomics datasets or high-resolution medical imaging features, RF’s feature sampling mechanism may inadvertently amplify spurious correlations between noncausal variables, particularly when strong feature interdependencies exist [35]. This could lead to overfitting of training-specific noise patterns, compromising generalizability to external cohorts—a critical concern for clinical deployment.

However, the support vector machine model demonstrated lower performance, which may vary due to differences in study design or sample selection. The support vector machine radiomics model, which maximized the margin between the training patterns and the decision boundary [36], performed well in predicting vulnerable plaques with AUC values of 0.977 for the training cohort [37]. Notably, the pericoronary adipose tissue (PCAT)–model showed better sensitivity (0.80, 95% CI 0.71-0.87) compared with models focused on the 3 major epicardial coronary arteries (LAD, LCX, and RCA), while among them the RCA-model performed better specificity (0.74, 95% CI 0.65-0.82). Beyond that, the combined use of learning algorithms has increasingly become a foundational approach in modeling. Militello et al [18] found that the optimal model was the RF enhanced with mutual information, achieving an AUROC of 0.820±0.076.

Wang et al [38] used various ML models based on lesion characteristics to predict the risk of acute coronary syndrome events, finding that the XGBoost (extreme gradient boosting) model outperformed traditional methods with an AUROC of 0.918 (95% CI 0.86-0.97). Additionally, the ANN demonstrated excellent performance in predicting myocardial ischemia, achieving an AUC value of 0.721 (95% CI 0.53-0.88), using 26 radiomics features computed from CTA [39]. The integration of PCAT radiomics features based on CCTA, clinical indices, and imaging parameters into column line drawings has been proposed as a promising tool for predicting MACEs [19,40]. For instance, patients with acute MI exhibited a distinct PCAT radiomic phenotype compared with those with stable or non-CAD [40]. Consequently, radiomics-based models outperformed PCAT attenuation-based models in accurately identifying patients with MI. Beyond PCAT, radiomic nomograms also demonstrated superior performance in predicting myocardial ischemia compared with stenosis and radiomics signature alone, as they effectively identified populations at higher risk for myocardial ischemia based on myocardial segments [41].

Artificial features extracted by radiomic methods need to be processed by advanced statistical techniques or ML algorithms to ensure feature robustness and reduce data dimensionality [21,42]. This is essential for providing a unified framework for feature extraction and classification, and various challenges have emerged in the process of extracting and selecting radiomic features. For instance, usually experienced operators are assigned to perform segmentation, and most studies do not consider the variability introduced by different physicians, algorithms, software, or variations in different breathing cycles. Importantly, the combination of clinical features and radiomic features has been proven to enhance diagnostic performance compared with relying solely on either type of feature.

Some studies preprocess variables with zero variance before feature selection. To ensure repeatability and minimize redundancy, metrics such as Spearman correlation coefficient, intraclass correlation coefficient, and near-zero variance were calculated. Features with a Spearman correlation coefficient greater than 0.8 or 0.9 were excluded. Standardization using z score standardization, hierarchical clustering, and techniques such as the least absolute shrinkage and selection operator and the maximum relevance minimum redundancy were subsequently applied to further refine the feature set. In our study, 2 studies constructed radiomic signatures to consolidate multiple features into a single value, and another 2 studies used rad-score to evaluate combined or individual models. Huang et al [19] used 4 supervised methods for irrelevancy reduction and 6 unsupervised methods for redundancy reduction. Militello et al [18] considered 3 feature selection methods: L1-, tree-, and mutual information–based.

Predictive or diagnostic ML models have used various plaque features, including minimum lumen area, luminal stenosis, percent atheroma volume, fibrous tissue, fibrotic fatty tissue, the Napkin-ring (NR) sign, necrotic core, calcium content, lesion length, plaque volume, proximal left anterior descending coronary artery lesion, and remodeling index [38,43-45]. Among these, the classification of noncalcified plaques is particularly critical in predicting cardiac events [46]. High-risk plaques (HRPs) identified by CCTA, as vital factors in identifying future risk of MACEs in the early period, are typically characterized by plaque burden ≥70%, minimum lumen area <4 mm², low-attenuation plaque, and positive remodeling, all of which have been independently associated with cardiovascular events [47]. Among those, low-attenuation plaque ≤60 Hounsfield unit and NR sign are the most powerful HRP-independent predictors for MACEs. In our manuscript, 1 of 10 adopted HRP into patients’ basic clinical data, corroborating evidence that HRP is an important factor in the pathogenesis of MACEs and has maximum use in the detection of CCTA. Li et al [12] in 2021 reported 4 vulnerable characteristics, including noncalcified plaque volume, NR sign, remodeling index, and spotty calcification, and claimed that compared with traditional plaque assessment, plaque radiomics analysis had greatly improved in identifying hemodynamically significant coronary stenosis.

Despite the radiomic features collected and discussed above, other important features should also be taken into consideration to predict MACEs. The index of microvascular resistance of the coronary artery, evaluating coronary microvascular dysfunction, can even lead to perioperative myocardial injury [48]. Nowadays, a novel and upgrading angiographically derived index of microcirculatory resistance was reported to be associated with better prediction of MACEs after percutaneous coronary intervention, with the best cutoff value of 25.17U to diagnose perioperative myocardial injury [49]. Fractional flow reserve (FFR), one of the hemodynamic parameters, especially using CCTA, namely FFRCT, played an important role in assessing and detecting stenosis of coronary arteries. Usually, an FFRCT<0.8 was considered a positive index, and a lower FFRCT score caused a higher risk for MACEs [50]. One study reported that measurement of CT-FFR 1 to 2 centimeters distal to the target lesion helped to avoid unnecessary injury while preserving diagnostic efficacy, and CT-FFR 1/2cm has been shown to provide better risk stratification for future adverse cardiac events [51]. It was also generally accepted that the clinical use and diagnostic results of FFRCT were independent of age [52]. The combined application of ML added a higher predictive value to FFRCT and provided a new research tool for future noninvasive diagnosis and treatment [53]. High-sensitivity C-reactive protein (hs-CRP), commonly representing the development of systemic inflammation, deeply interacted with the triglyceride glucose index, and its growing number suggested a higher risk of MACEs and coronary occlusion [54,55]. Kraaijenhof et al [56] in 2025 found that after 20 years of follow-up, the multivariable-adjusted hazard for hs-CRP predicting MACEs was 1.55 (95% CI 1.37-1.74). Both FFR and hs-CRP were considered as strong independent predictors of MACEs and CAD [57,58], and we hoped they could be included in our future follow-up studies. As a protective factor, low left ventricular ejection fraction was associated with high T1 (>1250 ms) in noninfarcted myocardium, which was an independent predictor of MACEs [59]. Patients with severely reduced left ventricular ejection fraction were recommended to undergo coronary artery bypass grafting rather than percutaneous coronary intervention to avoid higher mortality and MACE rates [60].

However, different voices are pointing toward radiomic features. Huang et al [42] designed a framework with a two-stream convolutional neural network feature extraction module to extract the radiomic features of CAD plaques. This study was conducted based on the limitations of radiomics methods in extracting features of PCAT and atherosclerotic plaques. The limitations were due to the limited artificially extracted features and the excessively complex statistical or ML methods for feature processing.

Class imbalance and imbalanced set ratio are worth noting in our manuscript. It has been reported that the precision-recall curve is more informative and powerful for imbalanced cases than ROC with more accurate classifier performance [61]. Though ROC plots are more widely used in binary classifiers, class imbalance may mislead the ROC plots to some degree. All 10 studies included in this study adopted the AUROC plots, without avoiding the bias caused by class imbalance. Moreover, another reason for model overfitting is the imbalance ratio between the training and testing groups. The most suitable ratio for LR models is expected to be 9:1, while high accuracy can still be achieved even when the testing set data is very small in RF models [62].

In addition to predictive performance, the use of ML algorithms plays a crucial role in their adoption and application in real-world scenarios. Many previously proposed models required users to write complex code, perform manual parameter tuning, or configure appropriate environments in specific programming languages, which could pose significant barriers for nonexpert users. Some algorithms even lacked user-friendly interfaces or did not offer software packages or online tools, further limiting accessibility. Therefore, the ML model is expected to develop into an independent software application or an internet-based web server, which will significantly enhance use by lowering the technical threshold for end users and thus become a practical tool [63]. Zhou et al [64] in 2023 established an intelligent diagnostic platform by using radiomics and a convolutional neural network–based DL algorithm, and it was proved to have high functionality and efficiency. This provides great guiding significance for the further development of our research, improving the prediction accuracy of MACEs by lowering the threshold for constructing a model combining radiomics and ML.

Limitations

Our study had several unavoidable limitations. First, despite comprehensive screening, we included only 10 original studies, which limited our ability to conduct various subgroup analyses on specific MACE endpoints and different plaque features. Second, high heterogeneity presented a significant constraint. Both the training and testing sets exhibited moderate to high heterogeneity, with the overall heterogeneity in the testing group being lower than that in the training group. Variations in sample sizes, execution settings, and selected ML models across different studies impacted the final results, particularly considering the differing ML algorithms used for diagnosis and the various feature selection methods used. Future research that incorporates more studies focusing on specific MACE endpoint subgroups, diverse ML algorithms, various plaque features, and extended follow-up periods may provide stronger evidence.

Our evaluation using the RQS indicated that the included studies demonstrated a high quality of scientific practice. Finally, we combined predictive and diagnostic ML models in our analysis because they shared a similar process for deriving conclusions about MACEs from radiomic features obtained via CCTA. It is important to note that the predictive models included in our analysis were retrospective studies, assessing outcomes from a more distant point in time to a more recent one, and the MACE endpoints they predicted also occurred in the past. Therefore, while these 2 types of models are fundamentally different, we chose to combine them to achieve a more comprehensive and complete result, supported by meta-regression analysis (P=.08 in the training group; P=.10 in the testing group).

Clinical Implications

Clinical studies found that the hybrid ML models outperformed most LR models and traditional CCTA scoring systems, especially RF. The key consideration was which features should be selected for the model’s variables, both clinical and anatomical, which would greatly affect its efficacy. By incorporating ML into clinical practice, health care professionals can significantly improve the risk stratification of patients with CAD. Second, the implementation of ML models will lead to more personalized clinical treatment strategies. In addition, the adoption of ML-based tools in diagnostic workflows can simplify clinical decision-making.

However, introducing ML models into clinical practice is not without challenges. It requires the integration of high-quality annotated datasets and must ensure that the models are interpretable to gain the trust of clinicians. Moreover, models must be validated in different patient populations and clinical environments to ensure that they will not be confounded by the complexities of real-world scenarios that can reduce their efficacy.

Conclusion

The application of ML models to coronary plaque features based on CCTA for the prediction of MACEs offers a promising avenue for improving cardiovascular care. The diagnostic value of ML models for predicting MACEs has been proven to be superior to general models based on basic feature extraction and integration from CCTA. In particular, LR-based ML diagnostic models hold significant clinical promise when combined with clinical features, and more clinical trial studies are needed for validation.

Acknowledgments

This work was supported by the Science and Technology Project of Nantong City (MS2023050), Engineering Research Center of Integration and Application of Digital Learning Technology, Ministry of Education (1311016), Jiangsu Students’ Platform for Innovation and Entrepreneurship Training Program (202410304163E and 202410304169T). This manuscript was written without the participation of artificial intelligence.

Data Availability

The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.

Authors' Contributions

WH supervised the study and took leadership responsibility for the research activity planning and execution. MY and LM were responsible for conceptualization, methodology, and data curation. MY wrote the original manuscript.

Conflicts of Interest

None declared.

Multimedia Appendix 1

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist.

DOCX File , 22 KB

Multimedia Appendix 2

Additional tables for this meta-analysis.

DOCX File , 49 KB

Multimedia Appendix 3

The full searching strategies.

DOC File , 35 KB

Multimedia Appendix 4

Additional figures for this meta-analysis.

DOCX File , 2034 KB

  1. Task Force Members, Montalescot G, Sechtem U, Achenbach S, Andreotti F, Arden C, et al. et al. 2013 ESC guidelines on the management of stable coronary artery disease: the task force on the management of stable coronary artery disease of the European Society of Cardiology. Eur Heart J. 2013;34(38):2949-3003. [FREE Full text] [CrossRef] [Medline]
  2. Kolossváry M, Karády J, Szilveszter B, Kitslaar P, Hoffmann U, Merkely B, et al. Radiomic features are superior to conventional quantitative computed tomographic metrics to identify coronary plaques with napkin-ring sign. Circ Cardiovasc Imaging. 2017;10(12):e006843. [FREE Full text] [CrossRef] [Medline]
  3. Lee SE, Hong Y, Hong J, Jung J, Sung JM, Andreini D, et al. Prediction of the development of new coronary atherosclerotic plaques with radiomics. J Cardiovasc Comput Tomogr. 2024;18(3):274-280. [CrossRef] [Medline]
  4. Hu W, Wu X, Dong D, Cui L, Jiang M, Zhang J, et al. Novel radiomics features from CCTA images for the functional evaluation of significant ischaemic lesions based on the coronary fractional flow reserve score. Int J Cardiovasc Imaging. 2020;36(10):2039-2050. [CrossRef] [Medline]
  5. Rundo L, Militello C. Image biomarkers and explainable AI: handcrafted features versus deep learned features. Eur Radiol Exp. 2024;8(1):130. [CrossRef] [Medline]
  6. Kolossváry M, Karády J, Kikuchi Y, Ivanov A, Schlett CL, Lu MT, et al. Radiomics versus visual and histogram-based assessment to identify atheromatous lesions at coronary CT angiography: an ex vivo study. Radiology. 2019;293(1):89-96. [FREE Full text] [CrossRef] [Medline]
  7. Task Force on the management of ST-segment elevation acute myocardial infarction, Van de Werf F, Bax J, Betriu A, Blomstrom-Lundqvist C, Crea F, et al. ESC guidelines on management of acute myocardial infarction in patients presenting with persistent ST-segment elevation. Rev Esp Cardiol. 2009;62(3):293, e1-293,247. [FREE Full text] [CrossRef] [Medline]
  8. Krievins DK, Zellans E, Latkovskis G, Kumsars I, Krievina AK, Jegere S, et al. Diagnosis and treatment of ischemia-producing coronary stenoses improves 5-year survival of patients undergoing major vascular surgery. J Vasc Surg. 2024;80(1):240-248. [FREE Full text] [CrossRef] [Medline]
  9. Nurmohamed NS, Danad I, Jukema RA, de Winter RW, de Groot RJ, Driessen RS, et al. CREDENCEPACIFIC-1 Investigators. Development and validation of a quantitative coronary CT angiography model for diagnosis of vessel-specific coronary ischemia. JACC Cardiovasc Imaging. 2024;17(8):894-906. [FREE Full text] [CrossRef] [Medline]
  10. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71. [FREE Full text] [CrossRef] [Medline]
  11. Qin L, Chen C, Gu S, Zhou M, Xu Z, Ge Y, et al. A radiomic approach to predict myocardial fibrosis on coronary CT angiography in hypertrophic cardiomyopathy. Int J Cardiol. 2021;337:113-118. [CrossRef] [Medline]
  12. Li L, Hu X, Tao X, Shi X, Zhou W, Hu H, et al. Radiomic features of plaques derived from coronary CT angiography to identify hemodynamically significant coronary stenosis, using invasive FFR as the reference standard. Eur J Radiol. 2021;140:109769. [CrossRef] [Medline]
  13. You H, Zhang R, Hu J, Sun Y, Li X, Hou J, et al. Performance of radiomics models based on coronary computed tomography angiography in predicting the risk of major adverse cardiovascular events within 3 years: a comparison between the pericoronary adipose tissue model and the epicardial adipose tissue model. Acad Radiol. 2023;30(3):390-401. [CrossRef] [Medline]
  14. Feng Y, Xu Z, Zhang L, Zhang Y, Xu H, Zhuang X, et al. Machine-learning-derived radiomics signature of pericoronary tissue in coronary CT angiography associates with functional ischemia. Front Physiol. 2022;13:980996. [FREE Full text] [CrossRef] [Medline]
  15. Wang Z, Zhang J, Zhang A, Sun Y, Su M, You H, Sen Li, et al. The role of epicardial and pericoronary adipose tissue radiomics in identifying patients with non-ST-segment elevation myocardial infarction from unstable angina. Heliyon. 2023;9(5):e15738. [FREE Full text] [CrossRef] [Medline]
  16. Jing M, Xi H, Sun J, Zhu H, Deng L, Han T, et al. Differentiation of acute coronary syndrome with radiomics of pericoronary adipose tissue. Br J Radiol. 2024;97(1156):850-858. [CrossRef] [Medline]
  17. Wang J, Xu J, Mao J, Fu S, Gu H, Wu N, et al. A novel hybrid machine learning model for auxiliary diagnosing myocardial ischemia. Front Cardiovasc Med. 2024;11:1327912. [FREE Full text] [CrossRef] [Medline]
  18. Militello C, Prinzi F, Sollami G, Rundo L, La Grutta L, Vitabile S. CT Radiomic Features and Clinical Biomarkers for Predicting Coronary Artery Disease. Cogn Comput. Feb 06, 2023;15(1):238-253. [CrossRef]
  19. Huang Z, Lam S, Lin Z, Zhou L, Pei L, Song A, et al. Predicting major adverse cardiac events using radiomics nomogram of pericoronary adipose tissue based on CCTA: a multi-center study. Med Phys. 2024;51(11):8348-8361. [CrossRef] [Medline]
  20. Zhang RR, You HR, Geng YY, Li XG, Sun Y, Hou J, et al. Predicting major adverse cardiovascular events within 3 years by optimization of radiomics model derived from pericoronary adipose tissue on coronary computed tomography angiography: a case-control study. BMC Med Imaging. 2024;24(1):117. [FREE Full text] [CrossRef] [Medline]
  21. Lambin P, Leijenaar RTH, Deist TM, Peerlings J, de Jong EE, van Timmeren J, et al. Radiomics: the bridge between medical imaging and personalized medicine. Nat Rev Clin Oncol. 2017;14(12):749-762. [CrossRef] [Medline]
  22. Moons KGM, Wolff RF, Riley RD, Whiting PF, Westwood M, Collins GS, et al. PROBAST: a tool to assess risk of bias and applicability of prediction model studies: explanation and elaboration. Ann Intern Med. 2019;170(1):W1-W33. [FREE Full text] [CrossRef] [Medline]
  23. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. 2015;350:g7594. [FREE Full text] [CrossRef] [Medline]
  24. Zhang Z, Yang L, Han W, Wu Y, Zhang L, Gao C, et al. Machine learning prediction models for gestational diabetes mellitus: meta-analysis. J Med Internet Res. 2022;24(3):e26634. [FREE Full text] [CrossRef] [Medline]
  25. Zamora J, Abraira V, Muriel A, Khan K, Coomarasamy A. Meta-DiSc: a software for meta-analysis of test accuracy data. BMC Med Res Methodol. 2006;6:31. [FREE Full text] [CrossRef] [Medline]
  26. Wolff RF, Moons KG, Riley RD, Whiting PF, Westwood M, Collins GS, et al. PROBAST Group†. PROBAST: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. 2019;170(1):51-58. [FREE Full text] [CrossRef] [Medline]
  27. Akinci D'Antonoli T, Cavallo AU, Vernuccio F, Stanzione A, Klontzas ME, Cannella R, et al. EuSoMII Radiomics Auditing Group. Reproducibility of radiomics quality score: an intra- and inter-rater reliability study. Eur Radiol. 2024;34(4):2791-2804. [FREE Full text] [CrossRef] [Medline]
  28. Tesche C, Bauer MJ, Baquet M, Hedels B, Straube F, Hartl S, et al. Improved long-term prognostic value of coronary CT angiography-derived plaque measures and clinical parameters on adverse cardiac outcome using machine learning. Eur Radiol. 2021;31(1):486-493. [CrossRef] [Medline]
  29. Li XN, Yin WH, Sun Y, Kang H, Luo J, Chen K, et al. Identification of pathology-confirmed vulnerable atherosclerotic lesions by coronary computed tomography angiography using radiomics analysis. Eur Radiol. 2022;32(6):4003-4013. [CrossRef] [Medline]
  30. Lin A, Kolossváry M, Cadet S, McElhinney P, Goeller M, Han D, et al. Radiomics-based precision phenotyping identifies unstable coronary plaques from computed tomography angiography. JACC Cardiovasc Imaging. 2022;15(5):859-871. [FREE Full text] [CrossRef] [Medline]
  31. Kim JN, Gomez-Perez L, Zimin VN, Makhlouf MHE, Al-Kindi S, Wilson DL, et al. Pericoronary adipose tissue radiomics from coronary computed tomography angiography identifies vulnerable plaques. Bioengineering (Basel). 2023;10(3):360. [FREE Full text] [CrossRef] [Medline]
  32. Mahendiran T, Thanou D, Senouf O, Meier D, Dayer N, Aminfar F, et al. Deep learning-based prediction of future myocardial infarction using invasive coronary angiography: a feasibility study. Open Heart. 2023;10(1):e002237. [FREE Full text] [CrossRef] [Medline]
  33. Ranganathan P, Pramesh CS, Aggarwal R. Common pitfalls in statistical analysis: logistic regression. Perspect Clin Res. 2017;8(3):148-151. [FREE Full text] [CrossRef] [Medline]
  34. Zhao J, Hou X, Pan M, Zhang H. Attention-based generative adversarial network in medical imaging: a narrative review. Comput Biol Med. 2022;149:105948. [CrossRef] [Medline]
  35. Nguyen TT, Huang JZ, Nguyen TT. Unbiased feature selection in learning random forests for high-dimensional data. ScientificWorldJournal. 2015;2015:471371. [FREE Full text] [CrossRef] [Medline]
  36. Boser BE, Guyon IM, Vapnik VN. A training algorithm for optimal margin classifiers. 1992. Presented at: COLT '92: Proceedings of the Fifth Annual Workshop on Computational Learning Theory; July 27-29, 1992:144-152; Pittsburgh, PA. [CrossRef]
  37. Zheng YL, Cai PY, Li J, Huang DH, Wang W, Li MM, et al. A novel radiomics-based technique for identifying vulnerable coronary plaques: a follow-up study. Coron Artery Dis. 2025;36(1):1-8. [CrossRef] [Medline]
  38. Wang Y, Chen H, Sun T, Li A, Wang S, Zhang J, et al. Risk predicting for acute coronary syndrome based on machine learning model with kinetic plaque features from serial coronary computed tomography angiography. Eur Heart J Cardiovasc Imaging. 2022;23(6):800-810. [CrossRef] [Medline]
  39. Yang YC, Wei XY, Tang XQ, Yin RH, Zhang M, Duan SF, et al. Exploring value of CT coronary imaging combined with machine-learning methods to predict myocardial ischemia. J Xray Sci Technol. 2022;30(4):767-776. [CrossRef] [Medline]
  40. Lin A, Kolossváry M, Yuvaraj J, Cadet S, McElhinney PA, Jiang C, et al. Myocardial infarction associates with a distinct pericoronary adipose tissue radiomic phenotype: a prospective case-control study. JACC Cardiovasc Imaging. 2020;13(11):2371-2383. [FREE Full text] [CrossRef] [Medline]
  41. Shu ZY, Cui SJ, Zhang YQ, Xu YY, Hung SC, Fu LP, et al. Predicting chronic myocardial ischemia using CCTA-based radiomics machine learning nomogram. J Nucl Cardiol. 2022;29(1):262-274. [CrossRef] [Medline]
  42. Huang Y, Yang J, Hou Y, Sun Q, Ma S, Feng C, et al. Automatic prediction of acute coronary syndrome based on pericoronary adipose tissue and atherosclerotic plaques. Comput Med Imaging Graph. 2023;108:102264. [CrossRef] [Medline]
  43. Yang S, Koo BK, Hoshino M, Lee JM, Murai T, Park J, et al. CT angiographic and plaque predictors of functionally significant coronary disease and outcome using machine learning. JACC Cardiovasc Imaging. 2021;14(3):629-641. [FREE Full text] [CrossRef] [Medline]
  44. Gu M, Mao Q, Wang H, Liang Y, Shen F, Cui H, et al. Coronary computed tomographic angiography-derived anatomic and hemodynamic plaque characteristics in prediction of cardiovascular events. Int J Cardiovasc Imaging. 2024;40(8):1641-1652. [CrossRef] [Medline]
  45. Kruk M, Wardziak Ł, Kolossvary M, Maurovich-Horvat P, Demkow M, Kępka C. Identification of noncalcified coronary plaque characteristics using machine learning radiomic analysis of non-contrast high-resolution computed tomography. Kardiol Pol. 2023;81(10):978-989. [FREE Full text] [CrossRef] [Medline]
  46. Yunus MM, Mohamed Yusof AK, Ab Rahman MZ, Koh XJ, Sabarudin A, Nohuddin PNE, et al. Automated classification of atherosclerotic radiomics features in coronary computed tomography angiography (CCTA). Diagnostics (Basel). 2022;12(7):1660. [FREE Full text] [CrossRef] [Medline]
  47. Yang S, Hoshino M, Koo B, Yonetsu T, Zhang J, Hwang D, et al. Relationship of plaque features at coronary CT to coronary hemodynamics and cardiovascular events. Radiology. 2022;305(3):578-587. [CrossRef] [Medline]
  48. Scarsini R, Shanmuganathan M, De Maria GL, Borlotti A, Kotronias RA, Burrage MK, et al. OxAMI Study Investigators. Coronary microvascular dysfunction assessed by pressure wire and CMR after STEMI predicts long-term outcomes. JACC Cardiovasc Imaging. 2021;14(10):1948-1959. [FREE Full text] [CrossRef] [Medline]
  49. Zhang B, Zhang Y, Zhang K, Hu K, Shi Z, Ma L. The coronary angiography-derived index of microcirculatory resistance predicts perioperative myocardial injury in stable coronary artery disease patients undergoing PCI. Heliyon. 2024;10(15):e35240. [FREE Full text] [CrossRef] [Medline]
  50. Wada S, Iwanaga Y, Nakai M, Nakao YM, Miyamoto Y, Noguchi T, et al. NADESICO Study Investigators. Combination of coronary CT angiography, FFRCT, and risk factors in the prediction of major adverse cardiovascular events in patients suspected CAD. Clin Cardiol. 2023;46(5):494-501. [FREE Full text] [CrossRef] [Medline]
  51. Nozaki YO, Fujimoto S, Kawaguchi YO, Aoshima C, Kamo Y, Sato H, et al. Prognostic value of the optimal measurement location of on-site CT-derived fractional flow reserve. J Cardiol. 2022;80(1):14-21. [FREE Full text] [CrossRef] [Medline]
  52. Anastasius M, Maggiore P, Huang A, Blanke P, Patel MR, Nørgaard BL, et al. The clinical utility of FFRCT stratified by age. J Cardiovasc Comput Tomogr. 2021;15(2):121-128. [CrossRef] [Medline]
  53. Qiao HY, Tang CX, Schoepf UJ, Tesche C, Bayer RR, Giovagnoli DA, et al. Impact of machine learning-based coronary computed tomography angiography fractional flow reserve on treatment decisions and clinical outcomes in patients with suspected coronary artery disease. Eur Radiol. 2020;30(11):5841-5851. [CrossRef] [Medline]
  54. Ma X, Chu H, Sun Y, Cheng Y, Zhang D, Zhou Y, et al. The effect of hsCRP on TyG index-associated cardiovascular risk in patients with acute coronary syndrome undergoing PCI. Sci Rep. 2024;14(1):18083. [FREE Full text] [CrossRef] [Medline]
  55. Bruno F, Adjibodou B, Obeid S, Kraler SC, Wenzl FA, Akhtar MM, et al. Occlusion of the infarct-related coronary artery presenting as acute coronary syndrome with and without ST-elevation: impact of inflammation and outcomes in a real-world prospective cohort. Eur Heart J Qual Care Clin Outcomes. 2023;9(6):564-574. [CrossRef] [Medline]
  56. Kraaijenhof JM, Nurmohamed NS, Nordestgaard AT, Reeskamp LF, Stroes ESG, Hovingh GK, et al. Low-density lipoprotein cholesterol, C-reactive protein, and lipoprotein(a) universal one-time screening in primary prevention: the EPIC-Norfolk study. Eur Heart J. 2025:ehaf209. [CrossRef] [Medline]
  57. Gabara L, Hinton J, Kira M, Saunders A, Shambrook J, Abbas A, et al. Derivation and validation of a novel functional FFR score incorporating the burden of coronary stenosis severity and flow impairment to predict clinical events. J Cardiovasc Comput Tomogr. 2024;18(1):33-42. [FREE Full text] [CrossRef] [Medline]
  58. Li Q, Song Y, Zhang Z, Xu J, Liu Z, Tang X, et al. The combined effect of triglyceride-glucose index and high-sensitivity C-reactive protein on cardiovascular outcomes in patients with chronic coronary syndrome: a multicenter cohort study. J Diabetes. 2024;16(8):e13589. [CrossRef] [Medline]
  59. Shanmuganathan M, Masi A, Burrage MK, Kotronias RA, Borlotti A, Scarsini R, et al. OxAMI Study Investigators. Acute response in the noninfarcted myocardium predicts long-term major adverse cardiac events after STEMI. JACC Cardiovasc Imaging. 2023;16(1):46-59. [FREE Full text] [CrossRef] [Medline]
  60. Sun LY, Gaudino M, Chen RJ, Bader Eddeen A, Ruel M. Long-term outcomes in patients with severely reduced left ventricular ejection fraction undergoing percutaneous coronary intervention vs coronary artery bypass grafting. JAMA Cardiol. 2020;5(6):631-641. [FREE Full text] [CrossRef] [Medline]
  61. Saito T, Rehmsmeier M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS One. 2015;10(3):e0118432. [FREE Full text] [CrossRef] [Medline]
  62. Sivakumar M, Parthasarathy S, Padmapriya T. Trade-off between training and testing ratio in machine learning for medical image processing. PeerJ Comput Sci. 2024;10:e2245. [CrossRef] [Medline]
  63. Park H, Nam Y, Kim JH, Choo J. HyperTendril: visual analytics for user-driven hyperparameter optimization of deep neural networks. IEEE Trans Vis Comput Graph. 2021;27(2):1407-1416. [CrossRef] [Medline]
  64. Zhou Z, Qian X, Hu J, Chen G, Zhang C, Zhu J, et al. An artificial intelligence-assisted diagnosis modeling software (AIMS) platform based on medical images and machine learning: a development and validation study. Quant Imaging Med Surg. 2023;13(11):7504-7522. [FREE Full text] [CrossRef] [Medline]


ANN: artificial neural network
AUC: area under the curve
AUROC: area under the receiver operating characteristic
CAD: coronary artery disease
CCTA: coronary computed tomography angiography
CT: computed tomography
DL: deep learning
DOR: diagnostic odds ratio
FFR: fractional flow reserve
HRP: high-risk plaque
hs-CRP: high-sensitivity C-reactive protein
LR: logistic regression
MACE: major adverse cardiovascular event
MEAT: epicardial adipose tissue model
MI: myocardial infarction
ML: machine learning
NLR: negative likelihood ratio
NR: napkin-ring
PCAT: peri-coronary adipose tissue
PICO: population, intervention, control, and outcomes
PLR: positive likelihood ratio
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
PROBAST: Prediction model Risk Of Bias Assessment Tool
RF: random forest
ROC: receiver operating characteristic
RQS: radiomics quality score
TRIPOD: Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis
XGBoost: extreme gradient boosting


Edited by J Sarvestan; submitted 16.11.24; peer-reviewed by Z Feng, C Militello; comments to author 07.04.25; revised version received 27.04.25; accepted 28.04.25; published 13.06.25.

Copyright

©Yuchen Ma, Mohan Li, Huiqun Wu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 13.06.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.