Abstract
Background: Machine learning (ML) models may offer greater clinical utility than conventional risk scores, such as the Thrombolysis in Myocardial Infarction (TIMI) and Global Registry of Acute Coronary Events (GRACE) risk scores. However, there is a lack of knowledge on whether ML or traditional models are better at predicting the risk of major adverse cardiovascular and cerebrovascular events (MACCEs) in patients with acute myocardial infarction (AMI) who have undergone percutaneous coronary interventions (PCI).
Objective: The aim of this study is to systematically review and critically appraise studies comparing the performance of ML models and conventional risk scores for predicting MACCEs in patients with AMI who have undergone PCI.
Methods: Nine academic and electronic databases including PubMed, CINAHL, Embase, Web of Science, Scopus, ACM, IEEE, Cochrane, and Google Scholar were systematically searched from January 1, 2010, to December 31, 2024. We included studies of patients with AMI who underwent PCI, and predicted MACCE risk using ML algorithms or conventional risk scores. We excluded conference abstracts, gray literature, reviews, case reports, editorials, qualitative studies, secondary data analyses, and non-English publications. Our systematic search yielded 10 retrospective studies, with a total sample size of 89,702 individuals. Three validation tools were used to assess the validity of the published prediction models. Most included studies were assessed as having a low overall risk of bias.
Results: The most frequently used ML algorithms were random forest (n=8) and logistic regression (n=6), while the most used conventional risk scores were GRACE (n=8) and TIMI (n=4). The most common MACCEs component was 1-year mortality (n=3), followed by 30-day mortality (n=2) and in-hospital mortality (n=2). Our meta-analysis demonstrated that ML-based models (area under the receiver operating characteristic curve: 0.88, 95% CI 0.86‐0.90; I²=97.8%; P<.001) outperformed conventional risk scores (area under the receiver operating characteristic curve: 0.79, 95% CI 0.75‐0.84; I²=99.6%; P<.001) in predicting mortality risk among patients with AMI who underwent PCI. Heterogeneity across studies was high. Publication bias was assessed using a funnel plot. The top-ranked predictors of mortality in both ML and conventional risk scores were age, systolic blood pressure, and Killip class.
Conclusions: This review demonstrated that ML-based models had superior discriminatory performance compared to conventional risk scores for predicting MACCEs in patients with AMI who had undergone PCI. The most commonly used predictors were confined to nonmodifiable clinical characteristics. Therefore, health care professionals should understand the advantages and limitations of ML algorithms and conventional risk scores before applying them in clinical practice. We highlight the importance of incorporating modifiable factors—including psychosocial and behavioral variables—into prediction models for MACCEs following PCI in patients with AMI. In addition, further multicenter prospective studies with external validation are required to address validation limitations.
Trial Registration: PROSPERO CRD42024557418; https://www.crd.york.ac.uk/PROSPERO/view/CRD42024557418
doi:10.2196/76215
Keywords
Introduction
Acute myocardial infarction (AMI) is associated with a greater risk of mortality due to sustained ischemia and necrosis of the myocardium, requiring aggressive treatment [
]. Percutaneous coronary intervention (PCI) can promptly reopen infarcted blood vessels and restore reperfusion in AMI patients [ ]. However, the risk of restenosis within 1 year after successful PCI remains high at around 10%‐20% [ ], potentially resulting in adverse health outcomes and increasing health care costs [ ]. Notably, major adverse cardiovascular and cerebrovascular events (MACCEs) are an important issue in managing patients with AMI undergoing PCI [ ]. MACCEs encompass composite outcomes such as cardiovascular-related death, hospitalization for unstable angina or heart failure, recurrent myocardial infarction, stroke, and coronary revascularization, including PCI and coronary artery bypass grafting [ ]. Therefore, promptly assessing high-risk patients and reducing the risk of MACCEs is crucial for quality of care.Conventional risk scores (CRS), such as the Thrombolysis in Myocardial Infarction (TIMI) and Global Registry of Acute Coronary Events (GRACE) scores, are commonly used for predicting MACCEs due to their long-established reliability and ease of application in patients with cardiovascular diseases [
]. However, these conventional scores have limitations, as they cannot capture the complex interplay of patient-specific characteristics, particularly nonlinear correlations within datasets [ , ].Machine learning (ML)–based prediction models have recently gained traction, offering the ability to detect subtle patterns that conventional GRACE and TIMI scores, which rely on logistic regression, may overlook in patients with AMI [
, ]. However, achieving higher accuracy with ML-based models requires large datasets and has the disadvantage of limited ability to clarify causal relationships between variables [ ]. Therefore, comparing the performance of ML-based models with CRS is essential to elucidate their similarities and differences in risk prediction [ ].Several systematic reviews have examined ML-based models for predicting adverse cardiac events after PCI [
, ], MACCEs in acute coronary syndrome, or outcomes in older adults who have undergone PCI [ ]. These reviews focused on ML and conventional statistical models for predicting negative health outcomes in individuals with coronary artery disease, both with and without PCI. Only 2 reviews have compared the performance of ML-based models with conventional statistical models for predicting readmission or mortality after MI [ , ]. However, these reviews did not focus on predicting MACCEs after PCI in AMI patients.This study aimed to appraise studies comparing the performance of ML-based models with CRS in predicting MACCEs after PCI in AMI patients. Furthermore, this study identifies common risk factors for mortality after PCI using both ML and CRS, which may aid in developing clinical guidelines and improving personalized risk assessment.
Methods
Study Design
This systematic review adhered to the Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies (CHARMS) [
] and the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) [ ] guidelines. The PRISMA checklist is available in . This review protocol was registered in the International Prospective Register of Systematic Reviews (PROSPERO) under the identifier CRD42024557418.Data Sources and Search Strategy
A comprehensive search was performed in multiple databases (PubMed, CINAHL, Embase, Web of Science, Scopus, ACM, IEEE, Cochrane, and Google Scholar), focusing on literature published between January 1, 2010, and December 31, 2024. The search strategy incorporated the use of both Medical Subject Headings (MeSH) and free-text terms related to acute coronary syndrome, AMI, PCI, MACCEs, machine learning, mortality, readmission, and prediction models. Detailed information on the search terms is shown in
.Study Selection and Data Extraction
EndNote 20 (Clarivate Plc) and Microsoft Excel 2019 (Microsoft Corp) were used to retrieve the full text of all retrieved peer-reviewed articles, remove duplicates, and manage the screening process. After applying these keyword-based filters and database classifications, 75,122 records were exported to EndNote 20 for further review. Duplicate references were then removed using both manual review and automated tools via EndNote 20 and Microsoft Excel 2019. Manual title and abstract screening (n=44,559) were then conducted by 2 independent reviewers (M-YY and GIH) based on predefined eligibility criteria. Two independent reviewers (M-YY and GIH) evaluated all potentially relevant full texts (n=444). We resolved discrepancies between reviewers through discussion or by involving a third reviewer (Y-JS) until a consensus was reached. Two authors (M-YY and GIH) performed data extraction, which was subsequently verified by a third author (Y-JS).
PICO (participant, intervention, comparison, and outcomes) was used to build eligibility criteria (
). We included studies that met the following criteria: (1) involved adult patients (18 y or older) diagnosed with AMI, including ST-segment elevation myocardial infarction and non–ST-segment elevation myocardial infarction; (2) included patients who underwent PCI; and (3) predicted MACCEs risk using ML algorithms and CRS with statistical methods. We excluded studies focused on confirming PCI procedure outcomes and studies that solely compared ML model performance. In addition, to ensure the consistent and accurate interpretation of machine learning terminology and research methodologies, while minimizing potential translation bias, the search was restricted to English-language publications. The publication period was restricted to studies published between 2010 and 2024 to capture the full range of research from the early emergence of clinical machine learning predictive models to the most recent advancements. To ensure consistent study selection and reduce potential bias, inclusion and exclusion criteria were applied systematically for study designs and publication types.Item | Inclusion criteria | Exclusion criteria |
Participants |
|
|
Intervention |
|
|
Comparison |
|
|
Outcomes |
|
|
Language |
|
|
Publication period |
|
|
Study designs |
|
|
Publication types |
|
|
aAMI: acute myocardial infarction.
bSTEMI: ST-segment elevation myocardial infarction.
cNSTEMI: non–ST-segment elevation myocardial infarction.
dMACCEs: major adverse cardiovascular and cerebrovascular events.
eML: machine learning.
fMACE: major adverse cardiovascular event.
Quality Appraisals and Risk of Bias Assessment
In our review, the completeness and reliability of the analyzed literature were assessed using 3 tools: the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis + AI (TRIPOD + AI) [
], the Prediction Model Risk of Bias Assessment Tool (PROBAST) [ ], and the CHARMS [ ].TRIPOD + AI is a checklist that outlines essential items for effectively reporting studies that develop or evaluate prediction models using ML or statistical approaches. The checklist consists of 27 primary items covering various sections: title, abstract, introduction, methods, open science practices, patient and public involvement, results, and discussion [
]. Compliance with the TRIPOD + AI checklist was evaluated by scoring each item 1 point if reported and 0 points if not reported. The risk of bias and clinical applicability of each included study was evaluated using the PROBAST, which comprises 4 domains and 20 signal questions to evaluate bias and applicability [ ]. Lastly, the CHARMS was also adopted [ ].Data Synthesis and Analysis
Data extraction was done by 3 reviewers independently. It included details such as the first author’s name, year of publication, country, sample size, study design, study characteristics (eg, ML algorithm and type of CRS), outcome measurement, predictors, and their performance. This review reported the area under the receiver operating characteristic curve (AUROC) as a measure of predictive accuracy and ML performance [
]. The AUROC values were reported with 95% CIs.Low, moderate, and high heterogeneity were considered when the value of the I² statistic was 25%, 50%, and 75%, respectively [
]. A random-effects model was used for the analysis to address heterogeneity in the pooled results and appropriately integrate effect sizes, using the inverse variance method for pooling [ ]. To assess publication bias, a funnel plot was generated for visual inspection of asymmetry. Egger’s regression test [ ] and Begg’s test [ ] were conducted to further evaluate the presence of potential bias. Additionally, the trim-and-fill method [ ] was applied to adjust for any bias detected by the funnel plot. Statistical analyses were performed using the “metagen” function in the “meta” package of R 4.2.2 (R Foundation for Statistical Computing) [ , ].Ethical Considerations
The institutional review board of Chung-Ang University (number 1041078‐20240611-HR-144) approved the study protocol.
Results
Characteristics of the Included Studies
We identified 75,122 records through a comprehensive search strategy, as presented in the PRISMA 2020 flowchart (
). After removing 30,547 duplicate records, 44,575 records remained for screening. Of these, 16 were excluded as they had been retracted due to issues such as ethical concerns or compromised data integrity. Subsequently, 44,559 records were screened by title and abstract, resulting in the selection of 444 full-text articles for further evaluation. Following full-text screening, an additional 393 articles were excluded for the following reasons: dissertations and theses (n=38); case reports (n=51); reviews or meta-analyses (n=42); full text not available (n=2); not related to machine learning (n=145); not focused on predicting MACCEs (n=62); and irrelevant to the study objective (n=53).
In addition, conference abstracts (n=285) corresponding to gray literature were classified through title and abstract verification, while dissertations and theses (n=38) were distinguished through full-text article verification. Conference abstracts and dissertations/theses were considered gray literature. Conference abstracts were excluded due to the lack of full methodological information and peer review, while dissertations were excluded depending on full-text availability and the ability to assess methodological quality.
Ultimately, 10 articles [
- ] comprising 89,702 samples were included in this review ( ). All articles were published between 2017 [ ] and 2024 [ , ]. Seven studies originated from Asian countries: 3 from Korea [ - ], 3 from China [ , , ], and 1 from Malaysia [ ]. All studies were retrospective, with sample sizes for model development ranging from 466 [ ] to 22,875 [ ]. Six studies used data from multicenter or national registries.First author | Year | Country | Data source | Study design | Sample size (n) | ML | algorithmsConventional risk score | Outcome variables | Main findings |
Shouval et al [ | ]2017 | Israel | 2006‐2013, multicenter registry | Retrospective | 2782 | RF | , LR , NB , ADT , PART , AdaBoostTIMI GRACE | and30-day mortality | RF, NB, and AdaBoost models showed similar performance to GRACE, and all models outperformed TIMI. |
Kim et al [ | ]2022 | Korea | 2005‐2008, multicenter registry | Retrospective | 10,813 | DNN | , GBM , GLMGRACE | 1-year MACCEs | The ML model outperformed GRACE with an accuracy of over 95%. |
Kwon et al [ | ]2019 | Korea | 2008‐2013, multicenter registry | Retrospective | 22,875 | RF, LR, DNN | TIMI, GRACE, and ACTION | 1-year mortality | The ML model was 30.9% more accurate than GRACE. |
Sherazi et al [ | ]2020 | Korea | 2005‐2008, multicenter registry | Retrospective | 8227 | RF, GBM, GLM, DNN | GRACE | 1-year mortality | The performance of the ML model improved by an average of 0.08 compared to GRACE. |
Aziz et al [ | ]2021 | Malaysia | 2006‐2016, national registry | Retrospective | 12,368 | RF, LR, SVM | TIMI | In-hospital mortality | The ML model outperformed TIMI. |
Bai et al [ | ]2021 | China | 2016‐2020, EHR from single hospital | Retrospective | 656 | RF, LR, KNN | , CatBoost, XGBoostGRACE | 1-year mortality | CatBoost model outperformed GRACE. |
Hadanny et al [ | ]2021 | Israel, England, and Wales | 2006‐2016, multicenter registry | Retrospective | 25,475 | RF | GRACE | 30-day mortality | The RF model showed higher discriminatory power than GRACE. |
Fang et al [ | ]2022 | China | 2018‐2022, EHR from single hospital | Retrospective | 466 | LR | TIMI | In-hospital MACCEs | The LR model predicted higher performance than TIMI. |
Liu et al [ | ]2024 | China | 2014‐2022, EHR from single hospital | Retrospective | 1363 | RF, LR, DT , SVM, XGBoost, AdaBoost | GRACE, KAMIR and ACEF | ,1-year readmission | The LR model showed the best performance. |
Shakhgeldyan et al [ | ]2024 | Russia | 2015‐2021, EHR from single hospital | Retrospective | 4677 | RF, MLR | , SGBGRACE | In-hospital mortality | The model built with important variables showed the highest accuracy. |
aML: machine learning.
bRF: random forest.
cLR: logistic regression.
dNB: naïve Bayes.
eADT: alternating decision trees.
fPART: pruned rules from classification trees.
gAdaBoost: adaptive boosting.
hTIMI: thrombolysis in myocardial infarction.
iGRACE: Global Registry of Acute Cardiac Events.
jDNN: deep neural network.
kGBM: gradient boosting machine.
lGLM: generalized linear model.
mMACCEs: major cardiovascular and cerebrovascular adverse events.
nACTION: Acute Coronary Treatment and Intervention Outcomes Network scores.
oSVM: support vector machine.
pEHR: electronic health record.
qKNN: k-nearest neighbor.
rXGBoost: extreme gradient boosting.
sDT: decision tree.
tKAMIR: Korea Acute Myocardial Infarction Registry.
uACEF: age, creatinine, and ejection fraction score.
vMLR: multinomial logistic regression.
wSGB: stochastic gradient boosting.
The most frequently used ML-based models were random forest (RF; n=8) and logistic regression (LR; n=6), while the most used CRS were GRACE (n=8) and TIMI (n=4). The most common MACCEs components were 1-year mortality (n=3), followed by 30-day mortality (n=2) and in-hospital mortality (n=2). All studies included in this review reported that ML-based models outperformed CRS, including GRACE and TIMI, demonstrating higher accuracy and discriminatory power.
Risk of Bias and Applicability of Studies
We critically appraised the 10 studies included in the review (
). The overall quality of the studies, as assessed using the TRIPOD + AI guidelines, showed that only 4 studies met more than 70% of the TRIPOD + AI criteria. Open science practices, which are included as key components of the TRIPOD + AI guidelines, are vital to health care prediction model research as they foster transparency, reproducibility, and interdisciplinary collaboration [ ]. However, only 2 studies [ , ] provided accessible and verifiable protocols.The PROBAST checklist was used to evaluate the quality of the included studies, specifically assessing the risk of bias and applicability concerns. According to the PROBAST checklist, the overall risk of bias was low in most studies included in this review. However, potential bias was identified in the outcome definition [
], and low applicability was observed due to unclear reporting and methodological concerns [ - ]. In the participant domain, all studies were rated as having a low risk of bias. In contrast, in the applicability domain, 2 studies [ , ] were identified as of high concern due to the inclusion of narrowly selected populations, which may not be representative of the intended target population.Based on the CHARMS, 2 studies exhibited a high risk of bias in at least 2 major categories [
, ]. Although all the studies demonstrated low risk of bias in most domains—including data source, participant selection, and outcome—attrition-related bias was common, with 7 studies failing to report participant losses or methods for handling missing data.Model Performance of Risk Prediction for MACCEs
and present the comparative performance of ML-based models and CRS in terms of accuracy and validation methods, respectively. If a study applied multiple ML-based models on the same dataset, the model with the highest predictive performance—based on accuracy or other relevant metrics—was selected for analysis to avoid unit-of-analysis errors. All performance metrics are rounded to 2 decimal places for consistency.
As shown in
, ML-based models used between 13 [ ] and 136 [ ] candidate variables. Only 2 studies conducted external validation [ , ], and 2 assessed model calibration [ , ]. The AUROC values for the best-performing ML-based models ranged from 0.77 (95% CI 0.74‐0.79) for ML-LR [ ] to 0.99 for SMOTEENN-CatBoost [ ]. summarizes the performance of the CRS models, which included between 4 [ ] and 23 [ ] significant variables. GRACE was the most commonly used CRS model (n=8), with AUROC values ranging from 0.65 [ ] to 0.88 [ ].Reference | Candidate variables (n) | Validation method | External validation | Calibration | Machine learning model performance | ||
Best model | AUROC | 95% CI | |||||
[23] | 54 | 10-fold | No | Not reported | RF | 0.91 | 0.84‐0.91 |
[24] | 51 | 10-fold | No | Not reported | DNN | 0.97 | Not reported |
[25] | 13 | Not reported | Yes | Not reported | DAMI | 0.91 | 0.90‐0.91 |
[26] | 69 | 4-fold | No | Not reported | DNN | 0.90 | Not reported |
[27] | 50 | 10-fold | No | Not reported | SVMvarImp-SBE-SVM | 0.88 | 0.85‐0.91 |
[28] | 41 | 10-fold | No | Not reported | SMOTEEN | -CatBoost0.99 | Not reported |
[29] | 32 | Not reported | Yes | Yes | RF | 0.91 | 0.89‐0.93 |
[30] | 39 | 10-fold | No | Yes | Nomogram | 0.83 | 0.79‐0.87 |
[31] | 96 | Not reported | No | Not reported | ML-LR | 0.77 | 0.74‐0.79 |
[32] | 136 | 10-fold | No | Not reported | MLR | 0.90 | 0.84‐0.96 |
aAUROC: area under the receiver operating characteristic curve.
bRF: random forest.
cDNN: deep neural network.
dDAMI: deep-learning-based risk stratification for the mortality of patients with acute myocardial infarction.
eSVMvarImp-SBE-SVM: support vector machine variable importance with sequential backward elimination and support vector machine classifier.
fSMOTEENN: hybrid sampling algorithm of synthetic minority oversampling technique (SMOTE) and edited nearest neighbor (ENN) algorithms.
gML-LR: constructed by logistic regression algorithm.
hMLR: multivariate logistic regression.
Reference | Significant variable (n) | Conventional risk score model performance | ||
Model | AUROC | 95% CI | ||
[23] | GRACE: 14, TIMI: 23 | GRACE | , TIMI0.87, 0.82 | Not reported |
[24] | 8 | GRACE | 0.76 | Not reported |
[25] | 13 | GRACE, TIMI, ACTION | 0.85, 0.78, 0.85 | 0.85‐0.86, 0.78‐0.79, 0.85‐0.86 |
[26] | 9 | GRACE | 0.81 | Not reported |
[27] | 9 | TIMI | 0.81 | 0.77‐0.80 |
[28] | Not reported | GRACE | 0.80 | Not reported |
[29] | 8 | GRACE | 0.88 | 0.87‐0.90 |
[30] | Not reported | TIMI | 0.70 | Not reported |
[31] | 4 | GRACE | 0.65 | 0.62‐0.67 |
[32] | 5 | GRACE | 0.83 | 0.78‐0.89 |
aAUROC: area under the receiver operating characteristic curve.
bGRACE: Global Registry of Acute Cardiac Events.
cTIMI: Thrombolysis in Myocardial Infarction.
dACTION: Acute Coronary Treatment and Intervention Outcomes Network scores.
Common Predictors of MACCEs and Mortality in ML and CRS
presents a detailed overview of the significant variables identified in both ML-based models and CRS models. The significant variables identified in ML and CRS models were categorized into 9 groups: sociodemographics, vital signs, electrocardiogram findings, laboratory findings, medical history, medication, angiographic findings, cardiac arrest, and others.
Category | The most important predictors of MACCEs | [study number]Common predictors of mortality | |
Machine learning models | Conventional risk score model | ||
Sociodemographics | Age [23,24,25,27,29,32], body mass index [25,26,27,29], gender [25,26,27], smoking history [23,24,26,27] | Age [23,24,25,30], weight [27] | Age [23,25] |
Vital signs | SBP | [23,25,27,29,32], DBP [23,27], MAP [29], heart rate [25,26,27,29,32]SBP [24,25,27,30], DBP [27], MAP [29], heart rate [24,26,30] | SBP [25,27], DBP [27], MAP [29], heart rate [26] |
Electrocardiogram findings | BBB | [27], ST elevation [25,27]BBB [30], ST elevation [24,30] | Not available |
Laboratory findings | ALT creatinine [23,24,25,26,29,32], CRP [25], CK-MB [25], eosinophils [32], glucose [23,25,26,27,29], hemoglobin [29], LDH [28], LDL [25,26], neutrophils [32], NT-proBNP [28,30], thrombocytes [32] | [28], cholesterol [29],Creatinine [23, 24, 26, 30], CRP [31], glucose [24,25] | Creatinine [26], glucose [25] |
Medical history | AHF | [10], CHF [23], CKD [26], DM [24,26,27,28], family history of IHD [24], malignant neoplasm [26], Killip class [24,25,26,27,29,30,32]CHF [27], hypertension [30], Killip class [24,25,26,30] | Killip class [25,26] |
Medication | Aspirin [27], beta-blockers [27], diuretics [26,27], insulin [27], Mucomyst [26], statins [24,27] | Beta-blockers [27], oral hypoglycemic agent statins [27] | Beta-blockers [27], statins [27] |
Angiographic findings | CABG | [24], cardiac catheterization [27], time from onset to PCI [29]2-vessel CAD | [24]Not available |
Cardiac arrest | IHCA | [24,30], OHCA [25]IHCA [24,30], OHCA [25] | OHCA [25] |
Others | Communication ability [31], discharge outcomes [31], early invasive therapy [26], nonweekday admissions [28] | Time to treatment [27] | Not available |
aMACCEs: major cardiovascular and cerebrovascular adverse events.
bSBP: systolic blood pressure.
cDBP: diastolic blood pressure.
dMAP: mean arterial pressure.
eBBB: bundle branch block.
fALT: alanine aminotransferase.
gCRP: c-reactive protein.
hCK-MB: creatine kinase myoglobin.
iLDH: lactate dehydrogenase.
jLDL: low density lipoprotein.
kNT-proBNP: N-terminal pro-brain natriuretic peptide.
lAHF: acute heart failure.
mCHF: chronic heart failure.
nCKD: chronic kidney disease.
oDM: diabetes mellitus.
pIHD: ischemic heart disease.
qCABG: coronary artery bypass grafting.
rPCI: percutaneous coronary intervention.
sCAD: coronary artery disease.
tIHCA: in-hospital cardiac arrest.
uOHCA: out-of-hospital cardiac arrest.
Our review identified 16 common predictors of MACCEs, including age, systolic blood pressure (SBP), diastolic blood pressure (DBP), mean arterial pressure (MAP), heart rate, bundle branch block and ST elevation (electrocardiogram findings), creatinine, C-reactive protein, glucose, chronic heart failure, Killip class, beta-blocker and statin use, and in-hospital and out-of-hospital cardiac arrest, all of which were significantly associated with a high risk of MACCEs in both ML-based models and CRS models.
Our review identified 10 common predictors of mortality in both models: age, SBP, DBP, MAP, heart rate, creatinine, glucose, Killip class, beta-blocker and statin use, and out-of-hospital cardiac arrest, all of which were related to a high risk of mortality. Age, SBP, and Killip class were the strongest predictors of mortality after PCI in patients with AMI (
and ).Meta-Analysis Results for Mortality Risk Prediction
This meta-analysis included 4 studies reporting AUROC values for mortality prediction, from which 12 AUROC values were obtained based on the validation method and mortality duration (
and ).

Using a random-effects model, the meta-analysis demonstrated that the ML-based models (AUROC: 0.88, 95% CI 0.86‐0.90; I²=97.5%; P<.001) outperformed CRS models (AUROC: 0.79, 95% CI 0.75‐0.84; I²=99.6%; P<.001) in predicting the mortality risk of patients with AMI who had undergone PCI. Substantial heterogeneity was observed across studies. A subgroup meta-analysis was also conducted based on mortality duration. ML-based models outperformed CRS models in both in-hospital (AUROC: 0.89, 95% CI 0.87‐0.90; I²=97.9%; P<.001 vs AUROC: 0.79, 95% CI 0.73‐0.85; I²=99.8%; P<.001) and 30-day mortality prediction (AUROC: 0.87, 95% CI 0.80‐0.94; I²=96.8%, P<.001 vs AUROC: 0.82, 95% CI 0.75‐0.89; I²=98.4%, P<.001), with high heterogeneity across studies.
To statistically confirm whether this observed performance difference was significant, a meta-regression analysis was conducted with model type (ML-based vs CRS) as the moderator. The results showed that ML-based models had significantly higher AUROC values than CRS models (β=.09, 95% CI 0.04‐0.13; P<.001). To further explore sources of heterogeneity, a meta-regression analysis was conducted using log-transformed sample size as a moderator. This revealed a statistically significant negative association between sample size and AUROC for ML-based models (β=–.04, 95% CI –0.07 to 0.01; P=.02). In contrast, no significant association was observed for CRS models (β=–.02, 95% CI –0.10 to 0.07; P=.66).
To assess the reliability and validity of the included studies, publication bias was evaluated. Funnel plots for both ML-based and CRS models showed asymmetry, prompting the use of the trim-and-fill method (
). After adjustment, pooled AUROC values slightly increased in both groups (ML-based models: 0.88 to 0.90; CRS models: 0.80 to 0.83). However, Egger’s regression test and Begg’s rank correlation test revealed no statistically significant evidence of publication bias for either ML-based models (Egger’s test: P=.17; Begg’s test: P=.32) or CRS models (Egger’s test: P=.53; Begg’s test: P=.63).Discussion
Principal Findings
In this review, we appraised studies that assessed the performance of ML-based models compared with CRS for MACCEs prediction after PCI in patients with AMI, using data from 89,702 patients across 10 retrospective studies conducted between 2017 and 2024. Our findings confirmed that ML algorithms outperform CRS methods, such as GRACE and TIMI, in predicting MACCEs risk. Notably, a meta-analysis of 4 studies demonstrated that ML-based models outperform CRS in predicting the mortality risk of AMI patients who had undergone PCI.
ML-based prediction models have gained attention as powerful tools that overcome the limitations of CRS by identifying subtle patterns within complex, multidimensional datasets [
, ]. Predictive models based on CRS rely on fixed assumptions and require prior variable selection, potentially leading to information loss from electronic health records [ ]. In other words, CRS may not adequately incorporate real-time data, including treatments and medication administration and may fail to identify changes in patient’s health status after hospital discharge [ , ], potentially resulting in lower accuracy or poorer performance compared to ML-based models. ML does not require preselection because less significant variables are inherently eliminated during model fitting. Additionally, it continuously improves prediction accuracy by learning from new data in real time [ , ]. ML-based models often require a larger number of input variables; however, many studies have shown that they outperform CRS models, particularly in time-sensitive prediction contexts [ , ]. This advantage is partly due to their ability to leverage variables measured closer to the outcome, which enhances predictive accuracy and highlights the clinical utility of using longitudinal in-hospital data. Although their superior predictive performance may sometimes stem from the inclusion of more input variables, recent advancements in feature selection and high-dimensional data processing now allow ML-based models to perform well with fewer clinically relevant predictors [ , ]. These techniques help identify and exclude redundant or low-value variables during model training, enabling the development of more parsimonious models that are easier to implement in clinical settings.Thus, when combined with CRS, ML-based models can serve complementary roles in developing more effective and reliable predictive models [
]. ML algorithms can rapidly analyze increasingly large datasets, identifying patterns and trends that may not be immediately evident to clinicians. This capability generates opportunities for earlier intervention in situations where CRS may be insufficient [ ]. Rather than replacing traditional risk scores, ML-based models can act as dynamic, real-time decision-support tools that complement static risk calculators [ ]. This integrated approach may improve clinical decision-making in complex scenarios—for example, in patients with prolonged hospital stays or evolving risk profiles.Our review specifically found that age, SBP, DBP, MAP, heart rate, creatinine and glucose levels, Killip class, beta-blocker and statin use, and out-of-hospital cardiac arrest were linked to a higher risk of mortality after PCI in patients with AMI, as well as with MACCEs prediction. Among CRS, these variables were primarily components of the GRACE score [
]. This finding suggests that incorporating CRS into ML-based prediction models may be beneficial for identifying high-risk groups. Moreover, age, SBP, and Killip class were among the top-ranked predictors of mortality in patients with AMI who had undergone PCI. These 3 variables can be easily used to predict MACCEs risk after PCI, even before hospital discharge.Despite the high accuracy and superior discriminative performance of ML-based models in predicting MACCEs, their generalizability remains limited. All included studies were retrospective, and only 2 studies conducted external validation [
, ]. Moreover, 7 of the 10 studies were conducted in Asian countries—including 3 from Korea [ - ]—and relied on registry-based data that may not fully reflect real-world clinical settings. Regional differences in cardiovascular risk profiles may further limit external validity [ ]. For example, Sim and Jeong [ ] reported that the risk factors for AMI in Korean patients differ from those in Western populations. This geographic concentration introduces potential biases, reducing the applicability of findings to Western or ethnically diverse populations. In addition, the lack of prospective cohort studies limits our ability to evaluate the real-time clinical utility and temporal robustness of ML-based predictions. To address these limitations, future research should prioritize multicenter, prospective studies across diverse populations to improve the generalizability and clinical relevance of ML-based models.One key barrier to the clinical implementation of ML-based models is the lack of transparency in their decision-making processes—a challenge often described as the “black box” problem [
, ]. This opacity generates concerns among clinicians, who must be able to understand and trust a model’s reasoning before applying it confidently in patient care. Given these concerns, explainable artificial intelligence—such as SHapley Additive exPlanations—is becoming increasingly important [ , ]. These techniques aim to produce ML-based models that are not only effective but also interpretable and trustworthy in clinical settings [ ]. By integrating explainable artificial intelligence methods into decision support tools, it is possible to bridge the gap between technical performance and clinical usability—enhancing both clinician confidence and real-world reliability.This review found that studies on MACCEs after PCI primarily focused on 30-day or 1-year mortality. The 1-year incidence of MACCEs after primary PCI was 10.8%, with most events (73.6%) occurring between 6 months and 1 year postdischarge [
]. Additionally, cardiac deaths accounted for 50% of overall mortality between 6 months and 2 years after AMI [ ]. Accordingly, further studies are needed to assess the occurrence of MACCEs over extended follow-up periods to enhance the accuracy of predictive risk models [ ].No significant difference was observed between ML-based models and CRS in predicting 30-day mortality. However, ML-based models consistently outperformed CRS for predicting in-hospital mortality. Although the performance of ML-based models slightly decreased when predicting 30-day mortality, their ability to predict long-term outcomes, particularly 1-year mortality, remained significantly superior to that of CRS. These findings suggest that although the predictive accuracy of ML-based models may slightly decline over longer prediction periods, they continue to offer a distinct advantage in forecasting long-term outcomes. Therefore, future research should incorporate longitudinal data and standardized validation methods to further strengthen the predictive accuracy and clinical applicability of ML-based models.
Substantial heterogeneity was observed across the included studies, aligning with previous findings on the complexity of ML-based prediction models [
]. The meta-regression analysis indicated that smaller sample sizes were significantly associated with higher AUROC values, suggesting possible overestimation due to internal validation or overfitting [ ]. However, even after accounting for sample size as a moderator, much of the heterogeneity remained unexplained, pointing to the likely influence of unmeasured factors such as feature engineering, data preprocessing, or institutional differences [ , ]. Future studies should leverage multicenter datasets representing diverse demographic and clinical profiles and conduct prospective external validation across institutions to evaluate real-world generalizability.Regarding the overall risk of bias assessed using the PROBAST checklist and CHARMS, most studies included in this review had a low risk of bias. However, only 4 studies met over 70% of the TRIPOD + AI criteria, indicating a need for improved adherence to established guidelines while developing predictive models. Consequently, future studies must adhere to rigorous methodological standards to ensure the validity of predictive models for MACCEs after PCI in patients with AMI. Furthermore, enabling data and code sharing during model development may enhance transparency and allow for independent validation of results.
Implications for Practice and Research
Our review highlights the superior performance of ML-based models in MACCEs prediction after PCI in patients with AMI compared with CRS. Several common predictors of MACCEs or mortality identified in both ML and CRS in this review may help researchers develop more accurate prediction models. However, most common predictors are limited to nonmodifiable clinical characteristics. Therefore, using longitudinal data, modifiable factors, including psychosocial and behavioral variables, should be incorporated into prediction models for MACCEs after PCI in patients with AMI. Health care professionals should understand the advantages and limitations of ML algorithms and CRS before applying them in clinical practice.
Limitations
This study has certain limitations. This review focused solely on identifying predictors of MACCEs after PCI in patients with AMI. Therefore, our results may not be generalizable to other populations. Despite conducting a comprehensive search across numerous databases to mitigate publication bias, residual bias may still be present, as the analysis was limited to studies published in peer-reviewed journals. Additionally, publication bias may have influenced the results, as studies showing superior ML performance were more likely to be published. Furthermore, all studies in this review were retrospective and had not been performed in clinical practice, posing a significant limitation to their clinical utility. Prospective studies using diverse datasets are needed to ascertain the clinical utility of predictive models and compare their performance with that of clinicians. Additionally, studies included in this review focused on structured data. Therefore, future studies should focus on collecting and integrating unstructured and structured data to develop more accurate risk prediction models.
Lastly, although a meta-regression was conducted to assess the effect of sample size on model performance, it was not feasible to explore all potential sources of heterogeneity due to limitations in the data reported by the included studies. For instance, detailed information on predictor selection strategies and institutional-level variations was often insufficient or inconsistently reported, making comprehensive moderator analysis challenging. Future studies with more standardized reporting may allow for a more nuanced exploration of heterogeneity.
Conclusions
Our findings indicate that ML algorithms outperform CRS in MACCEs prediction after PCI in patients with AMI. Our review suggests that integrating ML-based models with CRS may improve the precise identification of high-risk patients. Future studies should improve generalizability by including diverse populations and validating ML performance across various ethnicities, age groups, and disease profiles while considering CRS.
Acknowledgments
We would like to acknowledge the contributions of individuals who supported this work but did not meet the criteria for authorship. This work was supported by a grant from the National Research Foundation of Korea (NRF), funded by the Korean government (MSIT; grant number RS-2025-00562535). Generative AI tools were not used in the preparation of this manuscript.
Data Availability
The data underlying the findings of this study are available from the corresponding author upon request.
Authors' Contributions
M-YY: Conceptualization, data curation, visualization, validation, software, formal analysis, writing – original draft, writing – review & editing.
HYY: Conceptualization, formal analysis, data curation, writing – original draft, writing – review & editing.
GIH: Data curation, formal analysis, validation, software, visualization, formal analysis, writing – original draft, writing – review & editing.
E-JK: Formal analysis, validation, software, visualization, formal analysis, writing – original draft, writing – review & editing.
Y-JS: Conceptualization, methodology, validation, supervision, project administration, writing – original draft, writing – review & editing.
Conflicts of Interest
None declared.
References
- Kabiri A, Gharin P, Forouzannia SA, Ahmadzadeh K, Miri R, Yousefifard M. HEART versus GRACE score in predicting the outcomes of patients with acute coronary syndrome; a systematic review and meta-analysis. Arch Acad Emerg Med. 2023;11(1):e50. [CrossRef] [Medline]
- Azaza N, Baslaib FO, Al Rishani A, et al. Predictors of the development of major adverse cardiac events following percutaneous coronary intervention. Dubai Med J. 2022;5(2):117-121. [CrossRef]
- Deng W, Wang D, Wan Y, Lai S, Ding Y, Wang X. Prediction models for major adverse cardiovascular events after percutaneous coronary intervention: a systematic review. Front Cardiovasc Med. 2023;10:1287434. [CrossRef] [Medline]
- Sherazi SWA, Bae JW, Lee JY. A soft voting ensemble classifier for early prediction and diagnosis of occurrences of major adverse cardiovascular events for STEMI and NSTEMI during 2-year follow-up in patients with acute coronary syndrome. PLoS One. 2021;16(6):e0249338. [CrossRef] [Medline]
- Zaka A, Mutahar D, Gorcilov J, et al. Machine learning approaches for risk prediction after percutaneous coronary intervention: a systematic review and meta-analysis. Eur Heart J Digit Health. Jan 2025;6(1):23-44. [CrossRef] [Medline]
- Cho SM, Austin PC, Ross HJ, et al. Machine learning compared with conventional statistical models for predicting myocardial infarction readmission and mortality: a systematic review. Can J Cardiol. Aug 2021;37(8):1207-1214. [CrossRef] [Medline]
- Mohd Faizal AS, Thevarajah TM, Khor SM, Chang SW. A review of risk prediction models in cardiovascular disease: conventional approach vs. artificial intelligent approach. Comput Methods Programs Biomed. Aug 2021;207:106190. [CrossRef] [Medline]
- Błaziak M, Urban S, Wietrzyk W, et al. An artificial intelligence approach to guiding the management of heart failure patients using predictive models: a systematic review. Biomedicines. Sep 5, 2022;10(9):1-16. [CrossRef] [Medline]
- Jalali A, Hassanzadeh A, Najafi MS, et al. Predictors of major adverse cardiac and cerebrovascular events after percutaneous coronary intervention in older adults: a systematic review and meta-analysis. BMC Geriatr. Apr 12, 2024;24(1):337. [CrossRef] [Medline]
- Gupta AK, Mustafiz C, Mutahar D, et al. Machine learning vs traditional approaches to predict all-cause mortality for acute coronary syndrome: a systematic review and meta-analysis. Can J Cardiol. Feb 17, 2025;3:1-20. [CrossRef] [Medline]
- Moons KGM, de Groot JAH, Bouwmeester W, et al. Critical appraisal and data extraction for systematic reviews of prediction modelling studies: the CHARMS checklist. PLoS Med. Oct 2014;11(10):e1001744. [CrossRef] [Medline]
- Page MJ, McKenzie JE, Bossuyt PM, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. Mar 29, 2021;372:n71. [CrossRef] [Medline]
- Collins GS, Moons KGM, Dhiman P, et al. TRIPOD+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. BMJ. Apr 16, 2024;385:e078378. [CrossRef] [Medline]
- Wolff RF, Moons KGM, Riley RD, et al. PROBAST: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. Jan 1, 2019;170(1):51-58. [CrossRef] [Medline]
- Damen J, Hooft L, Schuit E, et al. Prediction models for cardiovascular disease risk in the general population: systematic review. BMJ. May 16, 2016;353:i2416. [CrossRef] [Medline]
- Steyerberg EW, Vickers AJ, Cook NR, et al. Assessing the performance of prediction models: a framework for traditional and novel measures. Epidemiology (Sunnyvale). Jan 2010;21(1):128-138. [CrossRef] [Medline]
- Higgins JPT, Thompson SG, Deeks JJ, Altman DG. Measuring inconsistency in meta-analyses. BMJ. Sep 6, 2003;327(7414):557-560. [CrossRef] [Medline]
- Schwarzer G. Meta-analysis in R. In: Egger M, Higgins JPT, Davey Smith G, editors. Systematic Reviews in Health Research: Meta‐Analysis in Context. Chichester; 2022:510-534. [CrossRef]
- Egger M, Davey Smith G, Schneider M, Minder C. Bias in meta-analysis detected by a simple, graphical test. BMJ. Sep 13, 1997;315(7109):629-634. [CrossRef] [Medline]
- Begg CB, Mazumdar M. Operating characteristics of a rank correlation test for publication bias. Biometrics. Dec 1994;50(4):1088-1101. [CrossRef] [Medline]
- Duval S, Tweedie R. Trim and fill: a simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis. Biometrics. Jun 2000;56(2):455-463. [CrossRef] [Medline]
- The R Project. URL: http://www.R-project.org [Accessed 2025-07-09]
- Shim SR, Kim SJ. Intervention meta-analysis: application and practice using R software. Epidemiol Health. 2019;41:e2019008. [CrossRef] [Medline]
- Shouval R, Hadanny A, Shlomo N, et al. Machine learning for prediction of 30-day mortality after ST elevation myocardial infraction: an Acute Coronary Syndrome Israeli Survey data mining study. Int J Cardiol. Nov 1, 2017;246:7-13. [CrossRef] [Medline]
- Kim YJ, Saqlian M, Lee JY. Deep learning–based prediction model of occurrences of major adverse cardiac events during 1-year follow-up after hospital discharge in patients with AMI using knowledge mining. Pers Ubiquit Comput. Apr 2022;26(2):259-267. [CrossRef]
- Kwon JM, Jeon KH, Kim HM, et al. Deep-learning-based risk stratification for mortality of patients with acute myocardial infarction. PLoS One. 2019;14(10):e0224502. [CrossRef] [Medline]
- Sherazi SWA, Jeong YJ, Jae MH, Bae JW, Lee JY. A machine learning-based 1-year mortality prediction model after hospital discharge for clinical patients with acute coronary syndrome. Health Informatics J. Jun 2020;26(2):1289-1304. [CrossRef] [Medline]
- Aziz F, Malek S, Ibrahim KS, et al. Short- and long-term mortality prediction after an acute ST-elevation myocardial infarction (STEMI) in Asians: a machine learning approach. PLoS One. 2021;16(8):e0254894. [CrossRef] [Medline]
- Bai Z, Lu J, Li T, et al. Clinical feature-based machine learning model for 1-year mortality risk prediction of ST-segment elevation myocardial infarction in patients with hyperuricemia: a retrospective study. Comput Math Methods Med. 2021;2021:7252280. [CrossRef] [Medline]
- Hadanny A, Shouval R, Wu J, et al. Predicting 30-day mortality after ST elevation myocardial infarction: machine learning- based random forest and its external validation using two independent nationwide datasets. J Cardiol. Nov 2021;78(5):439-446. [CrossRef] [Medline]
- Fang C, Chen Z, Zhang J, Jin X, Yang M. Construction and evaluation of nomogram model for individualized prediction of risk of major adverse cardiovascular events during hospitalization after percutaneous coronary intervention in patients with acute ST-segment elevation myocardial infarction. Front Cardiovasc Med. 2022;9:1050785. [CrossRef] [Medline]
- Liu Y, Du L, Li L, et al. Development and validation of a machine learning-based readmission risk prediction model for non-ST elevation myocardial infarction patients after percutaneous coronary intervention. Sci Rep. Jun 11, 2024;14(1):13393. [CrossRef] [Medline]
- Shakhgeldyan KI, Kuksin NS, Domzhalov IG, Rublev VY, Geltser BI. Interpretable machine learning for in-hospital mortality risk prediction in patients with ST-elevation myocardial infarction after percutaneous coronary interventions. Comput Biol Med. Mar 2024;170:107953. [CrossRef] [Medline]
- Wee CF, Tan CJW, Yau CE, et al. Accuracy of machine learning in predicting outcomes post-percutaneous coronary intervention: a systematic review. AsiaIntervention. Sep 2024;10(3):219-232. [CrossRef] [Medline]
- Simon S, Mandair D, Albakri A, et al. The impact of time horizon on classification accuracy: application of machine learning to prediction of incident coronary heart disease. JMIR Cardio. Nov 2, 2022;6(2):e38040. [CrossRef] [Medline]
- Dogan MV, Beach SRH, Simons RL, Lendasse A, Penaluna B, Philibert RA. Blood-based biomarkers for predicting the risk for five-year incident coronary heart disease in the Framingham Heart Study via machine learning. Genes (Basel). Dec 18, 2018;9(12):641. [CrossRef] [Medline]
- Murri R, Lenkowicz J, Masciocchi C, et al. A machine-learning parsimonious multivariable predictive model of mortality risk in patients with Covid-19. Sci Rep. Oct 27, 2021;11(1):21136. [CrossRef] [Medline]
- Ning Y, Li S, Ong MEH, et al. A novel interpretable machine learning system to generate clinical risk scores: an application for predicting early mortality or unplanned readmission in a retrospective cohort study. PLoS Digit Health. Jun 2022;1(6):e0000062. [CrossRef] [Medline]
- Dhillon SK, Ganggayah MD, Sinnadurai S, Lio P, Taib NA. Theory and practice of integrating machine learning and conventional statistics in medical data analysis. Diagnostics (Basel). Oct 18, 2022;12(10):1-25. [CrossRef] [Medline]
- Shin S, Austin PC, Ross HJ, et al. Machine learning vs. conventional statistical models for predicting heart failure readmission and mortality. ESC Heart Fail. Feb 2021;8(1):106-115. [CrossRef] [Medline]
- Ke J, Chen Y, Wang X, Wu Z, Chen F. Indirect comparison of TIMI, HEART and GRACE for predicting major cardiovascular events in patients admitted to the emergency department with acute chest pain: a systematic review and meta-analysis. BMJ Open. Aug 18, 2021;11(8):e048356. [CrossRef] [Medline]
- Sim DS, Jeong MH. Differences in the Korea Acute Myocardial Infarction Registry compared with Western registries. Korean Circ J. Nov 2017;47(6):811-822. [CrossRef] [Medline]
- Sariyar M, Holm J. Medical informatics in a tension between black-box AI and trust. In: IOS Press. 2022:41-44. [CrossRef]
- Wadden JJ. Defining the undefinable: the black box problem in healthcare artificial intelligence. J Med Ethics. Oct 2022;48(10):764-768. [CrossRef]
- Nundy S, Montgomery T, Wachter RM. Promoting trust between patients and physicians in the era of artificial intelligence. JAMA. Aug 13, 2019;322(6):497-498. [CrossRef] [Medline]
- Tjoa E, Guan C. A survey on explainable artificial intelligence (XAI): toward medical XAI. IEEE Trans Neural Netw Learning Syst. 2020;32(11):4793-4813. [CrossRef]
- Shah JA, Kumar R, Solangi BA, et al. One-year major adverse cardiovascular events among same-day discharged patients after primary percutaneous coronary intervention at a tertiary care cardiac centre in Karachi, Pakistan: a prospective observational study. BMJ Open. Apr 10, 2023;13(4):e067971. [CrossRef] [Medline]
- Pedersen F, Butrymovich V, Kelbæk H, et al. Short- and long-term cause of death in patients treated with primary PCI for STEMI. J Am Coll Cardiol. 2014;64(20):2101-2108. [CrossRef] [Medline]
- Savic L, Mrdovic I, Asanin M, Stankovic S, Krljanac G, Lasica R. Using the RISK-PCI score in the long-term prediction of major adverse cardiovascular events and mortality after primary percutaneous coronary intervention. J Interv Cardiol. 2019;2019:2679791. [CrossRef] [Medline]
- Tang R, Luo R, Tang S, Song H, Chen X. Machine learning in predicting antimicrobial resistance: a systematic review and meta-analysis. Int J Antimicrob Agents. 2022;60(5-6):106684. [CrossRef] [Medline]
- Zhang Z, Yang L, Han W, et al. Machine learning prediction models for gestational diabetes mellitus: meta-analysis. J Med Internet Res. Mar 16, 2022;24(3):e26634. [CrossRef] [Medline]
- van Kempen EJ, Post M, Mannil M, et al. Performance of machine learning algorithms for glioma segmentation of brain MRI: a systematic literature review and meta-analysis. Eur Radiol. Dec 2021;31(12):9638-9653. [CrossRef] [Medline]
Abbreviations
AI: Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis + Artificial Intelligence |
AMI: acute myocardial infarction |
AUROC: area under the receiver operating characteristic curve |
CHARMS: Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies |
GRACE: Global Registry of Acute Coronary Events |
MACCEs: major adverse cardiovascular and cerebrovascular events |
MeSH: Medical Subject Headings |
ML: machine learning |
PCI: percutaneous coronary interventions |
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
PROBAST: Prediction Model Risk of Bias Assessment Tool |
PROSPERO: International Prospective Register of Systematic Reviews |
TIMI: Thrombolysis in Myocardial Infarction |
Edited by Andrew Coristine; submitted 19.04.25; peer-reviewed by Aslan Erdogan, Tuncay Kiris; final revised version received 16.06.25; accepted 16.06.25; published 18.07.25.
Copyright© Min-Young Yu, Hae Young Yoo, Ga In Han, Eun-Jung Kim, Youn-Jung Son. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 18.7.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.