Systematic Reviews Open Access
Copyright ©The Author(s) 2025. Published by Baishideng Publishing Group Inc. All rights reserved.
World J Gastrointest Oncol. May 15, 2025; 17(5): 103804
Published online May 15, 2025. doi: 10.4251/wjgo.v17.i5.103804
Predicting gastric cancer survival using machine learning: A systematic review
Hong-Niu Wang, Fu-Qiang Wang, Wen-Qing Hu, Liang Zong, Department of Gastrointestinal Surgery, Changzhi People’s Hospital, The Affiliated Hospital of Changzhi Medical College, Changzhi 046000, Shanxi Province, China
Hong-Niu Wang, Graduate School of Medicine, Changzhi Medical College, Changzhi 046000, Shanxi Province, China
Jia-Hao An, Department of Graduate School of Medicine, Changzhi Medical College, Changzhi 046000, Shanxi Province, China
ORCID number: Hong-Niu Wang (0000-0003-2572-4868); Jia-Hao An (0009-0006-3018-4295); Wen-Qing Hu (0000-0003-3364-1034); Liang Zong (0000-0003-4139-4571).
Co-first authors: Hong-Niu Wang and Jia-Hao An.
Author contributions: Wang HN and An JH contributed equally to the preparation of the manuscript; Wang HN designed the review, collected and analyzed the data, and wrote the manuscript; An JH also designed the review, collected and analyzed the data, provided detailed explanations for the figures, and drafted the manuscript; Wang FQ, Hu WQ and Zong L reviewed and revised the manuscript. All authors have read and approved the final version of the manuscript.
Conflict-of-interest statement: The authors report no relevant conflicts of interest for this article.
PRISMA 2009 Checklist statement: All authors have read the PRISMA 2009 checklist, and the manuscript has been prepared and revised according to the PRISMA 2009 checklist.
Open Access: This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: https://creativecommons.org/Licenses/by-nc/4.0/
Corresponding author: Liang Zong, MD, PhD, Department of Gastrointestinal Surgery, Changzhi People’s Hospital, The Affiliated Hospital of Changzhi Medical College, No. 502 Changxing Middle Road, Changzhi 046000, Shanxi Province, China. 250537471@qq.com
Received: December 4, 2024
Revised: February 20, 2025
Accepted: February 26, 2025
Published online: May 15, 2025
Processing time: 162 Days and 15.1 Hours

Abstract
BACKGROUND

Gastric cancer (GC) has a poor prognosis, and the accurate prediction of patient survival remains a significant challenge in oncology. Machine learning (ML) has emerged as a promising tool for survival prediction, though concerns regarding model interpretability, reliance on retrospective data, and variability in performance persist.

AIM

To evaluate ML applications in predicting GC survival and to highlight key limitations in current methods.

METHODS

A comprehensive search of PubMed and Web of Science in November 2024 identified 16 relevant studies published after 2019. The most frequently used ML models were deep learning (37.5%), random forests (37.5%), support vector machines (31.25%), and ensemble methods (18.75%). The dataset sizes varied from 134 to 14177 patients, with nine studies incorporating external validation.

RESULTS

The reported area under the curve values were 0.669–0.980 for overall survival, 0.920–0.960 for cancer-specific survival, and 0.710–0.856 for disease-free survival. These results highlight the potential of ML-based models to improve clinical practice by enabling personalized treatment planning and risk stratification.

CONCLUSION

Despite challenges concerning retrospective studies and a lack of interpretability, ML models show promise; prospective trials and multidimensional data integration are recommended for improving their clinical applicability.

Key Words: Gastric cancer; Machine learning; Deep learning; Survival prediction; Artificial intelligence

Core Tip: Machine learning offers significant promise for predicting gastric cancer patients' survival, but challenges such as data quality, model interpretability, and generalizability must be addressed. This review highlights the importance of integrating diverse data types, robust data preprocessing, and advanced feature-selection techniques to improve prediction accuracy. While open-access and private datasets each have their advantages, ensuring the timeliness and relevance of data is essential for the development of clinically applicable models.



INTRODUCTION

Gastric cancer (GC) is the fifth most common cancer worldwide and the fifth leading cause of cancer-related deaths[1]. Due to its aggressive nature and significant heterogeneity, GC presents a major global health challenge[2]. Its high incidence and poor prognosis complicate treatment strategies[3,4]. Despite advances in surgical techniques and perioperative chemotherapy, the 5-year survival rate for patients with GC remains alarmingly low at 5%–20%[5]. Cancer survival is defined as the period from diagnosis to death due to the cancer[6]. Survival analyses are crucial for clinicians, researchers, patients, and policymakers, highlighting the urgent need for reliable predictive systems to assess prognoses and guide treatment decisions.

Artificial intelligence (AI) has made significant strides in the healthcare sector, and its roles are rapidly expanding. As AI technologies continue to evolve, the digital transformation of healthcare is becoming increasingly evident[7]. Machine learning (ML), a subset of AI, is particularly promising, with the potential to play a pivotal role in supporting both healthcare providers and patients[8].

Traditionally, survival prediction models have been based on Cox regression, which estimates the relative risk of an event occurring through a linear combination of covariates. However, the accuracy of these models depends on the careful selection of relevant variables by researchers. In parallel, ML techniques have shown significant potential in enhancing cancer prognosis prediction. ML includes various analytical methods such as random forests (RFs), ensemble methods, naïve Bayes classifiers, support vector machines (SVMs), neural networks (NNs), decision trees (DTs), and eXtreme Gradient Boosting (XGBoost), among others[9]. Several studies have demonstrated that ML approaches can outperform traditional methods in predictive accuracy[10].

As electronic medical record datasets continue to expand, their volume and complexity now exceed the capacity for human analysis. This has led to issues such as diagnostic errors, workflow inefficiencies, and inappropriate treatments in healthcare systems[11]. AI is increasingly being leveraged to address these challenges, given its ability to quickly process vast amounts of data and extract valuable insights. ML offers the advantage of reducing clinicians' workloads and minimizing human error. The high performance of ML makes it a promising tool for healthcare providers, as it will facilitate the development of predictive models to forecast cancer survival rates. The clinical prospects of ML in healthcare, particularly in cancer management, are vast and continue to evolve.

By integrating and analyzing multiple types of data (e.g., clinical, genomic, imaging, and histopathological information), ML models hold the potential to predict patient survival outcomes and assist in personalizing treatment strategies. By providing precise and timely predictions, ML models could help oncologists identify high-risk patients early, enabling the implementation of tailored interventions and targeted therapies. Moreover, the ability of ML models to continuously learn and adapt from new data ensures that these models can remain relevant and accurate, improving over time as more information becomes available.

The performance of ML algorithms is heavily dependent on the quality of the data used. It is known that poor data quality can lead to suboptimal model performance and inaccurate predictions[12]. However, there has been limited research exploring the impact of data quality and preprocessing steps on ML models. In addition, the data must be meticulously processed and made interpretable for machines. This requires a precise labeling of the data used to train the models, as any inaccuracies can significantly affect the model's prognostic predictions. Model interpretability remains a critical challenge, especially in deep learning models[13-15]. Solutions to these interpretability issues should be considered alongside the vast potential of AI in research.

Despite the growing potential of AI, the application of ML in GC remains underexplored. To promote the integration of ML into routine clinical practice, it is crucial to bridge this gap and gain a deeper understanding of ML's role and progress in GC research. This systematic review aims to provide a comprehensive overview of the application of ML in predicting survival outcomes in GC prognosis models.

MATERIALS AND METHODS
Search strategy

The search strategy applied in this review was designed around the key phrase "predicting GC survival using machine learning". Three thematic groups were initially established, comprising five key terms: "gastric cancer," "survival," and "machine learning/deep learning/artificial intelligence" along with their relevant synonyms. Two reviewers (Wang HN and An JH) independently conducted searches in two electronic databases—PubMed and Web of Science—covering studies published up until November 22, 2024. The search keywords were "gastric cancer," "gastric tumor," "survival," "machine learning," "deep learning," and "artificial intelligence". The full search strategy used in this systematic review is detailed in Supplementary Table 1. The study adhered to the PRISMA guidelines[16].

Inclusion and exclusion criteria

The following inclusion criteria were applied: (1) Peer-reviewed articles; (2) Studies that used ML algorithms to model GC survival; and (3) Published in English. The exclusion criteria were: (1) Studies that did not use ML models; (2) Articles not written in English; and (3) Reviews, case reports, conference abstracts, editorials, letters, or articles with abstract-only content or no full text. Two independent reviewers (Wang HN and An JH) screened all titles and abstracts. Any discrepancies were resolved through discussion with a third experienced reviewer.

Data extraction

The data extraction was performed independently by two reviewers (Wang HN and An JH), with discrepancies resolved by consultation with a third reviewer (Wang FQ). The literature screening was conducted using MS Office Excel 2021. Initially, titles and abstracts were reviewed to exclude irrelevant studies, followed by a full-text assessment to determine eligibility. The CHARMS (Checklist for critical Appraisal and data extraction for systematic Reviews of prediction Modelling Studies) checklist was also used[17]. The data extracted included the following categories (Supplementary Table 2).

Bibliographic information: First author, year of publication, country of study, type of prediction, and prediction outcomes.

Data characteristics: Data source, data type, sample size.

Data preparation and modeling process: Details on missing data handling, preprocessing algorithms and techniques, feature selection methods, types of predictive variables, and ML algorithms used.

Model construction and performance evaluation: Internal and external validation methods, evaluation metrics, calibration techniques, and hyperparameter tuning.

Predictive variables: Number and ranking of predictive variables, and their proportions.

Assessment of the risk of bias

The risk of bias in the included prediction models was assessed by two reviewers using the Prediction Model Risk of Bias Assessment Tool (PROBAST)[18]. The PROBAST is designed to evaluate both the risk of bias and applicability in prognostic and diagnostic models. It covers four domains: Participants, predictors, outcomes, and analysis. Each domain includes specific criteria for assessing low, high, and unclear risks of bias.

RESULTS

The comprehensive search of the PubMed and Web of Science databases identified 832 articles. After the exclusion of 269 duplicates, 563 unique records remained. The screening of the articles' titles and abstracts resulted in the exclusion of 521 studies that did not meet the inclusion criteria. Following a full-text assessment, 42 studies were considered for eligibility. Ultimately, 16 studies were included in the final analysis (Figure 1).

Figure 1
Figure 1 PRISMA flowchart of the study selection process.
Characteristics of the included studies

As summarized in Table 1, the included studies were published between 2019 and 2024. A detailed overview of these studies is provided in Table 2. Of the 16 studies, 14 focused on the overall survival (OS)[19-32], one addressed cancer-specific survival (CSS)[33], and four examined the disease-free survival (DFS)[20,21,26,34] of patients with GC.

Table 1 Extracted characteristics of the included articles.
Ref.
Country
Data source
Samples
Pre-processing
Feature selection
Survival
Data types
Machine learning algorithms
Validation
Evaluation
Hyperparameter tuning
Rahman et al[19] (2021)United KingdomNOGCA (2012-2018)2931YesYesOSClinicalRFInternalAUC, C-index, Brier-score, CalibrationNo
Chen et al[20]
(2019)
ChinaTANRICTCGA134YesYesOS, DFSClinical, molecularSVMInternalC-index, AUCYes
Tian et al[21]
(2024)
ChinaZhongshan Hospital1120YesNoOS, DFSClinical, imageDLInternal, externalC-index, AUCYes
Islam et al[29] (2024)United StatesFujian Medical University Union Hospital135YesYesOSImageRF, SVM, KNN, NBInternalAUC, accuracy,sensitivity, specificity, F1-scoreYes
Chen et al[22] (2024)ChinaTCGANot reportedYesYesOSClinical, molecularMultiple machine earningExternalC-Index, AUC, calibrationYes
Kuwayama et al[30] (2023)JapanChiba Cancer Center (2007-2016)1687YesYesOSClinicalLR, GB, DL, RFInternalAUC, accuracyNo
Zeng et al[27] (2024)ChinaSEER (2000-2019)11076YesYesOSClinicalRF, DLInternal, externalC-Index, AUC, Brier-score, CalibrationYes
Wu et al[25] (2024)ChinaSEER11414YesYesOSClinicalDL, RF
LR
Internal, externalC-Index, AUC, calibration, decision curve analysisYes
Li et al[26] (2022)ChinaNanfang Hospital (2004-2016)695YesYesOS, DFSClinicalSVMInternal, externalAUCNo
Aznar-Gimeno et al[32] (2024)Spain16 general hospitals (2023-2012)1246YesYesOSClinical, molecularRF, XGboost, DL, SVMExternalIndex, Brier-scoreYes
Jiang et al[34] (2022)ChinaNanfang Hospital (2005-2012)510NoNoDFSClinical CTDLInternal, externalC-index, AUC, Brier-score, CalibrationYes
Li et al[28] (2024)ChinaTCGA325YesYesOSClinical, molecularMultiple machine learningExternalC-index, AUC, CalibrationYes
Liao et al[33] (2024)ChinaSEER (2000–2019)775YesYesCSSClinicalMultiple machine learningInternalAUC, CalibrationYes
Wei et al[23] (2022)ChinaTCGA357YesYesOSClinical, molecular, imageMultiDeepCox-SCExternalC-index, AUCYes
Afrash et al[31] (2023)IranAyatollah Talleghani Hospital (2010-2017)974YesYesOSClinicalXGBoost, HGB, SVMInternalAccuracy, specificity, sensitivity, AUC, F1-scoreYes
Zeng et al[27] (2023)ChinaSEER (2000-2019)14177YesYesOSClinicalRF, DLInternalC-index, AUC, Brier-score, Calibration, IBSYes
Table 2 Classification of the features of the included articles.
CharacteristicsCategoriesNumber (n)


OS
CSS
DFS
Dataset sourcesHospitals6-3
SEER31-
TCGA4-1
NOGCA1--
TANRIC1-1
Dataset privacyPublic811
Private6-3
Data sourceSingle611
Multiple8-3
PreprocessingYes1413
No--1
Feature selectionYes1312
No1-2
ModelsOne5-4
Two or more91-
Models typeGB1--
HGB1--
KNN1--
LR2--
NB1--
RF6--
SVM5-2
XGboost2--
DL6-2
MultiDeepCox-SC1--
Ensemble learning21-
ValidationInternal1413
External8-2
EvaluationC-index10-3
AUC1314
Calibration611
Brier-score4-1
Accuracy3--
Specificity2--
Sensitivity2--
F1-score2--
IBS1--
Data typesClinical711
Image1--
Clinical + Image1-2
Clinical + Molecular4-1
Clinical + Molecular + Image1--
Assessment of the risk of bias

Bias risk and applicability assessments were conducted for the 16 studies included in this review. As shown in Table 3, three of the studies were classified as having a high risk of bias[21,22,34], five studies had a moderate risk of bias[20,25,30,31,33], and the remaining eight studies were considered to have a low risk of bias[19,23,24,26-29,32].

Table 3 Risk of bias and applicability assessment of included articles based on the Prediction Model Risk of Bias Assessment Tool criteria.
Ref.Risk of bias
Concern regarding applicability
Overall
Participant
Predictors
Outcomes
Analysis
Participant
Predictors
Outcomes
Risk of bias
Concern regarding applicability
Rahman et al[19] (2021)LowLowLowLowLowModerateLowLowModerate
Chen et al[20]
(2019)
LowLowModerateLowLowLowModerateModerateModerate
Tian et al[21]
(2024)
LowLowLowHighModerateLowLowHighModerate
Islam et al[29] (2024)LowLowLowLowLowLowLowLowLow
Chen et al[22] (2024)HighLowLowHighModerateLowLowHighModerate
Kuwayama et al[30] (2023)LowLowLowModerateModerateLowLowModerateModerate
Zeng et al[27] (2024)LowLowLowLowLowLowLowLowLow
Wu et al[25] (2024)ModerateLowLowLowModerateModerateLowModerateModerate
Li et al[26] (2022)LowLowLowLowModerateLowLowLowLow
Aznar-Gimeno et al[32] (2024)LowLowLowLowModerateModerateLowLowModerate
Jiang et al[34] (2022)LowLowLowHighLowLowLowHighLow
Li et al[28] (2024)LowLowLowLowLowLowLowLowLow
Liao et al[33] (2024)LowLowLowModerateLowModerateLowModerateModerate
Wei et al[23] (2022)LowLowLowLowLowLowLowLowLow
Afrash et al[31] (2023)ModerateLowLowModerateLowModerateLowModerateModerate
Zeng et al[27] (2023)LowLowLowLowModerateLowLowLowModerate
Main database information

Seven studies used datasets from hospitals and clinics[21,26,29-32,34], four studies accessed the Surveillance, Epidemiology, and End Results (SEER) database[24,25,27,33] and another four studies relied on The Cancer Genome Atlas (TCGA) database[20,22,23,28]. The National Oesophago-GC Audit database[19] and the TANRIC database (an open-access resource for interactive exploration of lncRNAs in cancer)[20] were used in one study each. Of the datasets, nine were publicly available[19,20,22-25,27,28,33], while seven were private[21,26,29-32,34]. The dataset sizes ranged from 134 to 14177 records, with seven studies reporting > 1000 records[19,21,24,25,27,30,32].

Data preprocessing

Data preprocessing techniques were applied in 15 of the studies, addressing missing data in 11 of them[19-21,23-27,31-33]. The strategies for handling missing values included listwise deletion, multiple imputation, and the K-nearest neighbor (KNN) algorithm. Feature selection was performed in all but one study[21,34], with 10 of the studies providing detailed descriptions of their methods[19,20,22-24,27,28,31,32]. Common feature selection algorithms included the Boruta method[19,31], relief forward selection[20,31], minimal-redundancy-maximal-relevance (mRMR)[31], Cox regression[22,23,27,28], RFs[24,33], survival analysis[32], and LASSO regression[23,31]. One study also applied data normalization to ensure uniform scaling across features[30].

Data modeling

Five of the 16 studies used a single ML algorithm to construct their models[19-21,26,34]. The other 11 studies employed multiple algorithms, comparing their results to identify the best-performing model[22-25,27-33]. The most frequently used algorithms were RFs, SVMs, deep learning, and ensemble learning. Hyperparameter tuning was incorporated into model training in 13 studies[20-25,27-29,31-34].

Model validation

Seven studies applied internal validation only, and the other nine used both internal and external validation[4,21-26,28,32]. Cross-validation was the most common internal validation method. Performance evaluation metrics varied across the studies. The area under the curve (AUC) ranged from 0.669 to 0.980 across the 15 studies. Five studies reported C-indices between 0.63 and 0.84. Two studies provided metrics including accuracy (0.8910–0.9200), specificity (0.8715-0.9000), sensitivity (0.8942-0.9400), and the F1-score (0.9080-0.9200). Detailed information on these metrics is presented in Table 4.

Table 4 Classification of the used evaluation indicators into types of survival from the lowest to the highest.
Evaluation methodOS

CSS

DFS


Min (%)
Max (%)
Min (%)
Max (%)
Min (%)
Max (%)
AUC66.9098.0092.0096.0071.0085.60
C-index63.000.84.00--65.4071.00
Brier-score13.700.25.00----
Accuracy89.100.92.00----
Specificity87.150.90.00----
Sensitivity89.420.94.00----
F1-score90.8092.00----
IBS14.2015.10----

Among the studies comparing multiple models, ensemble and hybrid approaches demonstrated superior performance in three studies[22,28,33]. Deep learning models exhibited optimal results in three other studies[24,25,27]. RF algorithms were observed to be the most effective in two studies[29,32], while XGBoost[31] and gradient boosting (GB)[30] models were identified as the best-performing algorithms in one study each.

Key variables

Clinical tabular data were used as input in 10 studies[19,21,24-27,30,31,33,34], with eight studies relying solely on these data[19,24-27,30,31,33]. Two studies incorporated both clinical and image-based data[21,34], and one study exclusively used image data[29]. Six studies employed molecular data for survival prediction[20,22,23,28,31,32]. A ranking of frequently used predictors is presented in Table 5. Common predictors included age, sex, ethnicity, cancer type, stage, grade, lymph node count, metastasis, histopathological features, tumor size, primary tumor site, metastasis status, American Society of Anesthesiologists classification, treatment modality, history of other cancer(s), and marital status. Six studies evaluated the contribution of specific predictors to survival outcomes[19,24,27,30,32,33].

Table 5 Predictive variables for survival types extracted from the articles.
Selected features
Number (n)
Percentage (%)
Age787.5
Stage787.5
Grade675.0
Treatment modality675.0
Primary tumor site562.5
Sex450.0
Tumor size450.0
Race337.5
Histopathology type337.5
Marital status337.5
Positive lymph node numbers225.0
Lymph node metastasis225.0
Metastasis status225.0
Regional nodes examined112.5
Lymph node dissection112.5
ASA grade112.5
History of other cancers112.5
Blood markers112.5
Lauren type112.5
Lymphovascular invasion112.5
Months from diagnosis to treatment112.5
Body weight112.5
DISCUSSION

This review synthesizes the application of ML algorithms in predicting the survival of patients with GC, providing insights into the potential of ML for guiding clinical decision-making in this field. From an initial pool of 832 articles, 16 studies were included for analysis. These studies offer both qualitative and quantitative evidence concerning the use of ML models to predict the survival of GC patients. Although the number of ML-based survival prediction studies in GC is still limited, this review reveals that the available research encompasses multiple survival outcomes, including patient OS, CSS, and DFS. The scarcity of studies of each type of survival outcome required their inclusion in a unified analysis.

Our PROBAST assessment revealed that studies with a high or moderate risk of bias tend to overestimate the model performance due to methodological limitations, such as insufficient external validation or incomplete handling of missing data. Although these studies contributed to the variability in the observed AUC/C-index values, their findings should be interpreted with caution. To further enhance the clinical applicability of ML, future research must prioritize robust validation frameworks and transparent reporting, particularly when integrating novel data types.

Data collection and databases

Among the 16 studies, seven used open-access databases, offering transparency in data preprocessing and model construction. Open-access datasets allow for collaborative validation and the optimization of models across different research groups. However, the use of older public databases may present challenges, as clinical practices evolve over time. Models trained on outdated data may lose clinical relevance, posing a problem that has been previously highlighted[35]. While open-access databases are valuable for survival prediction studies, their potential limitations emphasize the need for real-time data management and model timeliness. In contrast, private datasets require ethics approval and informed consent, which restricts their broader validation and comparison, thus limiting their contribution to model refinement.

Despite the advantages of public datasets, the generalizability of models built on these datasets to specific clinical settings remains a concern. Many publicly available datasets were collected years ago and may not fully reflect current clinical practices, making it imperative to develop methods for ensuring the continued relevance of predictive models.

The studies reviewed herein used datasets of varying sizes, with the largest comprising 14177 clinical samples[27], and the smallest including 134 samples[20]. ML algorithms are well-suited for handling multidimensional data, with the assumption that larger sample sizes improve model accuracy[36,37]. Among the studies examined in this review, those using image datasets generally had fewer records than those based on tabular data. Training models with small datasets can lead to overfitting, reducing the findings' generalizability[38]. Nevertheless, image datasets often yield more accurate models than tabular data, due to the advanced capabilities of image-processing algorithms[39]. These models include feature extraction, selection, transfer learning, fine-tuning, augmentation, and object detection[39-41]. In addition, convolutional NN (CNNs) have shown promising results for 3D image analysis [42]. The increasing use of medical image datasets for survival prediction suggests that combining larger image datasets with more advanced CNN architectures will produce more robust models.

In addition to image data, molecular markers such as gene expression and mutations have emerged as important factors in predicting the survival of patients with GC. Recent studies have integrated multiple data types to enhance prediction accuracy. As GC research advances, new variables related to survival have been identified[43,44]. highlighting the disease's complexity and the need to consider these factors in predictive models.

Data preprocessing

Data quality is crucial for the performance of ML algorithms in predictive modeling[45]. Medical datasets often contain noise, redundancy, outliers, missing data, and irrelevant variables, each of which can degrade a model's performance[46]. Proper data preprocessing—including reduction, cleaning, transformation, and integration—is essential to improve the accuracy of ML models. Missing data should be handled appropriately, with techniques such as single imputation, regression, and KNN-based imputation being commonly employed[47]. However, simple imputation methods may introduce bias, particularly with high-dimensional or large datasets[48]. More advanced approaches, such as multiple imputation, offer greater reliability by providing unbiased estimates of missing values and their standard errors[49]. Deleting missing data is another option, though doing so can distort data distributions and introduce bias[48]. Although deletion is straightforward, it should be used cautiously to avoid compromising the model accuracy[50]. It is crucial to avoid altering the data distribution prior to model training to maintain the reliability of the predictions.

Normalization and standardization are important preprocessing steps that reduce redundancy and enhance data consistency[51,52]. These techniques, when combined with outlier removal, improve model performance. Effective data wrangling is critical to ensuring that the model produces reliable outputs. However, many studies have failed to report the preprocessing steps taken, which can significantly affect their model's performance. To enhance the generalizability of ML models in clinical settings, future research should prioritize comprehensive data cleaning, the robust handling of missing values, and a clear reporting of preprocessing methods.

Feature selection

Feature selection is another crucial aspect of ML model development. Among the studies reviewed herein, two (12.5%) did not use feature selection techniques[21,34]. Feature selection improves predictive accuracy by identifying the most relevant features related to GC survival. Hyperparameter tuning and feature selection are commonly used to prevent overfitting and improve model precision. One study using various feature selection algorithms demonstrated that models trained on a full feature dataset performed poorly compared to those trained on selected features[30]. Incorporating effective feature selection and extraction methods is therefore essential for achieving optimal model performance. Future studies should explore and compare features that are significantly relevant to the prediction of survival among patients with GC. Identifying universally applicable features could enhance the accuracy of predictions across various types of GC.

ML algorithms

RF and deep learning are the most commonly used ML methods in the studies reviewed herein, at 37.5% and 37.5% respectively. The RF technique is widely recognized for its ability to address overfitting and improve model generalization compared to other tree-based methods such as DT, gradient boosting machines (GBMs), and XGBoost. Deep learning algorithms have shown significant promise, particularly in processing large, diverse datasets and detecting patterns in images, videos, text, and other forms of data[53]. In six of the 16 studies analyzed in the present review, deep learning was identified as the best-performing algorithm[21,23-25,27,34], and RF was recognized as the top algorithm in four studies[19,22,29,33]. RF has demonstrated excellent performance on small datasets and is capable of handling data with complex and large feature spaces[54]. DT, RF, and XGBoost are tree-based models. However, tree-based models are considered effective for detecting non-monotonic or nonlinear relationships between dependent and independent variables. While tree-based models offer numerous advantages, they also have some limitations. In particular, when training tree-based models on small datasets with highly correlated predictor variables, the detection of interactions between predictors may be hindered, potentially leading to overfitting. One of the primary advantages of RF over individual DT algorithms is its ability to reduce overfitting. The RF algorithm has gained widespread recognition and popularity in the ML field due to its superior performance compared to other tree-based methods, such as DT, GBM, and XGBoost. In a model comprising 16 lncRNA features, SVM performed the best[27]. Although SVM performs well on balanced datasets, its performance tends to decline on imbalanced datasets. Studies in which SVM underperformed did not address the concept of a dataset imbalance during the algorithm's application. Hybrid and ensemble models that combine the strengths of multiple learning algorithms have gained traction in the medical field[55]. These models offer feature selection capabilities and can improve computational efficiency, performance, and generalization. One research group reported that their model performed well in predicting longer survival times (10 years)[20]. However, no study has included ML models capable of predicting short-term, mid-term, and long-term survival in GC. Future research should focus on developing models that can accurately predict both short-term and long-term survival outcomes.

Model interpretation and validation

Interpretability is a critical issue for ML models[13]. Unlike traditional statistical models, ML algorithms are often considered "black boxes," which makes it difficult for researchers to understand the underlying prediction process and identify key variables affecting outcomes. Model performance is influenced not only by data quality, preprocessing, feature selection, and the suitability of the algorithm, but also by the validation strategy employed. Model validation is essential for assessing the performance of ML algorithms. External validation is considered the gold standard for evaluating the generalizability of models[56,57]. Of the studies reviewed herein, nine had external validation, which enhances the reliability of models by testing them on independent datasets. Many studies have used internal validation methods such as cross-validation or random splitting. Although cross-validation is useful for limited datasets[58], external validation ensures that the models used are applicable to diverse populations, each with unique characteristics that may influence survival outcomes. With the increasing availability of open-access datasets, external validation methods have become more feasible, enhancing the robustness of model evaluation.

This systematic review has several limitations. The variability in the studies included limits direct comparisons between them. Moreover, many studies did not provide detailed performance metrics such as specificity, sensitivity, or predictive values. Future research should aim to include comprehensive reports of preprocessing, feature selection, hyperparameter tuning, and model validation procedures, along with performance metrics. This will improve the transparency and reproducibility of ML-based survival prediction models for GC.

Another limitation of this review is the reliance on only two databases, PubMed and Web of Science, for the literature search. While both are well-established and widely used for their robust coverage of peer-reviewed biomedical literature, we recognize that incorporating additional databases, such as Scopus and Embase, could further enrich the comprehensiveness of our review. Those databases may include relevant studies not captured by our search. Nevertheless, we believe that PubMed and Web of Science offer a solid foundation for this review, given their extensive inclusion of high-quality, peer-reviewed articles. Future reviews may benefit from expanding the database selection to enhance the breadth of included studies.

CONCLUSION

Predicting the survival rates for patients with GC is critical for guiding treatment decisions. ML models have shown promising potential for the prediction of survival outcomes and are becoming increasingly applied in this field. The findings of this review highlight the growing use of ML algorithms based on clinical data, particularly from the SEER and TCGA databases. However, challenges remain, particularly related to data preprocessing, feature selection, and model validation. Addressing these challenges will improve the reliability and applicability of ML models for predicting the survival of patients with GC. Future research should focus on refining feature selection, optimizing the model choice, and enhancing model validation to better predict patient survival outcomes.

Footnotes

Provenance and peer review: Invited article; Externally peer reviewed.

Peer-review model: Single blind

Specialty type: Gastroenterology and hepatology

Country of origin: China

Peer-review report’s classification

Scientific Quality: Grade B, Grade B, Grade C

Novelty: Grade B, Grade B, Grade B

Creativity or Innovation: Grade B, Grade B, Grade C

Scientific Significance: Grade B, Grade B, Grade C

P-Reviewer: Qi XS; Yang MQ; Yang S S-Editor: Liu H L-Editor: Filipodia P-Editor: Zhang XD

References
1.  Bray F, Laversanne M, Sung H, Ferlay J, Siegel RL, Soerjomataram I, Jemal A. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74:229-263.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 5690]  [Cited by in RCA: 5862]  [Article Influence: 5862.0]  [Reference Citation Analysis (1)]
2.  Gao JP, Xu W, Liu WT, Yan M, Zhu ZG. Tumor heterogeneity of gastric cancer: From the perspective of tumor-initiating cell. World J Gastroenterol. 2018;24:2567-2581.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 72]  [Cited by in RCA: 98]  [Article Influence: 14.0]  [Reference Citation Analysis (4)]
3.  Karimi P, Islami F, Anandasabapathy S, Freedman ND, Kamangar F. Gastric cancer: descriptive epidemiology, risk factors, screening, and prevention. Cancer Epidemiol Biomarkers Prev. 2014;23:700-713.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1159]  [Cited by in RCA: 1302]  [Article Influence: 118.4]  [Reference Citation Analysis (0)]
4.  Rawicz-Pruszyński K, van Sandick JW, Mielko J, Ciseł B, Polkowski WP. Current challenges in gastric cancer surgery: European perspective. Surg Oncol. 2018;27:650-656.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 20]  [Cited by in RCA: 20]  [Article Influence: 2.9]  [Reference Citation Analysis (0)]
5.  Thrift AP, El-Serag HB. Burden of Gastric Cancer. Clin Gastroenterol Hepatol. 2020;18:534-542.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 432]  [Cited by in RCA: 931]  [Article Influence: 186.2]  [Reference Citation Analysis (1)]
6.  Vale-Silva LA, Rohr K. Long-term cancer survival prediction using multimodal deep learning. Sci Rep. 2021;11:13505.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 19]  [Cited by in RCA: 74]  [Article Influence: 18.5]  [Reference Citation Analysis (0)]
7.  Briganti G, Le Moine O. Artificial Intelligence in Medicine: Today and Tomorrow. Front Med (Lausanne). 2020;7:27.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 275]  [Cited by in RCA: 226]  [Article Influence: 45.2]  [Reference Citation Analysis (0)]
8.  Rajkomar A, Dean J, Kohane I. Machine Learning in Medicine. N Engl J Med. 2019;380:1347-1358.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1274]  [Cited by in RCA: 1532]  [Article Influence: 255.3]  [Reference Citation Analysis (0)]
9.  Nwanosike EM, Conway BR, Merchant HA, Hasan SS. Potential applications and performance of machine learning techniques and algorithms in clinical practice: A systematic review. Int J Med Inform. 2022;159:104679.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 8]  [Cited by in RCA: 56]  [Article Influence: 14.0]  [Reference Citation Analysis (0)]
10.  Alabi RO, Elmusrati M, Sawazaki-Calone I, Kowalski LP, Haglund C, Coletta RD, Mäkitie AA, Salo T, Leivo I, Almangush A. Machine learning application for prediction of locoregional recurrences in early oral tongue cancer: a Web-based prognostic tool. Virchows Arch. 2019;475:489-497.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 39]  [Cited by in RCA: 60]  [Article Influence: 10.0]  [Reference Citation Analysis (0)]
11.  Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. 2019;25:44-56.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2376]  [Cited by in RCA: 2642]  [Article Influence: 440.3]  [Reference Citation Analysis (0)]
12.  Noseworthy PA, Attia ZI, Brewer LC, Hayes SN, Yao X, Kapa S, Friedman PA, Lopez-Jimenez F. Assessing and Mitigating Bias in Medical Artificial Intelligence: The Effects of Race and Ethnicity on a Deep Learning Model for ECG Analysis. Circ Arrhythm Electrophysiol. 2020;13:e007988.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 55]  [Cited by in RCA: 118]  [Article Influence: 23.6]  [Reference Citation Analysis (0)]
13.  Guidotti R, Monreale A, Ruggieri S, Turini F, Giannotti F, Pedreschi D. A Survey of Methods for Explaining Black Box Models. ACM Comput Surv. 2019;51:1-42.  [PubMed]  [DOI]  [Full Text]
14.  Dembrower K, Liu Y, Azizpour H, Eklund M, Smith K, Lindholm P, Strand F. Comparison of a Deep Learning Risk Score and Standard Mammographic Density Score for Breast Cancer Risk Prediction. Radiology. 2020;294:265-272.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 50]  [Cited by in RCA: 86]  [Article Influence: 14.3]  [Reference Citation Analysis (0)]
15.  Wang H, Li Y, Khan SA, Luo Y. Prediction of breast cancer distant recurrence using natural language processing and knowledge-guided convolutional neural network. Artif Intell Med. 2020;110:101977.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 55]  [Cited by in RCA: 42]  [Article Influence: 8.4]  [Reference Citation Analysis (0)]
16.  Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gøtzsche PC, Ioannidis JP, Clarke M, Devereaux PJ, Kleijnen J, Moher D. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate healthcare interventions: explanation and elaboration. BMJ. 2009;339:b2700.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 12640]  [Cited by in RCA: 13169]  [Article Influence: 823.1]  [Reference Citation Analysis (0)]
17.  Moons KG, de Groot JA, Bouwmeester W, Vergouwe Y, Mallett S, Altman DG, Reitsma JB, Collins GS. Critical appraisal and data extraction for systematic reviews of prediction modelling studies: the CHARMS checklist. PLoS Med. 2014;11:e1001744.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1043]  [Cited by in RCA: 1165]  [Article Influence: 105.9]  [Reference Citation Analysis (0)]
18.  Moons KGM, Wolff RF, Riley RD, Whiting PF, Westwood M, Collins GS, Reitsma JB, Kleijnen J, Mallett S. PROBAST: A Tool to Assess Risk of Bias and Applicability of Prediction Model Studies: Explanation and Elaboration. Ann Intern Med. 2019;170:W1-W33.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 667]  [Cited by in RCA: 824]  [Article Influence: 137.3]  [Reference Citation Analysis (0)]
19.  Rahman SA, Maynard N, Trudgill N, Crosby T, Park M, Wahedally H, Underwood TJ, Cromwell DA; NOGCA Project Team and AUGIS. Prediction of long-term survival after gastrectomy using random survival forests. Br J Surg. 2021;108:1341-1350.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 11]  [Cited by in RCA: 17]  [Article Influence: 4.3]  [Reference Citation Analysis (0)]
20.  Chen T, Zhang C, Liu Y, Zhao Y, Lin D, Hu Y, Yu J, Li G. A gastric cancer LncRNAs model for MSI and survival prediction based on support vector machine. BMC Genomics. 2019;20:846.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 32]  [Cited by in RCA: 20]  [Article Influence: 3.3]  [Reference Citation Analysis (0)]
21.  Tian M, Yao Z, Zhou Y, Gan Q, Wang L, Lu H, Wang S, Zhou P, Dai Z, Zhang S, Sun Y, Tang Z, Yu J, Wang X. DeepRisk network: an AI-based tool for digital pathology signature and treatment responsiveness of gastric cancer using whole-slide images. J Transl Med. 2024;22:182.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
22.  Chen H, Zheng Z, Yang C, Tan T, Jiang Y, Xue W. Machine learning based intratumor heterogeneity signature for predicting prognosis and immunotherapy benefit in stomach adenocarcinoma. Sci Rep. 2024;14:23328.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
23.  Wei T, Yuan X, Gao R, Johnston L, Zhou J, Wang Y, Kong W, Xie Y, Zhang Y, Xu D, Yu Z. Survival prediction of stomach cancer using expression data and deep learning models with histopathological images. Cancer Sci. 2023;114:690-701.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 2]  [Reference Citation Analysis (0)]
24.  Zeng J, Song D, Li K, Cao F, Zheng Y. Deep learning model for predicting postoperative survival of patients with gastric cancer. Front Oncol. 2024;14:1329983.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
25.  Wu M, Yang X, Liu Y, Han F, Li X, Wang J, Guo D, Tang X, Lin L, Liu C. Development and validation of a deep learning model for predicting postoperative survival of patients with gastric cancer. BMC Public Health. 2024;24:723.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Reference Citation Analysis (0)]
26.  Li X, Zhai Z, Ding W, Chen L, Zhao Y, Xiong W, Zhang Y, Lin D, Chen Z, Wang W, Gao Y, Cai S, Yu J, Zhang X, Liu H, Li G, Chen T. An artificial intelligence model to predict survival and chemotherapy benefits for gastric cancer patients after gastrectomy development and validation in international multicenter cohorts. Int J Surg. 2022;105:106889.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Cited by in RCA: 1]  [Article Influence: 0.3]  [Reference Citation Analysis (0)]
27.  Zeng J, Li K, Cao F, Zheng Y. Development and validation of survival prediction model for gastric adenocarcinoma patients using deep learning: A SEER-based study. Front Oncol. 2023;13:1131859.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 4]  [Reference Citation Analysis (0)]
28.  Li F, Feng Q, Tao R. Machine learning-based cell death signature for predicting the prognosis and immunotherapy benefit in stomach adenocarcinoma. Medicine (Baltimore). 2024;103:e37314.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
29.  Islam W, Abdoli N, Alam TE, Jones M, Mutembei BM, Yan F, Tang Q. A Neoteric Feature Extraction Technique to Predict the Survival of Gastric Cancer Patients. Diagnostics (Basel). 2024;14.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
30.  Kuwayama N, Hoshino I, Mori Y, Yokota H, Iwatate Y, Uno T. Applying artificial intelligence using routine clinical data for preoperative diagnosis and prognosis evaluation of gastric cancer. Oncol Lett. 2023;26:499.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 3]  [Reference Citation Analysis (0)]
31.  Afrash MR, Mirbagheri E, Mashoufi M, Kazemi-Arpanahi H. Optimizing prognostic factors of five-year survival in gastric cancer patients using feature selection techniques with machine learning algorithms: a comparative study. BMC Med Inform Decis Mak. 2023;23:54.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 9]  [Reference Citation Analysis (0)]
32.  Aznar-Gimeno R, García-González MA, Muñoz-Sierra R, Carrera-Lasfuentes P, Rodrigálvarez-Chamarro MV, González-Muñoz C, Meléndez-Estrada E, Lanas Á, Del Hoyo-Alonso R. GastricAITool: A Clinical Decision Support Tool for the Diagnosis and Prognosis of Gastric Cancer. Biomedicines. 2024;12.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
33.  Liao T, Su T, Lu Y, Huang L, Wei WY, Feng LH. Random survival forest algorithm for risk stratification and survival prediction in gastric neuroendocrine neoplasms. Sci Rep. 2024;14:26969.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
34.  Jiang Y, Zhang Z, Yuan Q, Wang W, Wang H, Li T, Huang W, Xie J, Chen C, Sun Z, Yu J, Xu Y, Poultsides GA, Xing L, Zhou Z, Li G, Li R. Predicting peritoneal recurrence and disease-free survival from CT images in gastric cancer with multitask deep learning: a retrospective study. Lancet Digit Health. 2022;4:e340-e350.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 3]  [Cited by in RCA: 4]  [Article Influence: 1.3]  [Reference Citation Analysis (0)]
35.  Hickey GL, Grant SW, Murphy GJ, Bhabra M, Pagano D, McAllister K, Buchan I, Bridgewater B. Dynamic trends in cardiac surgery: why the logistic EuroSCORE is no longer suitable for contemporary cardiac surgery and implications for future risk models. Eur J Cardiothorac Surg. 2013;43:1146-1152.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 86]  [Cited by in RCA: 107]  [Article Influence: 8.2]  [Reference Citation Analysis (0)]
36.  Al-jarrah OY, Yoo PD, Muhaidat S, Karagiannidis GK, Taha K. Efficient Machine Learning for Big Data: A Review. Big Data Res. 2015;2:87-93.  [PubMed]  [DOI]  [Full Text]
37.  van der Ploeg T, Austin PC, Steyerberg EW. Modern modelling techniques are data hungry: a simulation study for predicting dichotomous endpoints. BMC Med Res Methodol. 2014;14:137.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 259]  [Cited by in RCA: 383]  [Article Influence: 34.8]  [Reference Citation Analysis (0)]
38.  Horenko I. On a Scalable Entropic Breaching of the Overfitting Barrier for Small Data Problems in Machine Learning. Neural Comput. 2020;32:1563-1579.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 11]  [Cited by in RCA: 11]  [Article Influence: 2.2]  [Reference Citation Analysis (0)]
39.  Zhang A, Xing L, Zou J, Wu JC. Shifting machine learning for healthcare from development to deployment and from models to data. Nat Biomed Eng. 2022;6:1330-1345.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 232]  [Cited by in RCA: 123]  [Article Influence: 41.0]  [Reference Citation Analysis (0)]
40.  Frid-Adar M, Diamant I, Klang E, Amitai M, Goldberger J, Greenspan H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing. 2018;321:321-331.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 703]  [Cited by in RCA: 309]  [Article Influence: 44.1]  [Reference Citation Analysis (0)]
41.  Hajiabadi M, Alizadeh Savareh B, Emami H, Bashiri A. Comparison of wavelet transformations to enhance convolutional neural network performance in brain tumor segmentation. BMC Med Inform Decis Mak. 2021;21:327.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 4]  [Reference Citation Analysis (0)]
42.  Alizadeh Savareh B, Emami H, Hajiabadi M, Ghafoori M, Majid Azimi S. Emergence of Convolutional Neural Network in Future Medicine: Why and How. A Review on Brain Tumor Segmentation. Polish J Med Phys Eng. 2018;24:43-53.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1]  [Cited by in RCA: 1]  [Article Influence: 0.1]  [Reference Citation Analysis (0)]
43.  Kim M, Seo AN. Molecular Pathology of Gastric Cancer. J Gastric Cancer. 2022;22:273-305.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 3]  [Cited by in RCA: 10]  [Article Influence: 3.3]  [Reference Citation Analysis (0)]
44.  Körfer J, Lordick F, Hacker UT. Molecular Targets for Gastric Cancer Treatment and Future Perspectives from a Clinical and Translational Point of View. Cancers (Basel). 2021;13.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 4]  [Cited by in RCA: 23]  [Article Influence: 5.8]  [Reference Citation Analysis (0)]
45.  Maharana K, Mondal S, Nemade B. A review: Data pre-processing and data augmentation techniques. Global Transit Proc. 2022;3:91-99.  [PubMed]  [DOI]  [Full Text]
46.  Razzaghi T, Roderick O, Safro I, Marko N. Multilevel Weighted Support Vector Machine for Classification on Healthcare Data with Missing Values. PLoS One. 2016;11:e0155119.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 55]  [Cited by in RCA: 34]  [Article Influence: 3.8]  [Reference Citation Analysis (0)]
47.  Donders AR, van der Heijden GJ, Stijnen T, Moons KG. Review: a gentle introduction to imputation of missing values. J Clin Epidemiol. 2006;59:1087-1091.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 1422]  [Cited by in RCA: 1462]  [Article Influence: 76.9]  [Reference Citation Analysis (0)]
48.  Emmanuel T, Maupong T, Mpoeleng D, Semong T, Mphago B, Tabona O. A survey on missing data in machine learning. J Big Data. 2021;8:140.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 83]  [Cited by in RCA: 143]  [Article Influence: 35.8]  [Reference Citation Analysis (0)]
49.  Nijman S, Leeuwenberg AM, Beekers I, Verkouter I, Jacobs J, Bots ML, Asselbergs FW, Moons K, Debray T. Missing data is poorly handled and reported in prediction model studies using machine learning: a literature review. J Clin Epidemiol. 2022;142:218-229.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 6]  [Cited by in RCA: 60]  [Article Influence: 15.0]  [Reference Citation Analysis (0)]
50.  Gyori NG, Palombo M, Clark CA, Zhang H, Alexander DC. Training data distribution significantly impacts the estimation of tissue microstructure with machine learning. Magn Reson Med. 2022;87:932-947.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 7]  [Cited by in RCA: 29]  [Article Influence: 7.3]  [Reference Citation Analysis (0)]
51.  Siraj MM, Rahmat NA, Din MM.   A Survey on Privacy Preserving Data Mining Approaches and Techniques. Proceedings of the 2019 8th International Conference on Software and Computer Applications 2019.  [PubMed]  [DOI]  [Full Text]
52.  Gal M, Rubinfeld DL. Data Standardization. New York University Law Review. 2018;94:737-770.  [PubMed]  [DOI]  [Full Text]
53.  Rajula HSR, Verlato G, Manchia M, Antonucci N, Fanos V. Comparison of Conventional Statistical Methods with Machine Learning in Medicine: Diagnosis, Drug Development, and Treatment. Medicina (Kaunas). 2020;56.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 56]  [Cited by in RCA: 204]  [Article Influence: 40.8]  [Reference Citation Analysis (0)]
54.  Khadse V, Mahalle PN, Biraris SV.   An Empirical Comparison of Supervised Machine Learning Algorithms for Internet of Things Data. 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA) 2018.  [PubMed]  [DOI]  [Full Text]
55.  Kazienko P, Lughofer E, Trawinski B. Editorial on the special issue “Hybrid and ensemble techniques in soft computing: recent advances and emerging trends”. Soft Comput. 2015;19:3353-3355.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 7]  [Cited by in RCA: 8]  [Article Influence: 0.8]  [Reference Citation Analysis (0)]
56.  Laupacis A. Clinical Prediction Rules. A Review and Suggested Modifications of Methodological Standards. JAMA. 1997;277:488-494.  [PubMed]  [DOI]  [Full Text]
57.  Vergouwe Y, Moons KG, Steyerberg EW. External validity of risk models: Use of benchmark values to disentangle a case-mix effect from incorrect coefficients. Am J Epidemiol. 2010;172:971-980.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 159]  [Cited by in RCA: 182]  [Article Influence: 12.1]  [Reference Citation Analysis (0)]
58.  Ramspek CL, Jager KJ, Dekker FW, Zoccali C, van Diepen M. External validation of prognostic models: what, why, how, when and where? Clin Kidney J. 2021;14:49-58.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 365]  [Cited by in RCA: 457]  [Article Influence: 114.3]  [Reference Citation Analysis (0)]