Minireviews Open Access
Copyright ©The Author(s) 2025. Published by Baishideng Publishing Group Inc. All rights reserved.
World J Gastrointest Oncol. May 15, 2025; 17(5): 104410
Published online May 15, 2025. doi: 10.4251/wjgo.v17.i5.104410
Research status and progress of deep learning in automatic esophageal cancer detection
Jing Chen, Xin Fan, Qiao-Liang Chen, Jian He, Department of Nuclear Medicine, Nanjing Drum Tower Hospital, Affiliated Hospital of Medical School, Nanjing University, Nanjing 210008, Jiangsu Province, China
Wei Ren, The Comprehensive Cancer Center of Drum Tower Hospital, Medical School of Nanjing University & Clinical Cancer Institute of Nanjing University, Nanjing 210008, Jiangsu Province, China
Qi Li, Department of Pathology, Nanjing Drum Tower Hospital, Nanjing 210008, Jiangsu Province, China
Dong Wang, Nanjing Center for Applied Mathematics, Nanjing 211135, Jiangsu Province, China
ORCID number: Xin Fan (0000-0002-9825-7909); Qi Li (0000-0003-2087-3610); Jian He (0000-0001-8140-4610).
Co-corresponding authors: Dong Wang and Jian He.
Author contributions: He J and Chen J conceived the idea for the manuscript; Chen J, Fan X, and Chen QL reviewed the literature and drafted the manuscript; Ren W and Li Q provided comprehensive and clinically relevant perspectives; He J and Wang D revised and finalized the manuscript, and contributed equally as co-corresponding authors; and all authors have read and approved the final version of the manuscript.
Supported by Funding for Clinical Trials from the Nanjing Drum Tower Hospital, Affiliated Hospital of Medical School, Nanjing University, No. 2021-LCYJ-MS-11.
Conflict-of-interest statement: All the authors report no relevant conflicts of interest for this article.
Open Access: This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: https://creativecommons.org/Licenses/by-nc/4.0/
Corresponding author: Jian He, MD, PhD, Associate Professor, Department of Nuclear Medicine, Nanjing Drum Tower Hospital, Affiliated Hospital of Medical School, Nanjing University, No. 321 Zhongshan Road, Nanjing 210008, Jiangsu Province, China. hjxueren@126.com
Received: December 28, 2024
Revised: February 28, 2025
Accepted: March 24, 2025
Published online: May 15, 2025
Processing time: 138 Days and 13.6 Hours

Abstract

Esophageal cancer (EC), a common malignant tumor of the digestive tract, requires early diagnosis and timely treatment to improve patient prognosis. Automated detection of EC using medical imaging has the potential to increase screening efficiency and diagnostic accuracy, thereby significantly improving long-term survival rates and the quality of life of patients. Recent advances in deep learning (DL), particularly convolutional neural networks, have demonstrated remarkable performance in medical imaging analysis. These techniques have shown significant progress in the automated identification of malignant tumors, quantitative analysis of lesions, and improvement in diagnostic accuracy and efficiency. This article comprehensively examines the research progress of DL in medical imaging for EC, covering various imaging modalities such as digital pathology, endoscopy, computed tomography, etc. It explores the clinical value and application prospects of DL in EC screening and diagnosis. Additionally, the article addresses several critical challenges that must be overcome for the clinical translation of DL techniques, including constructing high-quality datasets, promoting multimodal feature fusion, and optimizing artificial intelligence-clinical workflow integration. By providing a detailed overview of the current state of DL in EC imaging and highlighting the key challenges and future directions, this article aims to guide future research and facilitate the clinical implementation of DL technologies in EC management, ultimately contributing to better patient outcomes.

Key Words: Esophageal cancer; Artificial intelligence; Deep learning; Automatic detection; Medical imaging

Core Tip: Esophageal cancer (EC), a common malignant tumor, requires early detection for prognosis improvement. Deep learning (DL), particularly convolutional neural networks, has revolutionized EC diagnosis by enabling automated analysis of multimodal medical imaging, including digital pathology, endoscopy, and computed tomography. This article underscores the potential of DL to enhance screening accuracy and efficiency while addressing critical challenges such as constructing high-quality datasets, promoting multimodal feature fusion, validating model interpretability, and establishing dynamic evaluation systems. This article aims to guide future research and facilitate the clinical implementation of DL technologies in EC management.



INTRODUCTION

Esophageal cancer (EC) is a malignant tumor that originates from the mucosal epithelium of the esophagus and is primarily classified into two main types: Esophageal adenocarcinoma (EAC) and esophageal squamous cell carcinoma (ESCC). According to the International Agency for Research on Cancer, there were 511000 new cases of EC and 445000 deaths worldwide in 2022[1]. Among these, China accounts for nearly half of both the new cases and deaths. Although the incidence and mortality rates of EC have been decreasing in recent years[2], it remains a significant malignant tumor that poses a serious threat to the health of the Chinese population.

EC typically presents with non-specific symptoms in its early stage. Patients often seek medical attention only when they experience progressive dysphagia or persistent retrosternal pain, at which point they are frequently in the middle to late stages of the disease. As a result, over 50% of patients are not eligible for curative surgical therapy. Despite the increase in the age-standardized 5-year survival rate for EC in China from 27.8% to 33.4% between 2008 and 2017 due to advancements in surgical techniques and the introduction of neoadjuvant therapy[3], the overall survival rate remains suboptimal. The prognosis of EC is highly dependent on clinical staging; early-stage patients can achieve a 5-year survival rate of up to 95%[4], while the median survival time for advanced-stage patients is only 6 to 8 months, with a 5-year survival rate of less than 5%[5]. Therefore, increasing the early detection rate is crucial for improving the survival outcomes of patients with EC.

Due to the absence of specific biomarkers, the screening and diagnosis of EC primarily rely on imaging assessments, including endoscopy, computed tomography (CT), and other diagnostic modalities (Figure 1)[6]. Endoscopy combined with histopathological biopsy is currently the gold standard for diagnosing EC. However, due to the atypical morphological features of early EC and pre-cancerous lesions, the complexity of endoscopic procedures, and variations in physician expertise, diagnostic accuracy can vary significantly. Recently, the increasing utilization of chest CT has enhanced the potential for detecting early EC lesions. However, various factors, such as the insidious nature of EC lesions, low contrast, and indistinct borders, have been shown to significantly impact diagnostic accuracy. Some studies indicate that the accuracy of imaging radiologists in detecting EC on chest CT scans is less than 55%[7].

Figure 1
Figure 1 Common diagnostic modalities for esophageal cancer in current clinical practice. A: Endoscopy; B: Pathological sections stained with hematoxylin-eosin; C: Esophageal barium swallow; D: Plain computed tomography; E: 18F-fluorodeoxyglucose positron emission tomography; F: 18F-fluorodeoxyglucose positron emission tomography-computed tomography.

In recent years, advancements in artificial intelligence (AI), particularly in deep learning (DL), have highlighted the potential of computer aided diagnosis (CAD) systems for automatic tumor detection and diagnosis[8,9]. CAD systems leverage DL and other technologies to train on large datasets of medical images [including endoscopy, CT, magnetic resonance imaging (MRI), etc.], enabling them to identify tumor characteristics, automatically locate potential tumor regions or abnormal findings, and provide reliable auxiliary diagnostic information. Currently, CAD systems have been extensively researched for their role in enhancing the automated detection and assessment of various tumors, including lung cancer[10], breast cancer[11,12], and pancreatic cancer[13]. The effective integration of DL methodologies to develop high-performance CAD systems that assist clinicians in achieving more accurate and efficient diagnoses has become a significant area of research within the field of medical imaging analysis.

Although DL has been widely applied in lesion detection across various diseases, research on the automatic detection of EC remains relatively limited. Current AI research related to EC primarily focuses on the analysis of esophageal endoscopy images and digital pathological images, employing techniques such as automatic feature extraction and supervised learning to facilitate automatic detection, diagnosis, and differential diagnosis. This includes the early detection of EC, differentiation between esophageal developmental abnormalities and EC, and the automatic evaluation of infiltration depth. While large-scale endoscopic screening has demonstrated promising results in enhancing the early detection of EC[14], the widespread adoption of routine endoscopy remains challenging due to the insidious nature of the initial symptoms of EC, which limits its broader application. Therefore, it is particularly important to explore the application prospects of DL techniques in imaging examinations, especially in CT images. This paper reviews the research status and progress of DL in the detection of EC, analyzing its diagnostic efficiency of different imaging modalities, including endoscopy, pathology, CT, etc. It also examines the challenges faced by DL in clinical practice and provides insights into future directions for its development in the field of EC diagnosis.

DL DETECTION MODELS

Convolutional neural networks (CNNs), as the core architecture of many DL models, have become a fundamental basis for current research in semantic segmentation and object detection tasks. CNN-based object detection methods are primarily categorized into two-stage detection and single-stage detection, both of which continue to evolve to optimize the balance between model accuracy and computational efficiency.

Two-stage detection network

The pioneering work of Girshick et al[15] introduced the region-based CNN (R-CNN), marking a significant breakthrough in the field of object detection. R-CNN generates candidate regions using a selective search algorithm and independently performs feature extraction and classification, but it suffers from computational redundancy. Fast R-CNN addressed this issue by introducing the region of interest pooling layer, which enables feature map sharing and accelerates feature extraction, partially resolving the speed problem. The core innovation of Faster R-CNN lies in the construction of the Region Proposal Network, which generates candidate regions end-to-end, eliminating the need for selective search[16]. Although Faster R-CNN increases inference time, it significantly improves detection accuracy. Consequently, Faster R-CNN has become a landmark in the field of object detection and is widely applied to two-stage detection tasks. Faster R-CNN has demonstrated significant clinical potential. The latest optimized iteration of faster R-CNN achieved an accuracy of 95.32%, precision of 94.63%, specificity of 94.84%, and sensitivity of 96.23% in detecting lung nodules in CT scans[17]. This algorithm has also shown exceptional performance across diverse medical imaging tasks, including cervical cancer cytology screening[18,19], dermoscopic image analysis for skin cancer diagnosis[20], and automated segmentation and classification of brain tumors[21]. In EC detection[22], Faster R-CNN has been preliminarily explored and demonstrated promising efficacy.

Single-stage detection network

You Only Look Once (YOLO) formulates the object detection task as a regression problem by dividing the image into grids and predicting bounding boxes along with class probabilities for each grid cell, thereby achieving end-to-end object detection. A primary advantage of YOLO lies in its computational efficiency and high accuracy, as it eliminates the need for candidate region generation, making it particularly suitable for real-time applications. The inaugural version, YOLOv1, was proposed by Redmon et al[23] in 2015. Subsequent iterations have incorporated innovations such as multi-scale prediction and cross-layer connections, with each iteration advancing state-of-the-art computer vision. However, research on YOLO-based applications in EC detection remains in its exploratory phase[24]. The single shot multibox detector (SSD) algorithm efficiently integrates multi-scale feature maps with prior boxes, demonstrating significantly improved performance in EAC endoscopic image recognition compared to traditional approaches[25]. However, the SSD model’s performance stability is heavily dependent on sample size, which restricts its application in scenarios with limited training data[26].

THE STATUS OF DL IN DIFFERENT MODALITY IMAGES
Digital pathological images

In the field of AI-assisted diagnosis of EC pathology, DL has made groundbreaking progress from bench to bedside. Research teams have pioneered a quantitative evaluation system for grading dysplasia in Barrett’s esophagus (BE), a precancerous lesion of EAC. This system is based on computational morphometry of epithelial nuclei, showing that nuclear texture heterogeneity is an independent predictor of disease progression (P = 0.004). Additionally, they have provided histomorphological evidence for computer-aided diagnosis. Subsequently, Faghani et al[27] developed a three-class DL model using the YOLO framework, enabling precise classification of non-dysplasia, low-grade dysplasia, and high-grade dysplasia in hematoxylin-eosin stained pathological images. This model achieved 81.3% sensitivity and 100% specificity for low-grade dysplasia, with specificities exceeding 90% across other categories, thereby reinforcing the potential of DL in EC histopathological grading.

However, traditional pathological image classification requires pathologist review of each slide, a laborious process that hinders the scalability of DL technologies. To address this issue, Bouzid et al[28] proposed a weakly supervised DL framework based on multiple instance learning, using a dataset built from hematoxylin-eosin stained sections paired with routine pathology reports to enable large-scale screening programs for BE. By deploying a semi-automated workflow, the framework achieved area under the receiver operating characteristic curve values of 91.4% and 87.3% on internal and external test sets, respectively, encompassing 1866 patients. This reduced pathologists' workload by 48% while preserving diagnostic accuracy, significantly enhancing the efficiency of BE screening. The integration of AI models into clinical pathology workflows is an active area of research[29], encompassing the development of visualization interfaces to accelerate diagnostic assessments. For instance, AI-driven rapid assisted visual search tools and attention heatmap overlays on pathological images can reduce review time by highlighting regions of interest. Future AI-assisted diagnostic tools must undergo multicenter clinical trials to validate their clinical efficacy and system robustness, a critical step toward advancing next-generation intelligent pathology systems.

Endoscopic images

In gastrointestinal endoscopic imaging, DL has been extensively used for colon polyp detection, histological grading of gastric and colonic polyps, early detection of gastrointestinal tumors and diagnosis of Helicobacter pylori infection[30]. These innovations culminated in the first Food and Drug Administration-approved AI-assisted endoscopy system, GI Genius™. However, no DL-based systems specifically targeting EC have received regulatory approval, primarily due to the high heterogeneity and diverse morphological presentations of early EC and precursor lesions. These presentations include erythematous patches, erosions, plaques, mucosal roughness, localized thickening, and disorganized mucosal vasculature, which often mimic benign conditions such as esophagitis and gastric heterotopia, hampering research progress.

In recent years, CAD systems for early tumor detection in BE have received considerable research attention. However, most studies[31-33] are predominantly single-center, small-sample, and retrospective, and have inadequately accounted for confounding by benign lesions, limiting the generalizability of their findings. De Groof et al[34] developed a CAD system using 494364 annotated endoscopic images from five independent datasets to pre-train a hybrid ResNet-UNet model, achieving an overall specificity, sensitivity, and accuracy of 88%, 90%, and 89%, respectively, in differentiating neoplasia from non-dysplastic BE. The BONS-AI consortium further enhanced the robustness of the system by integrating retrospective and prospective data from 15 international endoscopy centers, demonstrating improved generalizability for clinical use[35]. Using advanced pre-training, this CAD system autonomously detects and classifies BE in ex vivo volumetric laser endomicroscopy images/videos. Neoplasia is highlighted by green bounding boxes, with the system achieving 90% sensitivity and 80% specificity for images, and 80% sensitivity and 82% specificity for videos. Endoscopists assisted by CAD showed significantly improved sensitivity for neoplasia detection (image: 74% to 88%, video: 67% to 79%). However, the real-time in vivo performance of volumetric laser endomicroscopy requires further validation to optimize clinical feasibility.

Current research in AI-assisted lesion detection focuses primarily on improving image segmentation and classification accuracy, with limited attention paid to optimizing detection speed and developing real-time algorithms. Horie et al[36] used the SSD framework to analyze 1118 images in 27 seconds, achieving 98% accuracy in differentiating between superficial and advanced EC. The model successfully detected all seven sub-centimeter (< 10 mm) tumors, demonstrating exceptional microlesion detection capability. Tang et al[37] developed a real-time deep CNN that reduced the processing time of a single white light image to 15 milliseconds through parameter optimization.

However, previous lesion detection models have often excluded benign esophageal lesions during training, resulting in dataset selection bias and an increased risk of false positives. Wang et al[24] addressed this gap by integrating benign lesions such as reflux esophagitis, fungal esophagitis and gastric heterotopia into their YOLOv5-based model. The system generates lesion suspicion heatmaps with color intensity positively correlating with malignancy probability, demonstrating ≥ 93% diagnostic accuracy and specificity across white light image, narrow band imaging and Lugol’s chromoendoscopy, comparable to expert endoscopist performance. Single frame analysis took only 0.17 seconds, highlighting the real-time clinical utility. However, this single-center retrospective study focused only on ESCC, excluding cases of EAC and BE, and relied exclusively on static images, without addressing poor quality data, mucosal artefacts or endoscopic motion effects. These limitations prevent clinical generalizability.

Yuan et al[38] conducted the first global randomized controlled trial to evaluate real-time AI support for the diagnosis of early ESCC in a real-world clinical setting, enrolling 11715 patients in 12 Chinese medical centers. The AI system, based on the YOLOv5 algorithm, was seamlessly integrated into standard endoscopy devices and annotated suspicious lesions with malignancy probabilities on the endoscopic screen via a single-screen interface without disrupting clinical workflow. The AI-assisted group showed a significantly lower miss rate than the conventional group (RR = 0.25, P = 0.079). In addition, Li et al[39] developed the ENDOANGEL-ELD system, which enables real-time localization of high-risk esophageal lesions during narrow band imaging using bounding boxes and attention heatmaps, achieving 89.7% sensitivity, 98.5% specificity and 98.2% accuracy. Nevertheless, a comprehensive evaluation of the clinical effectiveness and cost-effectiveness of AI systems remains imperative to confirm their ability to improve diagnostic outcomes in the early detection of EC.

CT images

Contrast-enhanced CT (CECT), a routine imaging modality for the diagnosis of EC, provides clear visualization of irregular esophageal wall thickening, focal or diffuse contrast enhancement, enlarged lymph nodes and periesophageal tissue invasion. Takeuchi et al[40] pioneered an AI diagnostic system using a fine-tuned VGG16 architecture (a deep CNN) trained on 1500 CECT and non-contrast CT images, including 457 EC cases and 1000 normal controls. The system achieved an accuracy of 84.2%, an F1 score of 0.742, a sensitivity of 71.7%, and a specificity of 90.0%. Gradient-weighted class activation mapping was implemented to generate lesion-localizing heatmaps to improve diagnostic interpretability. A major limitation is the 2-dimensional (2D) image-based modelling approach, whereas clinical practice relies on 3D spatial analysis, highlighting the need for future development of 3D convolutional networks. Chen et al[22] developed an EC detection system based on a modified Faster R-CNN framework incorporating Online Hard Example Mining. Evaluated on 1520 CECT images of the gastrointestinal tract, the system achieved an F1 score of 95.71%, a mean average precision of 92.15% and a detection speed of 5.3 seconds per image. The modified Faster R-CNN framework demonstrated superior performance to both the traditional Faster R-CNN and Inception-v2 models by effectively capturing multi-scale feature information through convolutional and parallel architecture designs. The integration of Online Hard Example Mining significantly improved detection accuracy. Yasaka et al[41] evaluated the efficacy of a DL model in assisting radiologists with different levels of expertise (one attending radiologist and three radiology residents) to detect EC on CECT images. The DL model significantly improved the area under the curve (AUC) across readers (pre-intervention: 0.96/0.93/0.96/0.93, post-intervention: 0.97/0.95/0.99/0.96), with statistically significant improvements observed for less experienced residents. Additionally, the DL model enhanced diagnostic confidence and performance among junior radiologists, leading to significant improvements in diagnostic efficiency.

Early studies demonstrate that esophageal wall thickening on non-contrast CT represents a significant imaging biomarker of EC, enabling radiologists to recommend confirmatory endoscopic evaluation based on this finding. Sui et al[7] developed an improved VB-Net segmentation network incorporating multi-scale feature information to achieve precise esophageal wall delineation on CT images. This model automatically quantifies esophageal wall thickness and localizes lesions, demonstrating 88.8% sensitivity and 90.9% specificity. Clinical validation also showed that AI-assisted radiologists (three practitioners) achieved a mean increase in sensitivity from 27.5% to 77.5% and an increase in accuracy from 53.6% to 75.8%, significantly reducing missed diagnosis rates. Lin et al[42] employed a dual-center retrospective dataset with the nnU-Net model for esophageal segmentation, followed by decision tree classification based on extracted radiomic features, achieving an AUC of 0.890. Using DL algorithms, the AUC for physician diagnosis improved from 0.855/0.820/0.930 to 0.910/0.955/0.965 (P < 0.01). However, current models have limitations. These include: (1) Overdependence on esophageal wall thickness, which fails to capture other morphological alterations (e.g., texture features), thereby limiting detection of small lesions; and (2) Insufficient discriminatory power to differentiate malignant lesions from benign esophageal pathologies such as esophagitis and leiomyomas.

Low-dose CT (LDCT) has been widely adopted in cancer screening due to its reduced radiation exposure. However, the inherent limitations of LDCT, including lower spatial resolution and elevated noise levels, introduce additional challenges for EC detection. The subtle anatomical features of the esophageal wall are often poorly visualized. Imaging artifacts from feeding tubes or stents may further obscure lesion boundaries, complicating precise margin delineation in EC. Traditional machine learning approaches (e.g., region-based segmentation and handcrafted feature extraction) rely heavily on pronounced gray-level intensity variations, rendering them inadequate to address the unique diagnostic complexities of EC. In contrast, DL techniques, particularly CNNs, demonstrate superior capability through hierarchical feature learning to capture complex tumor morphology and peritumoral tissue patterns. Even in scenarios with poorly defined lesion boundaries, CNNs maintain clinically acceptable localization and identification accuracy, highlighting their potential for clinical translation. However, research on automated EC detection using DL in LDCT remains exploratory.

To improve the diagnosis of EC on LDCT, the following approaches can be pursued: (1) Multimodal integration of CECT, MRI and positron emission tomography (PET) allows cross-modal feature fusion, using anatomical detail of CT, soft tissue characterization of MRI and metabolic profiling of PET to compensate for the inherent limitations of LDCT; (2) Advanced image processing techniques (e.g., noise suppression and contrast enhancement algorithms) can optimize LDCT image quality for improved DL inputs; and (3) The development of LDCT-specific neural networks incorporating multimodal data, clinical parameters and histopathological correlations could improve the detection of early malignancies. Systematic implementation of these approaches may accelerate the clinical translation of AI-based diagnostic systems.

PET/CT and barium swallow study

18F-fluorodeoxyglucose PET/CT is recommended for initial clinical staging of EC due to its superior capability in detecting distant metastases and enabling comprehensive assessment of tumor burden[43,44]. The integration of multimodal imaging techniques (e.g., PET/CT and PET/MRI) with DL frameworks has significant potential to advance EC diagnosis. However, current research on DL applications in PET/CT-based EC analysis remains limited, with existing studies mainly focusing on radiomics-driven staging prediction, treatment response monitoring and prognostic stratification[45,46]. In particular, automated lesion detection in PET/CT images represents a critical unmet need in this field.

Barium swallow imaging demonstrates high sensitivity for EC diagnosis by identifying imaging features including mucosal disruption, ulceration, filling defect and luminal narrowing[47], which are critical for detection. While Yang et al[48] pioneered CAD systems for this modality, early implementations based on traditional machine learning (e.g., support vector machines and k-nearest neighbors) required labor-intensive manual region of interest annotation and limited radiomic feature extraction. Zhang et al[49] introduced an automated DL system trained on five independent datasets, which achieved 90.3% accuracy, 92.5% sensitivity, and 88.7% specificity through probability-weighted lesion localization, while significantly reducing radiologist interpretation time and improving workflow efficiency. However, no subsequent studies have reported DL applications for EC detection in barium swallow imaging, highlighting a critical research gap in automated diagnostic solutions for this imaging modality.

CHALLENGES AND FUTURE PERSPECTIVES

Despite the demonstrated potential of DL in automated EC detection, current research efforts remain predominantly confined to experimental validation and proof-of-concept prototyping. Few implementations have achieved clinical adoption, with real-world diagnostic performance consistently underperforming experimental benchmarks. This section systematically analyzes the critical translational challenges and proposes strategic solutions to bridge the innovation-to-clinical translation gap.

Building high-quality datasets

A major challenge in training DL models is the lack of large, diverse patient populations and high-quality, annotated datasets. Current research relies predominantly on retrospective, single-center data, which often suffer from selection bias, fragmentation, and inconsistencies across sites. In addition, the quality of medical imaging data is affected by variations in acquisition protocols, imaging equipment and clinical settings. The lack of publicly available datasets for EC detection hinders the reproducibility and validation of models across healthcare institutions. These challenges significantly affect the generalizability and clinical applicability of models.

Federated learning, a decentralized approach to training, provides a viable solution by enabling multi-institutional collaboration while maintaining data privacy, thereby overcoming barriers to data sharing[50,51]. This approach not only improves model performance, but also allows for continuous updates and incremental learning, improving model adaptability to evolving clinical practice. The emergence of generative AI opens up new avenues for efficient data use, with research showing that generative adversarial networks can augment datasets and synthesize images, potentially reducing reliance on training data by more than 40% without compromising diagnostic accuracy[52]. In addition, innovative unsupervised and self-supervised learning methods facilitate the effective use of diverse imaging datasets, significantly reducing the annotation burden associated with clinical AI development. Health data, which are essential for medical AI, is increasingly seen as a valuable asset by healthcare organizations[51]. Since 2024, a number of global healthcare organizations, such as the Mayo Clinic, the National Health Service, Stanford Health Care and Xuanwu Hospital, have initiated the creation of specialized disease datasets and facilitated data sharing through trading mechanisms. The future development of medical AI will depend on the establishment of robust data production and distribution cycles, supported by standardized health data ecosystems and advanced technologies.

Promoting multimodal feature fusion

Health data are inherently multimodal, encompassing radiomic, pathological, and genomic information that is essential for managing EC. Current DL models primarily concentrate on single-modality analyses, such as CT or histopathology, which, while effective for specific tasks, fall short in integrating multimodal imaging techniques, including PET-CT or PET-MRI, alongside high-dimensional omics data. Early fusion methods relied on basic feature concatenation[53], which failed to fully exploit the correlations and complementary insights available between different modalities. Recent advancements have shifted towards transformer architectures that enable modality-agnostic learning[54], employing adaptive attention mechanisms and facilitating knowledge transfer across modalities through co-learning strategies. These frameworks promote adaptable input-output configurations and enhance model robustness via shared representation learning. This underscores the critical need for interdisciplinary collaboration to fully realize the clinical applications of these innovations.

Balancing model transparency and accuracy

A primary concern regarding DL models is the methodology by which they derive their conclusions. This inherent lack of interpretability poses challenges for their acceptance and practical use in clinical environments. The opacity of these models complicates the establishment of clinical trust and adherence to regulatory standards. While existing interpretability techniques, such as gradient-weighted class activation mapping, provide some insights, they often highlight correlations rather than causative factors, making it difficult to understand the rationale behind the model's decisions. This highlights the growing need for frameworks that enhance explainability, such as local interpretable model-agnostic explanations and Shapley Additive Explanations, which are designed to clarify localized decision-making processes. Furthermore, regulatory frameworks like the Food and Drug Administration’s AI/ML Software as Medical Device guidelines (2024) and the European Union’s AI Act (2024) mandate continuous monitoring and independent validation for high-risk medical AI systems after deployment. New protocols like TRIPOD-ML suggest the creation of dynamic validation ecosystems that encompass data drift detection, real-time performance assessment, and iterative model refinement to maintain clinical safety. Achieving a balance between interpretability and accuracy, alongside ensuring adaptive compliance, presents a significant challenge in the clinical application of AI technologies.

Establishing a standardized evaluation framework

AI models often show impressive results in retrospective studies; however, their practical application in clinical settings is hindered by significant validation challenges. These challenges arise from the variability in medical imaging, differing clinical workflows, and operator-dependent factors. Additionally, two unresolved methodological issues exacerbate this translational difficulty: The lack of agreement on optimal DL architectures and the absence of standardized evaluation protocols, which complicates comparisons across studies. A clear example of this standardization issue can be seen in the assessment of esophageal edge irregularities, where variations in inclusion criteria can lead to a 15% fluctuation in Dice score results[55]. Moreover, traditional performance metrics such as accuracy and AUC-receiver operating characteristic are often insufficient in clinical contexts, particularly in early cancer detection, where the stakes of false negatives (which may delay necessary interventions) are much higher than those of false positives (which could lead to unnecessary confirmatory tests). Recent methodological advancements have proposed the use of decision curve analysis to assess clinical net benefit, thus aligning model evaluation more closely with patient-centered considerations, including the differential weighting of misclassification risks and the implications for treatment pathways.

Enhancing cross-domain collaboration

The practical application of DL models in clinical settings is hindered by strict legal and privacy regulations, leading to higher operational costs. Privacy-preserving technologies like federated learning allow for model training across institutions without requiring the sharing of raw data[50]. Additionally, methods such as homomorphic encryption and differential privacy enhance secure data management. The complexities of health data ownership underscore the need for innovative governance structures. For example, Genomics England’s dynamic consent model empowers patients to control their data, whereas The Cancer Genome Atlas treats de-identified datasets as public resources for research. Achieving a balance between data privacy, regulatory compliance, and scientific progress demands collaborative efforts to establish adaptable legal frameworks and standardized technical protocols that foster the development of medical AI while safeguarding patient rights.

Optimizing AI-clinical workflow integration

Challenges in integrating DL models into clinical practice arise from ambiguous operational frameworks, which can lead to cognitive overload due to fragmented multi-platform workflows and complex parameter settings. Future implementations should focus on developing context-sensitive human-AI interfaces that utilize natural language processing and offer adaptive display options, such as visualizing lesions based on risk stratification. Furthermore, establishing closed-loop workflows that combine automated detection, quantitative analysis, and human validation could streamline the screening, confirmation, and reporting processes, thereby improving clinical adoption[56]. The high computational demands of DL models also pose challenges in low-resource healthcare settings, which often lack sophisticated computing infrastructure. This situation highlights the urgent need for effective interoperability between AI systems and hospital information networks to enhance decision support. Additionally, AI-driven early cancer detection requires substantial data storage and processing capabilities, raising concerns about cost and practicality in routine clinical applications. Successful clinical integration relies on securing regulatory approvals, ensuring compatibility with hospital information systems, and validating through prospective clinical trials. Addressing these challenges is essential for the effective implementation of AI technologies in healthcare environments.

CONCLUSION

DL holds significant promise for the early detection of EC; however, its application in clinical settings encounters various challenges. These include the need for high-quality datasets, effective fusion of multimodal features, balancing the tradeoff between explainability and accuracy, establishing standardized evaluation frameworks, enhancing cross-domain collaborations, and optimizing AI integration into clinical workflows. Currently, research predominantly concentrates on the identification of initial tumors, leaving three crucial areas insufficiently addressed: (1) The automated monitoring of disease progression over time; (2) The evaluation of treatment efficacy through repeated imaging analyses; and (3) The prediction of cancer recurrence. The development of new AI models that can adapt to individual patient data has the potential to fill these gaps by facilitating personalized treatment plans. This can be achieved through a comprehensive analysis of medical imaging, biopsy findings, and patient histories, along with the implementation of continuously updated risk assessments. Such advancements could transform the management of EC by enabling tailored therapeutic approaches and enhancing prognostic outcomes.

Footnotes

Provenance and peer review: Invited article; Externally peer reviewed.

Peer-review model: Single blind

Specialty type: Oncology

Country of origin: China

Peer-review report’s classification

Scientific Quality: Grade B, Grade B, Grade B, Grade C, Grade D

Novelty: Grade B, Grade B, Grade B, Grade B, Grade D

Creativity or Innovation: Grade B, Grade B, Grade B, Grade C, Grade C

Scientific Significance: Grade A, Grade B, Grade B, Grade B, Grade D

P-Reviewer: Ali SL; Dilek ON; Garg RK S-Editor: Wei YF L-Editor: A P-Editor: Zhao S

References
1.  Bray F, Laversanne M, Sung H, Ferlay J, Siegel RL, Soerjomataram I, Jemal A. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74:229-263.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 5690]  [Cited by in RCA: 5862]  [Article Influence: 5862.0]  [Reference Citation Analysis (1)]
2.  Chen R, Zheng R, Zhang S, Wang S, Sun K, Zeng H, Li L, Wei W, He J. Patterns and trends in esophageal cancer incidence and mortality in China: An analysis based on cancer registry data. J Natl Cancer Cent. 2023;3:21-27.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 17]  [Cited by in RCA: 24]  [Article Influence: 12.0]  [Reference Citation Analysis (0)]
3.  An L, Zheng R, Zeng H, Zhang S, Chen R, Wang S, Sun K, Li L, Wei W, He J. The survival of esophageal cancer by subtype in China with comparison to the United States. Int J Cancer. 2023;152:151-161.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 22]  [Reference Citation Analysis (0)]
4.  Duggan MA, Anderson WF, Altekruse S, Penberthy L, Sherman ME. The Surveillance, Epidemiology, and End Results (SEER) Program and Pathology: Toward Strengthening the Critical Relationship. Am J Surg Pathol. 2016;40:e94-e102.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 296]  [Cited by in RCA: 331]  [Article Influence: 36.8]  [Reference Citation Analysis (0)]
5.  Chen MY, Jie W, Zhou ZG, Wang J, Bai WW, Zhen CJ, Zhang P. [Advances and the status of radiotherapy in advanced esophageal cancer]. Zhonghua Fangshe Zhongliuxue Zazhi. 2023;32:562-566.  [PubMed]  [DOI]  [Full Text]
6.  Jayaprakasam VS, Yeh R, Ku GY, Petkovska I, Fuqua JL 3rd, Gollub M, Paroder V. Role of Imaging in Esophageal Cancer Management in 2020: Update for Radiologists. AJR Am J Roentgenol. 2020;215:1072-1084.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 8]  [Cited by in RCA: 29]  [Article Influence: 5.8]  [Reference Citation Analysis (0)]
7.  Sui H, Ma R, Liu L, Gao Y, Zhang W, Mo Z. Detection of Incidental Esophageal Cancers on Chest CT by Deep Learning. Front Oncol. 2021;11:700210.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 3]  [Cited by in RCA: 14]  [Article Influence: 3.5]  [Reference Citation Analysis (0)]
8.  Chinese Hospital Association Interventional Medical Centers Branch. [Interventional Medicine Center Association C. Expert consensus for diagnosis and treatment of esophageal cancer based on artificial intelligence platform]. Zhonghua Jieru Fangshexue Dianzi Zazhi. 2021;9:235-246.  [PubMed]  [DOI]  [Full Text]
9.  Tu JX, Ye HQ, Zhang XQ, Lin XT, Yang SL, Deng LF, Wu L. [Visualization Analysis of Artificial Intelligence in Global Esophageal Cancer Research, 2000-2022]. Zhongguo Quanke Yixue. 2023;26:760-768.  [PubMed]  [DOI]  [Full Text]
10.  Cellina M, Cacioppa LM, Cè M, Chiarpenello V, Costa M, Vincenzo Z, Pais D, Bausano MV, Rossini N, Bruno A, Floridi C. Artificial Intelligence in Lung Cancer Screening: The Future Is Now. Cancers (Basel). 2023;15:4344.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 29]  [Reference Citation Analysis (0)]
11.  Ahmad J, Akram S, Jaffar A, Ali Z, Bhatti SM, Ahmad A, Rehman SU. Deep learning empowered breast cancer diagnosis: Advancements in detection and classification. PLoS One. 2024;19:e0304757.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
12.  Rentiya ZS, Mandal S, Inban P, Vempalli H, Dabbara R, Ali S, Kaur K, Adegbite A, Intsiful TA, Jayan M, Odoma VA, Khan A. Revolutionizing Breast Cancer Detection With Artificial Intelligence (AI) in Radiology and Radiation Oncology: A Systematic Review. Cureus. 2024;16:e57619.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
13.  Ramaekers M, Viviers CGA, Janssen BV, Hellström TAE, Ewals L, van der Wulp K, Nederend J, Jacobs I, Pluyter JR, Mavroeidis D, van der Sommen F, Besselink MG, Luyer MDP; E/MTIC Oncology Collaborative Group. Computer-Aided Detection for Pancreatic Cancer Diagnosis: Radiological Challenges and Future Directions. J Clin Med. 2023;12.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 9]  [Cited by in RCA: 4]  [Article Influence: 2.0]  [Reference Citation Analysis (0)]
14.  Qu HT, Li Q, Hao L, Ni YJ, Luan WY, Yang Z, Chen XD, Zhang TT, Miao YD, Zhang F. Esophageal cancer screening, early detection and treatment: Current insights and future directions. World J Gastrointest Oncol. 2024;16:1180-1191.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in CrossRef: 18]  [Cited by in RCA: 10]  [Article Influence: 10.0]  [Reference Citation Analysis (1)]
15.  Girshick R, Donahue J, Darrell T, Malik J.   Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. 2014 IEEE Conference on Computer Vision and Pattern Recognition; Columbus, OH, United States; 2014: 580-587.  [PubMed]  [DOI]  [Full Text]
16.  Ren S, He K, Girshick R, Sun J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans Pattern Anal Mach Intell. 2017;39:1137-1149.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 14896]  [Cited by in RCA: 5291]  [Article Influence: 661.4]  [Reference Citation Analysis (0)]
17.  Lee JD, Hsu YT, Chien JC. Study of a Deep Convolution Network with Enhanced Region Proposal Network in the Detection of Cancerous Lung Tumors. Bioengineering (Basel). 2024;11:511.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
18.  Li X, Xu Z, Shen X, Zhou Y, Xiao B, Li TQ. Detection of Cervical Cancer Cells in Whole Slide Images Using Deformable and Global Context Aware Faster RCNN-FPN. Curr Oncol. 2021;28:3585-3601.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 43]  [Cited by in RCA: 24]  [Article Influence: 6.0]  [Reference Citation Analysis (0)]
19.  Wang W, Tian Y, Xu Y, Zhang XX, Li YS, Zhao SF, Bai YH. 3cDe-Net: a cervical cancer cell detection network based on an improved backbone network and multiscale feature fusion. BMC Med Imaging. 2022;22:130.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 1]  [Reference Citation Analysis (0)]
20.  Nawaz M, Mehmood Z, Nazir T, Naqvi RA, Rehman A, Iqbal M, Saba T. Skin cancer detection from dermoscopic images using deep learning and fuzzy k-means clustering. Microsc Res Tech. 2022;85:339-351.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 25]  [Cited by in RCA: 41]  [Article Influence: 10.3]  [Reference Citation Analysis (0)]
21.  Vinta SR, Chintalapati PV, Babu GR, Tamma R, Sai Chaitanya Kumar G. EDLNet: ensemble deep learning network model for automatic brain tumor classification and segmentation. J Biomol Struct Dyn. 2024;1-13.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
22.  Chen KB, Xuan Y, Lin AJ, Guo SH. Esophageal cancer detection based on classification of gastrointestinal CT images using improved Faster RCNN. Comput Methods Programs Biomed. 2021;207:106172.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 15]  [Cited by in RCA: 5]  [Article Influence: 1.3]  [Reference Citation Analysis (0)]
23.  Redmon J, Divvala SK, Girshick RB, Farhadi A.   You Only Look Once: Unified, Real-Time Object Detection. 2016 Preprint. Available from: arXiv:1506.02640.  [PubMed]  [DOI]  [Full Text]
24.  Wang SX, Ke Y, Liu YM, Liu SY, Song SB, He S, Zhang YM, Dou LZ, Liu Y, Liu XD, Wu HR, Su FX, Zhang FY, Zhang W, Wang GQ. [Establishment and clinical validation of an artificial intelligence YOLOv51 model for the detection of precancerous lesions and superficial esophageal cancer in endoscopic procedure]. Zhonghua Zhong Liu Za Zhi. 2022;44:395-401.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 1]  [Reference Citation Analysis (0)]
25.  Ghatwary N, Zolgharni M, Ye X. Early esophageal adenocarcinoma detection using deep learning methods. Int J Comput Assist Radiol Surg. 2019;14:611-621.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 38]  [Cited by in RCA: 54]  [Article Influence: 9.0]  [Reference Citation Analysis (0)]
26.  Hosseini F, Asadi F, Emami H, Harari RE. Machine learning applications for early detection of esophageal cancer: a systematic review. BMC Med Inform Decis Mak. 2023;23:124.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 12]  [Cited by in RCA: 5]  [Article Influence: 2.5]  [Reference Citation Analysis (0)]
27.  Faghani S, Codipilly DC, David Vogelsang, Moassefi M, Rouzrokh P, Khosravi B, Agarwal S, Dhaliwal L, Katzka DA, Hagen C, Lewis J, Leggett CL, Erickson BJ, Iyer PG. Development of a deep learning model for the histologic diagnosis of dysplasia in Barrett's esophagus. Gastrointest Endosc. 2022;96:918-925.e3.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Cited by in RCA: 22]  [Article Influence: 7.3]  [Reference Citation Analysis (0)]
28.  Bouzid K, Sharma H, Killcoyne S, Castro DC, Schwaighofer A, Ilse M, Salvatelli V, Oktay O, Murthy S, Bordeaux L, Moore L, O'Donovan M, Thieme A, Nori A, Gehrung M, Alvarez-Valle J. Enabling large-scale screening of Barrett's esophagus using weakly supervised deep learning in histopathology. Nat Commun. 2024;15:2026.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
29.  Pesecan CM, Stoicu-Tivadar L. Explaining Deep Learning Models Applied in Histopathology: Current Developments and the Path to Sustainability. Stud Health Technol Inform. 2024;316:1003-1007.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
30.  Stan-Ilie M, Sandru V, Constantinescu G, Plotogea OM, Rinja EM, Tincu IF, Jichitu A, Carasel AE, Butuc AC, Popa B. Artificial Intelligence-The Rising Star in the Field of Gastroenterology and Hepatology. Diagnostics (Basel). 2023;13:662.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
31.  Hussein M, González-Bueno Puyal J, Lines D, Sehgal V, Toth D, Ahmad OF, Kader R, Everson M, Lipman G, Fernandez-Sordo JO, Ragunath K, Esteban JM, Bisschops R, Banks M, Haefner M, Mountney P, Stoyanov D, Lovat LB, Haidry R. A new artificial intelligence system successfully detects and localises early neoplasia in Barrett's esophagus by using convolutional neural networks. United European Gastroenterol J. 2022;10:528-537.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 33]  [Cited by in RCA: 23]  [Article Influence: 7.7]  [Reference Citation Analysis (0)]
32.  Hashimoto R, Requa J, Dao T, Ninh A, Tran E, Mai D, Lugo M, El-Hage Chehade N, Chang KJ, Karnes WE, Samarasena JB. Artificial intelligence using convolutional neural networks for real-time detection of early esophageal neoplasia in Barrett's esophagus (with video). Gastrointest Endosc. 2020;91:1264-1271.e1.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 172]  [Cited by in RCA: 140]  [Article Influence: 28.0]  [Reference Citation Analysis (0)]
33.  Ebigbo A, Mendel R, Probst A, Manzeneder J, Prinz F, de Souza LA Jr, Papa J, Palm C, Messmann H. Real-time use of artificial intelligence in the evaluation of cancer in Barrett's oesophagus. Gut. 2020;69:615-616.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 84]  [Cited by in RCA: 121]  [Article Influence: 24.2]  [Reference Citation Analysis (0)]
34.  de Groof AJ, Struyvenberg MR, van der Putten J, van der Sommen F, Fockens KN, Curvers WL, Zinger S, Pouw RE, Coron E, Baldaque-Silva F, Pech O, Weusten B, Meining A, Neuhaus H, Bisschops R, Dent J, Schoon EJ, de With PH, Bergman JJ. Deep-Learning System Detects Neoplasia in Patients With Barrett's Esophagus With Higher Accuracy Than Endoscopists in a Multistep Training and Validation Study With Benchmarking. Gastroenterology. 2020;158:915-929.e4.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 172]  [Cited by in RCA: 216]  [Article Influence: 43.2]  [Reference Citation Analysis (0)]
35.  Fockens KN, Jong MR, Jukema JB, Boers TGW, Kusters CHJ, van der Putten JA, Pouw RE, Duits LC, Montazeri NSM, van Munster SN, Weusten BLAM, Alvarez Herrero L, Houben MHMG, Nagengast WB, Westerhof J, Alkhalaf A, Mallant-Hent RC, Scholten P, Ragunath K, Seewald S, Elbe P, Baldaque-Silva F, Barret M, Ortiz Fernández-Sordo J, Villarejo GM, Pech O, Beyna T, van der Sommen F, de With PH, de Groof AJ, Bergman JJ; Barrett's Oesophagus Imaging for Artificial Intelligence (BONS-AI) consortium. A deep learning system for detection of early Barrett's neoplasia: a model development and validation study. Lancet Digit Health. 2023;5:e905-e916.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 22]  [Cited by in RCA: 14]  [Article Influence: 7.0]  [Reference Citation Analysis (0)]
36.  Horie Y, Yoshio T, Aoyama K, Yoshimizu S, Horiuchi Y, Ishiyama A, Hirasawa T, Tsuchida T, Ozawa T, Ishihara S, Kumagai Y, Fujishiro M, Maetani I, Fujisaki J, Tada T. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointest Endosc. 2019;89:25-32.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 240]  [Cited by in RCA: 260]  [Article Influence: 43.3]  [Reference Citation Analysis (0)]
37.  Tang D, Wang L, Jiang J, Liu Y, Ni M, Fu Y, Guo H, Wang Z, An F, Zhang K, Hu Y, Zhan Q, Xu G, Zou X. A Novel Deep Learning System for Diagnosing Early Esophageal Squamous Cell Carcinoma: A Multicenter Diagnostic Study. Clin Transl Gastroenterol. 2021;12:e00393.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 11]  [Cited by in RCA: 23]  [Article Influence: 5.8]  [Reference Citation Analysis (0)]
38.  Yuan XL, Liu W, Lin YX, Deng QY, Gao YP, Wan L, Zhang B, Zhang T, Zhang WH, Bi XG, Yang GD, Zhu BH, Zhang F, Qin XB, Pan F, Zeng XH, Chaudhry H, Pang MY, Yang J, Zhang JY, Hu B. Effect of an artificial intelligence-assisted system on endoscopic diagnosis of superficial oesophageal squamous cell carcinoma and precancerous lesions: a multicentre, tandem, double-blind, randomised controlled trial. Lancet Gastroenterol Hepatol. 2024;9:34-44.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 26]  [Cited by in RCA: 11]  [Article Influence: 11.0]  [Reference Citation Analysis (0)]
39.  Li SW, Zhang LH, Cai Y, Zhou XB, Fu XY, Song YQ, Xu SW, Tang SP, Luo RQ, Huang Q, Yan LL, He SQ, Zhang Y, Wang J, Ge SQ, Gu BB, Peng JB, Wang Y, Fang LN, Wu WD, Ye WG, Zhu M, Luo DH, Jin XX, Yang HD, Zhou JJ, Wang ZZ, Wu JF, Qin QQ, Lu YD, Wang F, Chen YH, Chen X, Xu SJ, Tung TH, Luo CW, Ye LP, Yu HG, Mao XL. Deep learning assists detection of esophageal cancer and precursor lesions in a prospective, randomized controlled study. Sci Transl Med. 2024;16:eadk5395.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 14]  [Reference Citation Analysis (0)]
40.  Takeuchi M, Seto T, Hashimoto M, Ichihara N, Morimoto Y, Kawakubo H, Suzuki T, Jinzaki M, Kitagawa Y, Miyata H, Sakakibara Y. Performance of a deep learning-based identification system for esophageal cancer from CT images. Esophagus. 2021;18:612-620.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 9]  [Cited by in RCA: 23]  [Article Influence: 5.8]  [Reference Citation Analysis (0)]
41.  Yasaka K, Hatano S, Mizuki M, Okimoto N, Kubo T, Shibata E, Watadani T, Abe O. Effects of deep learning on radiologists' and radiology residents' performance in identifying esophageal cancer on CT. Br J Radiol. 2023;96:20220685.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 3]  [Cited by in RCA: 9]  [Article Influence: 4.5]  [Reference Citation Analysis (0)]
42.  Lin C, Guo Y, Huang X, Rao S, Zhou J. Esophageal cancer detection via non-contrast CT and deep learning. Front Med (Lausanne). 2024;11:1356752.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
43.  Bureau of Medical Administration, National Health Commission of the People’s Republic of China. [Standardization for diagnosis and treatment of esophageal cancer (2022 edition)]. Zhonghua Xiaohua Waike Zazhi. 2022;21:1247-1268.  [PubMed]  [DOI]  [Full Text]
44.  Expert Panels on Thoracic and Gastrointestinal Imaging, Raptis CA, Goldstein A, Henry TS, Porter KK, Catenacci D, Kelly AM, Kuzniewski CT, Lai AR, Lee E, Long JM, Martin MD, Morris MF, Sandler KL, Sirajuddin A, Surasi DS, Wallace GW, Kamel IR, Donnelly EF. ACR Appropriateness Criteria® Staging and Follow-Up of Esophageal Cancer. J Am Coll Radiol. 2022;19:S462-S472.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 2]  [Reference Citation Analysis (0)]
45.  Ma D, Zhou T, Chen J, Chen J. Radiomics diagnostic performance for predicting lymph node metastasis in esophageal cancer: a systematic review and meta-analysis. BMC Med Imaging. 2024;24:144.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
46.  Menon N, Guidozzi N, Chidambaram S, Markar SR. Performance of radiomics-based artificial intelligence systems in the diagnosis and prediction of treatment response and survival in esophageal cancer: a systematic review and meta-analysis of diagnostic accuracy. Dis Esophagus. 2023;36:doad034.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in RCA: 4]  [Reference Citation Analysis (0)]
47.  Levine MS, Chu P, Furth EE, Rubesin SE, Laufer I, Herlinger H. Carcinoma of the esophagus and esophagogastric junction: sensitivity of radiographic diagnosis. AJR Am J Roentgenol. 1997;168:1423-1426.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 69]  [Cited by in RCA: 71]  [Article Influence: 2.5]  [Reference Citation Analysis (0)]
48.  Yang F, Hamit M, Yan CB, Yao J, Kutluk A, Kong XM, Zhang SX. Feature Extraction and Classification on Esophageal X-Ray Images of Xinjiang Kazak Nationality. J Healthc Eng. 2017;2017:4620732.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 15]  [Cited by in RCA: 11]  [Article Influence: 1.4]  [Reference Citation Analysis (0)]
49.  Zhang P, She Y, Gao J, Feng Z, Tan Q, Min X, Xu S. Development of a Deep Learning System to Detect Esophageal Cancer by Barium Esophagram. Front Oncol. 2022;12:766243.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in RCA: 8]  [Reference Citation Analysis (0)]
50.  Sheller MJ, Edwards B, Reina GA, Martin J, Pati S, Kotrotsou A, Milchenko M, Xu W, Marcus D, Colen RR, Bakas S. Federated learning in medicine: facilitating multi-institutional collaborations without sharing patient data. Sci Rep. 2020;10:12598.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 582]  [Cited by in RCA: 337]  [Article Influence: 67.4]  [Reference Citation Analysis (0)]
51.  Hulsen T. Sharing Is Caring-Data Sharing Initiatives in Healthcare. Int J Environ Res Public Health. 2020;17:3046.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Full Text (PDF)]  [Cited by in Crossref: 86]  [Cited by in RCA: 51]  [Article Influence: 10.2]  [Reference Citation Analysis (0)]
52.  Islam S, Aziz MT, Nabil HR, Jim JR, Mridha MF, Kabir MM, Asai N, Shin J. Generative Adversarial Networks (GANs) in Medical Imaging: Advancements, Applications, and Challenges. IEEE Access. 2024;12:35728-35753.  [PubMed]  [DOI]  [Full Text]
53.  Baltrusaitis T, Ahuja C, Morency LP. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans Pattern Anal Mach Intell. 2019;41:423-443.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 955]  [Cited by in RCA: 475]  [Article Influence: 79.2]  [Reference Citation Analysis (0)]
54.  Shamshad F, Khan S, Zamir SW, Khan MH, Hayat M, Khan FS, Fu H. Transformers in medical imaging: A survey. Med Image Anal. 2023;88:102802.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Cited by in Crossref: 53]  [Cited by in RCA: 200]  [Article Influence: 100.0]  [Reference Citation Analysis (0)]
55.  Yan Y, Kehayias C, He J, Aerts HJWL, Fitzgerald KJ, Kann BH, Kozono DE, Guthier CV, Mak RH. Edge roughness quantifies impact of physician variation on training and performance of deep learning auto-segmentation models for the esophagus. Sci Rep. 2024;14:2536.  [RCA]  [PubMed]  [DOI]  [Full Text]  [Reference Citation Analysis (0)]
56.  Kumar P, Chauhan S, Awasthi LK. Artificial Intelligence in Healthcare: Review, Ethics, Trust Challenges & Future Research Directions. Eng Appl Artif Intel. 2023;120:105894.  [PubMed]  [DOI]  [Full Text]