This article is an open-access article which was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/
Author contributions: Alpsoy A and Yavuz A performed the data acquisition; Elpek GO designed the outline and coordinated the writing of the paper; all authors equally contributed to the writing of the paper and preparation of the tables.
Conflict-of-interest statement: There is no conflict of interest to disclose.
Open-Access: This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: https://creativecommons.org/Licenses/by-nc/4.0/
Corresponding author: Gulsum Ozlem Elpek, MD, Professor, Pathology, Akdeniz University Medical School, Dumlupınar bulvarı, Antalya 07070, Turkey. firstname.lastname@example.org
Received: December 6, 2021 Peer-review started: December 6, 2021 First decision: December 13, 2021 Revised: December 19, 2021 Accepted: December 27, 2021 Article in press: December 27, 2021 Published online: December 28, 2021
The integration of artificial intelligence (AI) has shown promising benefits in many fields of diagnostic histopathology, including for gastrointestinal cancers (GCs), such as tumor identification, classification, and prognosis prediction. In parallel, recent evidence suggests that AI may help reduce the workload in gastrointestinal pathology by automatically detecting tumor tissues and evaluating prognostic parameters. In addition, AI seems to be an attractive tool for biomarker/genetic alteration prediction in GC, as it can contain a massive amount of information from visual data that is complex and partially understandable by pathologists. From this point of view, it is suggested that advances in AI could lead to revolutionary changes in many fields of pathology. Unfortunately, these findings do not exclude the possibility that there are still many hurdles to overcome before AI applications can be safely and effectively applied in actual pathology practice. These include a broad spectrum of challenges from needs identification to cost-effectiveness. Therefore, unlike other disciplines of medicine, no histopathology-based AI application, including in GC, has ever been approved either by a regulatory authority or approved for public reimbursement. The purpose of this review is to present data related to the applications of AI in pathology practice in GC and present the challenges that need to be overcome for their implementation.
Core Tip: Recently, based on improvements in efficient computational power and learning capacities, various artificial intelligence applications, such as image-based diagnosis and prognosis prediction, have emerged in many fields of pathology. This review comprehensively summarizes the current status of artificial intelligence applications in gastrointestinal cancers. The present data are promising for the use of artificial intelligence to diagnose tumors, evaluate prognostic parameters, and detect biomarker/genetic alterations. However, many challenges hinder the implication of artificial intelligence models in real pathological practice. Therefore, these challenges and suggested solutions are also briefly presented to improve the accuracy and relevance of artificial intelligence in pathological practice, including in gastrointestinal cancers.
Citation: Alpsoy A, Yavuz A, Elpek GO. Artificial intelligence in pathological evaluation of gastrointestinal cancers. Artif Intell Gastroenterol 2021; 2(6): 141-156
Pathology is a medical specialty that performs morphological evaluations of organs, tissues, and cells to provide a definitive diagnosis of diseases and contributes to treatment by determining the critical parameters in their course. Although histopathological assessment under a light microscope is considered a cornerstone, especially in oncology, the search for more objective criteria to overwhelm the subjectivity related to interobserver and intraobserver variations and to diminish the increased workload and time consumption has led to the development of image analysis-based digital pathology (DP), which plays a crucial role in modern pathological practice[2,3].
Following the considerable advances of slide scanner technology that can quickly digitalize whole pathological slides at high resolution (whole-slide images, WSI), in 2017, the approval of the Philips IntelliSite whole-slide scanner (Philips Electronics, Amsterdam, Netherlands) by the Food and Drug Administration (FDA) in the United States allowed a comprehensive evolution in DP. This digitization not only facilitated the application of telepathology and created a valuable resource for education but also yielded the analysis of a large spectrum of morphological parameters and biomarkers/genetic alterations[5-7]. In addition, such digital images are constituted from matrices of numbers that contain much more information that is not accessible to the human eye[8,9]. Indeed, it may be possible to extract predictive and prognostic biomarkers from such digitized slides by computer-based image analysis. These methods are particularly of direct interest to ''computational pathology'', a relatively new pathology field driven by artificial intelligence (AI) that is expected to transform and improve the diagnosis and staging of cancers[3,10]. As a result, pathological AI models have evolved from expert systems to traditional machine learning (ML) and, finally, deep learning (DL). While the traditional supervised ML allows the production of data output from previously labeled training sets that can be corrected by the users, labeling big data can be time-consuming and challenging. In addition, the accuracy depends heavily on the quality of feature extraction. In contrast, unsupervised ML is a time-saving model because it provides automatic detection of patterns. However, input data that are not labeled by users pose challenges during interpretation, leading to varying results.
On the other hand, DL extracts features directly from the raw data and utilizes multiple layers of hidden data for the output[14-16]. Compared to expert systems and handcrafted ML models, DL models are simpler to conduct, have higher precision, and are more cost-effective[9,17] (Table 1). Furthermore, a considerable increase in computational processing capacity and the development of algorithms, such as convolutional neural networks (CNNs), fully CNNs, recurrent neural networks (RNNs), and generative adversarial networks, have resulted in numerous investigations on the application of DL-based AI in pathological practice[7,18,19]. The strengths and weaknesses of typical ML methods are summarized in Table 1.
Table 1 Strengths and weaknesses of machine learning methods in development of artificial intelligence models for gastrointestinal pathology.
Traditional ML (supervised)
Allows users to produce a data output from the previously labeled training set
Labeling big data can be time-consuming and challenging
Users can reflect domain knowledge features
Accuracy depends heavily on the quality of feature extraction
Traditional ML (unsupervised)
Users do not label any data or supervise the model
Input data is unknown and not labeled by users
Can detect patterns automatically
Users cannot get precise information regarding data sorting
Challenges during interpreting
Detects the important information and features without labeling
A large training data is required
High performance in image recognition
Lack of interpretability (black boxes)
Provides computational speed
Requires large amounts of labeled data for training
Automatically eliminates the background noise
High labeling cost
Can decide which information to remember from its past experience
Harder to train the model
A deep learning model for sequential data
High computational cost
Does not require detailed annotation
A large amount of training data is required
Can be applied to large data sets
High computational cost
Generates new realistic data resembling the original data
In addition, the use of AI in pathology has led to the emergence of many DL-based applications. Proscia, DeepLens, PathAI, and Inspirata are DL-based applications for the detection, diagnosis, and prognosis of several cancer subtypes[21-25]. In addition, Inspirata and PAIGE.AI are spending substantial time and resources on creating large libraries of digital WSI for use in training AI algorithms[21,24]. Interestingly, the landscape of DP is, in parallel, also undergoing important innovation and rapid changes.
It is also notable that some institutions are digitizing their entire pathology workflow, suggesting the routine use of AI-based systems in many areas of pathology soon[26,27]. Indeed, many studies have suggested that the integration of AI provides benefits for diagnosing and subtyping tumors, detecting histopathological parameters related to prognosis, and even identifying biomarker/genetic alterations in many fields of pathology. On the other hand, the existence of a broad spectrum of difficulties, from AI-based pathology laboratory infrastructures to the robustness of algorithms, indicates that there are still many obstacles to be resolved before introducing AI applications in real-life pathology practice. Nonetheless, AI-based approaches have the potential to contribute to pathological practice by improving workflows, eliminating simple errors, and increasing diagnostic reproducibility.
Regarding the gastrointestinal system, the accumulated data indicate that AI-based models might provide diagnostic assistance, prognosis prediction, and biomarker development for gastrointestinal cancer (GC). There have been few studies in the recent past that have addressed the effectiveness of AI models in GC[8,30]. However, effective implementation of these methods in real-life pathology practice requires further reviews comparing the results of previous studies and highlighting the challenges to be overcome.
This review presents recent data about the AI-based pathological evaluation of GC and current challenges for its implementation in gastrointestinal pathology practice with future directions to consider.
AI-BASED APPLICATIONS IN DIAGNOSIS OF GC
Recent studies on the use of AI models in the histopathological classification of gastric cancer are summarized in Table 2. Although the models used differ among studies, the results support that AI-based classification can be used in histopathological evaluations based on the accuracy and area under the curve (AUC) values determined. Different models are considered together in a few studies. For example, in a study where two DL-based methods were used to diagnose gastric cancer, the mean accuracy of both models was shown to be up to 89.7%. In another study that compared the classification results of experienced pathologists with those of the ML-based program created by NEC Corporation, in gastric biopsy specimens, the agreement rate for biopsy specimens negative for neoplastic lesions was found to be as high as 90.6%. More recently, Iizuka et al, who aimed to classify gastric biopsies as gastric adenocarcinoma, adenoma, or nonneoplastic mucosa by using AI algorithms based on CNNs and RNNs, revealed that the AUC for gastric adenocarcinoma classification was 0.9, supporting that AI-based models could be helpful in the diagnosis of gastric cancer. Although these results suggest that AI can be used to diagnose gastric cancer, it is difficult to relate these data to performance comparisons alone. In research, parameters such as the size of the dataset, resolution of detection, multisite validation, the number of categories to be classified, and most importantly, the presence of lesions other than malignancies that require diagnosis are also critical variables. In particular, the latter could be a potential limitation of AI-based models in actual practice. Indeed, a gastric biopsy is evaluated not only for malignancy but also for lesions such as gastritis and metaplasia. Therefore, an AI model used only for malignancy screening in gastric pathology will not reduce the pathologist's workload, as other findings also need to be reviewed.
Table 2 Artificial intelligence-based applications in gastric cancer.
Variable across tumors/gene alterations. Strongest relations in whole genome duplications
1Esophageal adenocarcinoma and Barrett’s esophagus.
2Gastric and colorectal cancers.
3Gastric, colorectal, esophageal, and liver cancers.
4Gastric, colorectal, and pancreatic cancers.
AI: Artificial intelligence; GastricNet: The deep learning framework; WSIs: Whole slide images; MIL: Multi-instance learning; AUC: Area under the curve; CNN: Convolutional neural networks; DL: Deep learning; ML: Machine learning; RNN: Recurrent neural networks; ANN: Artificial neural network; QUEEN technique: Quality assured efficient engineering of feedforward neural networks with supervised learning; NN: Neural network; MMHG: Multimodal hypergraph learning framework; SVM: Support vector machine.
AI applications have also been developed to diagnose colorectal cancer (CRC), which may allow classification of lesions as normal, hyperplasia, adenoma, adenocarcinoma, and histological subtypes of polyps or adenocarcinomas (Table 3). In an elegant study, Korbar et al observed that their AI models could classify five colorectal polyp types with a 93% accuracy. In another study, a created DL model was able to reclassify colorectal polyps in a manner comparable to those of the pathologist, even in datasets from other hospitals. From this perspective, the results of most studies are encouraging for the use of AI models in the diagnosis of CRC. However, this does not exclude the fact that comparing the performance of those models reliably necessitates a common task using a standardized dataset with standardized annotations because each model is derived from different datasets with different explanations and is focused on different tasks in current studies.
Table 3 Artificial intelligence-based applications in colorectal cancer.
Because gastric cancer has more complex and heterogeneous morphological features than CRC, most AI-based studies performed on these tumors focus on diagnosis rather than prognostication studies (Table 2). Nevertheless, there is some evidence showing that AI models can be helpful to evaluate histopathological parameters, such as differentiation and lymphovascular involvement, which are essential in determining the survival time[36-38], recurrence risk[39,40], metastasis[41-43], and, accordingly, treatment of gastric cancer. In the survival analysis, a higher predictive accuracy for overall survival and disease-free survival than the tumor-node-metastasis staging system defined by the American Joint Committee on Cancer by SVM application has been demonstrated. In addition, this method can also be used to predict adjuvant chemotherapeutic benefits, which can facilitate individualized therapy. Another study combining the demographics, pathological indicators, and physiological characteristics of the study group found that a method using a new multimodal hypergraph learning framework to improve the accuracy of survival prediction outperformed random forests and SVM in survival prediction. Furthermore, when the artificial neural network and Bayesian neural network (BNN) values were compared in survival estimation, it was shown that BNN was superior to the artificial neural network method.
The application of neural networks significantly improves the prediction of lymph node metastasis. In addition, in a study to determine the microenvironment that can predict tumor behavior, García et alobserved that a CNN model could be used to detect tumor-infiltrating lymphocytes (accuracy, 96.9%). However, the number of these studies should be increased to draw a better conclusion about the application of AI-based DP in the prognostication of gastric cancer.
In CRC, DL was found to be effective in predicting prognosis at all stages. For example, in a study where RNN analyzed tissue microarrays to predict 5-year disease-specific survival, the hazard ratio and AUC were determined to be 2.3 and 0.69, respectively. In another study, a 99% accuracy was observed in estimating the course of the disease using more than 1000 histological images collected from three institutions. Finally, in comparing five separate DL networks using 934 cases, Kather et al observed that the hazard ratio was 1.99 in determining overall survival. In studies investigating the microenvironment with AI-based models in these tumors, AUC values ranged from 0.91 to 0.99[47-49]. In another interesting study, Weis et al pointed out that detecting tumor bud hot spots with CNN may influence determining tumor budding, which plays a role in determining tumor behavior. The characteristics of these studies are briefly presented in Table 3. Although this needs to be supported and standardized by further comparative studies, all these findings suggest that AI can be applied for determining the behavior of CRC.
AI-BASED APPLICATIONS FOR GENETIC AND MOLECULAR TESTING IN GC
In routine practice, evaluating surgical and biopsy specimens of GI cancers is essential for identifying molecular biomarkers that predict the response to targeted therapies. This evaluation requires the use of immunohistochemistry or advanced molecular techniques.
The detection of genetic alterations called microsatellite instability (MSI), especially in CRC, is very important for treatment with immunomodulators[51-53]. In addition, it is possible to determine the MSI-related phenotype and identify conditions that require family information and close follow-up of the patient, such as Lynch syndrome. The revelation that some of the genetic events in these cancers are associated with certain morphological events has led to several attempts to use AI-based algorithms in WSIs. Furthermore, due to the large number of samples available, CRC was seen as a prototype for these studies. In this context, accumulated data indicate that AI-based models are influential in determining both MSI and other genotypic changes[47,55-57]. In particular, the DL algorithm developed by Echle et al to detect MSI in CRC using more than 8800 images recently showed an AUC of 0.96 in the multi-institution validation cohort (Table 3).
There have been other attempts to develop models that directly predict gene mutations from the WSI of gastric cancer. In addition, it has been observed that AI could also predict gene expression and RNA-seq data, and these models have remarkable potential for clinical translation[47,56,57] (Table 2).
However, further additional and prospective validation studies are necessary for GI cancers before applying AI in real life to reduce the molecular testing workload and allow testing in health care centers with limited resources.
CHALLENGES AND IMPLEMENTATION OF AI-BASED APPLICATIONS IN REAL-LIFE PRACTICE
In general, the need for a close review of the steps involved in ethics, design, financing, development, validation and regulation, implementation, and impact on the workforce in the application of AI in pathology has been highlighted.
From this perspective, although AI-based models are likely to play a critical role in gastrointestinal pathology, including GC, in the future, several problems similar to those in other fields of pathology need to be addressed to ensure implementation. Brief information about the difficulties encountered in applying AI models in pathology, including GC, and suggested solutions are presented in Table 4.
Table 4 Summary of challenges and suggested solutions in development process of artificial intelligence applications.
Lack of patient’s approval for commercial use
Approval for both research and product development
Design of AI models
Underestimation of end-users’ needs
Collaboration with skate holders
Optimization of data-sets
CNN: Large amounts of images
Augmentation techniques, transfer learning
Rare tumors: Limited number of images
Global data sharing
Variations in preanalytical and analytical phases
AI algorithms to standardize staining, color properties, and WSIs quality
Annotation of data-sets
Interobserver variations in diagnosis
Discrepancies among performances for trained algorithms
Presence of ground truth without objectivity
Multicenter evaluations that include many pathologists and data-set
Lack of current regulatory guidance specific for AI tools
New guidelines and regulations for safer and effective AI tools
Changes in work-flow
Selection of AI applications that will speed up the work-flow
IT infrastructure investment
Augmented microscopy directed to the cloud network service
The relative inexperience of pathologists
Training about AI, integration of AI in medical education
AI applications that lack interpretability ( Black-box)
Constructions of interpretable models, generating attention heat map
Lack of external quality assurance
Sheme for this purpose should be designed
The performance of AI algorithms should be assured for reporting
Although consent can be obtained from patients to use data for research purposes, a lack of approval for commercial use can cause problems in developing AI models. Some researchers argue that this can be resolved by developing a framework for global data sharing by obtaining approvals that convey the possibility of commercial use for research and product development.
Design of AI models
The primary expectation of AI in pathology is to fill gaps and address unmet needs in the daily workflow. These needs mainly include workload-intensive and repetitive procedures, such as calculating tumor necrosis, mitotic count, and lymph node metastases, and diagnosing lesions prone to interobserver variabilities. The main goal to consider in developing AI applications in pathology is to solve a real clinical need. However, the development of models for AI application in this field of medicine involves a variety of stakeholders, including not just pathologists but computer scientists, IT, and pharmaceutical companies, which inevitably leads to different expectations and perspectives. For example, some may have academic publishing purposes, while others may be profitable commercial products. Therefore, an expected solution in pathology may not meet the expectations in finance, leading to the company not preferring to develop. To overcome these challenges and develop AI algorithms that are effectively used in DP, GC, pathologists, academic professionals who can develop technology, and companies that will promote the product must collaborate in harmony.
Development of AI models
Once AI models are designed and built, their development requires an accurate definition of the output, straightforward design of the algorithm, collection of a large follow-up sample or even pilot data, data disclosure and processing, and statistical analysis.
From this perspective, high-quality dataset optimization can be considered one of the biggest obstacles to the development of AI in DP. CNNs require a large number, even thousands, of pathological image datasets, to perform adequately. Especially in rare tumors, the inability to obtain a very high number of images is quite limiting. To overcome this situation, the use of data augmentation techniques and learning methods is recommended. In contrast, Jones et al indicated that small-scale datasets of < 100 digital slides might be sufficient in the case of transfer learning. Recently, it was proposed to develop publicly available datasets for global data sharing. However, it cannot be ruled out that very few such datasets are available in pathology, partially due to privacy, copyright, and financial issues. Although The Cancer Genome Atlas provides many WSIs and associated molecular data, it does not contain enough cases for training AI applications for clinical practice[63,64]. Hartman et al pointed out that another potential source of datasets could be public challenges provided for developing DL algorithms.
Again, developing high performance in AI applications in DP requires training on large datasets, which can be affected by the preanalytical (variations in fixation protocols and variations in the thickness of tissue sections) and analytical (variations in staining techniques and scanning protocols) phases applied to acquire digital images[65,66]. Indeed, converting a glass slide to WSI is not a simple task, and color modifications may influence the accuracy of AI. For this purpose, several AI algorithms have emerged to standardize data in recent years, including staining and color properties[67-69]. In addition, several automated algorithms have also been provided to standardize WSI quality, which automatically detects regions of optimum quality and removes out-of-focus or artifact-related regions, such as DeepFocus[70,71].
Annotation of the dataset
The curation of the dataset should be followed by annotation, which is another complex task. The limits of this annotation are broad, depending on AI, ranging from classification at the slide level to labeling at the pixel level[7,30]. For pathologists, the task of annotating many images is a time-consuming, sometimes challenging effort that can affect the accuracy of the models being trained, especially when the task is complex, especially if, as in gastrointestinal pathology, the disease selected for diagnosis differs significantly among observers (e.g., intramucosal carcinomas) and if the accuracy of dataset descriptions cannot be warranted. Moreover, the trained algorithm may not produce the same performance in the dataset when used in other medical centers. Recently, many efforts have been made to solve the annotation problems that hinder the application of AI in pathology practice[67,73]. The data support that multi-instance learning (MIL) algorithms can be applied without detailed annotation. In particular, there is evidence that MIL can be effective when there is a large dataset and detailed annotations are impossible to obtain.
Validation and regulation
The preparation of the annotated dataset is followed by the model development process (preparation of the datasets for training, testing, and validation) and the selection of the learning method with the ML technique. In this context, the validation of AI-based technologies requires an evidence-based approach, and it is emphasized that analytical validation should also be considered in a laboratory-centered medical discipline, such as pathology[58,73]. Therefore, it is essential to establish steps and criteria for validating new tests according to the standards. For example, to validate the image analysis used to determine the expression of a biomarker, the technique can often be compared to a detailed manual tumor assessment. However, the performance of the AI technique compared with that of pathologists is not straightforward, given the intraobserver and interobserver variability. Today, there are difficulties associated with determining "ground truth" to AI applications. This situation leads to the need for repeated validation of the robustness and reproducibility of AI applications in large and variable patient groups.
There may be a relative lack of validation cohorts in the development of AI-based applications in DP. This shortcoming is also contributed by the potential limitation in sharing histopathological sections. Although the interobserver variability and subjectivity in the evaluations of pathologists also indicate the uncertainty of "ground truth" in this aspect, the best measure to overcome this obstacle may be multicenter evaluations that include more than one pathologist and dataset. From the perspective of GC, the lack of external validation in a substantial number of studies for AI applications may limit the practical use of AI.
Regulation of AI
Although appropriate regulations are necessary for the safe and effective use of AI in pathology, as highlighted by Allen, regulatory approval should be structured to define the risk-benefit balance, reduce potential harm, produce appropriate verification standards, and encourage innovation. On the other hand, the presence of various challenges should not be ignored in this regard.
Various regulatory authorities [such as the FDA, Centers for Medicare and Medicaid Services (CMS), and the European Union Conformité Européenne (EUCE)] are not yet fully prepared for the implementation of AI applications in clinical medicine. As a result, AI-based devices are being controlled by old and potentially outdated guidelines for testing medical devices.
Currently, in the United States, the FDA is working on new regulations to make AI-based devices safer and more effective. On the other hand, appropriate validation for all laboratory tests using human tissue prior to clinical application is required by CMS regardless of FDA approval, and this organization has no specific regulations to validate AI applications. Furthermore, the EUCE reported that in vitro diagnostic medical device directives will be replaced by in vitro diagnostic regulations in May 2022. In addition, it is necessary to take into account the regulatory trends of the country where AI is implemented.
The implementation of AI models in daily pathology practice depends on meeting specific requirements by overcoming various challenges. First, a laboratory infrastructure equipped to enable AI applications in a time frame that does not interfere with patient care is essential. Currently, many pathology laboratories only use tissue sections for diagnostic evaluations. However, the implementation of AI models will require new DP-related equipment, software, a specific data management system, data storage facilities, and, more importantly, a substantial investment to cover their cost. In addition, an institutional IT platform is required to enable practitioners to operate on-site and cloud-based computing systems. Thus, DP applications may require significant investment, hindering the implementation of these technologies. It has been demonstrated that augmented microscopy directly connected to the cloud network service can solve the whole slide scanner setup problem. The cloud-based AI application developed by GOOGLE can also aid in the search for morphologically similar features in a target image, regardless of the annotation status.
The relative inexperience of pathologists with AI-based technologies should not be overlooked. Therefore, pathologists need to improve their knowledge of both the installation of DP systems and the application of AI. Another problem is that, given the reported performance of some algorithms, automated AI models are believed to outperform pathologists, causing pathologists to be hesitant about these applications[79-81]. However, current results suggest that AI models are more likely to help improve the overall quality of pathological diagnosis and provide relevant additional information rather than replacing pathologists[82,83]. Indeed, there will always be a need for pathologists to audit technologies and control systems in AI implementation. Therefore, pathologists must be aware of the long-term risk-benefit balance of AI applications. Since current DL-based AI applications lack interpretability, it may be helpful to develop AI solutions that end-users can interpret, thus providing them with detailed explanations of how their predictions are made. Although DL's "black box" problem has not been fully resolved, several solutions have been reported, such as constructing an interpretable model, generating an attention heatmap, and constructing an external interpretive model[85-88].
While AI assistance in pathological diagnosis may reduce the opportunities for learning diagnostic skills during pathology training, resident pathologists should be trained and encouraged to learn the utility, limitations, and pitfalls of AI application as an adjunct method to improve the quality and precision of clinical diagnoses. Therefore, some reforms may be required in pathology training, starting with medical education followed by a pathology education program to address a more accurate and safer implementation of AI in pathology practice.
Like other clinical tests, quality assurance is an important issue for the effective use of AI in DP, and consequently, a scheme of external quality assurance for applications should be urgently prepared for its implementation. Furthermore, laboratory staff should be aware of the quality management system.
Beyond all this, the legal implications of signing a report prepared by a pathologist using AI should not be ignored. Therefore, to include AI findings in a pathological report, the performance of the algorithm must be assured. This legal issue also supports the notion that AI cannot replace pathologists but that AI can be used to support pathologists in clinical trials.
AI-based approaches have the potential to contribute to the pathological diagnosis and staging of GC by improving workflows, eliminating simple errors, and increasing diagnostic reproducibility. It is also the case that it encourages biomarker discovery by revealing impossible predictions using traditional visual methods. However, there are many hurdles to overcome, including infrastructure and the generalization of algorithms. Overcoming these obstacles requires the efforts of computer scientists, pathologists, and clinicians, who will deal with each challenge separately and cooperate in harmony. In this way, AI applications that are user-friendly, explainable, manageable, and cost-effective can play a crucial role in the development of pathological assessments to be used in the diagnosis, prognosis, and treatment of GC.
Provenance and peer review: Unsolicited article; Externally peer reviewed.
Food and Drug Administration.
IntelliSite Pathology Solution (PIPS, Philips Medical Systems) [cited 29 September 2020]. Available from: https://www.fda.gov/drugs/resources-informationapproved-drugs/intellisite-pathology-solution-pips-philips-medical-systems.
[PubMed] [DOI][Cited in This Article: ]
Bernstam EV, Shireman PK, Meric-Bernstam F, N Zozus M, Jiang X, Brimhall BB, Windham AK, Schmidt S, Visweswaran S, Ye Y, Goodrum H, Ling Y, Barapatre S, Becich MJ. Artificial intelligence in clinical and translational science: Successes, challenges and opportunities.Clin Transl Sci. 2021 epub ahead of print.
[PubMed] [DOI][Cited in This Article: ][Reference Citation Analysis (0)]
Lu F, Chen ZK, Yuan X, Li Q, Du ZD, Luo L, Zhang FY.
MMHG: Multi-modal hypergraph learning for overall survival after D2 gastrectomy for gastric cancer. Proceedings of the 15th Intl Conf on Dependable, Autonomic and Secure Computing, 15th Intl Conf on Pervasive Intelligence and Computing, 3rd Intl Conf on Big Data Intelligence and Computing and Cyber Science and Technology Congress; 2017 Nov 6-10; Orlando, FL, USA. California: IEEE Computer Society, 2017: 164-169.
[PubMed] [DOI][Cited in This Article: ][Cited by in F6Publishing: 1][Reference Citation Analysis (0)]
Kather JN, Krisam J, Charoentong P, Luedde T, Herpel E, Weis CA, Gaiser T, Marx A, Valous NA, Ferber D, Jansen L, Reyes-Aldasoro CC, Zörnig I, Jäger D, Brenner H, Chang-Claude J, Hoffmeister M, Halama N. Predicting survival from colorectal cancer histology slides using deep learning: A retrospective multicenter study.PLoS Med. 2019;16:e1002730.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 155][Cited by in F6Publishing: 114][Article Influence: 51.7][Reference Citation Analysis (0)]
Le DT, Uram JN, Wang H, Bartlett BR, Kemberling H, Eyring AD, Skora AD, Luber BS, Azad NS, Laheru D, Biedrzycki B, Donehower RC, Zaheer A, Fisher GA, Crocenzi TS, Lee JJ, Duffy SM, Goldberg RM, de la Chapelle A, Koshiji M, Bhaijee F, Huebner T, Hruban RH, Wood LD, Cuka N, Pardoll DM, Papadopoulos N, Kinzler KW, Zhou S, Cornish TC, Taube JM, Anders RA, Eshleman JR, Vogelstein B, Diaz LA Jr. PD-1 Blockade in Tumors with Mismatch-Repair Deficiency.N Engl J Med. 2015;372:2509-2520.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 4650][Cited by in F6Publishing: 2768][Article Influence: 664.3][Reference Citation Analysis (0)]
Mandal R, Samstein RM, Lee KW, Havel JJ, Wang H, Krishna C, Sabio EY, Makarov V, Kuo F, Blecua P, Ramaswamy AT, Durham JN, Bartlett B, Ma X, Srivastava R, Middha S, Zehir A, Hechtman JF, Morris LG, Weinhold N, Riaz N, Le DT, Diaz LA Jr, Chan TA. Genetic diversity of tumors with mismatch repair deficiency influences anti-PD-1 immunotherapy response.Science. 2019;364:485-491.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 173][Cited by in F6Publishing: 166][Article Influence: 57.7][Reference Citation Analysis (0)]
Echle A, Grabsch HI, Quirke P, van den Brandt PA, West NP, Hutchins GGA, Heij LR, Tan X, Richman SD, Krause J, Alwers E, Jenniskens J, Offermans K, Gray R, Brenner H, Chang-Claude J, Trautwein C, Pearson AT, Boor P, Luedde T, Gaisa NT, Hoffmeister M, Kather JN. Clinical-Grade Detection of Microsatellite Instability in Colorectal Tumors by Deep Learning.Gastroenterology. 2020;159:1406-1416.e11.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 39][Cited by in F6Publishing: 37][Article Influence: 19.5][Reference Citation Analysis (0)]
Kather JN, Heij LR, Grabsch HI, Loeffler C, Echle A, Muti HS, Krause J, Niehues JM, Sommer KAJ, Bankhead P, Kooreman LFS, Schulte JJ, Cipriani NA, Buelow RD, Boor P, Ortiz-Brüchle NN, Hanby AM, Speirs V, Kochanny S, Patnaik A, Srisuwananukorn A, Brenner H, Hoffmeister M, van den Brandt PA, Jäger D, Trautwein C, Pearson AT, Luedde T. Pan-cancer image-based detection of clinically actionable genetic alterations.Nat Cancer. 2020;1:789-799.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 52][Cited by in F6Publishing: 42][Article Influence: 26.0][Reference Citation Analysis (0)]
Cosatto E, Laquerre PF, Malon C, Graf HP, Saito A, Kiyuna T, Marugame A, Kamijo K.
Automated gastric cancer diagnosis on H and E-stained sections; training a classifier on a large scale with multiple instance machine learning. Proceedings of SPIE - Progress in Biomedical Optics and Imaging, MI: 2013.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 14][Cited by in F6Publishing: 2][Article Influence: 1.6][Reference Citation Analysis (0)]
United States Food and Drug Administration.
Proposed Regulatory Framework for Modifications to Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD). [cited 7 January 2021]. Available from: https://www.fda.gov/media/122535/download.
[PubMed] [DOI][Cited in This Article: ]
Medical Devices – Sector. [cited 7 January 2021]. Available from: https://ec.europa.eu/growth/sectors/medical-devices_en.
[PubMed] [DOI][Cited in This Article: ]
Deshpande S, Minhas F, Graham S, Rajpoot N.
SAFRON: Stitching across the frontier for generating colorectal cancer histology images. [cited 15 March 2021]. Available from: http:// arxiv. org/abs/ 2008. 04526.
[PubMed] [DOI][Cited in This Article: ]
Holzinger A, Malle B, Kieseberg P, Roth PM, Müller H, Reihs R, Zatloukal K.
Towards the augmented pathologist: Challenges of Explainable-AI in digital pathology. [cited 15 March 2021]. Available from: http:// arxiv. org/ abs/ 1712. 06657.
[PubMed] [DOI][Cited in This Article: ]
León F, Gélvez M, Jaimes Z, Gelvez T, Arguello H.
Supervised classification of histopathological images using convolutional neuronal networks for gastric cancer detection. 2019 XXII Symposium on Image, Signal Processing and Artificial Vision (STSIVA); 2019 Apr 24-26; Bucaramanga, Colombia. IEEE, 2019: 1-5.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 5][Cited by in F6Publishing: 3][Article Influence: 1.7][Reference Citation Analysis (0)]
Ponzio F, Macii E, Ficarra E, Di Cataldo S.
Colorectal Cancer Classification using Deep Convolutional Networks-An Experimental Study. Proceedings of the 11th International Joint Conference on Biomedical Engineering Systems and Technologies - Volume 2. Bioimaging, 2018: 58-66.
[PubMed] [DOI][Cited in This Article: ]
Skrede OJ, De Raedt S, Kleppe A, Hveem TS, Liestøl K, Maddison J, Askautrud HA, Pradhan M, Nesheim JA, Albregtsen F, Farstad IN, Domingo E, Church DN, Nesbakken A, Shepherd NA, Tomlinson I, Kerr R, Novelli M, Kerr DJ, Danielsen HE. Deep learning for prediction of colorectal cancer outcome: a discovery and validation study.Lancet. 2020;395:350-360.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 70][Cited by in F6Publishing: 34][Article Influence: 35.0][Reference Citation Analysis (0)]
Alom M, Yakopcic C, Taha T, Asari V.
Microscopic Nuclei Classification, Segmentation and Detection with improved Deep Convolutional Neural Network (DCNN) Approaches. 2018 Preprint. Available from: arXiv:1811.03447.
[PubMed] [DOI][Cited in This Article: ]
Sirinukunwattana K, Domingo E, Richman SD, Redmond KL, Blake A, Verrill C, Leedham SJ, Chatzipli A, Hardy C, Whalley CM, Wu CH, Beggs AD, McDermott U, Dunne PD, Meade A, Walker SM, Murray GI, Samuel L, Seymour M, Tomlinson I, Quirke P, Maughan T, Rittscher J, Koelzer VH. S: CORT consortium. Image-based consensus molecular subtype (imCMS) classification of colorectal cancer using deep learning.Gut. 2021;70:544-554.
[PubMed] [DOI][Cited in This Article: ][Cited by in Crossref: 48][Cited by in F6Publishing: 33][Article Influence: 48.0][Reference Citation Analysis (0)]