1
|
Jeong HJ, Seol A, Lee S, Lim H, Lee M, Oh SJ. Portable Ultrasound Bladder Volume Measurement Over Entire Volume Range Using a Deep Learning Artificial Intelligence Model in a Selected Cohort: A Proof of Principle Study. Neurourol Urodyn 2025. [PMID: 40384598 DOI: 10.1002/nau.70057] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2025] [Revised: 03/19/2025] [Accepted: 04/06/2025] [Indexed: 05/20/2025]
Abstract
OBJECTIVE We aimed to prospectively investigate whether bladder volume measured using deep learning artificial intelligence (AI) algorithms (AI-BV) is more accurate than that measured using conventional methods (C-BV) if using a portable ultrasound bladder scanner (PUBS). PATIENTS AND METHODS Patients who underwent filling cystometry because of lower urinary tract symptoms between January 2021 and July 2022 were enrolled. Every time the bladder was filled serially with normal saline from 0 mL to maximum cystometric capacity in 50 mL increments, C-BV was measured using PUBS. Ultrasound images obtained during this process were manually annotated to define the bladder contour, which was used to build a deep learning AI model. The true bladder volume (T-BV) for each bladder volume range was compared with C-BV and AI-BV for analysis. RESULTS We enrolled 250 patients (213 men and 37 women), and a deep learning AI model was established using 1912 bladder images. There was a significant difference between C-BV (205.5 ± 170.8 mL) and T-BV (190.5 ± 165.7 mL) (p = 0.001), but no significant difference between AI-BV (197.0 ± 161.1 mL) and T-BV (190.5 ± 165.7 mL) (p = 0.081). In bladder volume ranges of 101-150, 151-200, and 201-300 mL, there were significant differences in the percentage of volume differences between [C-BV and T-BV] and [AI-BV and T-BV] (p < 0.05), but no significant difference if converted to absolute values (p > 0.05). C-BV (R2 = 0.91, p < 0.001) and AI-BV (R2 = 0.90, p < 0.001) were highly correlated with T-BV. The mean difference between AI-BV and T-BV (6.5 ± 50.4) was significantly smaller than that between C-BV and T-BV (15.0 ± 50.9) (p = 0.001). CONCLUSION Following image pre-processing, deep learning AI-BV more accurately estimated true BV than conventional methods in this selected cohort on internal validation. Determination of the clinical relevance of these findings and performance in external cohorts requires further study. TRIAL REGISTRATION The clinical trial was conducted using an approved product for its approved indication, so approval from the Ministry of Food and Drug Safety (MFDS) was not required. Therefore, there is no clinical trial registration number.
Collapse
Affiliation(s)
- Hyun Ju Jeong
- Department of Medical Device Development, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Aeran Seol
- Department of Obstetrics and Gynecology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| | - Seungjun Lee
- Department of Obstetrics and Gynecology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
- Department of Obstetrics and Gynecology, Dankook University College of Medicine, Cheonan, Republic of Korea
| | - Hyunji Lim
- Department of Obstetrics and Gynecology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| | - Maria Lee
- Department of Obstetrics and Gynecology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| | - Seung-June Oh
- Department of Medical Device Development, Seoul National University College of Medicine, Seoul, Republic of Korea
- Department of Urology, Seoul National University College of Medicine, Seoul National University Hospital, Seoul, Republic of Korea
| |
Collapse
|
2
|
Tripathi S, Patel J, Mutter L, Dorfner FJ, Bridge CP, Daye D. Large language models as an academic resource for radiologists stepping into artificial intelligence research. Curr Probl Diagn Radiol 2025; 54:342-348. [PMID: 39672727 DOI: 10.1067/j.cpradiol.2024.12.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2024] [Accepted: 12/09/2024] [Indexed: 12/15/2024]
Abstract
BACKGROUND Radiologists increasingly use artificial intelligence (AI) to enhance diagnostic accuracy and optimize workflows. However, many lack the technical skills to effectively apply machine learning (ML) and deep learning (DL) algorithms, limiting the accessibility of these methods to radiology researchers who could otherwise benefit from them. Large language models (LLMs), such as GPT-4o, may serve as virtual advisors, offering tailored algorithm recommendations for specific research needs. This study evaluates GPT-4o's effectiveness as a recommender system to enhance radiologists' understanding and implementation of AI in research. INTERVENTION GPT-4o was used to recommend ML and DL algorithms based on specific details provided by researchers, including dataset characteristics, modality types, data sizes, and research objectives. The model acted as a virtual advisor, guiding researchers in selecting the most appropriate models for their studies. METHODS The study systematically evaluated GPT-4o's recommendations for clarity, task alignment, model diversity, and baseline selection. Responses were graded to assess the model's ability to meet the needs of radiology researchers. RESULTS GPT-4o effectively recommended appropriate ML and DL algorithms for various radiology tasks, including segmentation, classification, and regression in medical imaging. The model suggested a diverse range of established and innovative algorithms, such as U-Net, Random Forest, Attention U-Net, and EfficientNet, aligning well with accepted practices. CONCLUSION GPT-4o shows promise as a valuable tool for radiologists and early career researchers by providing clear and relevant AI and ML algorithm recommendations. Its ability to bridge the knowledge gap in AI implementation could democratize access to advanced technologies, fostering innovation and improving radiology research quality. Further studies should explore integrating LLMs into routine workflows and their role in ongoing professional development.
Collapse
Affiliation(s)
- Satvik Tripathi
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA
| | - Jay Patel
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA
| | - Liam Mutter
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA
| | - Felix J Dorfner
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA; Department of Radiology, Charité - Universitätsmedizin Berlin corporate member of Freie Universität Berlin and Humboldt Universität zu Berlin, Hindenburgdamm 30, 12203, Berlin, Germany
| | - Christopher P Bridge
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA
| | - Dania Daye
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA; Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, USA.
| |
Collapse
|
3
|
Godwin RC, Tung A, Berkowitz DE, Melvin RL. Transforming Physiology and Healthcare through Foundation Models. Physiology (Bethesda) 2025; 40:0. [PMID: 39832521 DOI: 10.1152/physiol.00048.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2024] [Revised: 11/30/2024] [Accepted: 01/08/2025] [Indexed: 01/22/2025] Open
Abstract
Recent developments in artificial intelligence (AI) may significantly alter physiological research and healthcare delivery. Whereas AI applications in medicine have historically been trained for specific tasks, recent technological advances have produced models trained on more diverse datasets with much higher parameter counts. These new, "foundation" models raise the possibility that more flexible AI tools can be applied to a wider set of healthcare tasks than in the past. This review describes how these newer models differ from conventional task-specific AI, which relies heavily on focused datasets and narrow, specific applications. By examining the integration of AI into diagnostic tools, personalized treatment strategies, biomedical research, and healthcare administration, we highlight how these newer models are revolutionizing predictive healthcare analytics and operational workflows. In addition, we address ethical and practical considerations associated with the use of foundation models by highlighting emerging trends, calling for changes to existing guidelines, and emphasizing the importance of aligning AI with clinical goals to ensure its responsible and effective use.
Collapse
Affiliation(s)
- Ryan C Godwin
- Department of Anesthesiology and Perioperative Medicine, University of Alabama at Birmingham, Birmingham, Alabama, United States
| | - Avery Tung
- Department of Anesthesia and Critical Care, University of Chicago, Chicago, Illinois, United States
| | - Dan E Berkowitz
- Department of Anesthesiology and Perioperative Medicine, University of Alabama at Birmingham, Birmingham, Alabama, United States
| | - Ryan L Melvin
- Department of Anesthesiology and Perioperative Medicine, University of Alabama at Birmingham, Birmingham, Alabama, United States
| |
Collapse
|
4
|
Chen L, Li DL, Zheng HF, Qiu CZ. Deep learning and radiomics-driven algorithm for automated identification of May-Thurner syndrome in Iliac CTV imaging. Front Med (Lausanne) 2025; 12:1526144. [PMID: 40365495 PMCID: PMC12069258 DOI: 10.3389/fmed.2025.1526144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Accepted: 03/24/2025] [Indexed: 05/15/2025] Open
Abstract
Objective This research aimed to create a dataset of Iliac CTV scans for automated May-Thurner syndrome (MTS) detection using deep learning and radiomics. In addition, it sought to establish an automated segmentation model for Iliac Vein CTV scans and construct a radiomic signature for MTS diagnosis. Methods We collected a dataset of 490 cases meeting specific inclusion and exclusion criteria, anonymized to comply with HIPAA regulations. Iliac Vein CTV scans were prepared with contrast agent administration, followed by image acquisition and evaluation. A deep learning-based segmentation model, UPerNet, was employed using 10-fold cross-validation. Radiomic features were extracted from the scans and used to construct a diagnostic radiomic signature. Statistical analysis, including Dice values and ROC analysis, was conducted to evaluate segmentation and diagnostic performance. Results The dataset consisted of 201 positive cases of MTS and 289 negative cases. The UPerNet segmentation model exhibited remarkable accuracy in identifying MTS regions. A Dice coefficient of 0.925 (95% confidence interval: 0.875-0.961) was observed, indicating the precision and reliability of our segmentation model. Radiomic analysis produced a diagnostic radiomic signature with significant clinical potential. ROC analysis demonstrated promising results, underscoring the efficacy of the developed model in distinguishing MTS cases. The radiomic signature demonstrated strong diagnostic capabilities for MTS. Within the training dataset, it attained a notable area under the curve (AUC) of 0.891, with a 95% confidence interval ranging from 0.825 to 0.956, showcasing its effectiveness. This diagnostic capability extended to the validation dataset, where the AUC remained strong at 0.892 (95% confidence interval: 0.793-0.991). These results highlight the accuracy of our segmentation model and the diagnostic value of our radiomic signature in identifying MTS cases. Conclusion This study presents a comprehensive approach to automate MTS detection from Iliac CTV scans, combining deep learning and radiomics. The results suggest the potential clinical utility of the developed model in diagnosing MTS, offering a non-invasive and efficient alternative to traditional methods.
Collapse
Affiliation(s)
- Lufeng Chen
- Department of Thoracic and Cardiovascular Surgery, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| | - Dong-Lin Li
- Department of Thoracic and Cardiovascular Surgery, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| | - Hua-Feng Zheng
- Department of Thoracic and Cardiovascular Surgery, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| | - Cheng-Zhi Qiu
- Department of General Surgery, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| |
Collapse
|
5
|
Kim H, Seo KH, Kim K, Shim J, Lee Y. Application and optimization of the U-Net++ model for cerebral artery segmentation based on computed tomographic angiography images. Eur J Radiol 2025; 188:112137. [PMID: 40367559 DOI: 10.1016/j.ejrad.2025.112137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2025] [Revised: 03/20/2025] [Accepted: 04/25/2025] [Indexed: 05/16/2025]
Abstract
Accurate segmentation of cerebral arteries on computed tomography angiography (CTA) images is essential for the diagnosis and management of cerebrovascular diseases, including ischemic stroke. This study implemented a deep learning-based U-Net++ model for cerebral artery segmentation in CTA images, focusing on optimizing pruning levels by analyzing the trade-off between segmentation performance and computational cost. Dual-energy CTA and direct subtraction CTA datasets were utilized to segment the internal carotid and vertebral arteries in close proximity to the bone. We implemented four pruning levels (L1-L4) in the U-Net++ model and evaluated the segmentation performance using accuracy, intersection over union, F1-score, boundary F1-score, and Hausdorff distance. Statistical analyses were conducted to assess the significance of segmentation performance differences across pruning levels. In addition, we measured training and inference times to evaluate the trade-off between segmentation performance and computational efficiency. Applying deep supervision improved segmentation performance across all factors. While the L4 pruning level achieved the highest segmentation performance, L3 significantly reduced training and inference times (by an average of 51.56 % and 22.62 %, respectively), while incurring only a small decrease in segmentation performance (7.08 %) compared to L4. These results suggest that L3 achieves an optimal balance between performance and computational cost. This study demonstrates that pruning levels in U-Net++ models can be optimized to reduce computational cost while maintaining effective segmentation performance. By simplifying deep learning models, this approach can improve the efficiency of cerebrovascular segmentation, contributing to faster and more accurate diagnoses in clinical settings.
Collapse
Affiliation(s)
- Hajin Kim
- Department of Health Science, General Graduate School of Gachon University, 191, Hambakmoero, Yeonsu-gu, Incheon 21936, Republic of Korea
| | - Kang-Hyeon Seo
- Department of Health Science, General Graduate School of Gachon University, 191, Hambakmoero, Yeonsu-gu, Incheon 21936, Republic of Korea
| | - Kyuseok Kim
- Institute of Human Convergence Health Science, Gachon University, 191, Hambakmoero, Yeonsu-gu, Incheon 21936, Republic of Korea
| | - Jina Shim
- Department of Radiotechnology, Wonkwang Health Science University, 514, Iksan-daero, Iksan-si, Jeonbuk-do 54538, Republic of Korea
| | - Youngjin Lee
- Department of Radiological Science, Gachon University, 191, Hambakmoero, Yeonsu-gu, Incheon 21936, Republic of Korea.
| |
Collapse
|
6
|
Song J, He M, Ren S, Shen B. An explainable dataset linking facial phenotypes and genes to rare genetic diseases. Sci Data 2025; 12:634. [PMID: 40234471 PMCID: PMC12000290 DOI: 10.1038/s41597-025-04922-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2024] [Accepted: 03/28/2025] [Indexed: 04/17/2025] Open
Abstract
Distinctive facial phenotypes serve as crucial diagnostic markers for many rare genetic diseases. Although AI-driven image recognition achieves high diagnostic accuracy, it often fails to explain its predictions. In this study, we present the Facial phenotype-Gene-Disease Dataset (FGDD), an explainable dataset collected from 509 research publications. It contains 1,147 data records encompassing 197 disease-causing genes, 437 facial phenotypes, and 211 disease entities, with 689 records having disease labels. Each data record represents a patient group and includes demographic information, variation information, and phenotype information. Baseline and explainability validations conducted on FGDD confirmed the dataset's effectiveness. FGDD supports the training of diagnostic models for rare genetic diseases while delivering explainable results, and provides a foundation for exploring intricate connections between genes, diseases, and facial phenotypes.
Collapse
Affiliation(s)
- Jie Song
- Department of Ophthalmology and Institutes for Systems Genetics, Frontiers Science Center for Disease-related Molecular Network, West China Hospital, Sichuan University, Chengdu, 610212, China
| | - Mengqiao He
- Department of Ophthalmology and Institutes for Systems Genetics, Frontiers Science Center for Disease-related Molecular Network, West China Hospital, Sichuan University, Chengdu, 610212, China
| | - Shumin Ren
- Department of Ophthalmology and Institutes for Systems Genetics, Frontiers Science Center for Disease-related Molecular Network, West China Hospital, Sichuan University, Chengdu, 610212, China.
| | - Bairong Shen
- Department of Ophthalmology and Institutes for Systems Genetics, Frontiers Science Center for Disease-related Molecular Network, West China Hospital, Sichuan University, Chengdu, 610212, China.
| |
Collapse
|
7
|
Laufer B, Abdulbaki Alshirbaji T, Docherty PD, Jalal NA, Krueger-Ziolek S, Moeller K. Tidal Volume Monitoring via Surface Motions of the Upper Body-A Pilot Study of an Artificial Intelligence Approach. SENSORS (BASEL, SWITZERLAND) 2025; 25:2401. [PMID: 40285091 PMCID: PMC12030981 DOI: 10.3390/s25082401] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/31/2025] [Revised: 03/25/2025] [Accepted: 04/08/2025] [Indexed: 04/29/2025]
Abstract
The measurement of tidal volumes via respiratory-induced surface movements of the upper body has been an objective in medical diagnostics for decades, but a real breakthrough has not yet been achieved. The improvement of measurement technology through new, improved sensor systems and the use of artificial intelligence have given this field of research a new dynamic in recent years and opened up new possibilities. Based on the measurement from a motion capture system, the respiration-induced surface motions of 16 test subjects were examined, and specific motion parameters were calculated. Subsequently, linear regression and a tailored convolutional neural network (CNN) were used to determine tidal volumes from an optimal set of motion parameters. The results showed that the linear regression approach, after individual calibration, could be used in clinical applications for 13/16 subjects (mean absolute error < 150 mL), while the CNN approach achieved this accuracy in 5/16 subjects. Here, the individual subject-specific calibration provides significant advantages for the linear regression approach compared to the CNN, which does not require calibration. A larger dataset may allow for greater confidence in the outcomes of the CNN approach. A CNN model trained on a larger dataset would improve performance and may enable clinical use. However, the database of 16 subjects only allows for low-risk use in home care or sports. The CNN approach can currently be used to monitor respiration in home care or competitive sports, while it has the potential to be used in clinical applications if based on a larger dataset that could be gradually built up. Thus, a CNN could provide tidal volumes, the missing parameter in vital signs monitoring, without calibration.
Collapse
Affiliation(s)
- Bernhard Laufer
- Institute of Technical Medicine (ITeM), Furtwangen University, 78056 Villingen-Schwenningen, Germany; (T.A.A.); (S.K.-Z.); (K.M.)
| | - Tamer Abdulbaki Alshirbaji
- Institute of Technical Medicine (ITeM), Furtwangen University, 78056 Villingen-Schwenningen, Germany; (T.A.A.); (S.K.-Z.); (K.M.)
- Innovation Center Computer Assisted Surgery (ICCAS), University of Leipzig, 04103 Leipzig, Germany;
| | - Paul David Docherty
- Department of Mechanical Engineering, University of Canterbury, Christchurch 8140, New Zealand;
| | - Nour Aldeen Jalal
- Innovation Center Computer Assisted Surgery (ICCAS), University of Leipzig, 04103 Leipzig, Germany;
| | - Sabine Krueger-Ziolek
- Institute of Technical Medicine (ITeM), Furtwangen University, 78056 Villingen-Schwenningen, Germany; (T.A.A.); (S.K.-Z.); (K.M.)
| | - Knut Moeller
- Institute of Technical Medicine (ITeM), Furtwangen University, 78056 Villingen-Schwenningen, Germany; (T.A.A.); (S.K.-Z.); (K.M.)
- Department of Mechanical Engineering, University of Canterbury, Christchurch 8140, New Zealand;
| |
Collapse
|
8
|
Shih YC, Ko CL, Wang SY, Chang CY, Lin SS, Huang CW, Cheng MF, Chen CM, Wu YW. Cross-institutional validation of a polar map-free 3D deep learning model for obstructive coronary artery disease prediction using myocardial perfusion imaging: insights into generalizability and bias. Eur J Nucl Med Mol Imaging 2025:10.1007/s00259-025-07243-w. [PMID: 40198356 DOI: 10.1007/s00259-025-07243-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2025] [Accepted: 03/24/2025] [Indexed: 04/10/2025]
Abstract
PURPOSE Deep learning (DL) models for predicting obstructive coronary artery disease (CAD) using myocardial perfusion imaging (MPI) have shown potential for enhancing diagnostic accuracy. However, their ability to maintain consistent performance across institutions and demographics remains uncertain. This study aimed to investigate the generalizability and potential biases of an in-house MPI DL model between two hospital-based cohorts. METHODS We retrospectively included patients from two medical centers in Taiwan who underwent stress/redistribution thallium-201 MPI followed by invasive coronary angiography within 90 days as the reference standard. A polar map-free 3D DL model trained on 928 MPI images from one center to predict obstructive CAD was tested on internal (933 images) and external (3234 images from the other center) validation sets. Diagnostic performance, assessed using area under receiver operating characteristic curves (AUCs), was compared between the internal and external cohorts, demographic groups, and with the performance of stress total perfusion deficit (TPD). RESULTS The model showed significantly lower performance in the external cohort compared to the internal cohort in both patient-based (AUC: 0.713 vs. 0.813) and vessel-based (AUC: 0.733 vs. 0.782) analyses, but still outperformed stress TPD (all p < 0.001). The performance was lower in patients who underwent treadmill stress MPI in the internal cohort and in patients over 70 years old in the external cohort. CONCLUSIONS This study demonstrated adequate performance but also limitations in the generalizability of the DL-based MPI model, along with biases related to stress type and patient age. Thorough validation is essential before the clinical implementation of DL MPI models.
Collapse
Affiliation(s)
- Yu-Cheng Shih
- Department of Nuclear Medicine, Far Eastern Memorial Hospital, New Taipei City, Taiwan
| | - Chi-Lun Ko
- Department of Biomedical Engineering, National Taiwan University, Taipei, Taiwan
- Department of Nuclear Medicine, National Taiwan University Hospital, Taipei, Taiwan
- College of Medicine, National Taiwan University, Taipei, Taiwan
| | - Shan-Ying Wang
- Department of Nuclear Medicine, Far Eastern Memorial Hospital, New Taipei City, Taiwan
- Electrical and Communication Engineering College, Yuan Ze University, Taoyuan, Taiwan
| | - Chen-Yu Chang
- Department of Biomedical Engineering, National Taiwan University, Taipei, Taiwan
| | - Shau-Syuan Lin
- Department of Biomedical Engineering, National Taiwan University, Taipei, Taiwan
| | - Cheng-Wen Huang
- Department of Biomedical Engineering, National Taiwan University, Taipei, Taiwan
| | - Mei-Fang Cheng
- Department of Nuclear Medicine, National Taiwan University Hospital, Taipei, Taiwan
- College of Medicine, National Taiwan University, Taipei, Taiwan
| | - Chung-Ming Chen
- Department of Biomedical Engineering, National Taiwan University, Taipei, Taiwan
| | - Yen-Wen Wu
- Department of Nuclear Medicine, Far Eastern Memorial Hospital, New Taipei City, Taiwan.
- Division of Cardiology, Cardiovascular Center, Far Eastern Memorial Hospital, No. 21, Sec. 2, Nanya S. Rd., Banqiao Dist, New Taipei City, 220216, Taiwan.
- School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan.
- Graduate Institute of Medicine, Yuan Ze University, Taoyuan City, Taiwan.
| |
Collapse
|
9
|
Gao S, Wang X, Xia Z, Zhang H, Yu J, Yang F. Artificial Intelligence in Dentistry: A Narrative Review of Diagnostic and Therapeutic Applications. Med Sci Monit 2025; 31:e946676. [PMID: 40195079 PMCID: PMC11992950 DOI: 10.12659/msm.946676] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 02/11/2025] [Indexed: 04/09/2025] Open
Abstract
Advancements in digital and precision medicine have fostered the rapid development of artificial intelligence (AI) applications, including machine learning, artificial neural networks (ANN), and deep learning, within the field of dentistry, particularly in imaging diagnosis and treatment. This review examines the progress of AI across various domains of dentistry, focusing on its role in enhancing diagnostics and optimizing treatment for oral diseases such as endodontic disease, periodontal disease, oral implantology, orthodontics, prosthodontic treatment, and oral and maxillofacial surgery. Additionally, it discusses the emerging opportunities and challenges associated with these technologies. The findings indicate that AI can be effectively utilized in numerous aspects of oral healthcare, including prevention, early screening, accurate diagnosis, treatment plan design assistance, treatment execution, follow-up monitoring, and prognosis assessment. However, notable challenges persist, including issues related to inaccurate data annotation, limited capability for fine-grained feature expression, a lack of universally applicable models, potential biases in learning algorithms, and legal risks pertaining to medical malpractice and data privacy breaches. Looking forward, future research is expected to concentrate on overcoming these challenges to enhance the accuracy and applicability of AI in diagnosing and treating oral diseases. This review aims to provide a comprehensive overview of the current state of AI in dentistry and to identify pathways for its effective integration into clinical practice.
Collapse
Affiliation(s)
- Sizhe Gao
- Department of Stomatology, Zhejiang Chinese Medical University, Hangzhou, Zhejiang, PR China
| | - Xianyun Wang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, Zhejiang, PR China
| | - Zhuoheng Xia
- Department of Stomatology, Zhejiang Chinese Medical University, Hangzhou, Zhejiang, PR China
| | - Huicong Zhang
- Center for Plastic and Reconstructive Surgery, Department of Stomatology, Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital, Hangzhou Medical College), Hangzhou, Zhejiang, PR China
| | - Jun Yu
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, Zhejiang, PR China
| | - Fan Yang
- Department of Stomatology, Zhejiang Chinese Medical University, Hangzhou, Zhejiang, PR China
| |
Collapse
|
10
|
Xiao X, Zhang J, Shao Y, Liu J, Shi K, He C, Kong D. Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges. SENSORS (BASEL, SWITZERLAND) 2025; 25:2361. [PMID: 40285051 PMCID: PMC12031589 DOI: 10.3390/s25082361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/13/2025] [Revised: 04/03/2025] [Accepted: 04/05/2025] [Indexed: 04/29/2025]
Abstract
The intricate imaging structures, artifacts, and noise present in ultrasound images and videos pose significant challenges for accurate segmentation. Deep learning has recently emerged as a prominent field, playing a crucial role in medical image processing. This paper reviews ultrasound image and video segmentation methods based on deep learning techniques, summarizing the latest developments in this field, such as diffusion and segment anything models as well as classical methods. These methods are classified into four main categories based on the characteristics of the segmentation methods. Each category is outlined and evaluated in the corresponding section. We provide a comprehensive overview of deep learning-based ultrasound image segmentation methods, evaluation metrics, and common ultrasound datasets, hoping to explain the advantages and disadvantages of each method, summarize its achievements, and discuss challenges and future trends.
Collapse
Affiliation(s)
- Xiaolong Xiao
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Jianfeng Zhang
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- Puyang Institute of Big Data and Artificial Intelligence, Puyang 457006, China
| | - Yuan Shao
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Jialong Liu
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Kaibing Shi
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
| | - Chunlei He
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
| | - Dexing Kong
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China
| |
Collapse
|
11
|
Ramezani F, Azimi H, Delfanian B, Amanollahi M, Saeidian J, Masoumi A, Farrokhpour H, Khalili Pour E, Khodaparast M. Classification of ocular surface diseases: Deep learning for distinguishing ocular surface squamous neoplasia from pterygium. Graefes Arch Clin Exp Ophthalmol 2025:10.1007/s00417-025-06804-x. [PMID: 40186633 DOI: 10.1007/s00417-025-06804-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 02/10/2025] [Accepted: 03/11/2025] [Indexed: 04/07/2025] Open
Abstract
PURPOSE Given the significance and potential risks associated with Ocular Surface Squamous Neoplasia (OSSN) and the importance of its differentiation from other conditions, we aimed to develop a Deep Learning (DL) model differentiating OSSN from pterygium (PTG) using slit photographs. METHODS A dataset comprising slit photographs of 162 patients including 77 images of OSSN and 85 images of PTG was assembled. After manual segmentation of the images, a Python-based transfer learning approach utilizing the EfficientNet B7 network was employed for automated image segmentation. GoogleNet, a pre-trained neural network was used to categorize the images into OSSN or PTG. To evaluate the performance of our DL model, K-Fold 10 Cross Validation was implemented, and various performance metrics were measured. RESULTS There was a statistically significant difference in mean age between the OSSN (63.23 ± 13.74 years) and PTG groups (47.18 ± 11.53) (P-value =.000). Furthermore, 84.41% of patients in the OSSN group and 80.00% of the patients in the PTG group were male. Our classification model, trained on automatically segmented images, demonstrated reliable performance measures in distinguishing OSSN from PTG, with an Area Under Curve (AUC) of 98%, sensitivity, F1 score, and accuracy of 94%, and a Matthews Correlation Coefficient (MCC) of 88%. CONCLUSIONS This study presents a novel DL model that effectively segments and classifies OSSN from PTG images with a relatively high accuracy. In addition to its clinical use, this model can be potentially used as a telemedicine application.
Collapse
Affiliation(s)
- Farshid Ramezani
- Clinical Research Development Center, Imam Khomeini, Mohammad Kermanshahi and Farabi Hospitals, Kermanshah University of Medical Sciences, Kermanshah, Iran
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran
| | - Hossein Azimi
- Faculty of Mathematical Sciences and Computer, Kharazmi University, No. 50, Taleghani Avenue, Tehran, Iran
| | - Behrouz Delfanian
- Faculty of Mathematical Sciences and Computer, Kharazmi University, No. 50, Taleghani Avenue, Tehran, Iran
| | - Mobina Amanollahi
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran
| | - Jamshid Saeidian
- Faculty of Mathematical Sciences and Computer, Kharazmi University, No. 50, Taleghani Avenue, Tehran, Iran
| | - Ahmad Masoumi
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran
| | - Hossein Farrokhpour
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran
| | - Elias Khalili Pour
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran.
- Retina Service, Farabi Eye Hospital, Tehran University of Medical Sciences, South Kargar Street, Qazvin Square, Qazvin Street, Tehran, Iran.
| | - Mehdi Khodaparast
- Translational Ophthalmology Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
12
|
ElShebiny T, Abdelrauof D, Elattar M, Motro M, Retrouvey JM, El-Dawlatly M, Mostafa Y, AlHazmi A, Palomo JM. A deep learning model for multiclass tooth segmentation on cone-beam computed tomography scans. Am J Orthod Dentofacial Orthop 2025:S0889-5406(25)00101-5. [PMID: 40186597 DOI: 10.1016/j.ajodo.2025.02.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2024] [Revised: 02/17/2025] [Accepted: 02/19/2025] [Indexed: 04/07/2025]
Abstract
INTRODUCTION Machine learning, a common artificial intelligence technology in medical image analysis, enables computers to learn statistical patterns from pairs of data and annotated labels. Supervised learning in machine learning allows the computer to predict how a specific anatomic structure should be segmented in new patients. This study aimed to develop and validate a deep learning algorithm that automatically creates 3-dimensional surface models of human teeth from a cone-beam computed tomography scan. METHODS A multiresolution dataset, including 216 × 272 × 272, 512 × 512 × 512, and 576 × 768 × 768. Ground truth labels for teeth segmentation were generated. Random partitioning was applied to allocate 140 patients to the training set, 40 to the validation set, and 30 scans for testing and model performance evaluation. Different evaluation metrics were used for assessment. RESULTS Our teeth identification model has achieved an accuracy of 87.92% ± 4.43% on the test set. The general (binary) teeth segmentation model achieved a notably higher accuracy, segmenting the teeth with 93.16% ± 1.18%. CONCLUSIONS The success of our model not only validates the efficacy of using artificial intelligence for dental imaging analysis but also sets a promising foundation for future advancements in automated and precise dental segmentation techniques.
Collapse
Affiliation(s)
- Tarek ElShebiny
- Department of Orthodontics, Case Western Reserve University, Cleveland, Ohio.
| | | | - Mustafa Elattar
- Department of Computer Science, School of Computer Science, Nile University, Giza, Egypt
| | - Melih Motro
- Department of Orthodontics and Dentofacial Orthopedics, Henry M. Goldman School of Dental Medicine, Boston, Mass
| | | | | | - Yehia Mostafa
- Department of Orthodontics, Future University, Cairo, Egypt
| | - Anwar AlHazmi
- Department of Preventive Dental Sciences, Jazan University, Jazan, Saudi Arabia
| | - Juan Martin Palomo
- Department of Orthodontics, Case Western Reserve University, Cleveland, Ohio
| |
Collapse
|
13
|
Lo Mastro A, Grassi E, Berritto D, Russo A, Reginelli A, Guerra E, Grassi F, Boccia F. Artificial intelligence in fracture detection on radiographs: a literature review. Jpn J Radiol 2025; 43:551-585. [PMID: 39538068 DOI: 10.1007/s11604-024-01702-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2024] [Accepted: 11/04/2024] [Indexed: 11/16/2024]
Abstract
Fractures are one of the most common reasons of admission to emergency department affecting individuals of all ages and regions worldwide that can be misdiagnosed during radiologic examination. Accurate and timely diagnosis of fracture is crucial for patients, and artificial intelligence that uses algorithms to imitate human intelligence to aid or enhance human performs is a promising solution to address this issue. In the last few years, numerous commercially available algorithms have been developed to enhance radiology practice and a large number of studies apply artificial intelligence to fracture detection. Recent contributions in literature have described numerous advantages showing how artificial intelligence performs better than doctors who have less experience in interpreting musculoskeletal X-rays, and assisting radiologists increases diagnostic accuracy and sensitivity, improves efficiency, and reduces interpretation time. Furthermore, algorithms perform better when they are trained with big data on a wide range of fracture patterns and variants and can provide standardized fracture identification across different radiologist, thanks to the structured report. In this review article, we discuss the use of artificial intelligence in fracture identification and its benefits and disadvantages. We also discuss its current potential impact on the field of radiology and radiomics.
Collapse
Affiliation(s)
- Antonio Lo Mastro
- Department of Radiology, University of Campania "Luigi Vanvitelli", Naples, Italy.
| | - Enrico Grassi
- Department of Orthopaedics, University of Florence, Florence, Italy
| | - Daniela Berritto
- Department of Clinical and Experimental Medicine, University of Foggia, Foggia, Italy
| | - Anna Russo
- Department of Radiology, University of Campania "Luigi Vanvitelli", Naples, Italy
| | - Alfonso Reginelli
- Department of Radiology, University of Campania "Luigi Vanvitelli", Naples, Italy
| | - Egidio Guerra
- Emergency Radiology Department, "Policlinico Riuniti Di Foggia", Foggia, Italy
| | - Francesca Grassi
- Department of Radiology, University of Campania "Luigi Vanvitelli", Naples, Italy
| | - Francesco Boccia
- Department of Radiology, University of Campania "Luigi Vanvitelli", Naples, Italy
| |
Collapse
|
14
|
Bhatt VD, Shah N, Bhatt DC, Dabir S, Sheth J, Berendschot TTJM, Erckens RJ, Webers CAB. Development and Evaluation of a Deep Learning Algorithm to Differentiate Between Membranes Attached to the Optic Disc on Ultrasonography. Clin Ophthalmol 2025; 19:939-947. [PMID: 40125481 PMCID: PMC11929409 DOI: 10.2147/opth.s501316] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2024] [Accepted: 03/10/2025] [Indexed: 03/25/2025] Open
Abstract
Purpose The purpose of this study was to create and test a deep learning algorithm that could identify and distinguish between membranes attached to optic disc [OD; retinal detachment (RD)/posterior vitreous detachment (PVD)] based on ocular ultrasonography (USG). Patients and Methods We obtained a database of B-scan ultrasonography from a high-volume imaging center. A transformer-based Vision Transformer (ViT) model was employed, pre-trained on ImageNet21K, to classify ultrasound B-scan images into healthy, RD, and PVD. Images were pre-processed using Hugging Face's AutoImage Processor for standardization. Labels were mapped to numerical values, and the dataset was split into training and validation (505 samples), and testing (212 samples) subsets to evaluate model performance. Alternate methods, such as ensemble strategies and object detection pipelines, were explored but showed limited improvement in classification accuracy. Results The AI model demonstrated high classification performance, achieving an accuracy of 98.21% for PVD, 97.22% for RD, and 95.83% for normal cases. Sensitivity was 98.21% for PVD, 96.55% for RD, and 92.86% for normal cases, while specificity reached 95.16%, 100%, and 95.42%, respectively. Despite the overall strong performance, some misclassification occurred, with seven instances of RD being incorrectly labeled as PVD. Conclusion We developed a transformer-based deep learning algorithm for ocular ultrasonography that accurately identifies membranes attached to the optic disc, distinguishing between RD (97.22% accuracy) and PVD (98.21% accuracy). Despite seven misclassifications, our model demonstrates robust performance and enhances diagnostic efficiency in high-volume imaging settings, thereby facilitating timely referrals and ultimately improving patient outcomes in urgent care scenarios. Overall, this promising innovation shows potential for clinical adoption.
Collapse
Affiliation(s)
| | - Nikhil Shah
- Pursuing Masters in Computer Science at Stevens Institute of Technology, Jersey City, NJ, USA
| | | | - Supriya Dabir
- Department of Retina, Rajan Eye Care Pvt Ltd, Chennai, India
| | - Jay Sheth
- Shantilal Shanghvi Eye Institute, Mumbai, India
| | | | - Roel J Erckens
- University Eye Clinic Maastricht, Maastricht, the Netherlands
| | | |
Collapse
|
15
|
Cadham CJ, Reicher J, Muelly M, Hutton DW. Cost-effectiveness of novel diagnostic tools for idiopathic pulmonary fibrosis in the United States. BMC Health Serv Res 2025; 25:385. [PMID: 40089758 PMCID: PMC11909868 DOI: 10.1186/s12913-025-12506-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Accepted: 03/01/2025] [Indexed: 03/17/2025] Open
Abstract
OBJECTIVES Novel non-invasive machine learning algorithms may improve accuracy and reduce the need for biopsy when diagnosing idiopathic pulmonary fibrosis (IPF). We conducted a cost-effectiveness analysis of diagnostic strategies for IPF. METHODS We developed a decision analytic model to evaluate diagnostic strategies for IPF in the United States. To assess the full spectrum of costs and benefits, we compared four interventions: a machine learning diagnostic algorithm, a genomic classifier, a biopsy-all strategy, and a treat-all strategy. The analysis was conducted from the health sector perspective with a lifetime horizon. The primary outcome measures were costs, Quality-Adjusted Life-Years (QALYs) gained, and Incremental Cost-Effectiveness Ratios (ICERs) based on the average of 10,000 probabilistic runs of the model. RESULTS Compared to a biopsy-all strategy the machine learning algorithm and genomic classifer reduced diagnostic-related costs by $14,876 and $3,884, respectively. Use of the machine learning algorithm consistently reduced diagnostic costs. When including downstream treatment costs and benefits of anti-fibrotic treatment, the machine learning algorithm had an ICER of $331,069 per QALY gained compared to the biopsy-all strategy. The genomic classifier had a higher ICER of $390,043 per QALY gained, while the treat-all strategy had the highest ICER of $3,245,403 per QALY gained. Results were sensitive to changes in various input parameters including IPF treatment costs, sensitivity and specificity of novel screening tools, and the rate of additional diagnostics following inconclusive results. High treatment costs were found to drive overall cost regardless of the diagnostic method. As treatment costs lowered, the supplemental diagnostic tools became increasingly cost-effective. CONCLUSIONS Novel tools for diagnosing IPF reduced diagnostic costs, while overall incremental cost-effectiveness ratios were high due to treatment costs. New IPF diagnosis approaches may become more favourable with lower-cost treatments for IPF.
Collapse
Affiliation(s)
- Christopher J Cadham
- Department of Health Management and Policy, School of Public Health, University of Michigan, 1420 Washington Heights, Ann Arbor, MI, 48109-2013, USA.
| | | | | | - David W Hutton
- Department of Health Management and Policy, School of Public Health, University of Michigan, 1420 Washington Heights, Ann Arbor, MI, 48109-2013, USA
| |
Collapse
|
16
|
Mirfendereski P, Li GY, Pearson AT, Kerr AR. Artificial intelligence and the diagnosis of oral cavity cancer and oral potentially malignant disorders from clinical photographs: a narrative review. FRONTIERS IN ORAL HEALTH 2025; 6:1569567. [PMID: 40130020 PMCID: PMC11931071 DOI: 10.3389/froh.2025.1569567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2025] [Accepted: 02/25/2025] [Indexed: 03/26/2025] Open
Abstract
Oral cavity cancer is associated with high morbidity and mortality, particularly with advanced stage diagnosis. Oral cavity cancer, typically squamous cell carcinoma (OSCC), is often preceded by oral potentially malignant disorders (OPMDs), which comprise eleven disorders with variable risks for malignant transformation. While OPMDs are clinical diagnoses, conventional oral exam followed by biopsy and histopathological analysis is the gold standard for diagnosis of OSCC. There is vast heterogeneity in the clinical presentation of OPMDs, with possible visual similarities to early-stage OSCC or even to various benign oral mucosal abnormalities. The diagnostic challenge of OSCC/OPMDs is compounded in the non-specialist or primary care setting. There has been significant research interest in technology to assist in the diagnosis of OSCC/OPMDs. Artificial intelligence (AI), which enables machine performance of human tasks, has already shown promise in several domains of medical diagnostics. Computer vision, the field of AI dedicated to the analysis of visual data, has over the past decade been applied to clinical photographs for the diagnosis of OSCC/OPMDs. Various methodological concerns and limitations may be encountered in the literature on OSCC/OPMD image analysis. This narrative review delineates the current landscape of AI clinical photograph analysis in the diagnosis of OSCC/OPMDs and navigates the limitations, methodological issues, and clinical workflow implications of this field, providing context for future research considerations.
Collapse
Affiliation(s)
- Payam Mirfendereski
- Departmment of Oral and Maxillofacial Pathology, Radiology, and Medicine, New York University College of Dentistry, New York, NY, United States
| | - Grace Y. Li
- Department of Medicine, Section of Hematology/Oncology, University of Chicago Medical Center, Chicago, IL, United States
| | - Alexander T. Pearson
- Department of Medicine, Section of Hematology/Oncology, University of Chicago Medical Center, Chicago, IL, United States
| | - Alexander Ross Kerr
- Departmment of Oral and Maxillofacial Pathology, Radiology, and Medicine, New York University College of Dentistry, New York, NY, United States
| |
Collapse
|
17
|
Blackman J, Veerapen R. On the practical, ethical, and legal necessity of clinical Artificial Intelligence explainability: an examination of key arguments. BMC Med Inform Decis Mak 2025; 25:111. [PMID: 40045339 PMCID: PMC11881432 DOI: 10.1186/s12911-025-02891-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2024] [Accepted: 01/22/2025] [Indexed: 03/09/2025] Open
Abstract
The necessity for explainability of artificial intelligence technologies in medical applications has been widely discussed and heavily debated within the literature. This paper comprises a systematized review of the arguments supporting and opposing this purported necessity. Both sides of the debate within the literature are quoted to synthesize discourse on common recurring themes and subsequently critically analyze and respond to it. While the use of autonomous black box algorithms is compellingly discouraged, the same cannot be said for the whole of medical artificial intelligence technologies that lack explainability. We contribute novel comparisons of unexplainable clinical artificial intelligence tools, diagnosis of idiopathy, and diagnoses by exclusion, to analyze implications on patient autonomy and informed consent. Applying a novel approach using comparisons with clinical practice guidelines, we contest the claim that lack of explainability compromises clinician due diligence and undermines epistemological responsibility. We find it problematic that many arguments in favour of the practical, ethical, or legal necessity of clinical artificial intelligence explainability conflate the use of unexplainable AI with automated decision making, or equate the use of clinical artificial intelligence with the exclusive use of clinical artificial intelligence.
Collapse
Affiliation(s)
- Justin Blackman
- Island Medical Program, Faculty of Medicine, University of British Columbia, University of Victoria, Victoria, BC, Canada.
| | - Richard Veerapen
- Island Medical Program, Faculty of Medicine, University of British Columbia, University of Victoria, Victoria, BC, Canada
- School of Health Information Science, University of Victoria, Victoria, BC, Canada
| |
Collapse
|
18
|
Zamanian M, Irannejad M, Abedi I, Saeb M, Roayaei M. Nested CNN architecture for three-dimensional dose distribution prediction in tomotherapy for prostate cancer. Strahlenther Onkol 2025; 201:306-315. [PMID: 39283345 DOI: 10.1007/s00066-024-02290-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2024] [Accepted: 07/29/2024] [Indexed: 02/21/2025]
Abstract
BACKGROUND The hypothesis of changing network layers to increase the accuracy of dose distribution prediction, instead of expanding their dimensions, which requires complex calculations, has been considered in our study. MATERIALS AND METHODS A total of 137 prostate cancer patients treated with the tomotherapy technique were categorized as 80% training and validating as well as 20% testing for the nested UNet and UNet architectures. Mean absolute error (MAE) was used to measure the dosimetry indices of dose-volume histograms (DVHs), and geometry indices, including the structural similarity index measure (SSIM), dice similarity coefficient (DSC), and Jaccard similarity coefficient (JSC), were used to evaluate the isodose volume (IV) similarity prediction. To verify a statistically significant difference, the two-way statistical Wilcoxon test was used at a level of 0.05 (p < 0.05). RESULTS Use of a nested UNet architecture reduced the predicted dose MAE in DVH indices. The MAE for planning target volume (PTV), bladder, rectum, and right and left femur were D98% = 1.11 ± 0.90; D98% = 2.27 ± 2.85, Dmean = 0.84 ± 0.62; D98% = 1.47 ± 12.02, Dmean = 0.77 ± 1.59; D2% = 0.65 ± 0.70, Dmean = 0.96 ± 2.82; and D2% = 1.18 ± 6.65, Dmean = 0.44 ± 1.13, respectively. Additionally, the greatest geometric similarity was observed in the mean SSIM for UNet and nested UNet (0.91 vs. 0.94, respectively). CONCLUSION The nested UNet network can be considered a suitable network due to its ability to improve the accuracy of dose distribution prediction compared to the UNet network in an acceptable time.
Collapse
Affiliation(s)
- Maryam Zamanian
- Department of Medical Physics, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Maziar Irannejad
- Department of Electrical Engineering, Islamic Azad University Najafabad Branch, Najafabad, Iran
| | - Iraj Abedi
- Department of Medical Physics, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
| | - Mohsen Saeb
- Department of Medical Physics, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
- Department of Radiation Oncology, Isfahan Seyedoshohada Hospital, Isfahan, Iran
| | - Mahnaz Roayaei
- Department of Radiation Oncology, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
| |
Collapse
|
19
|
Mdletshe S, Wang A. Enhancing medical imaging education: integrating computing technologies, digital image processing and artificial intelligence. J Med Radiat Sci 2025; 72:148-155. [PMID: 39508409 PMCID: PMC11909706 DOI: 10.1002/jmrs.837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Accepted: 10/18/2024] [Indexed: 11/15/2024] Open
Abstract
The rapid advancement of technology has brought significant changes to various fields, including medical imaging (MI). This discussion paper explores the integration of computing technologies (e.g. Python and MATLAB), digital image processing (e.g. image enhancement, segmentation and three-dimensional reconstruction) and artificial intelligence (AI) into the undergraduate MI curriculum. By examining current educational practices, gaps and limitations that hinder the development of future-ready MI professionals are identified. A comprehensive curriculum framework is proposed, incorporating essential computational skills, advanced image processing techniques and state-of-the-art AI tools, such as large language models like ChatGPT. The proposed curriculum framework aims to improve the quality of MI education significantly and better equip students for future professional practice and challenges while enhancing diagnostic accuracy, improving workflow efficiency and preparing students for the evolving demands of the MI field.
Collapse
Affiliation(s)
- Sibusiso Mdletshe
- Department of Anatomy and Medical Imaging, Faculty of Medical and Health SciencesThe University of AucklandAucklandNew Zealand
| | - Alan Wang
- Auckland Bioengineering InstituteThe University of AucklandAucklandNew Zealand
- Medical Imaging Research centre, Faculty of Medical and Health SciencesThe University of AucklandAucklandNew Zealand
- Centre for Co‐Created Ageing ResearchThe University of AucklandAucklandNew Zealand
- Centre for Brain ResearchThe University of AucklandAucklandNew Zealand
| |
Collapse
|
20
|
Maragno E, Ricchizzi S, Winter NR, Hellwig SJ, Stummer W, Hahn T, Holling M. Predictive modeling with linear machine learning can estimate glioblastoma survival in months based solely on MGMT-methylation status, age and sex. Acta Neurochir (Wien) 2025; 167:52. [PMID: 39992425 PMCID: PMC11850473 DOI: 10.1007/s00701-025-06441-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2024] [Accepted: 01/21/2025] [Indexed: 02/25/2025]
Abstract
PURPOSE Machine Learning (ML) has become an essential tool for analyzing biomedical data, facilitating the prediction of treatment outcomes and patient survival. However, the effectiveness of ML models heavily relies on both the choice of algorithms and the quality of the input data. In this study, we aimed to develop a novel predictive model to estimate individual survival for patients diagnosed with glioblastoma (GBM), focusing on key variables such as O6-Methylguanine-DNA Methyltransferase (MGMT) methylation status, age, and sex. METHODS To identify the optimal approach, we utilized retrospective data from 218 patients treated at our brain tumor center. The performance of the ML models was evaluated within repeated tenfold regression. The pipeline comprised five regression estimators, including both linear and non-linear algorithms. Permutation feature importance highlighted the feature with the most significant impact on the model. Statistical significance was assessed using a permutation test procedure. RESULTS The best machine learning algorithm achieved a mean absolute error (MAE) of 12.65 (SD = ± 2.18) and an explained variance (EV) of 7% (SD = ± 1.8%) with p < 0.001. Linear algorithms led to more accurate predictions than non-linear estimators. Feature importance testing indicated that age and positive MGMT-methylation influenced the predictions the most. CONCLUSION In summary, here we provide a novel approach allowing to predict GBM patient's survival in months solely based on key parameters such as age, sex and MGMT-methylation status and underscores MGMT-methylation status as key prognostic factor for GBM patients survival.
Collapse
Affiliation(s)
- Emanuele Maragno
- Department of Neurosurgery, University Hospital Münster, Albert-Schweitzer-Campus 1 A, 48149, Münster, Germany
| | - Sarah Ricchizzi
- Department of Neurosurgery, University Hospital Münster, Albert-Schweitzer-Campus 1 A, 48149, Münster, Germany
- Institute for Translational Psychiatry, University of Münster, Münster, Germany
| | - Nils Ralf Winter
- Institute for Translational Psychiatry, University of Münster, Münster, Germany
| | - Sönke Josua Hellwig
- Department of Neurosurgery, University Hospital Münster, Albert-Schweitzer-Campus 1 A, 48149, Münster, Germany
| | - Walter Stummer
- Department of Neurosurgery, University Hospital Münster, Albert-Schweitzer-Campus 1 A, 48149, Münster, Germany
| | - Tim Hahn
- Institute for Translational Psychiatry, University of Münster, Münster, Germany
| | - Markus Holling
- Department of Neurosurgery, University Hospital Münster, Albert-Schweitzer-Campus 1 A, 48149, Münster, Germany.
| |
Collapse
|
21
|
Ali M, Benfante V, Basirinia G, Alongi P, Sperandeo A, Quattrocchi A, Giannone AG, Cabibi D, Yezzi A, Di Raimondo D, Tuttolomondo A, Comelli A. Applications of Artificial Intelligence, Deep Learning, and Machine Learning to Support the Analysis of Microscopic Images of Cells and Tissues. J Imaging 2025; 11:59. [PMID: 39997561 PMCID: PMC11856378 DOI: 10.3390/jimaging11020059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2024] [Revised: 02/08/2025] [Accepted: 02/12/2025] [Indexed: 02/26/2025] Open
Abstract
Artificial intelligence (AI) transforms image data analysis across many biomedical fields, such as cell biology, radiology, pathology, cancer biology, and immunology, with object detection, image feature extraction, classification, and segmentation applications. Advancements in deep learning (DL) research have been a critical factor in advancing computer techniques for biomedical image analysis and data mining. A significant improvement in the accuracy of cell detection and segmentation algorithms has been achieved as a result of the emergence of open-source software and innovative deep neural network architectures. Automated cell segmentation now enables the extraction of quantifiable cellular and spatial features from microscope images of cells and tissues, providing critical insights into cellular organization in various diseases. This review aims to examine the latest AI and DL techniques for cell analysis and data mining in microscopy images, aid the biologists who have less background knowledge in AI and machine learning (ML), and incorporate the ML models into microscopy focus images.
Collapse
Affiliation(s)
- Muhammad Ali
- Ri.MED Foundation, Via Bandiera 11, 90133 Palermo, Italy; (M.A.); (G.B.)
- Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, Molecular and Clinical Medicine, University of Palermo, 90127 Palermo, Italy; (D.D.R.); (A.T.)
| | - Viviana Benfante
- Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, Molecular and Clinical Medicine, University of Palermo, 90127 Palermo, Italy; (D.D.R.); (A.T.)
- Advanced Diagnostic Imaging—INNOVA Project, Department of Radiological Sciences, A.R.N.A.S. Civico, Di Cristina e Benfratelli Hospitals, P.zza N. Leotta 4, 90127 Palermo, Italy;
- Pharmaceutical Factory, La Maddalena S.P.A., Via San Lorenzo Colli, 312/d, 90146 Palermo, Italy;
| | - Ghazal Basirinia
- Ri.MED Foundation, Via Bandiera 11, 90133 Palermo, Italy; (M.A.); (G.B.)
- Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, Molecular and Clinical Medicine, University of Palermo, 90127 Palermo, Italy; (D.D.R.); (A.T.)
| | - Pierpaolo Alongi
- Advanced Diagnostic Imaging—INNOVA Project, Department of Radiological Sciences, A.R.N.A.S. Civico, Di Cristina e Benfratelli Hospitals, P.zza N. Leotta 4, 90127 Palermo, Italy;
- Department of Biomedicine, Neuroscience and Advanced Diagnostics (BiND), University of Palermo, 90127 Palermo, Italy
| | - Alessandro Sperandeo
- Pharmaceutical Factory, La Maddalena S.P.A., Via San Lorenzo Colli, 312/d, 90146 Palermo, Italy;
| | - Alberto Quattrocchi
- Pathologic Anatomy Unit, Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, University of Palermo, 90127 Palermo, Italy; (A.Q.); (A.G.G.); (D.C.)
| | - Antonino Giulio Giannone
- Pathologic Anatomy Unit, Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, University of Palermo, 90127 Palermo, Italy; (A.Q.); (A.G.G.); (D.C.)
| | - Daniela Cabibi
- Pathologic Anatomy Unit, Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, University of Palermo, 90127 Palermo, Italy; (A.Q.); (A.G.G.); (D.C.)
| | - Anthony Yezzi
- Department of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA;
| | - Domenico Di Raimondo
- Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, Molecular and Clinical Medicine, University of Palermo, 90127 Palermo, Italy; (D.D.R.); (A.T.)
| | - Antonino Tuttolomondo
- Department of Health Promotion, Mother and Child Care, Internal Medicine and Medical Specialties, Molecular and Clinical Medicine, University of Palermo, 90127 Palermo, Italy; (D.D.R.); (A.T.)
| | - Albert Comelli
- Ri.MED Foundation, Via Bandiera 11, 90133 Palermo, Italy; (M.A.); (G.B.)
| |
Collapse
|
22
|
Scalia IG, Pathangey G, Abdelnabi M, Ibrahim OH, Abdelfattah FE, Pietri MP, Ibrahim R, Farina JM, Banerjee I, Tamarappoo BK, Arsanjani R, Ayoub C. Applications of Artificial Intelligence for the Prediction and Diagnosis of Cancer Therapy-Related Cardiac Dysfunction in Oncology Patients. Cancers (Basel) 2025; 17:605. [PMID: 40002200 PMCID: PMC11852369 DOI: 10.3390/cancers17040605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2025] [Revised: 02/04/2025] [Accepted: 02/06/2025] [Indexed: 02/27/2025] Open
Abstract
Cardiovascular diseases and cancer are the leading causes of morbidity and mortality in modern society. Expanding cancer therapies that have improved prognosis may also be associated with cardiotoxicity, and extended life span after survivorship is associated with the increasing prevalence of cardiovascular disease. As such, the field of cardio-oncology has been rapidly expanding, with an aim to identify cardiotoxicity and cardiac disease early in a patient who is receiving treatment for cancer or is in survivorship. Artificial intelligence is revolutionizing modern medicine with its ability to identify cardiac disease early. This article comprehensively reviews applications of artificial intelligence specifically applied to electrocardiograms, echocardiography, cardiac magnetic resonance imaging, and nuclear imaging to predict cardiac toxicity in the setting of cancer therapies, with a view to reduce early complications and cardiac side effects from cancer therapies such as chemotherapy, radiation therapy, or immunotherapy.
Collapse
Affiliation(s)
- Isabel G. Scalia
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Girish Pathangey
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Mahmoud Abdelnabi
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Omar H. Ibrahim
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Fatmaelzahraa E. Abdelfattah
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Milagros Pereyra Pietri
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Ramzi Ibrahim
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Juan M. Farina
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Imon Banerjee
- Department of Radiology, Mayo Clinic, Phoenix, AZ 85054, USA;
| | - Balaji K. Tamarappoo
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Reza Arsanjani
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| | - Chadi Ayoub
- Department of Cardiovascular Diseases, Mayo Clinic, Phoenix, AZ 85054, USA; (I.G.S.); (M.A.); (O.H.I.); (F.E.A.); (M.P.P.); (R.I.); (J.M.F.); (B.K.T.)
| |
Collapse
|
23
|
Gao R, Peng A, Duan Y, Chen M, Zheng T, Zhang M, Chen L, Sun H. Associations of Postencephalitic Epilepsy Using Multi-Contrast Whole Brain MRI: A Large Self-Supervised Vision Foundation Model Strategy. J Magn Reson Imaging 2025. [PMID: 39898495 DOI: 10.1002/jmri.29734] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2024] [Revised: 01/22/2025] [Accepted: 01/23/2025] [Indexed: 02/04/2025] Open
Abstract
BACKGROUND Postencephalitic epilepsy (PEE) is a severe neurological complication following encephalitis. Early identification of individuals at high risk for PEE is important for timely intervention. PURPOSE To develop a large self-supervised vision foundation model using a big dataset of multi-contrast head MRI scans, followed by fine-tuning with MRI data and follow-up outcomes from patients with PEE to develop a PEE association model. STUDY TYPE Retrospective. POPULATION Fifty-seven thousand six hundred twenty-one contrast-enhanced head MRI scans from 34,871 patients for foundation model construction, and head MRI scans from 144 patients with encephalitis (64 PEE, 80 N-PEE) for the PEE association model. FIELD STRENGTH/SEQUENCE 1.5-T, 3-T, T1-weighted imaging, T2-weighted imaging, fluid attenuated inversion recovery, T1-weighted contrast-enhanced imaging. ASSESSMENT The foundation model was developed using self-supervised learning and cross-contrast context recovery. Patients with encephalitis were monitored for a median of 3.7 years (range 0.7-7.5 years), with epilepsy diagnosed according to International League Against Epilepsy. Occlusion sensitivity mapping highlighted brain regions involved in PEE classifications. Model performance was compared with DenseNet without pre-trained weights. STATISTICAL TESTS Performance was assessed via confusion matrices, accuracy, sensitivity, specificity, precision, F1 score, and area under the receiver operating characteristic curve (AUC). The DeLong test evaluated AUC between the two models (P < 0.05 for statistical significance). RESULTS The PEE association model achieved accuracy, sensitivity, specificity, precision, F1 score, and AUC of 79.3% (95% CI: 0.71-0.92), 92.3% (95% CI: 0.80-1.00), 68.8% (95% CI: 0.55-0.87), 70.6% (95% CI: 0.61-0.90), 80.0% (95% CI: 0.71-0.93), and 81.0% (95% CI: 0.68-0.92), respectively. A significant AUC improvement was found compared to DenseNet (Delong test, P = 0.03). The association model focused on brain regions affected by encephalitis. DATA CONCLUSION Using extensive unlabeled data via self-supervised learning addressed the limitations of supervised tasks with limited data. The fine-tuned foundation model outperformed DenseNet, which was trained exclusively on task data. PLAIN LANGUAGE SUMMARY This research develops a model to assess the occurrence epilepsy after encephalitis, a severe brain inflammation condition. By using over 57,000 brain scans, the study trains a computer program to recognize patterns in brain images. The model analyzes whole-brain scans to identify areas commonly affected by the disease, such as the temporal and frontal lobes. It was tested on data from patients with encephalitis and showed better performance than older methods. The model can assess the risk of secondary epilepsy in patients with encephalitis, allowing doctors to intervene early and improve treatment outcomes for those affected by this condition. EVIDENCE LEVEL 4 TECHNICAL EFFICACY: Stage 1.
Collapse
Affiliation(s)
- Ronghui Gao
- Department of Radiology, West China Hospital, Sichuan University, Chengdu, China
| | - Anjiao Peng
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
| | - Yifei Duan
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
| | - Mengyao Chen
- Huaxi MR Research Center (HMRRC), West China Hospital, Sichuan University, Chengdu, China
| | - Tao Zheng
- IT Center, West China Hospital, Sichuan University, Chengdu, China
| | - Meng Zhang
- NVIDIA Corp, Beijing Representative Office, Beijing, China
| | - Lei Chen
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
- Pazhou Lab, Guangzhou, China
| | - Huaiqiang Sun
- Department of Radiology, West China Hospital, Sichuan University, Chengdu, China
- Huaxi MR Research Center (HMRRC), West China Hospital, Sichuan University, Chengdu, China
| |
Collapse
|
24
|
Mao Y, Imahori K, Fang W, Sugimori H, Kiuch S, Sutherland K, Kamishima T. Artificial Intelligence Quantification of Enhanced Synovium Throughout the Entire Hand in Rheumatoid Arthritis on Dynamic Contrast-Enhanced MRI. J Magn Reson Imaging 2025; 61:771-783. [PMID: 38807358 DOI: 10.1002/jmri.29463] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Revised: 05/13/2024] [Accepted: 05/14/2024] [Indexed: 05/30/2024] Open
Abstract
BACKGROUND Challenges persist in achieving automatic and efficient inflammation quantification using dynamic contrast-enhanced (DCE) MRI in rheumatoid arthritis (RA) patients. PURPOSE To investigate an automatic artificial intelligence (AI) approach and an optimized dynamic MRI protocol for quantifying disease activity in RA in whole hands while excluding arterial pixels. STUDY TYPE Retrospective. SUBJECTS Twelve RA patients underwent DCE-MRI with 27 phases for creating the AI model and tested on images with a variable number of phases from 35 RA patients. FIELD STRENGTH/SEQUENCE 3.0 T/DCE T1-weighted gradient echo sequence (mDixon, water image). ASSESSMENT The model was trained with various DCE-MRI time-intensity number of phases. Evaluations were conducted for similarity between AI segmentation and manual outlining in 51 ROIs with synovitis. The relationship between synovial volume via AI segmentation with rheumatoid arthritis magnetic resonance imaging scoring (RAMRIS) across whole hands was then evaluated. The reference standard was determined by an experienced musculoskeletal radiologist. STATISTICAL TEST Area under the curve (AUC) of receiver operating characteristic (ROC), Dice and Spearman's rank correlation coefficients, and interclass correlation coefficient (ICC). A P-value <0.05 was considered statistically significant. RESULTS A minimum of 15 phases (acquisition time at least 2.5 minutes) was found to be necessary. AUC ranged from 0.941 ± 0.009 to 0.965 ± 0.009. The Dice score was 0.557-0.615. Spearman's correlation coefficients between the AI model and ground truth were 0.884-0.927 and 0.736-0.831, for joint ROIs and whole hands, respectively. The Spearman's correlation coefficient for the additional test set between the model trained with 15 phases and RAMRIS was 0.768. CONCLUSION The AI-based classification model effectively identified synovitis pixels while excluding arteries. The optimal performance was achieved with at least 15 phases, providing a quantitative assessment of inflammatory activity in RA while minimizing acquisition time. EVIDENCE LEVEL 3 TECHNICAL EFFICACY: Stage 2.
Collapse
Affiliation(s)
- Yijun Mao
- Graduate School of Health Sciences, Hokkaido University, Sapporo, Hokkaido, Japan
| | - Kiko Imahori
- Graduate School of Health Sciences, Hokkaido University, Sapporo, Hokkaido, Japan
| | - Wanxuan Fang
- Graduate School of Health Sciences, Hokkaido University, Sapporo, Hokkaido, Japan
| | - Hiroyuki Sugimori
- Faculty of Health Sciences, Hokkaido University, Sapporo, Hokkaido, Japan
| | | | - Kenneth Sutherland
- Global Center for Biomedical Science and Engineering, Hokkaido University, Sapporo, Hokkaido, Japan
| | - Tamotsu Kamishima
- Faculty of Health Sciences, Hokkaido University, Sapporo, Hokkaido, Japan
| |
Collapse
|
25
|
Wang D, Chen R, Wang W, Yang Y, Yu Y, Liu L, Yang F, Cui S. Prediction of short-term adverse clinical outcomes of acute pulmonary embolism using conventional machine learning and deep Learning based on CTPA images. J Thromb Thrombolysis 2025; 58:331-339. [PMID: 39342072 DOI: 10.1007/s11239-024-03044-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 09/09/2024] [Indexed: 10/01/2024]
Abstract
To explore the predictive value of traditional machine learning (ML) and deep learning (DL) algorithms based on computed tomography pulmonary angiography (CTPA) images for short-term adverse outcomes in patients with acute pulmonary embolism (APE). This retrospective study enrolled 132 patients with APE confirmed by CTPA. Thrombus segmentation and texture feature extraction was performed using 3D-Slicer software. The least absolute shrinkage and selection operator (LASSO) algorithm was used for feature dimensionality reduction and selection, with optimal λ values determined using leave-one-fold cross-validation to identify texture features with non-zero coefficients. ML models (logistic regression, random forest, decision tree, support vector machine) and DL models (ResNet 50 and Vgg 19) were used to construct the prediction models. Model performance was evaluated using receiver operating characteristic (ROC) curves and the area under the curve (AUC). The cohort included 84 patients in the good prognosis group and 48 patients in the poor prognosis group. Univariate and multivariate logistic regression analyses showed that diabetes, RV/LV ≥ 1.0, and Qanadli index form independent risk factors predicting poor prognosis in patients with APE(P < 0.05). A total of 750 texture features were extracted, with 4 key features identified through screening. There was a weak positive correlation between texture features and clinical parameters. ROC curves analysis demonstrated AUC values of 0.85 (0.78-0.92), 0.76 (0.67-0.84), and 0.89 (0.83-0.95) for the clinical, texture feature, and combined models, respectively. In the ML models, the random forest model achieved the highest AUC (0.85), and the support vector machine model achieved the lowest AUC (0.62). And the AUCs for the DL models (ResNet 50 and Vgg 19) were 0.91 (95%CI: 0.90-0.92) and 0.94(95%CI: 0.93-0.95), respectively. Vgg 19 model demonstrated exceptional precision (0.93), recall (0.76), specificity (0.95) and F1 score (0.84). Both ML and DL models based on thrombus texture features from CTPA images demonstrated higher predictive efficacy for short-term adverse outcomes in patients with APE, especially the random forest and Vgg 19 models, potentially assisting clinical management in timely interventions to improve patient prognosis.
Collapse
Affiliation(s)
- Dawei Wang
- Department of Thoracic Surgery, The First Affiliated Hospital of Hebei North University, Zhangjiakou, Hebei, 075000, China
| | - Rong Chen
- Hebei North University, Zhangjiakou, Hebei, 075000, China
| | - Wenjiang Wang
- Hebei North University, Zhangjiakou, Hebei, 075000, China
| | - Yue Yang
- Hebei North University, Zhangjiakou, Hebei, 075000, China
| | - Yaxi Yu
- Hebei North University, Zhangjiakou, Hebei, 075000, China
| | - Lan Liu
- Department of Medical Imaging, The First Affiliated Hospital of Hebei North University, 12 Changqing Road, Zhangjiakou, Hebei, 075000, China
| | - Fei Yang
- Department of Medical Imaging, The First Affiliated Hospital of Hebei North University, 12 Changqing Road, Zhangjiakou, Hebei, 075000, China.
| | - Shujun Cui
- Department of Medical Imaging, The First Affiliated Hospital of Hebei North University, 12 Changqing Road, Zhangjiakou, Hebei, 075000, China
| |
Collapse
|
26
|
Burgon A, Zhang Y, Petrick N, Sahiner B, Cha KH, Samala RK. Bias Amplification to Facilitate the Systematic Evaluation of Bias Mitigation Methods. IEEE J Biomed Health Inform 2025; 29:1444-1454. [PMID: 39499598 DOI: 10.1109/jbhi.2024.3491946] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2024]
Abstract
The future of artificial intelligence (AI) safety is expected to include bias mitigation methods from development to application. The complexity and integration of these methods could grow in conjunction with advances in AI and human-AI interactions. Numerous methods are being proposed to mitigate bias, but without a structured way to compare their strengths and weaknesses. In this work, we present two approaches to systematically amplify subgroup performance bias. These approaches allow for the evaluation and comparison of the effectiveness of bias mitigation methods on AI models by varying the degrees of bias, and can be applied to any classification model. We used these approaches to compare four off-the-shelf bias mitigation methods. Both amplification approaches promote the development of learning shortcuts in which the model forms associations between patient attributes and AI output. We demonstrate these approaches in a case study, evaluating bias in the determination of COVID status from chest x-rays. The maximum achieved increase in performance bias, measured as a difference in predicted prevalence, was 72% and 32% for bias between subgroups related to patient sex and race, respectively. These changes in predicted prevalence were not accompanied by substantial changes in the differences in subgroup area under the receiver operating characteristic curves, indicating that the increased bias is due to the formation of learning shortcuts, not a difference in ability to distinguish positive and negative patients between subgroups.
Collapse
|
27
|
Qiu Q, Huang J, Yang Y, Zhao Y, Zhu X, Peng J, Zhu C, Liu S, Peng W, Sun J, Zhang X, Li M, Zhang X, Hu J, Xie Q, Feng Q, Zhang X. Automatic AI tool for opportunistic screening of vertebral compression fractures on chest frontal radiographs: A multicenter study. Bone 2025; 191:117330. [PMID: 39549901 DOI: 10.1016/j.bone.2024.117330] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/20/2024] [Revised: 11/09/2024] [Accepted: 11/10/2024] [Indexed: 11/18/2024]
Abstract
Vertebral compression fractures (VCFs) are the most common type of osteoporotic fractures, yet they are often clinically silent and undiagnosed. Chest frontal radiographs (CFRs) are frequently used in clinical practice and a portion of VCFs can be detected through this technology. This study aimed to develop an automatic artificial intelligence (AI) tool using deep learning (DL) model for the opportunistic screening of VCFs from CFRs. The datasets were collected from four medical centers, comprising 19,145 vertebrae (T6-T12) from 2735 patients. Patients from Center 1, 2 and 3 were divided into the training and internal testing datasets in an 8:2 ratio (n = 2361, with 16,527 vertebrae). Patients from Center 4 were used as the external test dataset (n = 374, with 2618 vertebrae). Model performance was assessed using sensitivity, specificity, accuracy and the area under the curve (AUC). A reader study with five clinicians of different experience levels was conducted with and without AI assistance. In the internal testing dataset, the model achieved a sensitivity of 83.0 % and an AUC of 0.930 at the fracture level. In the external testing dataset, the model demonstrated a sensitivity of 78.4 % and an AUC of 0.942 at the fracture level. The model's sensitivity outperformed that of five clinicians with different levels of experience. Notably, AI assistance significantly improved sensitivity at the patient level for both junior clinicians (from 56.1 % without AI to 81.6 % with AI) and senior clinicians (from 65.0 % to 85.6 %). In conclusion, the automatic AI tool significantly increases clinicians' sensitivity in diagnosing fractures on CFRs, showing great potential for the opportunistic screening of VCFs.
Collapse
Affiliation(s)
- Qianyi Qiu
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Junzhang Huang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China
| | - Yi Yang
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Yinxia Zhao
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Xiongfeng Zhu
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China
| | - Jiayou Peng
- Department of Radiology, Foshan Hospital of Traditional Chinese Medicine, Guangzhou University of Traditional Chinese Medicine, Foshan, China
| | - Cuiling Zhu
- Department of Radiology, Foshan Hospital of Traditional Chinese Medicine, Guangzhou University of Traditional Chinese Medicine, Foshan, China
| | - Shuxue Liu
- Zhongshan Hospital of Traditional Chinese Medicine Affiliated to Guangzhou University of Chinese Medicine, Zhongshan, China
| | - Weiqing Peng
- Zhongshan Hospital of Traditional Chinese Medicine Affiliated to Guangzhou University of Chinese Medicine, Zhongshan, China
| | - Junqi Sun
- Department of Radiology, Yuebei People's Hospital, Shaoguan, China
| | - Xinru Zhang
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - MianWen Li
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Xintao Zhang
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Jiaping Hu
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Qingling Xie
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China.
| | - Xiaodong Zhang
- Department of Medical Imaging, The Third Affiliated Hospital, Southern Medical University, Guangzhou, China.
| |
Collapse
|
28
|
Bai S, Shi L, Yang K. Deep learning in disease vector image identification. PEST MANAGEMENT SCIENCE 2025; 81:527-539. [PMID: 39422093 DOI: 10.1002/ps.8473] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/10/2024] [Revised: 09/25/2024] [Accepted: 09/26/2024] [Indexed: 10/19/2024]
Abstract
Vector-borne diseases (VBDs) represent a critical global public health concern, with approximately 80% of the world's population at risk of one or more VBD. Manual disease vector identification is time-consuming and expert-dependent, hindering disease control efforts. Deep learning (DL), widely used in image, text, and audio tasks, offers automation potential for disease vector identification. This paper explores the substantial potential of combining DL with disease vector identification. Our aim is to comprehensively summarize the current status of DL in disease vector identification, covering data collection, data preprocessing, model construction, evaluation methods, and applications in identification spanning from species classification to object detection and breeding site identification. We also discuss the challenges and possible prospects for DL in disease vector identification for further research. © 2024 Society of Chemical Industry.
Collapse
Affiliation(s)
- Shaowen Bai
- Key Laboratory of National Health and Family Planning Commission on Parasitic Disease Control and Prevention, Jiangsu Provincial Key Laboratory on Parasite and Vector Control Technology, Jiangsu Institute of Parasitic Diseases, Wuxi, China
- School of Public Health, Nanjing Medical University, Nanjing, China
| | - Liang Shi
- Key Laboratory of National Health and Family Planning Commission on Parasitic Disease Control and Prevention, Jiangsu Provincial Key Laboratory on Parasite and Vector Control Technology, Jiangsu Institute of Parasitic Diseases, Wuxi, China
- Fudan University School of Public Health, Shanghai, China
- Key Laboratory of Public Health Safety, Fudan University, Ministry of Education, Shanghai, China
- Fudan University Center for Tropical Disease Research, Shanghai, China
| | - Kun Yang
- Key Laboratory of National Health and Family Planning Commission on Parasitic Disease Control and Prevention, Jiangsu Provincial Key Laboratory on Parasite and Vector Control Technology, Jiangsu Institute of Parasitic Diseases, Wuxi, China
- School of Public Health, Nanjing Medical University, Nanjing, China
| |
Collapse
|
29
|
Canzone A, Belmonte G, Patti A, Vicari DSS, Rapisarda F, Giustino V, Drid P, Bianco A. The multiple uses of artificial intelligence in exercise programs: a narrative review. Front Public Health 2025; 13:1510801. [PMID: 39957989 PMCID: PMC11825809 DOI: 10.3389/fpubh.2025.1510801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2024] [Accepted: 01/13/2025] [Indexed: 02/18/2025] Open
Abstract
Background Artificial intelligence is based on algorithms that enable machines to perform tasks and activities that generally require human intelligence, and its use offers innovative solutions in various fields. Machine learning, a subset of artificial intelligence, concentrates on empowering computers to learn and enhance from data autonomously; this narrative review seeks to elucidate the utilization of artificial intelligence in fostering physical activity, training, exercise, and health outcomes, addressing a significant gap in the comprehension of practical applications. Methods Only Randomized Controlled Trials (RCTs) published in English were included. Inclusion criteria: all RCTs that use artificial intelligence to program, supervise, manage, or assist physical activity, training, exercise, or health programs. Only studies published from January 1, 2014, were considered. Exclusion criteria: all the studies that used robot-assisted, robot-supported, or robotic training were excluded. Results A total of 1772 studies were identified. After the first stage, where the duplicates were removed, 1,004 articles were screened by title and abstract. A total of 24 studies were identified, and finally, after a full-text review, 15 studies were identified as meeting all eligibility criteria for inclusion. The findings suggest that artificial intelligence holds promise in promoting physical activity across diverse populations, including children, adolescents, adults, older adult, and individuals with disabilities. Conclusion Our research found that artificial intelligence, machine learning and deep learning techniques were used: (a) as part of applications to generate automatic messages and be able to communicate with users; (b) as a predictive approach and for gesture and posture recognition; (c) as a control system; (d) as data collector; and (e) as a guided trainer.
Collapse
Affiliation(s)
- Alberto Canzone
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
- Department of Biomedical and Dental Sciences and Morphological and Functional Imaging, University of Messina, Messina, Italy
| | - Giacomo Belmonte
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
| | - Antonino Patti
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
| | - Domenico Savio Salvatore Vicari
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
- Department of Neurosciences, Biomedicine and Movement Sciences, University of Verona, Verona, Italy
| | - Fabio Rapisarda
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
| | - Valerio Giustino
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
| | - Patrik Drid
- Faculty of Sport and Physical Education, University of Novi Sad, Novi Sad, Serbia
| | - Antonino Bianco
- Sport and Exercise Sciences Research Unit, Department of Psychology, Educational Science and Human Movement, University of Palermo, Palermo, Italy
| |
Collapse
|
30
|
Luo L, Wang X, Lin Y, Ma X, Tan A, Chan R, Vardhanabhuti V, Chu WC, Cheng KT, Chen H. Deep Learning in Breast Cancer Imaging: A Decade of Progress and Future Directions. IEEE Rev Biomed Eng 2025; 18:130-151. [PMID: 38265911 DOI: 10.1109/rbme.2024.3357877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/26/2024]
Abstract
Breast cancer has reached the highest incidence rate worldwide among all malignancies since 2020. Breast imaging plays a significant role in early diagnosis and intervention to improve the outcome of breast cancer patients. In the past decade, deep learning has shown remarkable progress in breast cancer imaging analysis, holding great promise in interpreting the rich information and complex context of breast imaging modalities. Considering the rapid improvement in deep learning technology and the increasing severity of breast cancer, it is critical to summarize past progress and identify future challenges to be addressed. This paper provides an extensive review of deep learning-based breast cancer imaging research, covering studies on mammograms, ultrasound, magnetic resonance imaging, and digital pathology images over the past decade. The major deep learning methods and applications on imaging-based screening, diagnosis, treatment response prediction, and prognosis are elaborated and discussed. Drawn from the findings of this survey, we present a comprehensive discussion of the challenges and potential avenues for future research in deep learning-based breast cancer imaging.
Collapse
|
31
|
Cheng C, Wang Y, Zhao J, Wu D, Li H, Zhao H. Deep Learning and Radiomics in Triple-Negative Breast Cancer: Predicting Long-Term Prognosis and Clinical Outcomes. J Multidiscip Healthc 2025; 18:319-327. [PMID: 39866348 PMCID: PMC11762009 DOI: 10.2147/jmdh.s509004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2024] [Accepted: 01/11/2025] [Indexed: 01/28/2025] Open
Abstract
Triple-negative breast cancer (TNBC) is a unique breast cancer subtype characterized by the lack of estrogen receptor (ER), progesterone receptor (PR), and human epidermal growth factor receptor 2 (HER2) expression in tumor cells. TNBC represents about 15% to 20% of all breast cancers and is aggressive and highly malignant. Currently, TNBC diagnosis primarily depends on pathological examination, while treatment efficacy is assessed through imaging, biomarker detection, pathological evaluation, and clinical symptom improvement. Among these, biomarker detection and pathological assessments are invasive, time-intensive procedures that may be difficult for patients with severe comorbidities and high complication risks. Thus, there is an urgent need for new, supportive tools in TNBC diagnosis and treatment. Deep learning and radiomics techniques represent advanced machine learning methodologies and are also emerging outcomes in the medical-engineering field in recent years. They are extensions of conventional imaging diagnostic methods and have demonstrated tremendous potential in image segmentation, reconstruction, recognition, and classification. These techniques hold certain application prospects for the diagnosis of TNBC, assessment of treatment response, and long-term prognosis prediction. This article reviews recent progress in the application of deep learning, ultrasound, MRI, and radiomics for TNBC diagnosis and treatment, based on research from both domestic and international scholars.
Collapse
Affiliation(s)
- Chen Cheng
- Department of Ultrasound, Lianyungang Traditional Chinese Medicine Hospital, Lianyungang, 222004, People’s Republic of China
| | - Yan Wang
- Department of Ultrasound, Lianyungang Municipal Oriental Hospital, Lianyungang, 222046, People’s Republic of China
- Department of Ultrasound, Xuzhou Medical University Affiliated Hospital, Lianyungang, Jiangsu, 222061, People’s Republic of China
| | - Jine Zhao
- Department of Ultrasound, Donghai County People’s Hospital, Lianyungang, Jiangsu, 222300, People’s Republic of China
| | - Di Wu
- Department of Ultrasound, Lianyungang Traditional Chinese Medicine Hospital, Lianyungang, 222004, People’s Republic of China
| | - Honge Li
- Department of Ultrasound, the First People’s Hospital of Lianyungang, Lianyungang, Jiangsu, 222061, People’s Republic of China
| | - Hongyan Zhao
- Department of Ultrasound, Lianyungang Traditional Chinese Medicine Hospital, Lianyungang, 222004, People’s Republic of China
| |
Collapse
|
32
|
Zerunian M, Polidori T, Palmeri F, Nardacci S, Del Gaudio A, Masci B, Tremamunno G, Polici M, De Santis D, Pucciarelli F, Laghi A, Caruso D. Artificial Intelligence and Radiomics in Cholangiocarcinoma: A Comprehensive Review. Diagnostics (Basel) 2025; 15:148. [PMID: 39857033 PMCID: PMC11763775 DOI: 10.3390/diagnostics15020148] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2024] [Revised: 01/01/2025] [Accepted: 01/07/2025] [Indexed: 01/27/2025] Open
Abstract
Cholangiocarcinoma (CCA) is a malignant biliary system tumor and the second most common primary hepatic neoplasm, following hepatocellular carcinoma. CCA still has an extremely high unfavorable prognosis, regardless of type and location, and complete surgical resection remains the only curative therapeutic option; however, due to the underhanded onset and rapid progression of CCA, most patients present with advanced stages at first diagnosis, with only 30 to 60% of CCA patients eligible for surgery. Recent innovations in medical imaging combined with the use of radiomics and artificial intelligence (AI) can lead to improvements in the early detection, characterization, and pre-treatment staging of these tumors, guiding clinicians to make personalized therapeutic strategies. The aim of this review is to provide an overview of how radiological features of CCA can be analyzed through radiomics and with the help of AI for many different purposes, such as differential diagnosis, the prediction of lymph node metastasis, the defining of prognostic groups, and the prediction of early recurrence. The combination of radiomics with AI has immense potential. Still, its effectiveness in practice is yet to be validated by prospective multicentric studies that would allow for the development of standardized radiomics models.
Collapse
Affiliation(s)
- Marta Zerunian
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Tiziano Polidori
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Federica Palmeri
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Stefano Nardacci
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Antonella Del Gaudio
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Benedetta Masci
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Giuseppe Tremamunno
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Michela Polici
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
- PhD School in Translational Medicine and Oncology, Department of Medical and Surgical Sciences and Translational Medicine, Faculty of Medicine and Psychology, Sapienza University of Rome, 00189 Rome, Italy
| | - Domenico De Santis
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Francesco Pucciarelli
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Andrea Laghi
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| | - Damiano Caruso
- Department of Medical Surgical Sciences and Translational Medicine, Sapienza–University of Rome, Radiology Unit–Sant’Andrea University Hospital, 00189 Rome, Italy; (T.P.); (F.P.); (S.N.); (A.D.G.); (B.M.); (G.T.); (M.P.); (D.D.S.); (F.P.); (A.L.); (D.C.)
| |
Collapse
|
33
|
Shen L, Dai B, Dou S, Yan F, Yang T, Wu Y. Estimation of TP53 mutations for endometrial cancer based on diffusion-weighted imaging deep learning and radiomics features. BMC Cancer 2025; 25:45. [PMID: 39789538 PMCID: PMC11715916 DOI: 10.1186/s12885-025-13424-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2024] [Accepted: 01/01/2025] [Indexed: 01/12/2025] Open
Abstract
OBJECTIVES To construct a prediction model based on deep learning (DL) and radiomics features of diffusion weighted imaging (DWI), and clinical variables for evaluating TP53 mutations in endometrial cancer (EC). METHODS DWI and clinical data from 155 EC patients were included in this study, consisting of 80 in the training set, 35 in the test set, and 40 in the external validation set. Radiomics features, convolutional neural network-based DL features, and clinical variables were analyzed. Feature selection was performed using Mann-Whitney U test, LASSO regression, and SelectKBest. Prediction models were established by gaussian process (GP) and decision tree (DT) algorithms and evaluated by the area under the receiver operating characteristic curve (AUC), net reclassification index (NRI), calibration curves, and decision curve analysis (DCA). RESULTS Compared to the DL (AUCtraining = 0.830, AUCtest = 0.779, and AUCvalidation = 0.711), radiomics (AUCtraining = 0.810, AUCtest = 0.710, and AUCvalidation = 0.839), and clinical (AUCtraining = 0.780, AUCtest = 0.685, and AUCvalidation = 0.695) models, the combined model based on the GP algorithm, which consisted of four DL features, five radiomics features, and two clinical variables, not only demonstrated the highest diagnostic efficacy (AUCtraining = 0.949, AUCtest = 0.877, and AUCvalidation = 0.914) but also led to an improvement in risk reclassification of the TP53 mutation (NIRtraining = 66.38%, 56.98%, and 83.48%, NIRtest = 50.72%, 80.43%, and 89.49%, and NIRvalidation = 64.58%, 87.50%, and 120.83%, respectively). In addition, the combined model exhibited good agreement and clinical utility in calibration curves and DCA analyses, respectively. CONCLUSIONS A prediction model based on the GP algorithm and consisting of DL and radiomics features of DWI as well as clinical variables can effectively assess TP53 mutation in EC.
Collapse
Affiliation(s)
- Lei Shen
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China
| | - Bo Dai
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China
| | - Shewei Dou
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China
| | - Fengshan Yan
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China
| | - Tianyun Yang
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China
| | - Yaping Wu
- Department of Radiology, Henan Provincial People's Hospital & Zhengzhou University People's Hospital, Zhengzhou, Henan, China.
| |
Collapse
|
34
|
Li Y, Jiang S, Yang Z, Wang L, Wang L, Zhou Z. Data-Oriented Octree Inverse Hierarchical Order Aggregation Hybrid Transformer-CNN for 3D Medical Segmentation. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-024-01299-0. [PMID: 39777616 DOI: 10.1007/s10278-024-01299-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 09/06/2024] [Accepted: 10/07/2024] [Indexed: 01/11/2025]
Abstract
The hybrid CNN-transformer structures harness the global contextualization of transformers with the local feature acuity of CNNs, propelling medical image segmentation to the next level. However, the majority of research has focused on the design and composition of hybrid structures, neglecting the data structure, which enhance segmentation performance, optimize resource efficiency, and bolster model generalization and interpretability. In this work, we propose a data-oriented octree inverse hierarchical order aggregation hybrid transformer-CNN (nnU-OctTN), which focuses on delving deeply into the data itself to identify and harness potential. The nnU-OctTN employs the U-Net as a foundational framework, with the node aggregation transformer serving as the encoder. Data features are stored within an octree data structure with each node computed autonomously yet interconnected through a block-to-block local information exchange mechanism. Oriented towards multi-resolution feature data map learning, a cross-fusion module has been designed that associates the encoder and decoder in a staggered vertical and horizontal approach. Inspired by nnUNet, our framework automatically adapts network parameters to the dataset instead of using pre-trained weights for initialization. The nnU-OctTN method was evaluated on the BTCV, ACDC, and BraTS datasets and achieved excellent performance with dice score coefficient (DSC) 86.95, 92.82, and 90.61, respectively, demonstrating its generalizability and effectiveness. Cross-fusion module effectiveness and model scalability are validated through ablation experiments on BTCV and Kidney. Extensive qualitative and quantitative experimental results demonstrate that nnU-OctTN achieves high-quality 3D medical segmentation that has competitive performance against current state-of-the-art methods, providing a promising idea for clinical applications.
Collapse
Affiliation(s)
- Yuhua Li
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China
| | - Shan Jiang
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China.
| | - Zhiyong Yang
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China
| | - Lixiang Wang
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China
| | - Liwen Wang
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China
| | - Zeyang Zhou
- Mechanical Engineering Department, Tianjin University, No. 135, Yaguan Road, Haihe Education Park, Jinnan District, Tianjin City, 300350, China
| |
Collapse
|
35
|
Zhao M, Guo H, Cao X, Dai J, Wang Z, Zhao J, Peng C. Bibliometric analysis of research on the application of deep learning to ophthalmology. Quant Imaging Med Surg 2025; 15:852-866. [PMID: 39839016 PMCID: PMC11744151 DOI: 10.21037/qims-24-1340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2024] [Accepted: 12/13/2024] [Indexed: 01/23/2025]
Abstract
Background Recently, deep learning has become a popular area of research, and has revolutionized the diagnosis and prediction of ocular diseases, especially fundus diseases. This study aimed to conduct a bibliometric analysis of deep learning in the field of ophthalmology to describe international research trends and examine the current research directions. Methods This cross-sectional bibliometric analysis examined the development of research on deep learning in the field of ophthalmology and its sub-topics from 2015 to 2024. Visualization of similarities (VOS)-viewer was used to analyze and evaluate 3,055 articles. Data from the articles were collected on a specific date (September 11, 2024) and downloaded from the Web of Science Core Collection (WOSCC) in plain-text format. Results A total of 3,055 relevant articles on the WOSCC published from 2015 to 2024 were included in this analysis. The first article on the application of deep learning to ophthalmology was published in 2015, and the number of articles on the subject has grown significantly since 2019. China was the most productive country (n=1,187), followed by the United States (n=673). Sun Yat-sen University was the institution with the most publications. Cheng and Bogunovic were the most frequently published authors. The following four different clusters were identified based on a co-occurrence cluster analysis of high-frequency keywords: (I) deep learning for the segmentation and feature extraction of ophthalmic images; (II) deep learning for the automatic detection and classification of ophthalmic images; (III) application of deep learning to ophthalmic imaging techniques; and (IV) deep learning for the diagnosis and management of ophthalmic diseases. Conclusions The analysis of fundus images and the clinical application of deep learning techniques have emerged as prominent research areas in the field of ophthalmology. The substantial increase in publications and citations signifies the expanding impact and global collaboration in the application of deep learning research to ophthalmology. By identifying four distinct clusters representing sub-topics in deep learning ophthalmology research, this study contributes to the understanding of current trends and potential future advancements in the field.
Collapse
Affiliation(s)
- Min Zhao
- Department of Ophthalmology, the Fourth Affiliated Hospital of China Medical University, Shenyang, China
| | - Haoxin Guo
- Department of Information Center, the First Hospital of China Medical University, Shenyang, China
| | - Xindan Cao
- Department of Ophthalmology, the Fourth Affiliated Hospital of China Medical University, Shenyang, China
| | - Junshi Dai
- Department of Ophthalmology, the Fourth Affiliated Hospital of China Medical University, Shenyang, China
| | - Zhongqing Wang
- Department of Information Center, the First Hospital of China Medical University, Shenyang, China
| | - Jiangyue Zhao
- Department of Ophthalmology, the Fourth Affiliated Hospital of China Medical University, Shenyang, China
| | - Cheng Peng
- Department of Ophthalmology, the Fourth Affiliated Hospital of China Medical University, Shenyang, China
| |
Collapse
|
36
|
Ozturk L, Laclau C, Boulon C, Mangin M, Braz-Ma E, Constans J, Dari L, Le Hello C. Analysis of nailfold capillaroscopy images with artificial intelligence: Data from literature and performance of machine learning and deep learning from images acquired in the SCLEROCAP study. Microvasc Res 2025; 157:104753. [PMID: 39389419 DOI: 10.1016/j.mvr.2024.104753] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2024] [Revised: 09/04/2024] [Accepted: 10/06/2024] [Indexed: 10/12/2024]
Abstract
OBJECTIVE To evaluate the performance of machine learning and then deep learning to detect a systemic scleroderma (SSc) landscape from the same set of nailfold capillaroscopy (NC) images from the French prospective multicenter observational study SCLEROCAP. METHODS NC images from the first 100 SCLEROCAP patients were analyzed to assess the performance of machine learning and then deep learning in identifying the SSc landscape, the NC images having previously been independently and consensually labeled by expert clinicians. Images were divided into a training set (70 %) and a validation set (30 %). After features extraction from the NC images, we tested six classifiers (random forests (RF), support vector machine (SVM), logistic regression (LR), light gradient boosting (LGB), extreme gradient boosting (XGB), K-nearest neighbors (KNN)) on the training set with five different combinations of the images. The performance of each classifier was evaluated by the F1 score. In the deep learning section, we tested three pre-trained models from the TIMM library (ResNet-18, DenseNet-121 and VGG-16) on raw NC images after applying image augmentation methods. RESULTS With machine learning, performance ranged from 0.60 to 0.73 for each variable, with Hu and Haralick moments being the most discriminating. Performance was highest with the RF, LGB and XGB models (F1 scores: 0.75-0.79). The highest score was obtained by combining all variables and using the LGB model (F1 score: 0.79 ± 0.05, p < 0.01). With deep learning, performance reached a minimum accuracy of 0.87. The best results were obtained with the DenseNet-121 model (accuracy 0.94 ± 0.02, F1 score 0.94 ± 0.02, AUC 0.95 ± 0.03) as compared to ResNet-18 (accuracy 0.87 ± 0.04, F1 score 0.85 ± 0.03, AUC 0.87 ± 0.04) and VGG-16 (accuracy 0.90 ± 0.03, F1 score 0.91 ± 0.02, AUC 0.91 ± 0.04). CONCLUSION By using machine learning and then deep learning on the same set of labeled NC images from the SCLEROCAP study, the highest performances to detect SSc landscape were obtained with deep learning and in particular DenseNet-121. This pre-trained model could therefore be used to automatically interpret NC images in case of suspected SSc. This result nevertheless needs to be confirmed on a larger number of NC images.
Collapse
Affiliation(s)
- Lutfi Ozturk
- CHU de Saint-Etienne, Médecine Vasculaire et Thérapeutique, Saint-Etienne, France.
| | - Charlotte Laclau
- Université Jean Monnet, Laboratoire Hubert Curien, Saint-Etienne, France
| | | | | | - Etheve Braz-Ma
- Université Jean Monnet, Laboratoire Hubert Curien, Saint-Etienne, France
| | | | - Loubna Dari
- CHU St-André, Médecine Vasculaire, Bordeaux, France
| | - Claire Le Hello
- CHU de Saint-Etienne, Médecine Vasculaire et Thérapeutique, Saint-Etienne, France; Université Jean Monnet, CHU Saint-Etienne, Médecine Vasculaire et Thérapeutique, Mines Saint-Etienne, INSERM, SAINBIOSE U1059, Saint-Etienne, France
| |
Collapse
|
37
|
Maletz S, Balagurunathan Y, Murphy K, Folio L, Chima R, Zaheer A, Vadvala H. AI-powered innovations in pancreatitis imaging: a comprehensive literature synthesis. Abdom Radiol (NY) 2025; 50:438-452. [PMID: 39133362 DOI: 10.1007/s00261-024-04512-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2024] [Revised: 07/16/2024] [Accepted: 07/29/2024] [Indexed: 08/13/2024]
Abstract
Early identification of pancreatitis remains a significant clinical diagnostic challenge that impacts patient outcomes. The evolution of quantitative imaging followed by deep learning models has shown great promise in the non-invasive diagnosis of pancreatitis and its complications. We provide an overview of advancements in diagnostic imaging and quantitative imaging methods along with the evolution of artificial intelligence (AI). In this article, we review the current and future states of methodology and limitations of AI in improving clinical support in the context of early detection and management of pancreatitis.
Collapse
Affiliation(s)
- Sebastian Maletz
- University of South Florida Morsani College of Medicine, Tampa, USA
| | | | - Kade Murphy
- University of South Florida Morsani College of Medicine, Tampa, USA
| | - Les Folio
- University of South Florida Morsani College of Medicine, Tampa, USA
- Moffitt Cancer Center, Tampa, USA
| | - Ranjit Chima
- University of South Florida Morsani College of Medicine, Tampa, USA
- Moffitt Cancer Center, Tampa, USA
| | | | - Harshna Vadvala
- University of South Florida Morsani College of Medicine, Tampa, USA.
- Moffitt Cancer Center, Tampa, USA.
| |
Collapse
|
38
|
Chen H, Dreizin D, Gomez C, Zapaishchykova A, Unberath M. Interpretable Severity Scoring of Pelvic Trauma Through Automated Fracture Detection and Bayesian Inference. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:130-141. [PMID: 39037876 PMCID: PMC11783822 DOI: 10.1109/tmi.2024.3428836] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/24/2024]
Abstract
Pelvic ring disruptions result from blunt injury mechanisms and are potentially lethal mainly due to associated injuries and massive pelvic hemorrhage. The severity of pelvic fractures in trauma victims is frequently assessed by grading the fracture according to the Tile AO/OTA classification in whole-body Computed Tomography (CT) scans. Due to the high volume of whole-body CT scans generated in trauma centers, the overall information content of a single whole-body CT scan and low manual CT reading speed, an automatic approach to Tile classification would provide substantial value, e.g., to prioritize the reading sequence of the trauma radiologists or enable them to focus on other major injuries in multi-trauma patients. In such a high-stakes scenario, an automated method for Tile grading should ideally be transparent such that the symbolic information provided by the method follows the same logic a radiologist or orthopedic surgeon would use to determine the fracture grade. This paper introduces an automated yet interpretable pelvic trauma decision support system to assist radiologists in fracture detection and Tile grading. To achieve interpretability despite processing high-dimensional whole-body CT images, we design a neurosymbolic algorithm that operates similarly to human interpretation of CT scans. The algorithm first detects relevant pelvic fractures on CTs with high specificity using Faster-RCNN. To generate robust fracture detections and associated detection (un)certainties, we perform test-time augmentation of the CT scans to apply fracture detection several times in a self-ensembling approach. The fracture detections are interpreted using a structural causal model based on clinical best practices to infer an initial Tile grade. We apply a Bayesian causal model to recover likely co-occurring fractures that may have been rejected initially due to the highly specific operating point of the detector, resulting in an updated list of detected fractures and corresponding final Tile grade. Our method is transparent in that it provides fracture location and types, as well as information on important counterfactuals that would invalidate the system's recommendation. Our approach achieves an AUC of 0.89/0.74 for translational and rotational instability,which is comparable to radiologist performance. Despite being designed for human-machine teaming, our approach does not compromise on performance compared to previous black-box methods.
Collapse
|
39
|
Wu DY, Vo DT, Seiler SJ. For the busy clinical-imaging professional in an AI world: Gaining intuition about deep learning without math. J Med Imaging Radiat Sci 2025; 56:101762. [PMID: 39437625 DOI: 10.1016/j.jmir.2024.101762] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 07/25/2024] [Accepted: 08/25/2024] [Indexed: 10/25/2024]
Abstract
Medical diagnostics comprise recognizing patterns in images, tissue slides, and symptoms. Deep learning algorithms (DLs) are well suited to such tasks, but they are black boxes in various ways. To explain DL Computer-Aided Diagnostic (CAD) results and their accuracy to patients, to manage or drive the direction of future medical DLs, to make better decisions with CAD, etc., clinical professionals may benefit from hands-on, under-the-hood lessons about medical DL. For those who already have some high-level knowledge about DL, the next step is to gain a more-fundamental understanding of DLs, which may help illuminate inside the boxes. The objectives of this Continuing Medical Education (CME) article include:Better understanding can come from relatable medical analogies and personally experiencing quick simulations to observe deep learning in action, akin to the way clinicians are trained to perform other tasks. We developed readily-implementable demonstrations and simulation exercises. We framed the exercises using analogies to breast cancer, malignancy and cancer stage as example diagnostic applications. The simulations revealed a nuanced relationship between DL output accuracy and the quantity and nature of the data. The simulation results provided lessons-learned and implications for the clinical world. Although we focused on DLs for diagnosis, they are similar to DLs for treatment (e.g. radiotherapy) so that treatment providers may also benefit from this tutorial.
Collapse
Affiliation(s)
- Dolly Y Wu
- Volunteer Services, UT Southwestern Medical Center, Dallas, TX, USA.
| | - Dat T Vo
- Department of Radiation Oncology, UT Southwestern Medical Center, Dallas, TX, USA
| | - Stephen J Seiler
- Department of Radiology, UT Southwestern Medical Center, Dallas, TX, USA
| |
Collapse
|
40
|
Şahin E, Tatar OC, Ulutaş ME, Güler SA, Şimşek T, Utkan NZ, Cantürk NZ. Diagnostic Performance of Deep Learning Applications in Hepatocellular Carcinoma Detection Using Computed Tomography Imaging. THE TURKISH JOURNAL OF GASTROENTEROLOGY : THE OFFICIAL JOURNAL OF TURKISH SOCIETY OF GASTROENTEROLOGY 2024; 36:124-130. [PMID: 39760649 PMCID: PMC11851832 DOI: 10.5152/tjg.2024.24538] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/17/2024] [Accepted: 12/16/2024] [Indexed: 01/07/2025]
Abstract
Hepatocellular carcinoma (HCC) is a prevalent cancer that significantly contributes to mortality globally, primarily due to its late diagnosis. Early detection is crucial yet challenging. This study leverages the potential of deep learning (DL) technologies, employing the You Only Look Once (YOLO) architecture, to enhance the detection of HCC in computed tomography (CT) images, aiming to improve early diagnosis and thereby patient outcomes. We used a dataset of 1290 CT images from 122 patients, segmented according to a standard 70:20:10 split for training, validation, and testing phases. The YOLO-based DL model was trained on these images, with subsequent phases for validation and testing to assess the model's diagnostic capabilities comprehensively. The model exhibited exceptional diagnostic accuracy, with a precision of 0.97216, recall of 0.919, and an overall accuracy of 95.35%, significantly surpassing traditional diagnostic approaches. It achieved a specificity of 95.83% and a sensitivity of 94.74%, evidencing its effectiveness in clinical settings and its potential to reduce the rate of missed diagnoses and unnecessary interventions. The implementation of the YOLO architecture for detecting HCC in CT scans has shown substantial promise, indicating that DL models could soon become a standard tool in oncological diagnostics. As artificial intelligence technology continues to evolve, its integration into healthcare systems is expected to advance the accuracy and efficiency of diagnostics in oncology, enhancing early detection and treatment strategies and potentially improving patient survival rates.
Collapse
Affiliation(s)
- Enes Şahin
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
| | - Ozan Can Tatar
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
- Department of Information Systems Engineering, Kocaeli University Faculty of Technology, Kocaeli, Türkiye
| | - Mehmet Eşref Ulutaş
- University of Health Science, Gaziantep City Hospital, General Surgery, Gaziantep, Türkiye
| | - Sertaç Ata Güler
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
| | - Turgay Şimşek
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
| | - Nihat Zafer Utkan
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
| | - Nuh Zafer Cantürk
- Department of General Surgery, Kocaeli University Faculty of Medicine, Kocaeli, Türkiye
| |
Collapse
|
41
|
Wang T, Chen R, Fan N, Zang L, Yuan S, Du P, Wu Q, Wang A, Li J, Kong X, Zhu W. Machine Learning and Deep Learning for Diagnosis of Lumbar Spinal Stenosis: Systematic Review and Meta-Analysis. J Med Internet Res 2024; 26:e54676. [PMID: 39715552 PMCID: PMC11704645 DOI: 10.2196/54676] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2023] [Revised: 10/23/2024] [Accepted: 11/11/2024] [Indexed: 12/25/2024] Open
Abstract
BACKGROUND Lumbar spinal stenosis (LSS) is a major cause of pain and disability in older individuals worldwide. Although increasing studies of traditional machine learning (TML) and deep learning (DL) were conducted in the field of diagnosing LSS and gained prominent results, the performance of these models has not been analyzed systematically. OBJECTIVE This systematic review and meta-analysis aimed to pool the results and evaluate the heterogeneity of the current studies in using TML or DL models to diagnose LSS, thereby providing more comprehensive information for further clinical application. METHODS This review was performed under the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines using articles extracted from PubMed, Embase databases, and Cochrane Library databases. Studies that evaluated DL or TML algorithms assessment value on diagnosing LSS were included, while those with duplicated or unavailable data were excluded. Quality Assessment of Diagnostic Accuracy Studies 2 was used to estimate the risk of bias in each study. The MIDAS module and the METAPROP module of Stata (StataCorp) were used for data synthesis and statistical analyses. RESULTS A total of 12 studies with 15,044 patients reported the assessment value of TML or DL models for diagnosing LSS. The risk of bias assessment yielded 4 studies with high risk of bias, 3 with unclear risk of bias, and 5 with completely low risk of bias. The pooled sensitivity and specificity were 0.84 (95% CI: 0.82-0.86; I2=99.06%) and 0.87 (95% CI 0.84-0.90; I2=98.7%), respectively. The diagnostic odds ratio was 36 (95% CI 26-49), the positive likelihood ratio (LR+) was 6.6 (95% CI 5.1-8.4), and the negative likelihood ratio (LR-) was 0.18 (95% CI 0.16-0.21). The summary receiver operating characteristic curves, the area under the curve of TML or DL models for diagnosing LSS of 0.92 (95% CI 0.89-0.94), indicating a high diagnostic value. CONCLUSIONS This systematic review and meta-analysis emphasize that despite the generally satisfactory diagnostic performance of artificial intelligence systems in the experimental stage for the diagnosis of LSS, none of them is reliable and practical enough to apply in real clinical practice. Further efforts, including optimization of model balance, widely accepted objective reference standards, multimodal strategy, large dataset for training and testing, external validation, and sufficient and scientific report, should be made to bridge the distance between current TML or DL models and real-life clinical applications in future studies. TRIAL REGISTRATION PROSPERO CRD42024566535; https://tinyurl.com/msx59x8k.
Collapse
Affiliation(s)
- Tianyi Wang
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Ruiyuan Chen
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Ning Fan
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Lei Zang
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Shuo Yuan
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Peng Du
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Qichao Wu
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Aobo Wang
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Jian Li
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Xiaochuan Kong
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| | - Wenyi Zhu
- Beijing Chaoyang Hospital, Capital Medical University, Beijing, China
| |
Collapse
|
42
|
Carvalho J, Abdollahzadeh A, Ferrari RJ. Editorial: Deep learning and neuroimage processing in understanding neurological diseases. Front Comput Neurosci 2024; 18:1523973. [PMID: 39749286 PMCID: PMC11693727 DOI: 10.3389/fncom.2024.1523973] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2024] [Accepted: 12/09/2024] [Indexed: 01/04/2025] Open
Affiliation(s)
- Joana Carvalho
- Laboratory of Visual Neuroscience, Faculty of Psychology and Educational Sciences, University of Coimbra, Coimbra, Portugal
- Laboratory of Preclinical MRI, Champalimaud Research, Champalimaud Centre for the Unknown, Lisbon, Portugal
| | - Ali Abdollahzadeh
- A. I. Virtanen Institute for Molecular Sciences, University of Eastern Finland, Kuopio, Finland
- Center for Biomedical Imaging, Department of Radiology, New York University School of Medicine, New York, NY, United States
| | - Ricardo José Ferrari
- Biomedical Image Processing (BIP) Lab, Department of Computing, Federal University of São Carlos, São Carlos, Brazil
| |
Collapse
|
43
|
Lerch L, Huber LS, Kamath A, Pöllinger A, Pahud de Mortanges A, Obmann VC, Dammann F, Senn W, Reyes M. DreamOn: a data augmentation strategy to narrow the robustness gap between expert radiologists and deep learning classifiers. FRONTIERS IN RADIOLOGY 2024; 4:1420545. [PMID: 39758512 PMCID: PMC11696537 DOI: 10.3389/fradi.2024.1420545] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/20/2024] [Accepted: 11/22/2024] [Indexed: 01/07/2025]
Abstract
Purpose Successful performance of deep learning models for medical image analysis is highly dependent on the quality of the images being analysed. Factors like differences in imaging equipment and calibration, as well as patient-specific factors such as movements or biological variability (e.g., tissue density), lead to a large variability in the quality of obtained medical images. Consequently, robustness against the presence of noise is a crucial factor for the application of deep learning models in clinical contexts. Materials and methods We evaluate the effect of various data augmentation strategies on the robustness of a ResNet-18 trained to classify breast ultrasound images and benchmark the performance against trained human radiologists. Additionally, we introduce DreamOn, a novel, biologically inspired data augmentation strategy for medical image analysis. DreamOn is based on a conditional generative adversarial network (GAN) to generate REM-dream-inspired interpolations of training images. Results We find that while available data augmentation approaches substantially improve robustness compared to models trained without any data augmentation, radiologists outperform models on noisy images. Using DreamOn data augmentation, we obtain a substantial improvement in robustness in the high noise regime. Conclusions We show that REM-dream-inspired conditional GAN-based data augmentation is a promising approach to improving deep learning model robustness against noise perturbations in medical imaging. Additionally, we highlight a gap in robustness between deep learning models and human experts, emphasizing the imperative for ongoing developments in AI to match human diagnostic expertise.
Collapse
Affiliation(s)
- Luc Lerch
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
| | - Lukas S. Huber
- Cognition, Perception and Research Methods, Department of Psychology, University of Bern, Bern, Switzerland
- Neural Information Processing Group, Department of Computer Science, University of Tübingen, Tübingen, Germany
| | - Amith Kamath
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Alexander Pöllinger
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Aurélie Pahud de Mortanges
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Verena C. Obmann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Florian Dammann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Walter Senn
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
| | - Mauricio Reyes
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
- Department of Radiation Oncology, University Hospital Bern, University of Bern, Bern, Switzerland
| |
Collapse
|
44
|
Nie W, Jiang Y, Yao L, Zhu X, AL-Danakh AY, Liu W, Chen Q, Yang D. Prediction of bladder cancer prognosis and immune microenvironment assessment using machine learning and deep learning models. Heliyon 2024; 10:e39327. [PMID: 39687145 PMCID: PMC11647853 DOI: 10.1016/j.heliyon.2024.e39327] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2024] [Revised: 10/03/2024] [Accepted: 10/11/2024] [Indexed: 12/18/2024] Open
Abstract
Bladder cancer (BCa) is a heterogeneous malignancy characterized by distinct immune subtypes, primarily due to differences in tumor-infiltrating immune cells and their functional characteristics. Therefore, understanding the tumor immune microenvironment (TIME) landscape in BCa is crucial for prognostic prediction and guiding precision therapy. In this study, we integrated 10 machine learning algorithms to develop an immune-related machine learning signature (IRMLS) and subsequently created a deep learning model to detect the IRMLS subtype based on pathological images. The IRMLS proved to be an independent prognostic factor for overall survival (OS) and demonstrated robust and stable performance (p < 0.01). The high-risk group exhibited an immune-inflamed phenotype, associated with poorer prognosis and higher levels of immune cell infiltration. We further investigated the cancer immune cycle and mutation landscape within the IRMLS model, confirming that the high-risk group is more sensitive to immune checkpoint immunotherapy (ICI) and adjuvant chemotherapy with cisplatin (p = 2.8e-10), docetaxel (p = 8.8e-13), etoposide (p = 1.8e-07), and paclitaxel (p = 6.2e-13). In conclusion, we identified and validated a machine learning-based molecular characteristic, IRMLS, which reflects various aspects of the BCa biological process and offers new insights into personalized precision therapy for BCa patients.
Collapse
Affiliation(s)
- Weihao Nie
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
| | - Yiheng Jiang
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
| | - Luhan Yao
- School of Information and Communication Engineering, Dalian University of Technology, Dalian, China
| | - Xinqing Zhu
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
| | - Abdullah Y. AL-Danakh
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
| | - Wenlong Liu
- School of Information and Communication Engineering, Dalian University of Technology, Dalian, China
| | - Qiwei Chen
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
- Zhongda Hospital, Medical School, Advanced Institute for Life and Health, Southeast University, Nanjing, 210096, China
| | - Deyong Yang
- Department of Urology, First Affiliated Hospital of Dalian Medical University, Dalian, 116021, China
| |
Collapse
|
45
|
Maertens A, Brykman S, Hartung T, Gafita A, Bai H, Hoelzer D, Skoudis E, Paller CJ. Navigating the unseen peril: safeguarding medical imaging in the age of AI. Front Artif Intell 2024; 7:1400732. [PMID: 39717840 PMCID: PMC11665297 DOI: 10.3389/frai.2024.1400732] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2024] [Accepted: 11/13/2024] [Indexed: 12/25/2024] Open
Abstract
In response to the increasing significance of artificial intelligence (AI) in healthcare, there has been increased attention - including a Presidential executive order to create an AI Safety Institute - to the potential threats posed by AI. While much attention has been given to the conventional risks AI poses to cybersecurity, and critical infrastructure, here we provide an overview of some unique challenges of AI for the medical community. Above and beyond obvious concerns about vetting algorithms that impact patient care, there are additional subtle yet equally important things to consider: the potential harm AI poses to its own integrity and the broader medical information ecosystem. Recognizing the role of healthcare professionals as both consumers and contributors to AI training data, this article advocates for a proactive approach in understanding and shaping the data that underpins AI systems, emphasizing the need for informed engagement to maximize the benefits of AI while mitigating the risks.
Collapse
Affiliation(s)
- Alexandra Maertens
- Center for Alternatives to Animal Testing (CAAT), Johns Hopkins Bloomberg School of Public Health, Baltimore, MD, United States
| | - Steve Brykman
- Independent Creative Technologist, Boston, MA, United States
| | - Thomas Hartung
- Center for Alternatives to Animal Testing (CAAT), Johns Hopkins Bloomberg School of Public Health, Baltimore, MD, United States
- CAAT Europe, University of Konstanz, Konstanz, Germany
| | - Andrei Gafita
- Radiology and Radiological Science, Johns Hopkins University School of Medicine, Baltimore, MD, United States
| | - Harrison Bai
- Radiology and Radiological Science, Johns Hopkins University School of Medicine, Baltimore, MD, United States
| | - David Hoelzer
- SANS Technology Institute, Rockville, MD, United States
| | - Ed Skoudis
- SANS Technology Institute, Rockville, MD, United States
| | - Channing Judith Paller
- Department of Oncology, Johns Hopkins University School of Medicine, Baltimore, MD, United States
| |
Collapse
|
46
|
Wang C, Yu Z, Long Z, Zhao H, Wang Z. A few-shot diabetes foot ulcer image classification method based on deep ResNet and transfer learning. Sci Rep 2024; 14:29877. [PMID: 39622873 PMCID: PMC11612188 DOI: 10.1038/s41598-024-80691-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2024] [Accepted: 11/21/2024] [Indexed: 12/06/2024] Open
Abstract
Diabetes foot ulcer (DFU) is one of the common complications of diabetes patients, which may lead to infection, necrosis and even amputation. Therefore, early diagnosis, classification of severity and related treatment are crucial for the patients. Current DFU classification methods often require experienced doctors to manually classify the severity, which is time-consuming and low accuracy. The objective of the study is to propose a few-shot DFU image classification method based on deep residual neural network and transfer learning. Considering the difficulty in obtaining clinical DFU images, it is a few-shot problem. Therefore, the methods include: (1) Data augmentation of the original DFU images by using geometric transformations and random noise; (2) Deep ResNet models selection based on different convolutional layers comparative experiments; (3) DFU classification model training with transfer learning by using the selected pre-trained ResNet model and fine tuning. To verify the proposed classification method, the experiments were performed with the original and augmented datasets by separating three classifications: zero grade, mild grade, severe grade. (1) The datasets were augmented from the original 146 to 3000 image datasets and the corresponding DFU classification's average accuracy from 0.9167 to 0.9867; (2) Comparative experiments were conducted with ResNet18, ResNet34, ResNet50, ResNet101, ResNet152 by using 3000 image datasets, and the average accuracy/loss is 0.9325/0.2927, 0.9276/0.3234, 0.9901/0.1356, 0.9865/0.1427, 0.9790/0.1583 respectively; (3) Based on the augmented 3000 image datasets, it was achieved 0.9867 average accuracy with the DFU classification model, which was trained by the pre-trained ResNet50 and hyper-parameters. The experimental results indicated that the proposed few-shot DFU image classification method based on deep ResNet and transfer learning got very high accuracy, and it is expected to be suitable for low-cost and low-computational terminal equipment, aiming at helping clinical DFU classification efficiently and auxiliary diagnosis.
Collapse
Affiliation(s)
- Cheng Wang
- Shandong Academy of Intelligent Computing Technology, Shandong Institutes of Industrial Technology (SDIIT), Jinan, 250000, China.
- Beijing Key Laboratory of Mobile Computing and Pervasive Device, Institute of Computing Technology (ICT), Beijing, China.
- College of Education Science, Yan'an University, Yan'an, 716000, China.
| | - Zhen Yu
- Qilu University of Technology (Shandong Academy of Sciences), Jinan, 25000, China
| | - Zhou Long
- Beijing Key Laboratory of Mobile Computing and Pervasive Device, Institute of Computing Technology (ICT), Beijing, China
| | - Hui Zhao
- Orthopedics Department, Beijing Chao-Yang Hospital, Capital Medical University, Beijing, 100020, China
| | - Zhenwei Wang
- Orthopedics Department, Beijing Chao-Yang Hospital, Capital Medical University, Beijing, 100020, China.
| |
Collapse
|
47
|
Chrysafi P, Lam B, Carton S, Patell R. From Code to Clots: Applying Machine Learning to Clinical Aspects of Venous Thromboembolism Prevention, Diagnosis, and Management. Hamostaseologie 2024; 44:429-445. [PMID: 39657652 DOI: 10.1055/a-2415-8408] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2024] Open
Abstract
The high incidence of venous thromboembolism (VTE) globally and the morbidity and mortality burden associated with the disease make it a pressing issue. Machine learning (ML) can improve VTE prevention, detection, and treatment. The ability of this novel technology to process large amounts of high-dimensional data can help identify new risk factors and better risk stratify patients for thromboprophylaxis. Applications of ML for VTE include systems that interpret medical imaging, assess the severity of the VTE, tailor treatment according to individual patient needs, and identify VTE cases to facilitate surveillance. Generative artificial intelligence may be leveraged to design new molecules such as new anticoagulants, generate synthetic data to expand datasets, and reduce clinical burden by assisting in generating clinical notes. Potential challenges in the applications of these novel technologies include the availability of multidimensional large datasets, prospective studies and clinical trials to ensure safety and efficacy, continuous quality assessment to maintain algorithm accuracy, mitigation of unwanted bias, and regulatory and legal guardrails to protect patients and providers. We propose a practical approach for clinicians to integrate ML into research, from choosing appropriate problems to integrating ML into clinical workflows. ML offers much promise and opportunity for clinicians and researchers in VTE to translate this technology into the clinic and directly benefit the patients.
Collapse
Affiliation(s)
- Pavlina Chrysafi
- Department of Medicine, Mount Auburn Hospital, Harvard Medical School, Cambridge, Massachusetts, United States
| | - Barbara Lam
- Division of Hemostasis and Thrombosis, Department of Medicine, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, Massachusetts, United States
- Division of Clinical Informatics, Department of Medicine, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, Massachusetts, United States
| | - Samuel Carton
- Department of Computer Science, College of Engineering and Physical Sciences, University of New Hampshire, Durham, New Hampshire, United States
| | - Rushad Patell
- Division of Hemostasis and Thrombosis, Department of Medicine, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, Massachusetts, United States
| |
Collapse
|
48
|
Wu DY, Vo DT, Seiler SJ. Long overdue national big data policies hinder accurate and equitable cancer detection AI systems. J Med Imaging Radiat Sci 2024; 55:101387. [PMID: 38443215 DOI: 10.1016/j.jmir.2024.02.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Revised: 02/04/2024] [Accepted: 02/09/2024] [Indexed: 03/07/2024]
Affiliation(s)
- Dolly Y Wu
- Volunteer Services, UT Southwestern Medical Center, Dallas, TX, USA.
| | - Dat T Vo
- Department of Radiation Oncology, UT Southwestern Medical Center, Dallas, TX, USA
| | - Stephen J Seiler
- Department of Radiology, UT Southwestern Medical Center, Dallas, TX, USA
| |
Collapse
|
49
|
Jeng PH, Yang CY, Huang TR, Kuo CF, Liu SC. Harnessing AI for precision tonsillitis diagnosis: a revolutionary approach in endoscopic analysis. Eur Arch Otorhinolaryngol 2024; 281:6555-6563. [PMID: 39230610 DOI: 10.1007/s00405-024-08938-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2024] [Accepted: 08/19/2024] [Indexed: 09/05/2024]
Abstract
BACKGROUND Diagnosing and treating tonsillitis pose no significant challenge for otolaryngologists; however, it can increase the infection risk for healthcare professionals amidst the coronavirus pandemic. In recent years, with the advancement of artificial intelligence (AI), its application in medical imaging has also thrived. This research is to identify the optimal convolutional neural network (CNN) algorithm for accurate diagnosis of tonsillitis and early precision treatment. METHODS Semi-supervised learning with pseudo-labels used for self-training was adopted to train our CNN, with the algorithm including UNet, PSPNet, and FPN. A total of 485 pharyngoscopic images from 485 participants were included, comprising healthy individuals (133 cases), patients with the common cold (295 cases), and patients with tonsillitis (57 cases). Both color and texture features from 485 images are extracted for analysis. RESULTS UNet outperformed PSPNet and FPN in accurately segmenting oropharyngeal anatomy automatically, with average Dice coefficient of 97.74% and a pixel accuracy of 98.12%, making it suitable for enhancing the diagnosis of tonsillitis. The normal tonsils generally have more uniform and smooth textures and have pinkish color, similar to the surrounding mucosal tissues, while tonsillitis, particularly the antibiotic-required type, shows white or yellowish pus-filled spots or patches, and shows more granular or lumpy texture in contrast, indicating inflammation and changes in tissue structure. After training with 485 cases, our algorithm with UNet achieved accuracy rates of 93.75%, 97.1%, and 91.67% in differentiating the three tonsil groups, demonstrating excellent results. CONCLUSION Our research highlights the potential of using UNet for fully automated semantic segmentation of oropharyngeal structures, which aids in subsequent feature extraction, machine learning, and enables accurate AI diagnosis of tonsillitis. This innovation shows promise for enhancing both the accuracy and speed of tonsillitis assessments.
Collapse
Affiliation(s)
- Po-Hsuan Jeng
- Department of Otolaryngology-Head and Neck Surgery Tri-Service General Hospital, National Defense Medical Center, No. 325, Sec. 2, Cheng-Gong Road, Neihu District, Taipei, Taiwan 114, Republic of China
- Graduate Institute of Medical Science, National Defense Medical Center, Taipei, Taiwan
| | - Chien-Yi Yang
- Division of General Surgery, Department of Surgery Tri-Service General Hospital Songshan Branch, National Defense Medical Center, Taipei, Taiwan, Republic of China
| | - Tien-Ru Huang
- Department of Otolaryngology-Head and Neck Surgery Tri-Service General Hospital, National Defense Medical Center, No. 325, Sec. 2, Cheng-Gong Road, Neihu District, Taipei, Taiwan 114, Republic of China
| | - Chung-Feng Kuo
- Department of Material Science & Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan, Republic of China
| | - Shao-Cheng Liu
- Department of Otolaryngology-Head and Neck Surgery Tri-Service General Hospital, National Defense Medical Center, No. 325, Sec. 2, Cheng-Gong Road, Neihu District, Taipei, Taiwan 114, Republic of China.
| |
Collapse
|
50
|
Yakolli N, Shivanna DB, Rao RS, Patil S, Ronsivalle V, Cicciù M, Minervini G. Diagnosis of odontogenic keratocysts and non-keratocysts using edge attention convolution neural network. Minerva Dent Oral Sci 2024; 73:303-311. [PMID: 39327988 DOI: 10.23736/s2724-6329.24.04874-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/28/2024]
Abstract
BACKGROUND The study's objective was to develop an automated method for a histopathology recognition model for odontogenic keratocysts (OKC) and non-keratocyst (Non-KC) in jaw cyst sections stained with hematoxylin (H) and eosin (E) on a tiny bit of incisional biopsy prior to surgery. METHODS This hastens the speed and precision of diagnosis to patients. Also, navigates the clinicians with the therapeutic doctrine. To build such a system and to increase the accuracy of the existing models, the edge attention CNN model with Keras functional API was implemented which efficiently analyzes the texture information of the images. Approximately 2861 microscopic images at a 40X magnification were taken from 54 OKC, 23 Dentigerous cysts (DC), and 20 Radicular cysts. RESULTS The model was trained using both RGB and canny edge-detected images. The model gave a good accuracy of 96.8%, which is suitable for real-time. Histopathological images are better analyzed through textural features. The proposed edge attention CNN highlights the edges, making texture analysis more precise. CONCLUSIONS The suggested method will work for OKC and Non-KC diagnosis automation systems. The use of a whole slide imaging scanner has the potential to increase accuracy and remove human bias.
Collapse
Affiliation(s)
- Nivedan Yakolli
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, Ramaiah University of Applied Sciences, Bengaluru, India
| | - Divya B Shivanna
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, Ramaiah University of Applied Sciences, Bengaluru, India
| | - Roopa S Rao
- Department of Oral Pathology and Microbiology, Faculty of Dental Sciences, Ramaiah University of Applied Sciences, Bengaluru, India
| | - Shankargouda Patil
- College of Dental Medicine, Roseman University of Health Sciences, South Jordan, UT, USA
| | - Vincenzo Ronsivalle
- Department of General Surgery and Surgical-Medical Specialties, School of Dentistry, University of Catania, Catania, Italy
| | - Marco Cicciù
- Department of General Surgery and Surgical-Medical Specialties, School of Dentistry, University of Catania, Catania, Italy
| | - Giuseppe Minervini
- Multidisciplinary Department of Medical-Surgical and Odontostomatological Specialties, University of Campania "Luigi Vanvitelli", Naples, Italy -
- Saveetha Dental College and Hospitals, Saveetha Institute of Medical and Technical Sciences (SIMATS), Saveetha University, Chennai, Tamil Nadu, India
| |
Collapse
|