1
|
Salimi M, Houshi S, Gholamrezanezhad A, Vadipour P, Seifi S. Radiomics-based machine learning in prediction of response to neoadjuvant chemotherapy in osteosarcoma: A systematic review and meta-analysis. Clin Imaging 2025; 123:110494. [PMID: 40349577 DOI: 10.1016/j.clinimag.2025.110494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2025] [Revised: 05/03/2025] [Accepted: 05/07/2025] [Indexed: 05/14/2025]
Abstract
BACKGROUND AND AIMS Osteosarcoma (OS) is the most common primary bone malignancy, and neoadjuvant chemotherapy (NAC) improves survival rates. However, OS heterogeneity results in variable treatment responses, highlighting the need for reliable, non-invasive tools to predict NAC response. Radiomics-based machine learning (ML) offers potential for identifying imaging biomarkers to predict treatment outcomes. This systematic review and meta-analysis evaluated the accuracy and reliability of radiomics models for predicting NAC response in OS. METHODS A systematic search was conducted in PubMed, Embase, Scopus, and Web of Science up to November 2024. Studies using radiomics-based ML for NAC response prediction in OS were included. Pooled sensitivity, specificity, and AUC for training and validation cohorts were calculated using bivariate random-effects modeling, with clinical-combined models analyzed separately. Quality assessment was performed using the QUADAS-2 tool, radiomics quality score (RQS), and METRICS scores. RESULTS Sixteen studies were included, with 63 % using MRI and 37 % using CT. Twelve studies, comprising 1639 participants, were included in the meta-analysis. Pooled metrics for training cohorts showed an AUC of 0.93, sensitivity of 0.89, and specificity of 0.85. Validation cohorts achieved an AUC of 0.87, sensitivity of 0.81, and specificity of 0.82. Clinical-combined models outperformed radiomics-only models. The mean RQS score was 9.44 ± 3.41, and the mean METRICS score was 60.8 % ± 17.4 %. CONCLUSION Radiomics-based ML shows promise for predicting NAC response in OS, especially when combined with clinical indicators. However, limitations in external validation and methodological consistency must be addressed.
Collapse
Affiliation(s)
- Mohsen Salimi
- Research Center of Thoracic Oncology (RCTO), National Research Institute of Tuberculosis and Lung Diseases (NRITLD), Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Shakiba Houshi
- Isfahan Neurosciences Research Center, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Ali Gholamrezanezhad
- Department of Radiology, Keck School of Medicine, University of Southern California (USC), Los Angeles, CA, USA
| | - Pouria Vadipour
- Western Michigan University Homer Stryker M.D. School of Medicine, Kalamazoo, MI, USA
| | - Sharareh Seifi
- Research Center of Thoracic Oncology (RCTO), National Research Institute of Tuberculosis and Lung Diseases (NRITLD), Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| |
Collapse
|
2
|
Lareyre F, Wanhainen A, Raffort J. Artificial Intelligence-Powered Technologies for the Management of Vascular Diseases: Building Guidelines and Moving Forward Evidence Generation. J Endovasc Ther 2025; 32:541-544. [PMID: 37464795 DOI: 10.1177/15266028231187599] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/20/2023]
Affiliation(s)
- Fabien Lareyre
- Department of Vascular Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
| | - Anders Wanhainen
- Section of Vascular Surgery, Department of Surgical Sciences, Uppsala University, Uppsala, Sweden
- Department of Surgical and Perioperative Sciences, Umeå University, Umeå, Sweden
| | - Juliette Raffort
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
- 3IA Institute, Université Côte d'Azur, Nice, France
- Department of Clinical Biochemistry, University Hospital of Nice, Nice, France
| |
Collapse
|
3
|
Fatania K, Frood R, Mistry H, Short SC, O'Connor J, Scarsbrook AF, Currie S. Impact of intensity standardisation and ComBat batch size on clinical-radiomic prognostic models performance in a multi-centre study of patients with glioblastoma. Eur Radiol 2025; 35:3354-3366. [PMID: 39607450 PMCID: PMC12081554 DOI: 10.1007/s00330-024-11168-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2024] [Revised: 08/12/2024] [Accepted: 09/30/2024] [Indexed: 11/29/2024]
Abstract
PURPOSE To assess the effect of different intensity standardisation techniques (ISTs) and ComBat batch sizes on radiomics survival model performance and stability in a heterogenous, multi-centre cohort of patients with glioblastoma (GBM). METHODS Multi-centre pre-operative MRI acquired between 2014 and 2020 in patients with IDH-wildtype unifocal WHO grade 4 GBM were retrospectively evaluated. WhiteStripe (WS), Nyul histogram matching (HM), and Z-score (ZS) ISTs were applied before radiomic feature (RF) extraction. RFs were realigned using ComBat and minimum batch size (MBS) of 5, 10, or 15 patients. Cox proportional hazards models for overall survival (OS) prediction were produced using five different selection strategies and the impact of IST and MBS was evaluated using bootstrapping. Calibration, discrimination, relative explained variation, and model fit were assessed. Instability was evaluated using 95% confidence intervals (95% CIs), feature selection frequency and calibration curves across the bootstrap resamples. RESULTS One hundred ninety-five patients were included. Median OS = 13 (95% CI: 12-14) months. Twelve to fourteen unique MRI protocols were used per MRI sequence. HM and WS produced the highest relative increase in model discrimination, explained variation and model fit but IST choice did not greatly impact on stability, nor calibration. Larger ComBat batches improved discrimination, model fit, and explained variation but higher MBS (reduced sample size) reduced stability (across all performance metrics) and reduced calibration accuracy. CONCLUSION Heterogenous, real-world GBM data poses a challenge to the reproducibility of radiomics. ComBat generally improved model performance as MBS increased but reduced stability and calibration. HM and WS tended to improve model performance. KEY POINTS Question ComBat harmonisation of RFs and intensity standardisation of MRI have not been thoroughly evaluated in multicentre, heterogeneous GBM data. Findings The addition of ComBat and ISTs can improve discrimination, relative model fit, and explained variance but degrades the calibration and stability of survival models. Clinical relevance Radiomics risk prediction models in real-world, multicentre contexts could be improved by ComBat and ISTs, however, this degrades calibration and prediction stability and this must be thoroughly investigated before patients can be accurately separated into different risk groups.
Collapse
Affiliation(s)
- Kavi Fatania
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, England, UK.
- Leeds Institute of Medical Research, University of Leeds, Leeds, UK.
| | - Russell Frood
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, England, UK
- Leeds Institute of Medical Research, University of Leeds, Leeds, UK
| | - Hitesh Mistry
- Division of Cancer Sciences, University of Manchester, Manchester, UK
| | - Susan C Short
- Leeds Institute of Medical Research, University of Leeds, Leeds, UK
- Department of Oncology, Leeds Teaching Hospitals NHS Trust, England, UK
| | - James O'Connor
- Division of Cancer Sciences, University of Manchester, Manchester, UK
- Department of Radiology, The Christie Hospital, Manchester, UK
- Division of Radiotherapy and Imaging, Institute of Cancer Research, London, UK
| | - Andrew F Scarsbrook
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, England, UK
- Leeds Institute of Medical Research, University of Leeds, Leeds, UK
| | - Stuart Currie
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, England, UK
- Leeds Institute of Medical Research, University of Leeds, Leeds, UK
| |
Collapse
|
4
|
van der Linden LR, Vavliakis I, de Groot TM, Jutte PC, Doornberg JN, Lozano-Calderon SA, Groot OQ. Artificial Intelligence in bone Metastases: A systematic review in guideline adherence of 92 studies. J Bone Oncol 2025; 52:100682. [PMID: 40337637 PMCID: PMC12056386 DOI: 10.1016/j.jbo.2025.100682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 02/09/2025] [Accepted: 04/15/2025] [Indexed: 05/09/2025] Open
Abstract
Background The last decade has witnessed a surge in artificial intelligence (AI). With bone metastases becoming more prevalent, there is an increasing call for personalized treatment options, a domain where AI can greatly contribute. However, integrating AI into clinical settings has proven to be difficult. Therefore, we aimed to provide an overview of AI modalities for treating bone metastases and recommend implementation-worthy models based on TRIPOD, CLAIM, and UPM scores. Methods This systematic review included 92 studies on AI models in bone metastases between 2008 and 2024. Using three assessment tools we provided a reliable foundation for recommending AI modalities fit for clinical use (TRIPOD or CLAIM ≥ 70 % and UPM score ≥ 10). Results Most models focused on survival prediction (44/92;48%), followed by imaging studies (37/92;40%). Median TRIPOD completeness was 70% (IQR 64-81%), CLAIM completeness was 57% (IQR 48-67%), and UPM score was 7 (IQR 5-9). In total, 10% (9/92) AI modalities were deemed fit for clinical use. Conclusion Transparent reporting, utilizing the aforementioned three evaluation tools, is essential for effectively integrating AI models into clinical practice, as currently, only 10% of AI models for bone metastases are deemed fit for clinical use. Such transparency ensures that both patients and clinicians can benefit from clinically useful AI models, potentially enhancing AI-driven personalized cancer treatment.
Collapse
Affiliation(s)
- Lotte R. van der Linden
- Department of Orthopaedic Surgery, University Medical Center Groningen, Groningen, the Netherlands
- Department of Orthopaedic Surgery, Massachusetts General Hospital, Boston, MA, USA
| | - Ioannis Vavliakis
- Department of Orthopaedic Surgery, University Medical Center Groningen, Groningen, the Netherlands
| | - Tom M. de Groot
- Department of Orthopaedic Surgery, University Medical Center Groningen, Groningen, the Netherlands
- Department of Orthopaedic Surgery, Massachusetts General Hospital, Boston, MA, USA
| | - Paul C. Jutte
- Department of Orthopaedic Surgery, University Medical Center Groningen, Groningen, the Netherlands
| | - Job N. Doornberg
- Department of Orthopaedic Surgery, University Medical Center Groningen, Groningen, the Netherlands
| | | | - Olivier Q. Groot
- Department of Orthopaedic Surgery, Massachusetts General Hospital, Boston, MA, USA
- Department of Orthopaedic Surgery, University Medical Center Utrecht, Utrecht, the Netherlands
| |
Collapse
|
5
|
Pareto D, Naval-Baudin P, Pons-Escoda A, Bargalló N, Garcia-Gil M, Majós C, Rovira À. Image analysis research in neuroradiology: bridging clinical and technical domains. Neuroradiology 2025:10.1007/s00234-025-03633-x. [PMID: 40434412 DOI: 10.1007/s00234-025-03633-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Accepted: 04/20/2025] [Indexed: 05/29/2025]
Abstract
PURPOSE Advancements in magnetic resonance imaging (MRI) analysis over the past decades have significantly reshaped the field of neuroradiology. The ability to extract multiple quantitative measures from each MRI scan, alongside the development of extensive data repositories, has been fundamental to the emergence of advanced methodologies such as radiomics and artificial intelligence (AI). This educational review aims to delineate the importance of image analysis, highlight key paradigm shifts, examine their implications, and identify existing constraints that must be addressed to facilitate integration into clinical practice. Particular attention is given to aiding junior neuroradiologists in navigating this complex and evolving landscape. METHODS A comprehensive review of the available analysis toolboxes was conducted, focusing on major technological advancements in MRI analysis, the evolution of data repositories, and the rise of AI and radiomics in neuroradiology. Stakeholders within the field were identified and their roles examined. Additionally, current challenges and barriers to clinical implementation were analyzed. RESULTS The analysis revealed several pivotal shifts, including the transition from qualitative to quantitative imaging, the central role of large datasets in developing AI tools, and the growing importance of interdisciplinary collaboration. Key stakeholders-including academic institutions, industry partners, regulatory bodies, and clinical practitioners-were identified, each playing a distinct role in advancing the field. However, significant barriers remain, particularly regarding standardization, data sharing, regulatory approval, and integration into clinical workflows. CONCLUSIONS While advancements in MRI analysis offer tremendous potential to enhance neuroradiology practice, realizing this potential requires overcoming technical, regulatory, and practical barriers. Education and structured support for junior neuroradiologists are essential to ensure they are well-equipped to participate in and drive future developments. A coordinated effort among stakeholders is crucial to facilitate the seamless translation of these technological innovations into everyday clinical practice.
Collapse
Affiliation(s)
- Deborah Pareto
- Neuroradiology Section, Radiology Department (IDI), Vall Hebron University Hospital, Psg Vall Hebron 119-129, 08035, Barcelona, Spain.
- Neuroradiology Group, Vall Hebron Research Institute, Barcelona, Spain.
| | - Pablo Naval-Baudin
- Neuroradiology Section, Department of Radiology, Hospital Universitari de Bellvitge, L'Hospitalet de Llobregat, Spain
- Institut Diagnòstic Per La Imatge (IDI), Centre Bellvitge, L'Hospitalet de Llobregat, Spain
- Diagnostic Imaging and Nuclear Medicine Research Group, Bellvitge Biomedical Research Institute (IDIBELL), L'Hospitalet de Llobregat, Spain
| | - Albert Pons-Escoda
- Neuroradiology Section, Department of Radiology, Hospital Universitari de Bellvitge, L'Hospitalet de Llobregat, Spain
- Institut Diagnòstic Per La Imatge (IDI), Centre Bellvitge, L'Hospitalet de Llobregat, Spain
- Diagnostic Imaging and Nuclear Medicine Research Group, Bellvitge Biomedical Research Institute (IDIBELL), L'Hospitalet de Llobregat, Spain
| | - Núria Bargalló
- Neuroradiology Section, Radiology Department, Diagnostic Image Center, Hospital Clinic of Barcelona, University of Barcelona, Barcelona, Spain
| | - María Garcia-Gil
- Institut Diagnòstic Per La Imatge (IDI), Serveis Corporatius, Parc Sanitaria Pere Virgili, Barcelona, Spain
| | - Carlos Majós
- Neuroradiology Section, Department of Radiology, Hospital Universitari de Bellvitge, L'Hospitalet de Llobregat, Spain
- Institut Diagnòstic Per La Imatge (IDI), Centre Bellvitge, L'Hospitalet de Llobregat, Spain
- Diagnostic Imaging and Nuclear Medicine Research Group, Bellvitge Biomedical Research Institute (IDIBELL), L'Hospitalet de Llobregat, Spain
| | - Àlex Rovira
- Neuroradiology Section, Radiology Department (IDI), Vall Hebron University Hospital, Psg Vall Hebron 119-129, 08035, Barcelona, Spain
- Neuroradiology Group, Vall Hebron Research Institute, Barcelona, Spain
| |
Collapse
|
6
|
Palkovics D, Molnar B, Pinter C, García-Mato D, Diaz-Pinto A, Windisch P, Ramseier CA. Automatic deep learning segmentation of mandibular periodontal bone topography on cone-beam computed tomography images. J Dent 2025:105813. [PMID: 40373868 DOI: 10.1016/j.jdent.2025.105813] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2025] [Revised: 04/21/2025] [Accepted: 05/09/2025] [Indexed: 05/17/2025] Open
Abstract
OBJECTIVES This study evaluated the performance of a multi-stage Segmentation Residual Network (SegResNet)-based deep learning (DL) model for the automatic segmentation of cone-beam computed tomography (CBCT) images of patients with stage III and IV periodontitis. METHODS Seventy pre-processed CBCT scans from patients undergoing periodontal rehabilitation were used for training and validation. The model was tested on 10 CBCT scans independent from the training dataset by comparing results with semi-automatic (SA) segmentations. Segmentation accuracy was assessed using the Dice similarity coefficient (DSC), Intersection over Union (IoU), and Hausdorff distance 95th percentile (HD95). Linear periodontal measurements were performed on four tooth surfaces to assess the validity of the DL segmentation in the periodontal region. RESULTS The DL model achieved a mean DSC of 0.9650 ± 0.0097, with an IoU of 0.9340 ± 0.0180 and HD95 of 0.4820 mm ± 0.1269 mm, showing strong agreement with SA segmentation. Linear measurements revealed high statistical correlations between the mesial, distal, and lingual surfaces, with intraclass correlation coefficients (ICC) of 0.9442 (p<0.0001), 0.9232 (p<0.0001), and 0.9598(p<0.0001), respectively, while buccal measurements revealed lower consistency, with an ICC of 0.7481 (p<0.0001). The DL method reduced the segmentation time by 47 times compared to the SA method. CONCLUSIONS Acquired 3D models may enable precise treatment planning in cases where conventional diagnostic modalities are insufficient. However, the robustness of the model must be increased to improve its general reliability and consistency at the buccal aspect of the periodontal region. CLINICAL SIGNIFICANCE This study presents a DL model for the CBCT-based segmentation of periodontal defects, demonstrating high accuracy and a 47-fold time reduction compared to SA methods, thus improving the feasibility of 3D diagnostics for advanced periodontitis.
Collapse
Affiliation(s)
- Daniel Palkovics
- Department of Periodontology, Semmelweis University, Szentkirályi utca 47. 4th floor, 1088 Budapest, Hungary; Department of Periodontology, University of Bern, Freiburgstrasse 7. 3010 Bern, Switzerland; Dent.AI Medical Imaging Ltd., Irinyi József utca 31, 1111 Budapest, Hungary..
| | - Balint Molnar
- Department of Periodontology, Semmelweis University, Szentkirályi utca 47. 4th floor, 1088 Budapest, Hungary; Dent.AI Medical Imaging Ltd., Irinyi József utca 31, 1111 Budapest, Hungary..
| | - Csaba Pinter
- Dent.AI Medical Imaging Ltd., Irinyi József utca 31, 1111 Budapest, Hungary.; Empresa de Base Technológica Internacional de Canarias, S.L. (EBATINCA), Calle Triana, 60, Piso 3, Oficina B, 35002 Las Palmas De Gran Canaria, Spain.
| | - David García-Mato
- Dent.AI Medical Imaging Ltd., Irinyi József utca 31, 1111 Budapest, Hungary..
| | - Andres Diaz-Pinto
- School of Biomedical Engineering & Imaging Sciences, King's College London, St Thomas' Campus, St Thomas' Hospital, Westminster Bridge Road, London SE1 7EH, UK.
| | - Peter Windisch
- Department of Periodontology, Semmelweis University, Szentkirályi utca 47. 4th floor, 1088 Budapest, Hungary; Dent.AI Medical Imaging Ltd., Irinyi József utca 31, 1111 Budapest, Hungary..
| | - Christoph A Ramseier
- Department of Periodontology, University of Bern, Freiburgstrasse 7. 3010 Bern, Switzerland.
| |
Collapse
|
7
|
Lacaita PG, Galijasevic M, Swoboda M, Gruber L, Scharll Y, Barbieri F, Widmann G, Feuchtner GM. The Accuracy of ChatGPT-4o in Interpreting Chest and Abdominal X-Ray Images. J Pers Med 2025; 15:194. [PMID: 40423065 DOI: 10.3390/jpm15050194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2025] [Revised: 04/28/2025] [Accepted: 05/08/2025] [Indexed: 05/28/2025] Open
Abstract
Background/Objectives: Large language models (LLMs), such as ChatGPT, have emerged as potential clinical support tools to enhance precision in personalized patient care, but their reliability in radiological image interpretation remains uncertain. The primary aim of our study was to evaluate the diagnostic accuracy of ChatGPT-4o in interpreting chest X-rays (CXRs) and abdominal X-rays (AXRs) by comparing its performance to expert radiology findings, whilst secondary aims were diagnostic confidence and patient safety. Methods: A total of 500 X-rays, including 257 CXR (51.4%) and 243 AXR (48.5%), were analyzed. Diagnoses made by ChatGPT-4o were compared to expert interpretations. Confidence scores (1-4) were assigned and responses were evaluated for patient safety. Results: ChatGPT-4o correctly identified 345 of 500 (69%) pathologies (95% CI: 64.81-72.9). For AXRs 175 of 243 (72.02%) pathologies were correctly diagnosed (95% CI: 66.06-77.28), while for CXRs 170 of 257 (66.15%) were accurate (95% CI: 60.16-71.66). The highest detection rates among CXRs were observed for pulmonary edema, tumor, pneumonia, pleural effusion, cardiomegaly, and emphysema, and lower rates were observed for pneumothorax, rib fractures, and enlarged mediastinum. AXR performance was highest for intestinal obstruction and foreign bodies, and weaker for pneumoperitoneum, renal calculi, and diverticulitis. Confidence scores were higher for AXRs (mean 3.45 ± 1.1) than CXRs (mean 2.48 ± 1.45). All responses (100%) were considered to be safe for the patient. Interobserver agreement was high (kappa = 0.920), and reliability (second prompt) was moderate (kappa = 0.750). Conclusions: ChatGPT-4o demonstrated moderate accuracy for the interpretation of X-rays, being higher for AXRs compared to CXRs. Improvements are required for its use as efficient clinical support tool.
Collapse
Affiliation(s)
- Pietro G Lacaita
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Malik Galijasevic
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Michael Swoboda
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Leonhard Gruber
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Yannick Scharll
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Fabian Barbieri
- Department of Cardiology, Angiology and Intensive Care Medicine, Deutsches Herzzentrum Charite, 10117 Berlin, Germany
| | - Gerlig Widmann
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| | - Gudrun M Feuchtner
- Department Radiology, Innsbruck Medical University, 6020 Innsbruck, Austria
| |
Collapse
|
8
|
Mushcab H, Al Ramis M, AlRujaib A, Eskandarani R, Sunbul T, AlOtaibi A, Obaidan M, Al Harbi R, Aljabri D. Application of Artificial Intelligence in Cardio-Oncology Imaging for Cancer Therapy-Related Cardiovascular Toxicity: Systematic Review. JMIR Cancer 2025; 11:e63964. [PMID: 40344203 PMCID: PMC12083731 DOI: 10.2196/63964] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2024] [Revised: 03/02/2025] [Accepted: 03/03/2025] [Indexed: 05/11/2025] Open
Abstract
Background Artificial intelligence (AI) is a revolutionary tool yet to be fully integrated into several health care sectors, including medical imaging. AI can transform how medical imaging is conducted and interpreted, especially in cardio-oncology. Objective This study aims to systematically review the available literature on the use of AI in cardio-oncology imaging to predict cardiotoxicity and describe the possible improvement of different imaging modalities that can be achieved if AI is successfully deployed to routine practice. Methods We conducted a database search in PubMed, Ovid MEDLINE, Cochrane Library, CINAHL, and Google Scholar from inception to 2023 using the AI research assistant tool (Elicit) to search for original studies reporting AI outcomes in adult patients diagnosed with any cancer and undergoing cardiotoxicity assessment. Outcomes included incidence of cardiotoxicity, left ventricular ejection fraction, risk factors associated with cardiotoxicity, heart failure, myocardial dysfunction, signs of cancer therapy-related cardiovascular toxicity, echocardiography, and cardiac magnetic resonance imaging. Descriptive information about each study was recorded, including imaging technique, AI model, outcomes, and limitations. Results The systematic search resulted in 7 studies conducted between 2018 and 2023, which are included in this review. Most of these studies were conducted in the United States (71%), included patients with breast cancer (86%), and used magnetic resonance imaging as the imaging modality (57%). The quality assessment of the studies had an average of 86% compliance in all of the tool's sections. In conclusion, this systematic review demonstrates the potential of AI to enhance cardio-oncology imaging for predicting cardiotoxicity in patients with cancer. Conclusions Our findings suggest that AI can enhance the accuracy and efficiency of cardiotoxicity assessments. However, further research through larger, multicenter trials is needed to validate these applications and refine AI technologies for routine use, paving the way for improved patient outcomes in cancer survivors at risk of cardiotoxicity.
Collapse
Affiliation(s)
- Hayat Mushcab
- Research Office, Johns Hopkins Aramco Healthcare, Medical Access Road 1, Dhahran, Saudi Arabia, 966 556373411
| | | | - Abdulrahman AlRujaib
- College of Medicine, Royal College of Surgeons in Ireland-Bahrain, Muharraq, Bahrain
| | | | - Tamara Sunbul
- Health Informatics Department, Johns Hopkins Aramco Healthcare, Dhahran, Saudi Arabia
| | - Anwar AlOtaibi
- Research Office, Johns Hopkins Aramco Healthcare, Medical Access Road 1, Dhahran, Saudi Arabia, 966 556373411
| | | | - Reman Al Harbi
- College of Medicine, University of Jeddah, Jeddah, Saudi Arabia
| | - Duaa Aljabri
- College of Public Health, Imam Abdulrahman bin Faisal University, Dammam, Saudi Arabia
| |
Collapse
|
9
|
Familiar AM, Khalili N, Khalili N, Schuman C, Grove E, Viswanathan K, Seidlitz J, Alexander-Bloch A, Zapaishchykova A, Kann BH, Vossough A, Storm PB, Resnick AC, Kazerooni AF, Nabavizadeh A. Empowering Data Sharing in Neuroscience: A Deep Learning Deidentification Method for Pediatric Brain MRIs. AJNR Am J Neuroradiol 2025; 46:964-972. [PMID: 39532533 DOI: 10.3174/ajnr.a8581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2024] [Accepted: 11/07/2024] [Indexed: 11/16/2024]
Abstract
BACKGROUND AND PURPOSE Privacy concerns, such as identifiable facial features within brain scans, have hindered the availability of pediatric neuroimaging data sets for research. Consequently, pediatric neuroscience research lags adult counterparts, particularly in rare disease and under-represented populations. The removal of face regions (image defacing) can mitigate this; however, existing defacing tools often fail with pediatric cases and diverse image types, leaving a critical gap in data accessibility. Given recent National Institutes of Health data sharing mandates, novel solutions are a critical need. MATERIALS AND METHODS To develop an artificial intelligence (AI)-powered tool for automatic defacing of pediatric brain MRIs, deep learning methodologies (nnU-Net) were used by using a large, diverse multi-institutional data set of clinical radiology images. This included multiparametric MRIs (T1-weighted [T1W], T1W-contrast-enhanced, T2-weighted [T2W], T2W-FLAIR) with 976 total images from 208 patients with brain tumor (Children's Brain Tumor Network, CBTN) and 36 clinical control patients (Scans with Limited Imaging Pathology, SLIP) ranging in age from 7 days to 21 years old. RESULTS Face and ear removal accuracy for withheld testing data were the primary measure of model performance. Potential influences of defacing on downstream research usage were evaluated with standard image processing and AI-based pipelines. Group-level statistical trends were compared between original (nondefaced) and defaced images. Across image types, the model had high accuracy for removing face regions (mean accuracy, 98%; n=98 subjects/392 images), with lower performance for removal of ears (73%). Analysis of global and regional brain measures (SLIP cohort) showed minimal differences between original and defaced outputs (mean r S = 0.93, all P < .0001). AI-generated whole brain and tumor volumes (CBTN cohort) and temporalis muscle metrics (volume, cross-sectional area, centile scores; SLIP cohort) were not significantly affected by image defacing (all r S > 0.9, P < .0001). CONCLUSIONS The defacing model demonstrates efficacy in removing facial regions across multiple MRI types and exhibits minimal impact on downstream research usage. A software package with the trained model is freely provided for wider use and further development (pediatric-auto-defacer; https://github.com/d3b-center/pediatric-auto-defacer-public). By offering a solution tailored to pediatric cases and multiple MRI sequences, this defacing tool will expedite research efforts and promote broader adoption of data sharing practices within the neuroscience community.
Collapse
Affiliation(s)
- Ariana M Familiar
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Neda Khalili
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Nastaran Khalili
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Cassidy Schuman
- School of Engineering and Applied Science (C.S., E.G.), University of Pennsylvania, Philadelphia, Pennsylvania
| | - Evan Grove
- School of Engineering and Applied Science (C.S., E.G.), University of Pennsylvania, Philadelphia, Pennsylvania
| | - Karthik Viswanathan
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Jakob Seidlitz
- Department of Child and Adolescent Psychiatry and Behavioral Science (J.S., A.A.-B.), The Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Psychiatry (J.S., A.A.-B.), University of Pennsylvania, Philadelphia, Pennsylvania
| | - Aaron Alexander-Bloch
- Department of Child and Adolescent Psychiatry and Behavioral Science (J.S., A.A.-B.), The Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Psychiatry (J.S., A.A.-B.), University of Pennsylvania, Philadelphia, Pennsylvania
- Lifespan Brain Institute at the Children's Hospital of Philadelphia and University of Pennsylvania (A.A.-B.), Philadelphia, Pennsylvania
| | - Anna Zapaishchykova
- Artificial Intelligence in Medicine (AIM) Program (A.Z., B.H.K.), Mass General Brigham, Harvard Medical School, Boston, Massachusetts
- Department of Radiation Oncology (A.Z., B.H.K.), Dana-Farber Cancer Institute and Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts
| | - Benjamin H Kann
- Artificial Intelligence in Medicine (AIM) Program (A.Z., B.H.K.), Mass General Brigham, Harvard Medical School, Boston, Massachusetts
- Department of Radiation Oncology (A.Z., B.H.K.), Dana-Farber Cancer Institute and Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts
| | - Arastoo Vossough
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Division of Radiology (A.V.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Radiology, Perelman School of Medicine (A.V., A.N.), University of Pennsylvania, Philadelphia, Pennsylvania
| | - Phillip B Storm
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Adam C Resnick
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
| | - Anahita Fathi Kazerooni
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Neurosurgery (A.M.F., Neda K., Nastaran K., K.V., P.B.S., A.C.R., A.F.K), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- AI2D Center for AI and Data Science for Integrated Diagnostics (A.F.K.), University of Pennsylvania, Philadelphia, Pennsylvania
| | - Ali Nabavizadeh
- From the Center for Data-Driven Discovery in Biomedicine (Db) (A.M.F., Neda K., Nastaran K., K.V., A.V., P.B.S., A.C.R., A.F.K., A.N.), Children's Hospital of Philadelphia, Philadelphia, Pennsylvania
- Department of Radiology, Perelman School of Medicine (A.V., A.N.), University of Pennsylvania, Philadelphia, Pennsylvania
| |
Collapse
|
10
|
Goulet M, Duguay-Drouin P, Mascolo-Fortin J, Mégrourèche J, Octave N, Tsui JMG. Clinical Application of Deep Learning-Assisted Needles Reconstruction in Prostate Ultrasound Brachytherapy. Int J Radiat Oncol Biol Phys 2025; 122:199-207. [PMID: 39800329 DOI: 10.1016/j.ijrobp.2024.12.026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Revised: 11/12/2024] [Accepted: 12/25/2024] [Indexed: 02/04/2025]
Abstract
PURPOSE High dose rate (HDR) prostate brachytherapy (BT) procedure requires image-guided needle insertion. Given that general anesthesia is often employed during the procedure, minimizing overall planning time is crucial. In this study, we explore the clinical feasibility and time-saving potential of artificial intelligence (AI)-driven auto-reconstruction of transperineal needles in the context of ultrasound (US)-guided prostate BT planning. METHODS AND MATERIALS This study included a total of 102 US-planned BT images from a single institution and split into 3 groups: 50 for model training and validation, 11 to evaluate reconstruction accuracy (test set), and 41 to evaluate the AI tool in a clinical implementation (clinical set). Reconstruction accuracy for the test set was evaluated by comparing the performance of AI-derived and manually reconstructed needles from 5 medical physicists on the 3D-US scans after treatment. The needle total reconstruction time for the clinical set was defined as the timestamp difference from scan acquisition to the start of dose calculations and was compared with values recorded before the clinical implementation of the AI-assisted tool. RESULTS A mean error of (0.44 ± 0.32) mm was found between the AI-reconstructed and the human consensus needle positions in the test set, with 95.0% of AI needle points falling below 1 mm from their human-made counterparts. Post-hoc analysis showed that only one of the human observers' reconstructions were significantly different from the others including the AIs. In the clinical set, the AI algorithm achieved a true positive reconstruction rate of 93.4% with only 4.5% of these needles requiring manual corrections from the planner before dosimetry. Total time required to perform AI-assisted catheter reconstruction on clinical cases was on average 15.2 min lower (P < .01) compared with procedure without AI assistance. CONCLUSIONS This study demonstrates the feasibility of an AI-assisted needle reconstructing tool for 3D-US-based HDR prostate BT. This is a step toward treatment planning automation and increased efficiency in HDR prostate BT.
Collapse
Affiliation(s)
- Mathieu Goulet
- Département de radio-oncologie, CISSS de Chaudière-Appalaches, Lévis, Québec, Canada.
| | | | - Julia Mascolo-Fortin
- Département de radio-oncologie, CISSS de Chaudière-Appalaches, Lévis, Québec, Canada
| | - Julien Mégrourèche
- Département de radio-oncologie, CISSS de Chaudière-Appalaches, Lévis, Québec, Canada
| | - Nadia Octave
- Département de radio-oncologie, CISSS de Chaudière-Appalaches, Lévis, Québec, Canada
| | - James Man Git Tsui
- Department of Radiation Oncology, McGill University Health Center, Montreal, Québec, Canada
| |
Collapse
|
11
|
Yao J, Alabousi A, Mironov O. Evaluation of a BERT Natural Language Processing Model for Automating CT and MRI Triage and Protocol Selection. Can Assoc Radiol J 2025; 76:265-272. [PMID: 38832645 DOI: 10.1177/08465371241255895] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/05/2024] Open
Abstract
Purpose: To evaluate the accuracy of a Bidirectional Encoder Representations for Transformers (BERT) Natural Language Processing (NLP) model for automating triage and protocol selection of cross-sectional image requisitions. Methods: A retrospective study was completed using 222 392 CT and MRI studies from a single Canadian university hospital database (January 2018-September 2022). Three hundred unique protocols (116 CT and 184 MRI) were included. A BERT model was trained, validated, and tested using an 80%-10%-10% stratified split. Naive Bayes (NB) and Support Vector Machine (SVM) machine learning models were used as comparators. Models were assessed using F1 score, precision, recall, and area under the receiver operating characteristic curve (AUROC). The BERT model was also assessed for multi-class protocol suggestion and subgroups based on referral location, modality, and imaging section. Results: BERT was superior to SVM for protocol selection (F1 score: BERT-0.901 vs SVM-0.881). However, was not significantly different from SVM for triage prediction (F1 score: BERT-0.844 vs SVM-0.845). Both models outperformed NB for protocol and triage. BERT had superior performance on minority classes compared to SVM and NB. For multiclass prediction, BERT accuracy was up to 0.991 for top-5 protocol suggestion, and 0.981 for top-2 triage suggestion. Emergency department patients had the highest F1 scores for both protocol (0.957) and triage (0.986), compared to inpatients and outpatients. Conclusion: The BERT NLP model demonstrated strong performance in automating the triage and protocol selection of radiology studies, showing potential to enhance radiologist workflows. These findings suggest the feasibility of using advanced NLP models to streamline radiology operations.
Collapse
Affiliation(s)
- Jason Yao
- Department of Radiology, McMaster University, Hamilton, ON, Canada
| | - Abdullah Alabousi
- Department of Radiology, McMaster University, Hamilton, ON, Canada
- St Joseph's Healthcare Hamilton, Hamilton, ON, Canada
| | - Oleg Mironov
- Department of Radiology, McMaster University, Hamilton, ON, Canada
- St Joseph's Healthcare Hamilton, Hamilton, ON, Canada
| |
Collapse
|
12
|
Yi PH, Bachina P, Bharti B, Garin SP, Kanhere A, Kulkarni P, Li D, Parekh VS, Santomartino SM, Moy L, Sulam J. Pitfalls and Best Practices in Evaluation of AI Algorithmic Biases in Radiology. Radiology 2025; 315:e241674. [PMID: 40392092 DOI: 10.1148/radiol.241674] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/22/2025]
Abstract
Despite growing awareness of problems with fairness in artificial intelligence (AI) models in radiology, evaluation of algorithmic biases, or AI biases, remains challenging due to various complexities. These include incomplete reporting of demographic information in medical imaging datasets, variability in definitions of demographic categories, and inconsistent statistical definitions of bias. To guide the appropriate evaluation of AI biases in radiology, this article summarizes the pitfalls in the evaluation and measurement of algorithmic biases. These pitfalls span the spectrum from the technical (eg, how different statistical definitions of bias impact conclusions about whether an AI model is biased) to those associated with social context (eg, how different conventions of race and ethnicity impact identification or masking of biases). Actionable best practices and future directions to avoid these pitfalls are summarized across three key areas: (a) medical imaging datasets, (b) demographic definitions, and (c) statistical evaluations of bias. Although AI bias in radiology has been broadly reviewed in the recent literature, this article focuses specifically on underrecognized potential pitfalls related to the three key areas. By providing awareness of these pitfalls along with actionable practices to avoid them, exciting AI technologies can be used in radiology for the good of all people.
Collapse
Affiliation(s)
- Paul H Yi
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Preetham Bachina
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Beepul Bharti
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Sean P Garin
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Adway Kanhere
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Pranav Kulkarni
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - David Li
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Vishwa S Parekh
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Samantha M Santomartino
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Linda Moy
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| | - Jeremias Sulam
- From the Department of Radiology, St Jude Children's Research Hospital, 262 Danny Thomas Pl, Memphis, TN 38105-3678 (P.H.Y.); Johns Hopkins University School of Medicine, Baltimore, Md (P.B.); Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Md (B.B., J.S.); Uniformed Services University of the Health Sciences, Bethesda, Md (S.P.G.); Institute for Health Computing, University of Maryland School of Medicine, Baltimore, Md (A.K., P.K.); Department of Medical Imaging, Western University Schulich School of Medicine & Dentistry, London, Ontario, Canada (D.L.); Department of Diagnostic and Interventional Imaging, McGovern Medical School at The University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Tex (V.S.P.); Drexel University School of Medicine, Philadelphia, Pa (S.M.S.); and Department of Radiology, New York University Grossman School of Medicine, New York, NY (L.M.)
| |
Collapse
|
13
|
Bai G, Huo S, Wang G, Tian S. Artificial intelligence radiomics in the diagnosis, treatment, and prognosis of gynecological cancer: a literature review. Transl Cancer Res 2025; 14:2508-2532. [PMID: 40386259 PMCID: PMC12079260 DOI: 10.21037/tcr-2025-618] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2025] [Accepted: 04/18/2025] [Indexed: 05/20/2025]
Abstract
Background and Objective Gynecological cancer is the most common cancer that affects women's quality of life and well-being. Artificial intelligence (AI) technology enables us to exploit high-dimensional imaging data for precision oncology. Tremendous progress has been made with AI radiomics in cancers such as lung and breast cancers. Herein, we performed a literature review on AI radiomics in the management of gynecological cancer. Methods A search was performed in the databases of PubMed, Embase, and Web of Science for original articles written in English up to 10 September 2024, using the terms "gynecological cancer", "cervical cancer", "endometrial cancer", "ovarian cancer", AND "artificial intelligence", "AI", AND "radiomics". The included studies mainly focused on the current landscape of AI radiomics in the diagnosis, treatment, and prognosis of gynecological cancer. Key Content and Findings A total of 128 studies were included, with 86 studies focusing on tumor diagnosis (n=23) and characterization (n=63), 15 on treatment response prediction, and 27 on recurrence and survival prediction. AI radiomics has shown potential value in tumor diagnosis and characterization [tumor staging, histological subtyping, lymph node metastasis (LNM), lymphovascular space invasion (LVSI), myometrial invasion (MI), and other molecular or clinicopathological factors], chemotherapy or chemoradiotherapy response evaluation, and prognosis (disease recurrence or metastasis, and survival) prediction. However, most included studies were single-center and retrospective. There was substantial heterogeneity in methodology and results reporting. Conclusions AI radiomics has been increasingly adopted in the management of gynecological cancer. Further validation in large-scale datasets is needed before clinical translation.
Collapse
Affiliation(s)
- Gengshen Bai
- Department of Intervention, The Second People’s Hospital of Baiyin City, Baiyin, China
| | - Shiwen Huo
- Jiangsu Hengrui Pharmaceuticals Co., Ltd., Shanghai, China
| | - Guangcai Wang
- Department of Intervention, The Second People’s Hospital of Baiyin City, Baiyin, China
| | - Shijia Tian
- Department of Intervention, The Second People’s Hospital of Baiyin City, Baiyin, China
| |
Collapse
|
14
|
Williams SC, Duvaux D, Das A, Sinha S, Layard Horsfall H, Funnell JP, Hanrahan JG, Khan DZ, Muirhead W, Kitchen N, Vasconcelos F, Bano S, Stoyanov D, Grover P, Marcus HJ. Automated Operative Phase and Step Recognition in Vestibular Schwannoma Surgery: Development and Preclinical Evaluation of a Deep Learning Neural Network (IDEAL Stage 0). Neurosurgery 2025:00006123-990000000-01600. [PMID: 40304484 DOI: 10.1227/neu.0000000000003466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2024] [Accepted: 01/07/2025] [Indexed: 05/02/2025] Open
Abstract
BACKGROUND AND OBJECTIVES Machine learning (ML) in surgical video analysis offers promising prospects for training and decision support in surgery. The past decade has seen key advances in ML-based operative workflow analysis, though existing applications mostly feature shorter surgeries (<2 hours) with limited scene changes. The aim of this study was to develop and evaluate a ML model capable of automated operative workflow recognition for retrosigmoid vestibular schwannoma (VS) resection. In doing so, this project furthers previous research by applying workflow prediction platforms to lengthy (median >5 hours duration), data-heavy surgeries, using VS resection as an exemplar. METHODS A video dataset of 21 microscopic retrosigmoid VS resections was collected at a single institution over 3 years and underwent workflow annotation according to a previously agreed expert consensus (Approach, Excision, and Closure phases; and Debulking or Dissection steps within the Excision phase). Annotations were used to train a ML model consisting of a convolutional neural network and a recurrent neural network. 5-fold cross-validation was used, and performance metrics (accuracy, precision, recall, F1 score) were assessed for phase and step prediction. RESULTS Median operative video time was 5 hours 18 minutes (IQR 3 hours 21 minutes-6 hours 1 minute). The "Tumor Excision" phase accounted for the majority of each case (median 4 hours 23 minutes), whereas "Approach and Exposure" (28 minutes) and "Closure" (17 minutes) comprised shorter phases. The ML model accurately predicted operative phases (accuracy 81%, weighted F1 0.83) and dichotomized steps (accuracy 86%, weighted F1 0.86). CONCLUSION This study demonstrates that our ML model can accurately predict the surgical phases and intraphase steps in retrosigmoid VS resection. This demonstrates the successful application of ML in operative workflow recognition on low-volume, lengthy, data-heavy surgical videos. Despite this, there remains room for improvement in individual step classification. Future applications of ML in low-volume high-complexity operations should prioritize collaborative video sharing to overcome barriers to clinical translation.
Collapse
Affiliation(s)
- Simon C Williams
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | | | - Adrito Das
- UCL Hawkes Institute, University College London, London, UK
| | - Siddharth Sinha
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | - Hugo Layard Horsfall
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
- The Francis Crick Institute, London, UK
| | - Jonathan P Funnell
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | - John G Hanrahan
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | - Danyal Z Khan
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | - William Muirhead
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
- Institute of Neurology, Institute of Brain Repair and Rehabilitation, University College London, London, UK
| | - Neil Kitchen
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | | | - Sophia Bano
- UCL Hawkes Institute, University College London, London, UK
| | - Danail Stoyanov
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| | - Patrick Grover
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | - Hani J Marcus
- Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
- UCL Hawkes Institute, University College London, London, UK
| |
Collapse
|
15
|
Wang X, Zhu MX, Wang JF, Liu P, Zhang LY, Zhou Y, Lin XX, Du YD, He KL. Multivariable prognostic models for post-hepatectomy liver failure: An updated systematic review. World J Hepatol 2025; 17:103330. [PMID: 40308827 PMCID: PMC12038414 DOI: 10.4254/wjh.v17.i4.103330] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/18/2024] [Revised: 02/28/2025] [Accepted: 03/21/2025] [Indexed: 04/25/2025] Open
Abstract
BACKGROUND Partial hepatectomy continues to be the primary treatment approach for liver tumors, and post-hepatectomy liver failure (PHLF) remains the most critical life-threatening complication following surgery. AIM To comprehensively review the PHLF prognostic models developed in recent years and objectively assess the risk of bias in these models. METHODS This review followed the Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies and the Preferred Reporting Items for Systematic Reviews and Meta-Analyses guideline. Three databases were searched from November 2019 to December 2022, and references as well as cited literature in all included studies were manually screened in March 2023. Based on the defined inclusion criteria, articles on PHLF prognostic models were selected, and data from all included articles were extracted by two independent reviewers. The PROBAST was used to evaluate the quality of each included article. RESULTS A total of thirty-four studies met the eligibility criteria and were included in the analysis. Nearly all of the models (32/34, 94.1%) were developed and validated exclusively using private data sources. Predictive variables were categorized into five distinct types, with the majority of studies (32/34, 94.1%) utilizing multiple types of data. The area under the curve for the training models included ranged from 0.697 to 0.956. Analytical issues resulted in a high risk of bias across all studies included. CONCLUSION The validation performance of the existing models was substantially lower compared to the development models. All included studies were evaluated as having a high risk of bias, primarily due to issues within the analytical domain. The progression of modeling technology, particularly in artificial intelligence modeling, necessitates the use of suitable quality assessment tools.
Collapse
Affiliation(s)
- Xiao Wang
- Department of Hepatobiliary Surgery, Chinese PLA 970 Hospital, Yantai 264001, Shandong Province, China
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China
| | - Ming-Xiang Zhu
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China
- Medical School of Chinese PLA, Chinese PLA General Hospital, Beijing 100853, China
| | - Jun-Feng Wang
- Division of Pharmacoepidemiology and Clinical Pharmacology, Utrecht Institute for Pharmaceutical Sciences, Utrecht University, Utrecht 358 4CG, Netherlands
| | - Pan Liu
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China
| | - Li-Yuan Zhang
- China National Clinical Research Center for Neurological Diseases, Beijing 100853, China
| | - You Zhou
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China
- School of Medicine, Nankai University, Tianjin 300071, China
| | - Xi-Xiang Lin
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China
| | - Ying-Dong Du
- Department of Hepatobiliary Surgery, Chinese PLA 970 Hospital, Yantai 264001, Shandong Province, China
| | - Kun-Lun He
- Medical Big Data Research Center, Chinese PLA General Hospital, Beijing 100853, China.
| |
Collapse
|
16
|
Chew BH, Ngiam KY. Artificial intelligence tool development: what clinicians need to know? BMC Med 2025; 23:244. [PMID: 40275334 PMCID: PMC12023651 DOI: 10.1186/s12916-025-04076-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/16/2024] [Accepted: 04/11/2025] [Indexed: 04/26/2025] Open
Abstract
Digital medicine and smart healthcare will not be realised without the cognizant participation of clinicians. Artificial intelligence (AI) today primarily involves computers or machines designed to simulate aspects of human intelligence using mathematically designed neural networks, although early AI systems relied on a variety of non-neural network techniques. With the increased complexity of the neural layers, deep machine learning (ML) can self-learn and augment many human tasks that require decision-making on the basis of multiple sources of data. Clinicians are important stakeholders in the use of AI and ML tools. The review questions are as follows: What is the typical process of AI tool development in the full cycle? What are the important concepts and technical aspects of each step? This review synthesises a targeted literature review and reports and summarises online structured materials to present a succinct explanation of the whole development process of AI tools. The development of AI tools in healthcare involves a series of cyclical processes: (1) identifying clinical problems suitable for AI solutions, (2) forming project teams or collaborating with experts, (3) organising and curating relevant data, (4) establishing robust physical and virtual infrastructure, and computer systems' architecture that support subsequent stages, (5) exploring AI neural networks on open access platforms before making a new decision, (6) validating AI/ML models, (7) registration, (8) clinical deployment and continuous performance monitoring and (9) improving the AI ecosystem ensures its adaptability to evolving clinical needs. A sound understanding of this would help clinicians appreciate the development of AI tools and engage in codesigning, evaluating and monitoring the tools. This would facilitate broader use and closer regulation of AI/ML tools in healthcare settings.
Collapse
Affiliation(s)
- Boon-How Chew
- Department of Biomedical Informatics, Yong Loo Lin School of Medicine, National University of Singapore C/O NUHS Tower Block, Level 8, 1E Kent Ridge Road, Singapore, 119228, Singapore.
- Department of Family Medicine, Faculty of Medicine and Health Sciences, Universiti Putra Malaysia, Serdang, Selangor, 43400, Malaysia.
| | - Kee Yuan Ngiam
- Department of Biomedical Informatics, Yong Loo Lin School of Medicine, National University of Singapore C/O NUHS Tower Block, Level 8, 1E Kent Ridge Road, Singapore, 119228, Singapore
- Department of Surgery, Division of General Surgery (Thyroid and Endocrine Surgery), National University of Singapore, University Surgical Cluster, National University Hospital National University Health System Corporate Office, Singapore, Singapore
| |
Collapse
|
17
|
S Krishna N, Garza-Frias E, Dasegowda G, Kaviani P, Karout L, Fahimi R, Bizzo B, Dreyer KJ, Kalra MK, Digumarthy S. Generalizability of AI-based image segmentation and centering estimation algorithm: a multi-region, multi-center, and multi-scanner study. RADIATION PROTECTION DOSIMETRY 2025; 201:441-449. [PMID: 40197806 DOI: 10.1093/rpd/ncaf018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 12/22/2024] [Accepted: 04/03/2025] [Indexed: 04/10/2025]
Abstract
We created and validated an open-access AI algorithm (AIc) for assessing image segmentation and patient centering in a multi-body-region, multi-center, and multi-scanner study. Our study included 825 head, chest, and abdomen-pelvis CT from 275 patients (153 females, 128 males; mean age 67 ± 14 years) scanned at five academic and community hospitals. CT images were processed with the AIc to determine vertical and horizontal centering at the skull base (head CT), carina (chest CT), and L2-L3 disc (abdomen CT). We manually measured the vertical and horizontal off-centering. We found strong correlations between AIc and manual estimate of off-centering in both the vertical (head, r = 0.93; chest, r = 0.94; abdomen, and r = 0.95) and horizontal directions (head CT, r = 0.85; chest, r = 0.85; abdomen, r = 0.8) and across age groups (r = 0.70-0.97), gender (r = 0.81-0.96), and multiple scanners from the five sites (r = 0.74-0.99). The AIc area under the receiver operating characteristic curve for centered and off-centered CT exams ranged from 0.72 (head) to 0.99 (chest). Therefore, our study showed that positron-emission tomography/CT (PET/CT) examinations commonly exhibit significant off-centering, particularly with vertical deviations often exceeding 30 mm and horizontal deviations between 10 and 30 mm. In addition, it demonstrated that our AI model can effectively assess both vertical and horizontal off-centering, although it performs better at estimating vertical off-centering.
Collapse
Affiliation(s)
- Neal S Krishna
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
- University of Connecticut School of Medicine, 200 Academic Wy, Farmington, CT 06032, United States
| | - Emiliano Garza-Frias
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Giridhar Dasegowda
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
- University of Arkansas for Medical Sciences, 4301 West Markham Street, Little Rock, AR 72205, United States
| | - Parisa Kaviani
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Lina Karout
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Roshan Fahimi
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Bernardo Bizzo
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Keith J Dreyer
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Mannudeep K Kalra
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| | - Subba Digumarthy
- Department of Radiology, Massachusetts General Hospital and Harvard Medical School, 75 Blossom Court, Suite 248, Boston, MA 02114, United States
| |
Collapse
|
18
|
Villagran Asiares A, Vitadello T, Velarde OM, Schachoff S, Ibrahim T, Nekolla SG. Can multiparametric FDG-PET/MRI analysis really enhance the prediction of myocardial recovery after CTO revascularization? A machine learning study. Z Med Phys 2025:S0939-3889(25)00038-8. [PMID: 40268665 DOI: 10.1016/j.zemedi.2025.03.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2024] [Revised: 03/15/2025] [Accepted: 03/28/2025] [Indexed: 04/25/2025]
Abstract
PURPOSE To comprehensively evaluate the effectiveness of FDG-PET/MRI multiparametric analysis in predicting myocardial wall motion recovery following revascularization of chronic coronary total occlusions (CTO), incorporating both traditional and machine learning approaches. METHODS This retrospective study assessed fluorine-18 fluorodeoxyglucose uptake (FDG), late gadolinium enhanced MR imaging (LGE), and MR wall motion abnormalities (WMA) of the left ventricle walls of a clinical cohort with 21 CTO patients (62 ± 9 years, 20 men). All patients were examined using a PET/MRI prior to revascularization and a follow-up cardiac MRI six months later. Prediction models for wall motion recovery after perfusion restoration were developed using linear and nonlinear algorithms as well as multiparametric variables. Performance and prediction explainability were evaluated in a 5x2 cross-validation framework, using ROC AUC and McNemar's test modified for clustered matched-pair data, and Shapley values. RESULTS Based on 79 CTO-subtended myocardial wall segments with wall motion abnormalities at baseline, the reference logistic regression model LGE + FDG obtained 0.55(SE = 0.07) in the clustered ROC AUC (cROC AUC) and 0.17(0.05) in the Global Absolute Shapley value. The reference outperformed FDG standalone in cROC AUC (-35(17) %, p < 0.0001), but not LGE standalone (11(12) %, p > 0.05). There were no statistically significant differences between the marginal probabilities of success of these three models. Moreover, no significant improvements (differences < 10 % in cROC AUC, and < 20 % in Global Absolute Shapley, p > 0.05) were found when using mixed effects logistic regression, decision tree, k-nearest neighbor, Naive Bayes, random forest, and support vector machine, with multiparametric combinations of FDG, LGE, and/or WMA. CONCLUSION In this clinical cohort, adding more complex interactions between PET/MRI imaging of cardiac function, infarct extension, and/or metabolism did not enhance the prediction of wall motion recovery after perfusion restoration. This finding raises the question whether multiparametric FDG-PET/MRI analysis has demonstrable benefits in risk stratification for CTO revascularization. Further studies with larger cohorts and external validation datasets are crucial to clarify this question and refine the role of multiparametric imaging in this context.
Collapse
Affiliation(s)
- Alberto Villagran Asiares
- Nuklearmedizinische Klinik und Poliklinik, TUM Klinikum, Klinikum rechts der Isar, School of Medicine and Health, Technical University of Munich, Munich, Germany; Klinik und Poliklinik für Kinder- und Jugendpsychiatrie, Psychosomatik und Psychotherapie. Klinikum der Ludwig-Maximilians-Universität München, Munich, Germany.
| | - Teresa Vitadello
- Klinik und Poliklinik für Innere Medizin I, TUM Klinikum, Klinikum rechts der Isar, School of Medicine and Health, Technical University of Munich, Munich, Germany
| | - Osvaldo M Velarde
- Biomedical Engineering Department, The City College of New York, New York, NY 10030, United States
| | - Sylvia Schachoff
- Nuklearmedizinische Klinik und Poliklinik, TUM Klinikum, Klinikum rechts der Isar, School of Medicine and Health, Technical University of Munich, Munich, Germany
| | - Tareq Ibrahim
- Klinik und Poliklinik für Innere Medizin I, TUM Klinikum, Klinikum rechts der Isar, School of Medicine and Health, Technical University of Munich, Munich, Germany
| | - Stephan G Nekolla
- Nuklearmedizinische Klinik und Poliklinik, TUM Klinikum, Klinikum rechts der Isar, School of Medicine and Health, Technical University of Munich, Munich, Germany; Deutsches Zentrum für Herz-Kreislauf-Forschung e.V., partner site Munich Heart Alliance, Munich, Germany
| |
Collapse
|
19
|
Park JS, Park SY, Moon JW, Kim K, Suh DI. Artificial Intelligence Models for Pediatric Lung Sound Analysis: Systematic Review and Meta-Analysis. J Med Internet Res 2025; 27:e66491. [PMID: 40249944 PMCID: PMC12048790 DOI: 10.2196/66491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2024] [Revised: 02/14/2025] [Accepted: 03/13/2025] [Indexed: 04/20/2025] Open
Abstract
BACKGROUND Pediatric respiratory diseases, including asthma and pneumonia, are major causes of morbidity and mortality in children. Auscultation of lung sounds is a key diagnostic tool but is prone to subjective variability. The integration of artificial intelligence (AI) and machine learning (ML) with electronic stethoscopes offers a promising approach for automated and objective lung sound. OBJECTIVE This systematic review and meta-analysis assess the performance of ML models in pediatric lung sound analysis. The study evaluates the methodologies, model performance, and database characteristics while identifying limitations and future directions for clinical implementation. METHODS A systematic search was conducted in Medline via PubMed, Embase, Web of Science, OVID, and IEEE Xplore for studies published between January 1, 1990, and December 16, 2024. Inclusion criteria are as follows: studies developing ML models for pediatric lung sound classification with a defined database, physician-labeled reference standard, and reported performance metrics. Exclusion criteria are as follows: studies focusing on adults, cardiac auscultation, validation of existing models, or lacking performance metrics. Risk of bias was assessed using a modified Quality Assessment of Diagnostic Accuracy Studies (version 2) framework. Data were extracted on study design, dataset, ML methods, feature extraction, and classification tasks. Bivariate meta-analysis was performed for binary classification tasks, including wheezing and abnormal lung sound detection. RESULTS A total of 41 studies met the inclusion criteria. The most common classification task was binary detection of abnormal lung sounds, particularly wheezing. Pooled sensitivity and specificity for wheeze detection were 0.902 (95% CI 0.726-0.970) and 0.955 (95% CI 0.762-0.993), respectively. For abnormal lung sound detection, pooled sensitivity was 0.907 (95% CI 0.816-0.956) and specificity 0.877 (95% CI 0.813-0.921). The most frequently used feature extraction methods were Mel-spectrogram, Mel-frequency cepstral coefficients, and short-time Fourier transform. Convolutional neural networks were the predominant ML model, often combined with recurrent neural networks or residual network architectures. However, high heterogeneity in dataset size, annotation methods, and evaluation criteria were observed. Most studies relied on small, single-center datasets, limiting generalizability. CONCLUSIONS ML models show high accuracy in pediatric lung sound analysis, but face limitations due to dataset heterogeneity, lack of standard guidelines, and limited external validation. Future research should focus on standardized protocols and the development of large-scale, multicenter datasets to improve model robustness and clinical implementation.
Collapse
Affiliation(s)
- Ji Soo Park
- Department of Pediatrics, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Sa-Yoon Park
- The Institute of Convergence Medicine with Innovative Technology, Seoul National University Hospital, Seoul, Republic of Korea
- Department of Physiology, College of Korean Medicine, Wonkwang University, Iksan, Republic of Korea
| | - Jae Won Moon
- Department of Pediatrics, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Kwangsoo Kim
- The Institute of Convergence Medicine with Innovative Technology, Seoul National University Hospital, Seoul, Republic of Korea
- Department of Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Dong In Suh
- Department of Pediatrics, Seoul National University College of Medicine, Seoul, Republic of Korea
| |
Collapse
|
20
|
Zhang C, Gao X, Zheng X, Xie J, Feng G, Bao Y, Gu P, He C, Wang R, Tian J. A fully automated, expert-perceptive image quality assessment system for whole-body [18F]FDG PET/CT. EJNMMI Res 2025; 15:42. [PMID: 40249445 PMCID: PMC12008089 DOI: 10.1186/s13550-025-01238-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2024] [Accepted: 04/05/2025] [Indexed: 04/19/2025] Open
Abstract
BACKGROUND The quality of clinical PET/CT images is critical for both accurate diagnosis and image-based research. However, current image quality assessment (IQA) methods predominantly rely on handcrafted features and region-specific analyses, thereby limiting automation in whole-body and multicenter evaluations. This study aims to develop an expert-perceptive deep learning-based IQA system for [18F]FDG PET/CT to tackle the lack of automated, interpretable assessments of clinical whole-body PET/CT image quality. METHODS This retrospective multicenter study included clinical whole-body [18F]FDG PET/CT scans from 718 patients. Automated identification and localization algorithms were applied to select predefined pairs of PET and CT slices from whole-body images. Fifteen experienced experts, trained to conduct blinded slice-level subjective assessments, provided average visual scores as reference standards. Using the MANIQA framework, the developed IQA model integrates the Vision Transformer, Transposed Attention, and Scale Swin Transformer Blocks to categorize PET and CT images into five quality classes. The model's correlation, consistency, and accuracy with expert evaluations on both PET and CT test sets were statistically analysed to assess the system's IQA performance. Additionally, the model's ability to distinguish high-quality images was evaluated using receiver operating characteristic (ROC) curves. RESULTS The IQA model demonstrated high accuracy in predicting image quality categories and showed strong concordance with expert evaluations of PET/CT image quality. In predicting slice-level image quality across all body regions, the model achieved an average accuracy of 0.832 for PET and 0.902 for CT. The model's scores showed substantial agreement with expert assessments, achieving average Spearman coefficients (ρ) of 0.891 for PET and 0.624 for CT, while the average Intraclass Correlation Coefficient (ICC) reached 0.953 for PET and 0.92 for CT. The PET IQA model demonstrated strong discriminative performance, achieving an area under the curve (AUC) of ≥ 0.88 for both the thoracic and abdominal regions. CONCLUSIONS This fully automated IQA system provides a robust and comprehensive framework for the objective evaluation of clinical image quality. Furthermore, it demonstrates significant potential as an impartial, expert-level tool for standardised multicenter clinical IQA.
Collapse
Affiliation(s)
- Cong Zhang
- Medical School of Chinese PLA, Beijing, China
- Department of Nuclear Medicine, The First Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Xin Gao
- Shanghai Universal Medical Imaging Diagnostic Center, Shanghai, China
| | - Xuebin Zheng
- Department of Scientific Research, Shanghai Aitrox Technology Corporation Limited, Shanghai, China
| | - Jun Xie
- Department of Scientific Research, Shanghai Aitrox Technology Corporation Limited, Shanghai, China
| | - Gang Feng
- Shanghai Universal Medical Imaging Diagnostic Center, Shanghai, China
| | - Yunchao Bao
- Department of Scientific Research, Shanghai Aitrox Technology Corporation Limited, Shanghai, China
| | - Pengchen Gu
- Department of Scientific Research, Shanghai Aitrox Technology Corporation Limited, Shanghai, China
| | - Chuan He
- Department of Scientific Research, Shanghai Aitrox Technology Corporation Limited, Shanghai, China
| | - Ruimin Wang
- Medical School of Chinese PLA, Beijing, China.
| | - Jiahe Tian
- Medical School of Chinese PLA, Beijing, China.
| |
Collapse
|
21
|
Bian Y, Wang L, Li J, Yang X, Wang E, Li Y, Liu Y, Xiang L, Yang Q. Quantitative Ischemic Lesions of Portable Low-Field Strength MRI Using Deep Learning-Based Super-Resolution. Stroke 2025. [PMID: 40235448 DOI: 10.1161/strokeaha.124.050540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2024] [Revised: 03/07/2025] [Accepted: 03/24/2025] [Indexed: 04/17/2025]
Abstract
BACKGROUND Deep learning-based synthetic super-resolution magnetic resonance imaging (SynthMRI) may improve the quantitative lesion performance of portable low-field strength magnetic resonance imaging (LF-MRI). The aim of this study is to evaluate whether SynthMRI improves the diagnostic performance of LF-MRI in assessing ischemic lesions. METHODS We retrospectively included 178 stroke patients and 104 healthy controls with both LF-MRI and high-field strength magnetic resonance imaging (HF-MRI) examinations. Using HF-MRI as the ground truth, the deep learning-based super-resolution framework (SCUNet) was pretrained using large-scale open-source data sets to generate SynthMRI images from LF-MRI images. Participants were split into a training set (64.2%) to fine-tune the pretrained SCUNet, and a testing set (35.8%) to evaluate the performance of SynthMRI. Sensitivity and specificity of LF-MRI and SynthMRI were assessed. Agreement with HF-MRI for Alberta Stroke Program Early Computed Tomography Score in the anterior and posterior circulation (diffusion-weighted imaging-Alberta Stroke Program Early Computed Tomography Score and diffusion-weighted imaging-posterior circulation Alberta Stroke Program Early Computed Tomography Score) was evaluated using intraclass correlation coefficients (ICCs). Agreement with HF-MRI for lesion volume and mean apparent diffusion coefficient (ADC) within lesions was assessed using both ICCs and Pearson correlation coefficients. RESULTS SynthMRI demonstrated significantly higher sensitivity and specificity than LF-MRI (89.0% [83.3%-94.6%] versus 77.1% [69.5%-84.7%]; P<0.001 and 91.3% [84.7%-98.0%] versus 71.0% [60.3%-81.7%]; P<0.001, respectively). The ICCs of diffusion-weighted imaging-Alberta Stroke Program Early Computed Tomography Score between SynthMRI and HF-MRI were also better than that between LF-MRI and HF-MRI (0.952 [0.920-0.972] versus 0.797 [0.678-0.876], P<0.001). For lesion volume and mean apparent diffusion coefficient within lesions, SynthMRI showed significantly higher agreement (P<0.001) with HF-MRI (ICC>0.85, r>0.78) than LF-MRI (ICC>0.45, r>0.35). Furthermore, for lesions during various poststroke phases, SynthMRI exhibited significantly higher agreement with HF-MRI than LF-MRI during the early hyperacute and subacute phases. CONCLUSIONS SynthMRI demonstrates high agreement with HF-MRI in detecting and quantifying ischemic lesions and is better than LF-MRI, particularly for lesions during the early hyperacute and subacute phases.
Collapse
Affiliation(s)
- Yueyan Bian
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Long Wang
- Subtle Medical, Shanghai, China (L.W., L.X.)
| | - Jin Li
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Xiaoxu Yang
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Erling Wang
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Yingying Li
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Yuehong Liu
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
| | - Lei Xiang
- Subtle Medical, Shanghai, China (L.W., L.X.)
| | - Qi Yang
- Department of Radiology, Beijing Chaoyang Hospital, Capital Medical University, China (Y.B., J.L., X.Y., E.W., Y. Li, Y. Liu, Q.Y.)
- Laboratory for Clinical Medicine, Capital Medical University, Beijing, China (Q.Y.)
| |
Collapse
|
22
|
O'Rourke S, Xu S, Carrero S, Drebin HM, Felman A, Ko A, Misseldine A, Mouchtaris SG, Musialowicz B, Wong TT, Zech JR. AI as teacher: effectiveness of an AI-based training module to improve trainee pediatric fracture detection. Skeletal Radiol 2025:10.1007/s00256-025-04927-0. [PMID: 40227327 DOI: 10.1007/s00256-025-04927-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/17/2025] [Revised: 04/03/2025] [Accepted: 04/04/2025] [Indexed: 04/15/2025]
Abstract
OBJECTIVE Prior work has demonstrated that AI access can help residents more accurately detect pediatric fractures. We wished to evaluate the effectiveness of an unsupervised AI-based training module as a pediatric fracture detection educational tool. MATERIALS AND METHODS Two hundred forty radiographic examinations from throughout the pediatric upper extremity were split into two groups of 120 examinations. A previously developed open-source deep learning fracture detection algorithm ( www.childfx.com ) was used to annotate radiographs. Four medical students and four PGY-2 radiology residents first evaluated 120 examinations for fracture without AI assistance and subsequently reviewed AI annotations on these cases via a training module. They then interpreted 120 different examinations without AI assistance. Pre- and post-intervention fracture detection accuracy was evaluated using a chi-squared test. RESULTS Overall resident fracture detection accuracy significantly improved from 71.3% pre-intervention to 77.5% post-intervention (p = 0.032). Medical student fracture detection accuracy was not significantly changed from 56.3% pre-intervention to 57.3% post-intervention (p = 0.794). Eighty-eight percent of responding participants (7/8) would recommend this model of learning. CONCLUSION We found that a tailored AI-based training module increased resident accuracy for detecting pediatric fractures by 6.2%. Medical student accuracy was not improved, likely due to their limited background familiarity with the task. AI offers a scalable method for automatically generating annotated teaching cases covering varied pathology, allowing residents to efficiently learn from simulated experience.
Collapse
Affiliation(s)
- Sean O'Rourke
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - Sophia Xu
- Department of Medical Education, Columbia University Vagelos College of Physicians and Surgeons, New York, NY, USA
| | - Stephanie Carrero
- Department of Medical Education, Columbia University Vagelos College of Physicians and Surgeons, New York, NY, USA
| | - Harrison M Drebin
- Department of Medical Education, Columbia University Vagelos College of Physicians and Surgeons, New York, NY, USA
| | - Ariel Felman
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - Andrew Ko
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - Adam Misseldine
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - Sofia G Mouchtaris
- Department of Medical Education, Columbia University Vagelos College of Physicians and Surgeons, New York, NY, USA
| | - Brett Musialowicz
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - Tony T Wong
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA
| | - John R Zech
- Department of Radiology, Columbia University Irving Medical Center, 622 W 168 St, New York, NY, 10032, USA.
| |
Collapse
|
23
|
Kennedy E, Vadlamani S, Lindsey HM, Peterson KS, Dams O'Connor K, Agarwal R, Amiri HH, Andersen RK, Babikian T, Baron DA, Bigler ED, Caeyenberghs K, Delano-Wood L, Disner SG, Dobryakova E, Eapen BC, Edelstein RM, Esopenko C, Genova HM, Geuze E, Goodrich-Hunsaker NJ, Grafman J, Håberg AK, Hodges CB, Hoskinson KR, Hovenden ES, Irimia A, Jahanshad N, Jha RM, Keleher F, Kenney K, Koerte IK, Liebel SW, Livny A, Løvstad M, Martindale SL, Max JE, Mayer AR, Meier TB, Menefee DS, Mohamed AZ, Mondello S, Monti MM, Morey RA, Newcombe V, Newsome MR, Olsen A, Pastorek NJ, Pugh MJ, Razi A, Resch JE, Rowland JA, Russell K, Ryan NP, Scheibel RS, Schmidt AT, Spitz G, Stephens JA, Tal A, Talbert LD, Tartaglia MC, Taylor BA, Thomopoulos SI, Troyanskaya M, Valera EM, van der Horn HJ, Van Horn JD, Verma R, Wade BSC, Walker WC, Ware AL, Werner JK, Yeates KO, Zafonte RD, Zeineh MM, Zielinski B, Thompson PM, Hillary FG, Tate DF, Wilde EA, Dennis EL. Linking Symptom Inventories Using Semantic Textual Similarity. J Neurotrauma 2025. [PMID: 40200899 DOI: 10.1089/neu.2024.0301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/10/2025] Open
Abstract
An extensive library of symptom inventories has been developed over time to measure clinical symptoms of traumatic brain injury (TBI), but this variety has led to several long-standing issues. Most notably, results drawn from different settings and studies are not comparable. This creates a fundamental problem in TBI diagnostics and outcome prediction, namely that it is not possible to equate results drawn from distinct tools and symptom inventories. Here, we present an approach using semantic textual similarity (STS) to link symptoms and scores across previously incongruous symptom inventories by ranking item text similarities according to their conceptual likeness. We tested the ability of four pretrained deep learning models to screen thousands of symptom description pairs for related content-a challenging task typically requiring expert panels. Models were tasked to predict symptom severity across four different inventories for 6,607 participants drawn from 16 international data sources. The STS approach achieved 74.8% accuracy across five tasks, outperforming other models tested. Correlation and factor analysis found the properties of the scales were broadly preserved under conversion. This work suggests that incorporating contextual, semantic information can assist expert decision-making processes, yielding broad gains for the harmonization of TBI assessment.
Collapse
Affiliation(s)
- Eamonn Kennedy
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
- Division of Epidemiology, University of Utah, Salt Lake City, Utah, USA
| | - Shashank Vadlamani
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- Division of Epidemiology, University of Utah, Salt Lake City, Utah, USA
| | - Hannah M Lindsey
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
| | - Kelly S Peterson
- Division of Epidemiology, University of Utah, Salt Lake City, Utah, USA
- Veterans Health Administration (VHA) Office of Analytics and Performance Integration (VHA), Washington, District of Columbia, USA
| | - Kristen Dams O'Connor
- Department of Rehabilitation and Human Performance, Icahn School of Medicine at Mount Sinai, New York, New York, USA
- Department of Neurology, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Ronak Agarwal
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
| | - Houshang H Amiri
- Neuroscience Research Center, Institute of Neuropharmacology, Kerman University of Medical Sciences, Kerman, Iran
- Department of Radiology and Nuclear Medicine, Vrije Universiteit Amsterdam, Amsterdam UMC Location VUmc, Amsterdam, The Netherlands
| | - Raeda K Andersen
- Crawford Research Institute, Shepherd Center, Atlanta, Georgia, USA
- Department of Sociology, Georgia State University, Atlanta, Georgia, USA
| | - Talin Babikian
- Department of Psychiatry and Biobehavioral Sciences, Semel Institute for Neuroscience and Human Behavior, UCLA, Los Angeles, California, USA
- UCLA BrainSPORT Program, Los Angeles, California, USA
| | - David A Baron
- Department of Psychiatry, Center for Behavioral Health and Sport, Western University of Health Sciences, Lebanon, Pomona, California, USA
| | - Erin D Bigler
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- Department of Psychology and Neuroscience Center, Brigham Young University, Provo, Utah, USA
- Department of Psychiatry, University of Utah, Salt Lake City, Utah, USA
| | - Karen Caeyenberghs
- Cognitive Neuroscience Unit, School of Psychology, Deakin University, Geelong, Australia
| | - Lisa Delano-Wood
- VA San Diego Healthcare System; Center of Stress and Mental Health, and Department of Psychiatry, UC San Diego School of Medicine, San Diego, California, USA
- Department of Psychiatry, UC San Diego School of Medicine, San Diego, California, USA
| | - Seth G Disner
- Minneapolis VA Health Care System, Minneapolis, Minnesota, USA
- Department of Psychiatry, University of Minnesota, Minneapolis, Minnesota, USA
| | - Ekaterina Dobryakova
- Center for Traumatic Brain Injury, Kessler Foundation, East Hanover, New Jersey, USA
- Rutgers New Jersey Medical School, Newark, New Jersey, USA
| | - Blessen C Eapen
- VA Greater Los Angeles Health Care System, Los Angeles, California, USA
- Division of Physical Medicine and Rehabilitation, David Geffen School of Medicine at UCLA, Los Angeles, California, USA
| | - Rachel M Edelstein
- Department of Psychology, University of Virginia, Charlottesville, Virginia, USA
| | - Carrie Esopenko
- Department of Rehabilitation and Human Performance, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Helen M Genova
- Center for Autism Research, Kessler Foundation, East Hanover, New Jersey, USA
| | - Elbert Geuze
- Ministry of Defence, Brain Research and Innovation Centre, Utrecht, The Netherlands
| | | | - Jordan Grafman
- Shirley Ryan Ability Lab, Chicago, Illinois, USA
- Northwestern University, Chicago, Illinois, USA
| | - Asta K Håberg
- Faculty of Medicine and Health Sciences, Department of Neuromedicine and Movement Science, Norwegian University of Science and Technology (NTNU), Trondheim, Norway
- Department for Physical Health and Aging, Norwegian Institute of Public Health, Oslo, Norway
| | - Cooper B Hodges
- Department of Psychology, Brigham Young University, Provo, Utah, USA
| | - Kristen R Hoskinson
- Center for Biobehavioral Health, Abigail Wexner Research Institute at Nationwide Children's Hospital, Columbus, Ohio, USA
- Department of Pediatrics, The Ohio State University College of Medicine, Columbus, Ohio, USA
| | - Elizabeth S Hovenden
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
| | - Andrei Irimia
- Ethel Percy Andrus Gerontology Center, Leonard Davis School of Gerontology, University of Southern California, Los Angeles, California, USA
- Alfred E. Mann Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, California, USA
- Department of Quantitative and Computational Biology, Dornsife College of Arts and Sciences, University of Southern California, Los Angeles, California, USA
| | - Neda Jahanshad
- Imaging Genetics Center, Stevens Neuroimaging & Informatics Institute, Keck School of Medicine of USC, Marina del Rey, California, USA
| | - Ruchira M Jha
- Departments of Neurology, Translational Neuroscience, Barrow Neurological Institute, Phoenix, Arizona, USA
| | - Finian Keleher
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
| | - Kimbra Kenney
- Department of Neurology, Uniformed Services University of the Health Sciences, Bethesda, Maryland, USA
- National Intrepid Center of Excellence, Walter Reed National Military Medical Center, Bethesda, Maryland, USA
| | - Inga K Koerte
- Psychiatry Neuroimaging Laboratory, Brigham & Women's Hospital, Boston, Massachusetts, USA
- Department of Child and Adolescent Psychiatry, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
| | - Spencer W Liebel
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
| | - Abigail Livny
- Division of Diagnostic Imaging, Sheba Medical Center, Tel-Aviv, Israel
- Faculty of Medicine, Tel-Aviv University, Tel-Aviv, Israel
| | - Marianne Løvstad
- Sunnaas Rehabilitation Centre, Nesodden, Norway
- University of Oslo, Oslo, Norway
| | - Sarah L Martindale
- W. G. (Bill) Hefner VA Healthcare System, Salisbury, North Carolina, USA
- Department of Physiology & Pharmacology, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
| | - Jeffrey E Max
- Department of Psychiatry, UC San Diego, La Jolla, California, USA
| | | | - Timothy B Meier
- Department of Neurosurgery, Medical College of Wisconsin, Milwaukee, Wisconsin, USA
| | - Deleene S Menefee
- Michael E. DeBakey Veterans Affairs Medical Center, Houston, Texas, USA
- The Menning Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, Texas, USA
| | - Abdalla Z Mohamed
- Thompson Institute, University of the Sunshine Coast, Birtinya, Australia
| | - Stefania Mondello
- Department of Biomedical and Dental Sciences and Morphofunctional Imaging, University of Messina, Messina, Italy
| | - Martin M Monti
- Department of Psychology, University of California Los Angeles, Los Angeles, California, USA
- Department of Neurosurgery, Brain Injury Research Center (BIRC), University of California Los Angeles, Los Angeles, California, USA
| | - Rajendra A Morey
- Brain Imaging and Analysis Center, Duke University, Durham, North Carolina, USA
- Department of Psychiatry and Behavioral Sciences, Duke University, Durham, North Carolina, USA
| | | | - Mary R Newsome
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
| | - Alexander Olsen
- Department of Psychology, NTNU, Trondheim, Norway
- Clinic of Rehabilitation, St. Olavs Hospital, Trondheim University Hospital, Trondheim, Norway
- NorHEAD - Norwegian Centre for Headache Research, NTNU, Trondheim, Norway
| | - Nicholas J Pastorek
- Michael E. DeBakey Veterans Affairs Medical Center, Houston, Texas, USA
- H. Ben Taub Department of Physical Medicine and Rehabilitation, Baylor College of Medicine, Houston, Texas, USA
| | - Mary Jo Pugh
- Information Decision-Enhancement and Analytic Sciences Center, VA Salt Lake City, Salt Lake City, Utah, USA
- Department of Medicine, University of Utah School of Medicine, Salt Lake City, Utah, USA
| | - Adeel Razi
- Turner Institute for Brain and Mental Health, Monash University, Melbourne, Australia
- Wellcome Centre for Human Neuroimaging, University College London, London, United Kingdom
- CIFAR Azrieli Global Scholars Program, CIFAR, Toronto, Canada
| | - Jacob E Resch
- Department of Kinesiology, University of Virginia, Charlottesville, Virginia, USA
| | - Jared A Rowland
- W. G. (Bill) Hefner VA Healthcare System, Salisbury, North Carolina, USA
- Department of Neurobiology and Anatomy, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
- Mid-Atlantic Mental Illness Research and Education Center, Durham, North Carolina, USA
| | - Kelly Russell
- Department of Pediatrics and Child Health, University of Manitoba, Winnipeg, Canada
- Children's Hospital Research Institute of Manitoba, Winnipeg, Canada
| | - Nicholas P Ryan
- Cognitive Neuroscience Unit, School of Psychology, Deakin University, Geelong, Australia
- Brain and Mind Research, Murdoch Children's Research Institute, Parkville, Australia
| | - Randall S Scheibel
- Michael E. DeBakey Veterans Affairs Medical Center, Houston, Texas, USA
- H. Ben Taub Department of Physical Medicine and Rehabilitation, Baylor College of Medicine, Houston, Texas, USA
| | - Adam T Schmidt
- Department of Psychological Sciences, Texas Tech University, Lubbock, Texas, USA
- Center of Excellence For Translational Neuroscience and Therapeutics, Lubbock, Texas, USA
| | - Gershon Spitz
- Monash-Epworth Rehabilitation Research Centre, School of Psychological Sciences, Monash University, Melbourne, Australia
- Department of Neuroscience, Central Clinical School, Monash University, Melbourne, Australia
| | - Jaclyn A Stephens
- College of Health and Human Sciences, Colorado State University, Fort Collins, Colorado, USA
- Molecular Cellular Integrative Neuroscience Program, Colorado State University, Fort Collins, Colorado, USA
| | - Assaf Tal
- Biomedical Engineering, Tel Aviv University, Tel Aviv Israel
| | - Leah D Talbert
- Department of Psychology, Brigham Young University, Provo, Utah, USA
| | - Maria Carmela Tartaglia
- University Health Network, Toronto, Canada
- Centre for Research in Neurodegenerative Diseases, University of Toronto, Toronto, Canada
| | - Brian A Taylor
- Department of Imaging Physics, The University of Texas M. D. Anderson Cancer Center, Houston, Texas, USA
| | - Sophia I Thomopoulos
- Imaging Genetics Center, Stevens Neuroimaging & Informatics Institute, Keck School of Medicine of USC, Marina del Rey, California, USA
| | - Maya Troyanskaya
- Michael E. DeBakey Veterans Affairs Medical Center, Houston, Texas, USA
- H. Ben Taub Department of Physical Medicine and Rehabilitation, Baylor College of Medicine, Houston, Texas, USA
| | - Eve M Valera
- Department of Psychiatry, Harvard Medical School, Boston, Massachusetts, USA
- Massachusetts General Hospital, Boston, Massachusetts, USA
| | | | - John D Van Horn
- Department of Psychology and School of Data Science, University of Virginia, Charlottesville, Virginia, USA
| | - Ragini Verma
- Department of Radiology, University of Pennsylvania, Pennsylvania, USA
- Cohen Veterans Bioscience, New York City, New York, USA
| | - Benjamin S C Wade
- Division of Neuropsychiatry and Neuromodulation, Massachusetts General Hospital and Harvard Medical School, Boston, Massachusetts, USA
| | - Willian C Walker
- Department of Physical Medicine & Rehabilitation, Virginia Commonwealth University, Richmond, Virginia, USA
- Richmond Veterans Affairs (VA) Medical Center, Central Virginia VA Health Care System, Richmond, Virginia, USA
| | - Ashley L Ware
- Department of Psychology, Georgia State University, Atlanta, Georgia, USA
- Department of Neurology, University of Utah, Salt Lake City, Utah, USA
| | - J Kent Werner
- Department of Neurology, Uniformed Services University of the Health Sciences, Bethesda, Maryland, USA
| | - Keith Owen Yeates
- Department of Psychology, Alberta Children's Hospital Research Institute, and the Hotchkiss Brain Institute, University of Calgary, Calgary, Canada
| | - Ross D Zafonte
- Spaulding Rehabilitation Hospital, Boston, Massachusetts, USA
- Department of Physical Medicine & Rehabilitation, Harvard Medical School, Boston, Massachusetts, USA
| | | | - Brandon Zielinski
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- Departments of Pediatrics, Neurology, and Neuroscience, University of Florida, Gainesville, Florida, USA
- Departments of Pediatrics, and Radiology, University of Utah, Salt Lake City, Utah, USA
| | - Paul M Thompson
- Imaging Genetics Center, Stevens Neuroimaging & Informatics Institute, Keck School of Medicine of USC, Marina del Rey, California, USA
- Departments of Neurology, Pediatrics, Psychiatry, Radiology, Engineering, and Ophthalmology, USC, Los Angeles, California, USA
| | - Frank G Hillary
- Department of Psychology, Penn State University, State College, Pennsylvania, USA
- Department of Neurology, Hershey Medical Center, State College, Pennsylvania, USA
| | - David F Tate
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
| | - Elisabeth A Wilde
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
| | - Emily L Dennis
- Department of Neurology, University of Utah School of Medicine, Salt Lake City, Utah, USA
- George E. Wahlen Veterans Affairs Medical Center, Salt Lake City, Utah, USA
| |
Collapse
|
24
|
Tourdias T, Bani-Sadr A, Lecler A. Can generative T2*-weighted images replace true T2*-weighted images in brain MRI? Diagn Interv Imaging 2025:S2211-5684(25)00071-3. [PMID: 40204535 DOI: 10.1016/j.diii.2025.03.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2025] [Accepted: 03/31/2025] [Indexed: 04/11/2025]
Affiliation(s)
- Thomas Tourdias
- Univ. Bordeaux, INSERM, Neurocentre Magendie, U1215, 33000 Bordeaux, France; CHU de Bordeaux, Neuroimagerie Diagnostique et Thérapeutique, 33000 Bordeaux, France.
| | - Alexandre Bani-Sadr
- Department of Neuroradiology, Neurological Hospital, Hospices Civils de Lyon, 69029 Bron, France; Univ. Lyon 1, CREATIS Laboratory, CNRS 5220 - UMR U1294, 69100 Villeurbanne, France
| | - Augustin Lecler
- Université Paris Cité, Faculté de Médecine, 75006 Paris, France; Department of Neuroradiology, Hôpital Fondation Adolphe de Rothschild, 75019 Paris, France
| |
Collapse
|
25
|
Chen YC, Chen L, Lai YL, Chang WT, Lee SY. AI-Driven Detection and Measurement of Keratinized Gingiva in Dental Photographs: Validation Using Reference Retainers. J Clin Periodontol 2025. [PMID: 40195567 DOI: 10.1111/jcpe.14164] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 02/27/2025] [Accepted: 03/23/2025] [Indexed: 04/09/2025]
Abstract
AIM To evaluate a deep learning (DL) model for detecting keratinized gingiva (KG) in dental photographs and validate its clinical applicability using reference retainers for calibration. MATERIALS AND METHODS A total of 576 sextant photographs were selected from 32 subjects, each with three sets of photographs: iodine-stained, unstained and line-marked retainers. Relative keratinized gingiva width (rKGW) was measured using visual, functional and histochemical staining methods with reference retainers. A pre-trained DeepLabv3 model with ResNet50 backbone was fine-tuned to predict KG areas, which were then applied to the photographs with line-marked retainers for subsequent rKGW measurement. RESULTS The AI model achieved a Dice coefficient of 93.30% and an accuracy of 93.32%. Using histochemical measurements as gold standards, the absolute differences in rKGW of AI measurements were statistically insignificant with visual (p = 0.935) and functional (p = 0.979) measurements. The adjusted difference between AI and histochemical measurements was 0.377 mm. AI closely matched histochemical measurements in the maxillary anterior region (0.011 mm, p = 0.903) but was significantly higher in the maxillary posterior region (0.327 mm, p < 0.05). CONCLUSIONS The proposed AI model is the first to reliably identify full-mouth KG, validated thoroughly using reference retainers. However, predictions for posterior teeth warrant further improvement.
Collapse
Affiliation(s)
- Ya-Chi Chen
- Department of Stomatology, Taipei Veterans General Hospital, Taipei, Taiwan
- School of Dentistry, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Ling Chen
- Institute of Hospital and Health Care Administration, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Yu-Lin Lai
- Department of Stomatology, Taipei Veterans General Hospital, Taipei, Taiwan
- School of Dentistry, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Wei-Ting Chang
- Department of Stomatology, Taipei Veterans General Hospital, Taipei, Taiwan
- School of Dentistry, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Shyh-Yuan Lee
- Department of Stomatology, Taipei Veterans General Hospital, Taipei, Taiwan
- School of Dentistry, National Yang Ming Chiao Tung University, Taipei, Taiwan
- Department of Dentistry, Taipei City Hospital, Taipei, Taiwan
- Oral Medicine Innovation Center, National Yang Ming Chiao Tung University, Taipei, Taiwan
| |
Collapse
|
26
|
Liu S, McCoy AB, Wright A. Improving large language model applications in biomedicine with retrieval-augmented generation: a systematic review, meta-analysis, and clinical development guidelines. J Am Med Inform Assoc 2025; 32:605-615. [PMID: 39812777 PMCID: PMC12005634 DOI: 10.1093/jamia/ocaf008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2024] [Revised: 12/17/2024] [Accepted: 01/03/2025] [Indexed: 01/16/2025] Open
Abstract
OBJECTIVE The objectives of this study are to synthesize findings from recent research of retrieval-augmented generation (RAG) and large language models (LLMs) in biomedicine and provide clinical development guidelines to improve effectiveness. MATERIALS AND METHODS We conducted a systematic literature review and a meta-analysis. The report was created in adherence to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses 2020 analysis. Searches were performed in 3 databases (PubMed, Embase, PsycINFO) using terms related to "retrieval augmented generation" and "large language model," for articles published in 2023 and 2024. We selected studies that compared baseline LLM performance with RAG performance. We developed a random-effect meta-analysis model, using odds ratio as the effect size. RESULTS Among 335 studies, 20 were included in this literature review. The pooled effect size was 1.35, with a 95% confidence interval of 1.19-1.53, indicating a statistically significant effect (P = .001). We reported clinical tasks, baseline LLMs, retrieval sources and strategies, as well as evaluation methods. DISCUSSION Building on our literature review, we developed Guidelines for Unified Implementation and Development of Enhanced LLM Applications with RAG in Clinical Settings to inform clinical applications using RAG. CONCLUSION Overall, RAG implementation showed a 1.35 odds ratio increase in performance compared to baseline LLMs. Future research should focus on (1) system-level enhancement: the combination of RAG and agent, (2) knowledge-level enhancement: deep integration of knowledge into LLM, and (3) integration-level enhancement: integrating RAG systems within electronic health records.
Collapse
Affiliation(s)
- Siru Liu
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37212, United States
- Department of Computer Science, Vanderbilt University, Nashville, TN 37212, United States
| | - Allison B McCoy
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37212, United States
| | - Adam Wright
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37212, United States
- Department of Medicine, Vanderbilt University Medical Center, Nashville, TN 37212, United States
| |
Collapse
|
27
|
Malhi BS, Lo J, Toto-Brocchi M, Avval AH, Ma Y, Du J. Quantitative magnetic resonance imaging in Alzheimer's disease: a narrative review. Quant Imaging Med Surg 2025; 15:3641-3664. [PMID: 40235823 PMCID: PMC11994541 DOI: 10.21037/qims-24-1602] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2024] [Accepted: 02/25/2025] [Indexed: 04/17/2025]
Abstract
Background and Objective Alzheimer's disease (AD) is a complex neurodegenerative disorder characterized by progressive cognitive decline and is traditionally associated with grey matter pathology. Recent research highlights the significance of white matter and myelin damage in AD, presenting a paradigm shift in understanding the disease. The aim of this study was to summarize current advancements in magnetic resonance imaging (MRI) techniques and their applications in assessing myelin and brain pathology in AD with a special focus on ultrashort echo time (UTE) based techniques, alongside the role of artificial intelligence (AI) in enhancing diagnostic accuracy. Methods Between April and May 2024, we conducted a literature search using Google Scholar, Web of Science, and PubMed, focusing on publications from 1990 to 2024. Search terms included "Quantitative imaging", "Alzheimer's MRI", "T1ρ Alzheimer's", "MT imaging Alzheimer's", and "myelin water fraction Alzheimer's". We included quantitative MRI studies involving AD brains and excluded volumetric analyses, non-quantitative studies, non-English reports, non-peer-reviewed studies, and animal research. Key Content and Findings Quantitative MRI techniques, including T1, T1ρ, magnetization transfer ratio (MTR), T2, T2*, susceptibility, myelin water fraction (MWF), and non-aqueous myelin proton density (PD) were described. These biomarkers represent different pathophysiological elements of brain damage and may have distinct functions at different phases of the disease. The role of AI in enhancing diagnostic accuracy is also discussed. Conclusions In conclusion, integrating advanced MRI techniques and AI offers promising avenues for understanding and diagnosing AD. The focus on myelin damage and white matter integrity underscores the importance of comprehensive imaging approaches. Continued research and development are essential to address current challenges and improve clinical practice in AD diagnostics.
Collapse
Affiliation(s)
| | - James Lo
- Department of Radiology, University of California, San Diego, La Jolla, CA, USA
- Department of Bioengineering, University of California, San Diego, La Jolla, CA, USA
| | - Marco Toto-Brocchi
- Department of Radiology, University of California, San Diego, La Jolla, CA, USA
| | | | - Yajun Ma
- Department of Radiology, University of California, San Diego, La Jolla, CA, USA
| | - Jiang Du
- Department of Radiology, University of California, San Diego, La Jolla, CA, USA
| |
Collapse
|
28
|
Kawata N, Iwao Y, Matsuura Y, Higashide T, Okamoto T, Sekiguchi Y, Nagayoshi M, Takiguchi Y, Suzuki T, Haneishi H. Generation of short-term follow-up chest CT images using a latent diffusion model in COVID-19. Jpn J Radiol 2025; 43:622-633. [PMID: 39585556 PMCID: PMC11953082 DOI: 10.1007/s11604-024-01699-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2024] [Accepted: 11/02/2024] [Indexed: 11/26/2024]
Abstract
PURPOSE Despite a global decrease in the number of COVID-19 patients, early prediction of the clinical course for optimal patient care remains challenging. Recently, the usefulness of image generation for medical images has been investigated. This study aimed to generate short-term follow-up chest CT images using a latent diffusion model in patients with COVID-19. MATERIALS AND METHODS We retrospectively enrolled 505 patients with COVID-19 for whom the clinical parameters (patient background, clinical symptoms, and blood test results) upon admission were available and chest CT imaging was performed. Subject datasets (n = 505) were allocated for training (n = 403), and the remaining (n = 102) were reserved for evaluation. The image underwent variational autoencoder (VAE) encoding, resulting in latent vectors. The information consisting of initial clinical parameters and radiomic features were formatted as a table data encoder. Initial and follow-up latent vectors and the initial table data encoders were utilized for training the diffusion model. The evaluation data were used to generate prognostic images. Then, similarity of the prognostic images (generated images) and the follow-up images (real images) was evaluated by zero-mean normalized cross-correlation (ZNCC), peak signal-to-noise ratio (PSNR), and structural similarity (SSIM). Visual assessment was also performed using a numerical rating scale. RESULTS Prognostic chest CT images were generated using the diffusion model. Image similarity showed reasonable values of 0.973 ± 0.028 for the ZNCC, 24.48 ± 3.46 for the PSNR, and 0.844 ± 0.075 for the SSIM. Visual evaluation of the images by two pulmonologists and one radiologist yielded a reasonable mean score. CONCLUSIONS The similarity and validity of generated predictive images for the course of COVID-19-associated pneumonia using a diffusion model were reasonable. The generation of prognostic images may suggest potential utility for early prediction of the clinical course in COVID-19-associated pneumonia and other respiratory diseases.
Collapse
Affiliation(s)
- Naoko Kawata
- Department of Respirology, Graduate School of Medicine, Chiba University, 1-8-1, Inohana, Chuo-Ku, Chiba-Shi, Chiba, 260-8677, Japan.
- Graduate School of Science and Engineering, Chiba University, Chiba, 263-8522, Japan.
| | - Yuma Iwao
- Center for Frontier Medical Engineering, Chiba University, 1-33, Yayoi-Cho, Inage-Ku, Chiba-Shi, Chiba, 263-8522, Japan
- Institute for Quantum Medical Science, National Institutes for Quantum Science and Technology, 4-9-1, Anagawa, Inage-Ku, Chiba-Shi, Chiba, 263-8555, Japan
| | - Yukiko Matsuura
- Department of Respiratory Medicine, Chiba Aoba Municipal Hospital, 1273-2 Aoba-Cho, Chuo-Ku, Chiba-Shi, Chiba, 260-0852, Japan
| | - Takashi Higashide
- Department of Radiology, Chiba University Hospital, 1-8-1, Inohana, Chuo-Ku, Chiba-Shi, Chiba, 260-8677, Japan
- Department of Radiology, Japanese Red Cross Narita Hospital, 90-1, Iida-Cho, Narita-Shi, Chiba, 286-8523, Japan
| | - Takayuki Okamoto
- Center for Frontier Medical Engineering, Chiba University, 1-33, Yayoi-Cho, Inage-Ku, Chiba-Shi, Chiba, 263-8522, Japan
| | - Yuki Sekiguchi
- Graduate School of Science and Engineering, Chiba University, Chiba, 263-8522, Japan
| | - Masaru Nagayoshi
- Department of Respiratory Medicine, Chiba Aoba Municipal Hospital, 1273-2 Aoba-Cho, Chuo-Ku, Chiba-Shi, Chiba, 260-0852, Japan
| | - Yasuo Takiguchi
- Department of Respiratory Medicine, Chiba Aoba Municipal Hospital, 1273-2 Aoba-Cho, Chuo-Ku, Chiba-Shi, Chiba, 260-0852, Japan
| | - Takuji Suzuki
- Department of Respirology, Graduate School of Medicine, Chiba University, 1-8-1, Inohana, Chuo-Ku, Chiba-Shi, Chiba, 260-8677, Japan
| | - Hideaki Haneishi
- Center for Frontier Medical Engineering, Chiba University, 1-33, Yayoi-Cho, Inage-Ku, Chiba-Shi, Chiba, 263-8522, Japan
| |
Collapse
|
29
|
Xu HL, Gong TT, Song XJ, Chen Q, Bao Q, Yao W, Xie MM, Li C, Grzegorzek M, Shi Y, Sun HZ, Li XH, Zhao YH, Gao S, Wu QJ. Artificial Intelligence Performance in Image-Based Cancer Identification: Umbrella Review of Systematic Reviews. J Med Internet Res 2025; 27:e53567. [PMID: 40167239 PMCID: PMC12000792 DOI: 10.2196/53567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2023] [Revised: 07/30/2024] [Accepted: 11/11/2024] [Indexed: 04/02/2025] Open
Abstract
BACKGROUND Artificial intelligence (AI) has the potential to transform cancer diagnosis, ultimately leading to better patient outcomes. OBJECTIVE We performed an umbrella review to summarize and critically evaluate the evidence for the AI-based imaging diagnosis of cancers. METHODS PubMed, Embase, Web of Science, Cochrane, and IEEE databases were searched for relevant systematic reviews from inception to June 19, 2024. Two independent investigators abstracted data and assessed the quality of evidence, using the Joanna Briggs Institute (JBI) Critical Appraisal Checklist for Systematic Reviews and Research Syntheses. We further assessed the quality of evidence in each meta-analysis by applying the Grading of Recommendations, Assessment, Development, and Evaluation (GRADE) criteria. Diagnostic performance data were synthesized narratively. RESULTS In a comprehensive analysis of 158 included studies evaluating the performance of AI algorithms in noninvasive imaging diagnosis across 8 major human system cancers, the accuracy of the classifiers for central nervous system cancers varied widely (ranging from 48% to 100%). Similarities were observed in the diagnostic performance for cancers of the head and neck, respiratory system, digestive system, urinary system, female-related systems, skin, and other sites. Most meta-analyses demonstrated positive summary performance. For instance, 9 reviews meta-analyzed sensitivity and specificity for esophageal cancer, showing ranges of 90%-95% and 80%-93.8%, respectively. In the case of breast cancer detection, 8 reviews calculated the pooled sensitivity and specificity within the ranges of 75.4%-92% and 83%-90.6%, respectively. Four meta-analyses reported the ranges of sensitivity and specificity in ovarian cancer, and both were 75%-94%. Notably, in lung cancer, the pooled specificity was relatively low, primarily distributed between 65% and 80%. Furthermore, 80.4% (127/158) of the included studies were of high quality according to the JBI Critical Appraisal Checklist, with the remaining studies classified as medium quality. The GRADE assessment indicated that the overall quality of the evidence was moderate to low. CONCLUSIONS Although AI shows great potential for achieving accelerated, accurate, and more objective diagnoses of multiple cancers, there are still hurdles to overcome before its implementation in clinical settings. The present findings highlight that a concerted effort from the research community, clinicians, and policymakers is required to overcome existing hurdles and translate this potential into improved patient outcomes and health care delivery. TRIAL REGISTRATION PROSPERO CRD42022364278; https://www.crd.york.ac.uk/PROSPERO/view/CRD42022364278.
Collapse
Affiliation(s)
- He-Li Xu
- Department of Clinical Epidemiology, Shengjing Hospital of China Medical University, Shenyang, China
- Clinical Research Center, Shengjing Hospital of China Medical University, Shenyang, China
- Liaoning Key Laboratory of Precision Medical Research on Major Chronic Disease, Shengjing Hospital of China Medical University, Shenyang, China
| | - Ting-Ting Gong
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Xin-Jian Song
- Department of Clinical Epidemiology, Shengjing Hospital of China Medical University, Shenyang, China
- Clinical Research Center, Shengjing Hospital of China Medical University, Shenyang, China
- Liaoning Key Laboratory of Precision Medical Research on Major Chronic Disease, Shengjing Hospital of China Medical University, Shenyang, China
| | - Qian Chen
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Qi Bao
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
- Department of Epidemiology, School of Public Health, China Medical University, Shenyang, China
| | - Wei Yao
- Department of Clinical Epidemiology, Shengjing Hospital of China Medical University, Shenyang, China
- Clinical Research Center, Shengjing Hospital of China Medical University, Shenyang, China
- Liaoning Key Laboratory of Precision Medical Research on Major Chronic Disease, Shengjing Hospital of China Medical University, Shenyang, China
| | - Meng-Meng Xie
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Chen Li
- Microscopic Image and Medical Image Analysis Group, College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Marcin Grzegorzek
- Institute for Medical Informatics, University of Luebeck, Luebeck, Germany
| | - Yu Shi
- Department of Radiology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Hong-Zan Sun
- Department of Radiology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Xiao-Han Li
- Department of Pathology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Yu-Hong Zhao
- Department of Clinical Epidemiology, Shengjing Hospital of China Medical University, Shenyang, China
- Clinical Research Center, Shengjing Hospital of China Medical University, Shenyang, China
- Liaoning Key Laboratory of Precision Medical Research on Major Chronic Disease, Shengjing Hospital of China Medical University, Shenyang, China
| | - Song Gao
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
| | - Qi-Jun Wu
- Clinical Research Center, Shengjing Hospital of China Medical University, Shenyang, China
- Liaoning Key Laboratory of Precision Medical Research on Major Chronic Disease, Shengjing Hospital of China Medical University, Shenyang, China
- Department of Obstetrics and Gynecology, Shengjing Hospital of China Medical University, Shenyang, China
- Department of Clinical Epidemiology, Shengjing Hospital of China Medical University, Shenyang, Liaoning, China
- NHC Key Laboratory of Advanced Reproductive Medicine and Fertility (China Medical University), National Health Commission, Shenyang, China
| |
Collapse
|
30
|
Jensen CG, Rasmussen BS, Overgaard S, Varnum C, Haubro MH, Jensen J. A deep learning algorithm for radiographic measurements of the hip versus human CT measurements: An intermodality agreement study. Acta Radiol Open 2025; 14:20584601251330554. [PMID: 40162114 PMCID: PMC11948554 DOI: 10.1177/20584601251330554] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Revised: 02/25/2025] [Accepted: 03/09/2025] [Indexed: 04/02/2025] Open
Abstract
Background Hip dysplasia (HD) is a prevalent cause of non-traumatic hip pain, which may result in osteoarthritis. Radiological measurements of HD exhibit variability based on reader and imaging modality, why it is important to know the agreement between different measurement methods. Purpose To estimate agreement between measurements of lateral center edge angle (LCEA) and acetabular inclination angle (AIA) made, respectively, on Computed Tomography (CT) scans by humans and radiographs analyzed by an algorithm. To estimate impact of pelvic rotation on agreement between CT and radiographic measurements. Material and Methods CT measurements were retrospectively extracted from 172 radiology reports. Radiographs were analyzed using an algorithm. Bland-Altman analysis assessed agreement between CT and radiographic measurements. Regression analyses estimated impact of pelvic rotation on inter-modality agreement. Results Mean measured bias (95% confidence interval [CI]) between CT and radiographs for LCEA of right/left hip was 5.53° (95% CI: 4.81 to 6.24) and 5.13 (95% CI: 4.43 to 5.83), respectively. Corresponding values for right/left AIA were 1.08 (95% CI: 0.49 to 1.67) and -0.03 (95% CI: -0.60 to 0.05). Pelvic rotation affected right LCEA and AIA measurements, with a change in obturator foramen index of, respectively, 0.35 and 0.6 resulting in approximately 2° change in values. Conclusion There was a significant difference in agreement of 5° between CT and radiographs for the LCEA bilaterally. The difference for the AIA was between 0 and 1°, probably of little clinical significance. Pelvic rotation slightly affected bias of the right LCEA, suggesting minimal clinical impact of a slightly rotated pelvis.
Collapse
Affiliation(s)
- Christian Greve Jensen
- Department of Clinical Research, University of Southern Denmark Odense, Denmark
- Research and Innovation Unit of Radiology, University of Southern Denmark Odense, Denmark
| | - Benjamin Schnack Rasmussen
- Department of Clinical Research, University of Southern Denmark Odense, Denmark
- Research and Innovation Unit of Radiology, University of Southern Denmark Odense, Denmark
- CAI-X (Centre for Clinical Artificial Intelligence), Odense University Hospital, University of Southern Denmark Odense, Denmark
- Department of Radiology, Odense University Hospital Odense, Denmark
| | - Søren Overgaard
- Department of Orthopaedic Surgery and Traumatology, Copenhagen University Hospital, Copenhagen, Denmark
- Department of Clinical Medicine, Faculty of Health and Medical Sciences, University of Copenhagen Copenhagen, Denmark
| | - Claus Varnum
- Department of Orthopedic Surgery, Lillebaelt Hospital, Vejle, University Hospital of Southern Denmark Vejle, Denmark
- Department of Regional Health Research, University of Southern Denmark Odense, Denmark
| | - Martin Haagen Haubro
- Department of Orthopedic Surgery and Traumatology, Odense University Hospital Odense, Denmark
| | - Janni Jensen
- Research and Innovation Unit of Radiology, University of Southern Denmark Odense, Denmark
- Department of Radiology, Odense University Hospital Odense, Denmark
| |
Collapse
|
31
|
Spaanderman DJ, Marzetti M, Wan X, Scarsbrook AF, Robinson P, Oei EHG, Visser JJ, Hemke R, van Langevelde K, Hanff DF, van Leenders GJLH, Verhoef C, Grünhagen DJ, Niessen WJ, Klein S, Starmans MPA. AI in radiological imaging of soft-tissue and bone tumours: a systematic review evaluating against CLAIM and FUTURE-AI guidelines. EBioMedicine 2025; 114:105642. [PMID: 40118007 PMCID: PMC11976239 DOI: 10.1016/j.ebiom.2025.105642] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2024] [Revised: 02/14/2025] [Accepted: 02/27/2025] [Indexed: 03/23/2025] Open
Abstract
BACKGROUND Soft-tissue and bone tumours (STBT) are rare, diagnostically challenging lesions with variable clinical behaviours and treatment approaches. This systematic review aims to provide an overview of Artificial Intelligence (AI) methods using radiological imaging for diagnosis and prognosis of these tumours, highlighting challenges in clinical translation, and evaluating study alignment with the Checklist for AI in Medical Imaging (CLAIM) and the FUTURE-AI international consensus guidelines for trustworthy and deployable AI to promote the clinical translation of AI methods. METHODS The systematic review identified literature from several bibliographic databases, covering papers published before 17/07/2024. Original research published in peer-reviewed journals, focused on radiology-based AI for diagnosis or prognosis of primary STBT was included. Exclusion criteria were animal, cadaveric, or laboratory studies, and non-English papers. Abstracts were screened by two of three independent reviewers to determine eligibility. Included papers were assessed against the two guidelines by one of three independent reviewers. The review protocol was registered with PROSPERO (CRD42023467970). FINDINGS The search identified 15,015 abstracts, from which 325 articles were included for evaluation. Most studies performed moderately on CLAIM, averaging a score of 28.9 ± 7.5 out of 53, but poorly on FUTURE-AI, averaging 5.1 ± 2.1 out of 30. INTERPRETATION Imaging-AI tools for STBT remain at the proof-of-concept stage, indicating significant room for improvement. Future efforts by AI developers should focus on design (e.g. defining unmet clinical need, intended clinical setting and how AI would be integrated in clinical workflow), development (e.g. building on previous work, training with data that reflect real-world usage, explainability), evaluation (e.g. ensuring biases are evaluated and addressed, evaluating AI against current best practices), and the awareness of data reproducibility and availability (making documented code and data publicly available). Following these recommendations could improve clinical translation of AI methods. FUNDING Hanarth Fonds, ICAI Lab, NIHR, EuCanImage.
Collapse
Affiliation(s)
- Douwe J Spaanderman
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands.
| | - Matthew Marzetti
- Department of Medical Physics, Leeds Teaching Hospitals NHS Trust, UK; Leeds Biomedical Research Centre, University of Leeds, UK
| | - Xinyi Wan
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Andrew F Scarsbrook
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, UK; Leeds Institute of Medical Research, University of Leeds, UK
| | - Philip Robinson
- Department of Radiology, Leeds Teaching Hospitals NHS Trust, UK
| | - Edwin H G Oei
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Jacob J Visser
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Robert Hemke
- Department of Radiology and Nuclear Medicine, Amsterdam UMC, Amsterdam, the Netherlands
| | | | - David F Hanff
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Geert J L H van Leenders
- Department of Pathology, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Cornelis Verhoef
- Department of Surgical Oncology, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Dirk J Grünhagen
- Department of Surgical Oncology, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Wiro J Niessen
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands; Faculty of Medical Sciences, University of Groningen, Groningen, the Netherlands
| | - Stefan Klein
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| | - Martijn P A Starmans
- Department of Radiology and Nuclear Medicine, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands; Department of Pathology, Erasmus MC Cancer Institute, University Medical Center Rotterdam, Rotterdam, the Netherlands
| |
Collapse
|
32
|
Cronin P, Nasser OMH, Rawson JV. Currently Available Radiology-Specific Reporting Guidelines. Acad Radiol 2025; 32:1798-1805. [PMID: 39880692 DOI: 10.1016/j.acra.2025.01.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Revised: 12/01/2024] [Accepted: 01/14/2025] [Indexed: 01/31/2025]
Abstract
The aim of this paper is to contextualize and review reporting guidelines available at the EQUATOR Network that are most relevant to radiology-specific investigations. Eight EQUATOR Network reporting guidelines for the clinical area of radiology, not including the subspecialized areas of imaging of the cardiovascular, neurologic, and oncologic diseases are reviewed and discussed. The reporting guidelines are for diagnostic and therapeutic clinical research. Why the reporting guideline was development, by whom, their aims and what they hope to achieve are discussed. A table summarizes what the reporting guideline is provided for; an acronym if present is given; a full bibliographic reference with PMID number; the reporting guideline website URL or link; the study design and section of the report that the guideline applies to; and the date that the reporting guideline was last updated.
Collapse
Affiliation(s)
- Paul Cronin
- Emory Department of Radiology and Imaging Science, Division of Cardiothoracic Imaging, Emory University, Atlanta, Georgia (P.C.).
| | - Omar Msto Hussain Nasser
- Harvard Medical School, Boston, Massachusetts (O.M.H.N., J.V.R.); Department of Radiology, Beth Israel Medical Center, Boston, Massachusetts (O.M.H.N., J.V.R.)
| | - James V Rawson
- Harvard Medical School, Boston, Massachusetts (O.M.H.N., J.V.R.); Department of Radiology, Beth Israel Medical Center, Boston, Massachusetts (O.M.H.N., J.V.R.)
| |
Collapse
|
33
|
Lee CF, Lin J, Huang YL, Chen ST, Chou CT, Chen DR, Wu WP. Deep learning-based breast MRI for predicting axillary lymph node metastasis: a systematic review and meta-analysis. Cancer Imaging 2025; 25:44. [PMID: 40165212 PMCID: PMC11956454 DOI: 10.1186/s40644-025-00863-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2024] [Accepted: 03/12/2025] [Indexed: 04/02/2025] Open
Abstract
BACKGROUND To perform a systematic review and meta-analysis that assesses the diagnostic performance of deep learning algorithms applied to breast MRI for predicting axillary lymph nodes metastases in patients of breast cancer. METHODS A systematic literature search in PubMed, MEDLINE, and Embase databases for articles published from January 2004 to February 2025. Inclusion criteria were: patients with breast cancer; deep learning using MRI images was applied to predict axillary lymph nodes metastases; sufficient data were present; original research articles. Quality Assessment of Diagnostic Accuracy Studies-AI and Checklist for Artificial Intelligence in Medical Imaging was used to assess the quality. Statistical analysis included pooling of diagnostic accuracy and investigating between-study heterogeneity. A summary receiver operating characteristic curve (SROC) was performed. R statistical software (version 4.4.0) was used for statistical analyses. RESULTS A total of 10 studies were included. The pooled sensitivity and specificity were 0.76 (95% CI, 0.67-0.83) and 0.81 (95% CI, 0.74-0.87), respectively, with both measures having moderate between-study heterogeneity (I2 = 61% and 60%, respectively; p < 0.01). The SROC analysis yielded a weighted AUC of 0.788. CONCLUSION This meta-analysis demonstrates that deep learning algorithms applied to breast MRI offer promising diagnostic performance for predicting axillary lymph node metastases in breast cancer patients. Incorporating deep learning into clinical practice may enhance decision-making by providing a non-invasive method to more accurately predict lymph node involvement, potentially reducing unnecessary surgeries.
Collapse
Affiliation(s)
- Chia-Fen Lee
- Department of Radiology, Changhua Christian Hospital, Changhua, Taiwan
- Department of Biomedical Imaging and Radiological Sciences, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Joseph Lin
- Division of General Surgery, Changhua Christian Hospital, Changhua, Taiwan
- Comprehensive Breast Cancer Center, Changhua Christian Hospital, Changhua, Taiwan
- Division of Breast Surgery, Yuanlin Christian Hospital, Yuanlin, Taiwan
| | - Yu-Len Huang
- Department of Computer Science, Tunghai University, Taichung, Taiwan
| | - Shou-Tung Chen
- Division of General Surgery, Changhua Christian Hospital, Changhua, Taiwan
- Comprehensive Breast Cancer Center, Changhua Christian Hospital, Changhua, Taiwan
| | - Chen-Te Chou
- Department of Radiology, Changhua Christian Hospital, Changhua, Taiwan
| | - Dar-Ren Chen
- Division of General Surgery, Changhua Christian Hospital, Changhua, Taiwan
- Comprehensive Breast Cancer Center, Changhua Christian Hospital, Changhua, Taiwan
| | - Wen-Pei Wu
- Department of Radiology, Changhua Christian Hospital, Changhua, Taiwan.
- Department of Biomedical Imaging and Radiological Sciences, National Yang Ming Chiao Tung University, Taipei, Taiwan.
- Department of Medical Imaging, Changhua Christian Hospital, 135 Nanxiao Street, Changhua, 500, Taiwan.
| |
Collapse
|
34
|
Wang L, Zhang P, Feng Y, Lv W, Min X, Liu Z, Li J, Feng Z. Identification of testicular cancer with T2-weighted MRI-based radiomics and automatic machine learning. BMC Cancer 2025; 25:563. [PMID: 40155850 PMCID: PMC11951623 DOI: 10.1186/s12885-025-13844-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2024] [Accepted: 02/28/2025] [Indexed: 04/01/2025] Open
Abstract
BACKGROUND Distinguishing between benign and malignant testicular lesions on clinical magnetic resonance imaging (MRI) is crucial for guiding treatment planning. However, conventional MRI-based radiomics to identify testicular cancer requires expert machine learning knowledge. This study aims to investigate the potential of utilizing automatic machine learning (AutoML) based on MRI to diagnose testicular lesions without the need for expert algorithm optimization. METHODS Retrospective preoperative MRI scans from 115 patients diagnosed with testicular disease through pathology were obtained. A total of 1781 radiomics features were extracted from each lesion on the T2-weighted images. Intraclass and interclass correlation coefficients were used to evaluate the intra-observer and interobserver agreements for each radiomics feature. We developed an AutoML method based on the tree-based pipeline optimization tool (TPOT) algorithm to construct a discriminant model. The best pipeline was determined through 100 repeated operations using a 5-fold cross-validation algorithm in TPOT. The model was evaluated for accuracy, sensitivity, and specificity using the area under the curve (AUC) value of the receiver operating characteristic (ROC) curve. Shapley Additive exPlanations were used to illustrate the optimization results. RESULTS Utilizing the TPOT method, 100 diagnostic models were developed to identify testicular lesions. The best model was determined based on the highest AUC in the training cohort. The prediction model yielded AUC values of 0.989 (95% confidence interval [CI]: 0.985-0.993) and 0.909 (95% CI: 0.893-0.923) in the training and testing cohorts, respectively. CONCLUSIONS AutoML, based on the TPOT algorithm, holds potential as a noninvasive method for effectively discriminating between benign and malignant testicular lesions.
Collapse
Affiliation(s)
- Liang Wang
- Computer Center, Tongji Hospital, Tongji Medical College, Uazhong University of Science and Technology, Wuhan, China
| | - PeiPei Zhang
- Department of Radiology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Yanhui Feng
- Computer Center, Tongji Hospital, Tongji Medical College, Uazhong University of Science and Technology, Wuhan, China
- School of Medicine and Health Management, Huazhong University of Science and Technology, Wuhan, China
| | - Wenzhi Lv
- Department of Radiology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- Britton Chance Center, MoE Key Laboratory for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics-Huazhong University of Science and Technology, Wuhan, China
| | - Xiangde Min
- Department of Radiology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Zhiyong Liu
- School of Medicine and Health Management, Huazhong University of Science and Technology, Wuhan, China
| | - Jin Li
- Computer Center, Tongji Hospital, Tongji Medical College, Uazhong University of Science and Technology, Wuhan, China.
| | - Zhaoyan Feng
- Department of Radiology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
| |
Collapse
|
35
|
Pehrson LM, Petersen J, Panduro NS, Lauridsen CA, Carlsen JF, Darkner S, Nielsen MB, Ingala S. AI-Guided Delineation of Gross Tumor Volume for Body Tumors: A Systematic Review. Diagnostics (Basel) 2025; 15:846. [PMID: 40218196 PMCID: PMC11988838 DOI: 10.3390/diagnostics15070846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2025] [Revised: 03/14/2025] [Accepted: 03/18/2025] [Indexed: 04/14/2025] Open
Abstract
Background: Approximately 50% of all oncological patients undergo radiation therapy, where personalized planning of treatment relies on gross tumor volume (GTV) delineation. Manual delineation of GTV is time-consuming, operator-dependent, and prone to variability. An increasing number of studies apply artificial intelligence (AI) techniques to automate such delineation processes. Methods: To perform a systematic review comparing the performance of AI models in tumor delineations within the body (thoracic cavity, esophagus, abdomen, and pelvis, or soft tissue and bone). A retrospective search of five electronic databases was performed between January 2017 and February 2025. Original research studies developing and/or validating algorithms delineating GTV in CT, MRI, and/or PET were included. The Checklist for Artificial Intelligence in Medical Imaging (CLAIM) and Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis statement and checklist (TRIPOD) were used to assess the risk, bias, and reporting adherence. Results: After screening 2430 articles, 48 were included. The pooled diagnostic performance from the use of AI algorithms across different tumors and topological areas ranged 0.62-0.92 in dice similarity coefficient (DSC) and 1.33-47.10 mm in Hausdorff distance (HD). The algorithms with the highest DSC deployed an encoder-decoder architecture. Conclusions: AI algorithms demonstrate a high level of concordance with clinicians in GTV delineation. Translation to clinical settings requires the building of trust, improvement in performance and robustness of results, and testing in prospective studies and randomized controlled trials.
Collapse
Affiliation(s)
- Lea Marie Pehrson
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Department of Clinical Medicine, University of Copenhagen, 2100 Copenhagen, Denmark
- Department of Computer Science, University of Copenhagen, 2100 Copenhagen, Denmark
| | - Jens Petersen
- Department of Computer Science, University of Copenhagen, 2100 Copenhagen, Denmark
- Department of Oncology, Rigshospitalet, 2100 Copenhagen, Denmark
| | - Nathalie Sarup Panduro
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Department of Clinical Medicine, University of Copenhagen, 2100 Copenhagen, Denmark
| | - Carsten Ammitzbøl Lauridsen
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Radiography Education, University College Copenhagen, 2200 Copenhagen, Denmark
| | - Jonathan Frederik Carlsen
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Department of Clinical Medicine, University of Copenhagen, 2100 Copenhagen, Denmark
| | - Sune Darkner
- Department of Computer Science, University of Copenhagen, 2100 Copenhagen, Denmark
| | - Michael Bachmann Nielsen
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Department of Clinical Medicine, University of Copenhagen, 2100 Copenhagen, Denmark
| | - Silvia Ingala
- Department of Diagnostic Radiology, Copenhagen University Hospital Rigshospitalet, 2100 Copenhagen, Denmark
- Cerebriu A/S, 1434 Copenhagen, Denmark
- Department of Diagnostic Radiology, Copenhagen University Hospital Herlev and Gentofte, 2730 Herlev, Denmark
| |
Collapse
|
36
|
Salimi M, Vadipour P, Bahadori AR, Houshi S, Mirshamsi A, Fatemian H. Predicting hemorrhagic transformation in acute ischemic stroke: a systematic review, meta-analysis, and methodological quality assessment of CT/MRI-based deep learning and radiomics models. Emerg Radiol 2025:10.1007/s10140-025-02336-3. [PMID: 40133723 DOI: 10.1007/s10140-025-02336-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2025] [Accepted: 03/19/2025] [Indexed: 03/27/2025]
Abstract
Acute ischemic stroke (AIS) is a major cause of mortality and morbidity, with hemorrhagic transformation (HT) as a severe complication. Accurate prediction of HT is essential for optimizing treatment strategies. This review assesses the accuracy and utility of deep learning (DL) and radiomics in predicting HT through imaging, regarding clinical decision-making for AIS patients. A literature search was conducted across five databases (Pubmed, Scopus, Web of Science, Embase, IEEE) up to January 23, 2025. Studies involving DL or radiomics-based ML models for predicting HT in AIS patients were included. Data from training, validation, and clinical-combined models were extracted and analyzed separately. Pooled sensitivity, specificity, and AUC were calculated with a random-effects bivariate model. For the quality assessment of studies, the Methodological Radiomics Score (METRICS) and QUADAS-2 tool were used. 16 studies consisting of 3,083 individual participants were included in the meta-analysis. The pooled AUC for training cohorts was 0.87, sensitivity 0.80, and specificity 0.85. For validation cohorts, AUC was 0.87, sensitivity 0.81, and specificity 0.86. Clinical-combined models showed an AUC of 0.93, sensitivity 0.84, and specificity 0.89. Moderate to severe heterogeneity was noted and addressed. Deep-learning models outperformed radiomics models, while clinical-combined models outperformed deep learning-only and radiomics-only models. The average METRICS score was 62.85%. No publication bias was detected. DL and radiomics models showed great potential in predicting HT in AIS patients. However, addressing methodological issues-such as inconsistent reference standards and limited external validation-is essential for the clinical implementation of these models.
Collapse
Affiliation(s)
- Mohsen Salimi
- School of Medicine, Shiraz University of Medical Sciences, Shiraz, Iran.
| | - Pouria Vadipour
- Western Michigan University Homer Stryker M.D. School of Medicine, Kalamazoo, MI, USA
| | - Amir Reza Bahadori
- School of Medicine, Shiraz University of Medical Sciences, Shiraz, Iran
- Iranian Center of Neurological Research, Neuroscience Institute, Tehran University of Medical Sciences, Tehran, Iran
| | - Shakiba Houshi
- Isfahan Neurosciences Research Center, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Ali Mirshamsi
- School of Medicine, Shiraz University of Medical Sciences, Shiraz, Iran
| | - Hossein Fatemian
- School of Medicine, Shiraz University of Medical Sciences, Shiraz, Iran
| |
Collapse
|
37
|
Lam NFD, Cai J, Ng KH. Artificial intelligence and its potential integration with the clinical practice of diagnostic imaging medical physicists: a review. Phys Eng Sci Med 2025:10.1007/s13246-025-01535-z. [PMID: 40126762 DOI: 10.1007/s13246-025-01535-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2024] [Accepted: 03/12/2025] [Indexed: 03/26/2025]
Abstract
Current clinical practice in imaging medical physics is concerned with quality assurance, image processing and analysis, radiation dosimetry, risk assessment and radiation protection, and in-house training and research. Physicist workloads are projected to increase as medical imaging technologies become more sophisticated. Artificial intelligence (AI) is a rising technology with potential to assist medical physicists in their work. Exploration of AI integration into imaging medical physicist workloads is limited. In this review paper, we provide an overview of AI techniques, outline their potential usage in imaging medical physics, and discuss the limitations and challenges to clinical adoption of AI technologies.
Collapse
Affiliation(s)
| | - Jing Cai
- Department of Health Technology and Informatics, Hong Kong Polytechnic University, Hung Hom, Hong Kong
| | - Kwan Hoong Ng
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, Kuala Lumpur, Malaysia.
- Faculty of Medicine and Health Sciences, UCSI University, Springhill, Negri Sembilan, Malaysia.
| |
Collapse
|
38
|
Goto M, Futamura Y, Makishima H, Saito T, Sakamoto N, Iijima T, Tamaki Y, Okumura T, Sakurai T, Sakurai H. Development of a deep learning-based model to evaluate changes during radiotherapy using cervical cancer digital pathology. JOURNAL OF RADIATION RESEARCH 2025; 66:144-156. [PMID: 40051384 PMCID: PMC11932348 DOI: 10.1093/jrr/rraf004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/22/2024] [Revised: 12/20/2024] [Accepted: 01/24/2025] [Indexed: 03/25/2025]
Abstract
This study aims to create a deep learning-based classification model for cervical cancer biopsy before and during radiotherapy, visualize the results on whole slide images (WSIs), and explore the clinical significance of obtained features. This study included 95 patients with cervical cancer who received radiotherapy between April 2013 and December 2020. Hematoxylin-eosin stained biopsies were digitized to WSIs and divided into small tiles. Our model adopted the feature extractor of DenseNet121 and the classifier of the support vector machine. About 12 400 tiles were used for training the model and 6000 tiles for testing. The model performance was assessed on a per-tile and per-WSI basis. The resultant probability was defined as radiotherapy status probability (RSP) and its color map was visualized on WSIs. Survival analysis was performed to examine the clinical significance of the RSP. In the test set, the trained model had an area under the receiver operating characteristic curve of 0.76 per-tile and 0.95 per-WSI. In visualization, the model focused on viable tumor components and stroma in tumor biopsies. While survival analysis failed to show the prognostic impact of RSP during treatment, cases with low RSP at diagnosis had prolonged overall survival compared to those with high RSP (P = 0.045). In conclusion, we successfully developed a model to classify biopsies before and during radiotherapy and visualized the result on slide images. Low RSP cases before treatment had a better prognosis, suggesting that tumor morphologic features obtained using the model may be useful for predicting prognosis.
Collapse
Affiliation(s)
- Masaaki Goto
- Department of Radiation Oncology & Proton Medical Research Center, Institute of Medicine, University of Tsukuba, 2-1-1 Amakubo, Tsubuka, Ibaraki 305-8576, Japan
- Department of Radiation Oncology, Japan Red Cross Medical Center, 4-1-22 Hiroo, Shibuya, Tokyo 150-8935, Japan
| | - Yasunori Futamura
- Center for Artificial Intelligence Research, University of Tsukuba, 1-1-1 Tennodai, Tsubuka, Ibaraki 305-8577, Japan
- Department of Computer Science, University of Tsukuba, 1-1-1 Tennodai, Tsubuka, Ibaraki 305-8577, Japan
| | - Hirokazu Makishima
- Department of Radiation Oncology & Proton Medical Research Center, Institute of Medicine, University of Tsukuba, 2-1-1 Amakubo, Tsubuka, Ibaraki 305-8576, Japan
- QST Hospital, National Institute for Quantum Science and Technology, 4-9-1 Anagawa, Inage, Chiba 263-8555, Japan
| | - Takashi Saito
- Department of Radiation Oncology & Proton Medical Research Center, Institute of Medicine, University of Tsukuba, 2-1-1 Amakubo, Tsubuka, Ibaraki 305-8576, Japan
| | - Noriaki Sakamoto
- Department of Diagnostic Pathology, Institute of Medicine, University of Tsukuba, 1-1-1 Tennodai, Tsubuka, Ibaraki 305-8575, Japan
| | - Tatsuo Iijima
- Department of Diagnostic Pathology, Ibaraki Prefectural Central Hospital, 6528 Koibuchi, Kasama, Ibaraki 309-1793, Japan
| | - Yoshio Tamaki
- Department of Radiation Oncology, Fukushima Rosai Hospital, 3 Numaziri, Uchigotsuzuramachi, Iwaki, Fukushima 973-8403, Japan
| | - Toshiyuki Okumura
- Department of Radiation Oncology, Ibaraki Prefectural Central Hospital, 6528 Koibuchi, Kasama, Ibaraki 309-1793, Japan
| | - Tetsuya Sakurai
- Center for Artificial Intelligence Research, University of Tsukuba, 1-1-1 Tennodai, Tsubuka, Ibaraki 305-8577, Japan
- Department of Computer Science, University of Tsukuba, 1-1-1 Tennodai, Tsubuka, Ibaraki 305-8577, Japan
| | - Hideyuki Sakurai
- Department of Radiation Oncology & Proton Medical Research Center, Institute of Medicine, University of Tsukuba, 2-1-1 Amakubo, Tsubuka, Ibaraki 305-8576, Japan
| |
Collapse
|
39
|
Vasilev Y, Rumyantsev D, Vladzymyrskyy A, Omelyanskaya O, Pestrenin L, Shulkin I, Nikitin E, Kapninskiy A, Arzamasov K. Evolution of an Artificial Intelligence-Powered Application for Mammography. Diagnostics (Basel) 2025; 15:822. [PMID: 40218172 PMCID: PMC11988740 DOI: 10.3390/diagnostics15070822] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2025] [Revised: 03/19/2025] [Accepted: 03/20/2025] [Indexed: 04/14/2025] Open
Abstract
Background: The implementation of radiological artificial intelligence (AI) solutions remains challenging due to limitations in existing testing methodologies. This study assesses the efficacy of a comprehensive methodology for performance testing and monitoring of commercial-grade mammographic AI models. Methods: We utilized a combination of retrospective and prospective multicenter approaches to evaluate a neural network based on the Faster R-CNN architecture with a ResNet-50 backbone, trained on a dataset of 3641 mammograms. The methodology encompassed functional and calibration testing, coupled with routine technical and clinical monitoring. Feedback from testers and radiologists was relayed to the developers, who made updates to the AI model. The test dataset comprised 112 medical organizations, representing 10 manufacturers of mammography equipment and encompassing 593,365 studies. The evaluation metrics included the area under the curve (AUC), accuracy, sensitivity, specificity, technical defects, and clinical assessment scores. Results: The results demonstrated significant enhancement in the AI model's performance through collaborative efforts among developers, testers, and radiologists. Notable improvements included functionality, diagnostic accuracy, and technical stability. Specifically, the AUC rose by 24.7% (from 0.73 to 0.91), the accuracy improved by 15.6% (from 0.77 to 0.89), sensitivity grew by 37.1% (from 0.62 to 0.85), and specificity increased by 10.7% (from 0.84 to 0.93). The average proportion of technical defects declined from 9.0% to 1.0%, while the clinical assessment score improved from 63.4 to 72.0. Following 2 years and 9 months of testing, the AI solution was integrated into the compulsory health insurance system. Conclusions: The multi-stage, lifecycle-based testing methodology demonstrated substantial potential in software enhancement and integration into clinical practice. Key elements of this methodology include robust functional and diagnostic requirements, continuous testing and updates, systematic feedback collection from testers and radiologists, and prospective monitoring.
Collapse
Affiliation(s)
- Yuriy Vasilev
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
| | - Denis Rumyantsev
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
| | - Anton Vladzymyrskyy
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
- Department of Information Technology and Medical Data Processing, I.M. Sechenov First Moscow State Medical University (Sechenov University), 119991 Moscow, Russia
| | - Olga Omelyanskaya
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
| | - Lev Pestrenin
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
| | - Igor Shulkin
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
| | - Evgeniy Nikitin
- Celsus (Medical Screening Systems), Viktorenko St., Bldg. 11, Room 21N, 125167 Moscow, Russia; (E.N.); (A.K.)
| | - Artem Kapninskiy
- Celsus (Medical Screening Systems), Viktorenko St., Bldg. 11, Room 21N, 125167 Moscow, Russia; (E.N.); (A.K.)
| | - Kirill Arzamasov
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Health Care Department, 127051 Moscow, Russia; (Y.V.); (A.V.); (O.O.); (L.P.); (I.S.); (K.A.)
- Department of Artificial Intelligence Technologies, MIREA—Russian Technological University, 119454 Moscow, Russia
| |
Collapse
|
40
|
Mei H, Chen H, Zheng Q, Yang R, Wang N, Jiao P, Wang X, Chen Z, Liu X. Foundation Model and Radiomics-Based Quantitative Characterization of Perirenal Fat in Renal Cell Carcinoma Surgery. Acad Radiol 2025:S1076-6332(25)00199-0. [PMID: 40133088 DOI: 10.1016/j.acra.2025.03.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2025] [Revised: 03/02/2025] [Accepted: 03/02/2025] [Indexed: 03/27/2025]
Abstract
RATIONALE AND OBJECTIVES To quantitatively characterize the degree of perirenal fat adhesion using artificial intelligence in renal cell carcinoma. MATERIALS AND METHODS This retrospective study analyzed a total of 596 patients from three cohorts, utilizing corticomedullary phase computed tomography urography (CTU) images. The nnUNet v2 network combined with numerical computation was employed to segment the perirenal fat region. Pyradiomics algorithms and a computed tomography foundation model were used to extract features from CTU images separately, creating single-modality predictive models for identifying perirenal fat adhesion. By concatenating the Pyradiomics and foundation model features, an early fusion multimodal predictive signature was developed. The prognostic performance of the single-modality and multimodality models was further validated in two independent cohorts. RESULTS The nnUNet v2 segmentation model accurately segmented both kidneys. The neural network and thresholding approach effectively delineated the perirenal fat region. Single-modality models based on radiomic and computed tomography foundation features demonstrated a certain degree of accuracy in diagnosing and identifying perirenal fat adhesion, while the early feature fusion diagnostic model outperformed the single-modality models. Also, the perirenal fat adhesion score showed a positive correlation with surgical time and intraoperative blood loss. CONCLUSION AI-based radiomics and foundation models can accurately identify the degree of perirenal fat adhesion and have the potential to be used for surgical risk assessment.
Collapse
Affiliation(s)
- Haonan Mei
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Hui Chen
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Qingyuan Zheng
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Rui Yang
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Nanxi Wang
- School of Software Engineering, Hubei Open University, Wuhan, China (N.W.)
| | - Panpan Jiao
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Xiao Wang
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Zhiyuan Chen
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.)
| | - Xiuheng Liu
- Department of Urology, Renmin Hospital of Wuhan University, 238 Jiefang Road, Wuchang District, Wuhan, Hubei Province 430060, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.); Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, Hubei, China (H.M., H.C., Q.Z., R.Y., P.J., X.W., Z.C., X.L.).
| |
Collapse
|
41
|
Jiang M, Kong C, Lu S, Li Q, Chu C, Li W. Ovarian masses suggested for MRI examination: assessment of deep learning models based on non-contrast-enhanced MRI sequences for predicting malignancy. Abdom Radiol (NY) 2025:10.1007/s00261-025-04891-2. [PMID: 40116887 DOI: 10.1007/s00261-025-04891-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2025] [Revised: 03/04/2025] [Accepted: 03/10/2025] [Indexed: 03/23/2025]
Abstract
PURPOSE We aims to assessed and compare four deep learning(DL) models using non-contrast-enhanced magnetic resonance imaging(MRI) to differentiate benign from malignant ovarian tumors, considering diagnostic efficacy and associated development costs. METHODS 526 patients (327 benign lesions vs 199 malignant lesions) who were recommended for MRI due to suspected ovarian masses, confirmed with histopathology, were included in this retrospective study. A training cohort (n=367) and a validation cohort (n=159) were constructed. Based on the images of T1-weighted imaging (T1WI), T2-weighted imaging (T2WI), diffusion-weighted imaging (DWI), we evaluated the diagnostic performance of four DL models (ConvNeXt, FBNet, GhostNet, ResNet50) in distinguishing between benign and malignant ovarian tumors. Two radiologists with varying levels of experience independently reviewed all original non-contrast-enhanced MR images from the validation cohort to determine if each case was benign or malignant. The area under the receiver operating characteristic curve (AUC), confusion matrices, accuracy, sensitivity, specificity, positive predictive value(PPV) and negative predictive value(NPV) were used to compare performance. RESULTS The study of 526 ovarian mass patients (ages 1-92) evaluated four DL models for predicting malignant tumors, with AUCs ranging from 0.8091 to 0.8572 and accuracy between 81.1% and 85.5%. An experienced radiologist achieved 86.2% accuracy, slightly surpassing the DL models, while a less experienced radiologist had 69.2% accuracy. Resnet50 had the highest sensitivity (78.3%) and NPV (87.3%), while ConvNeXt excelled in specificity and PPV (100%). GhostNet and FBNet are more parameter-efficient than other models. CONCLUSION The four DL models effectively distinguished between benign and malignant ovarian tumors using non-contrast MRI. These models outperformed less experienced radiologists and were slightly less accurate than experienced ones. ResNet50 had the best predictive performance, while GhostNet was highly accurate with fewer parameters. Our study indicates that DL models based on non-contrast-enhanced MRI have the potential to assist in diagnosis.
Collapse
Affiliation(s)
- Meijiao Jiang
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Chui Kong
- School of Information Science and Technology, Fudan University, Shanghai, China
| | - Siwei Lu
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Qingwan Li
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Caiting Chu
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
| | - Wenhua Li
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
- Department of Radiology, Xinhua Hospital, Shanghai Jiao Tong University School of Medicine, Chongming Branch, Shanghai, China.
| |
Collapse
|
42
|
Pelcat A, Le Berre A, Ben Hassen W, Debacker C, Charron S, Thirion B, Legrand L, Turc G, Oppenheim C, Benzakoun J. Generative T2*-weighted images as a substitute for true T2*-weighted images on brain MRI in patients with acute stroke. Diagn Interv Imaging 2025:S2211-5684(25)00048-8. [PMID: 40113490 DOI: 10.1016/j.diii.2025.03.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2024] [Revised: 03/11/2025] [Accepted: 03/14/2025] [Indexed: 03/22/2025]
Abstract
PURPOSE The purpose of this study was to validate a deep learning algorithm that generates T2*-weighted images from diffusion-weighted (DW) images and to compare its performance with that of true T2*-weighted images for hemorrhage detection on MRI in patients with acute stroke. MATERIALS AND METHODS This single-center, retrospective study included DW- and T2*-weighted images obtained less than 48 hours after symptom onset in consecutive patients admitted for acute stroke. Datasets were divided into training (60 %), validation (20 %), and test (20 %) sets, with stratification by stroke type (hemorrhagic/ischemic). A generative adversarial network was trained to produce generative T2*-weighted images using DW images. Concordance between true T2*-weighted images and generative T2*-weighted images for hemorrhage detection was independently graded by two readers into three categories (parenchymal hematoma, hemorrhagic infarct or no hemorrhage), and discordances were resolved by consensus reading. Sensitivity, specificity and accuracy of generative T2*-weighted images were estimated using true T2*-weighted images as the standard of reference. RESULTS A total of 1491 MRI sets from 939 patients (487 women, 452 men) with a median age of 71 years (first quartile, 57; third quartile, 81; range: 21-101) were included. In the test set (n = 300), there were no differences between true T2*-weighted images and generative T2*-weighted images for intraobserver reproducibility (κ = 0.97 [95 % CI: 0.95-0.99] vs. 0.95 [95 % CI: 0.92-0.97]; P = 0.27) and interobserver reproducibility (κ = 0.93 [95 % CI: 0.90-0.97] vs. 0.92 [95 % CI: 0.88-0.96]; P = 0.64). After consensus reading, concordance between true T2*-weighted images and generative T2*-weighted images was excellent (κ = 0.92; 95 % CI: 0.91-0.96). Generative T2*-weighted images achieved 90 % sensitivity (73/81; 95 % CI: 81-96), 97 % specificity (213/219; 95 % CI: 94-99) and 95 % accuracy (286/300; 95 % CI: 92-97) for the diagnosis of any cerebral hemorrhage (hemorrhagic infarct or parenchymal hemorrhage). CONCLUSION Generative T2*-weighted images and true T2*-weighted images have non-different diagnostic performances for hemorrhage detection in patients with acute stroke and may be used to shorten MRI protocols.
Collapse
Affiliation(s)
- Antoine Pelcat
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France
| | - Alice Le Berre
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France
| | - Wagih Ben Hassen
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France
| | - Clement Debacker
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France
| | - Sylvain Charron
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France
| | - Bertrand Thirion
- INRIA, CEA, Université Paris-Saclay, MIND Team, 91400 Palaiseau, France
| | - Laurence Legrand
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France
| | - Guillaume Turc
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, Stroke Team, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neurology, 75014 Paris, France
| | - Catherine Oppenheim
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France
| | - Joseph Benzakoun
- Université Paris Cité, Institute of Psychiatry and Neuroscience of Paris (IPNP), INSERM U1266, IMA-BRAIN, 75014 Paris, France; GHU Paris Psychiatrie et Neurosciences, Hôpital Sainte Anne, Department of Neuroradiology, 75014 Paris, France.
| |
Collapse
|
43
|
Wang H, He Y, Wan L, Li C, Li Z, Li Z, Xu H, Tu C. Deep learning models in classifying primary bone tumors and bone infections based on radiographs. NPJ Precis Oncol 2025; 9:72. [PMID: 40074845 PMCID: PMC11904180 DOI: 10.1038/s41698-025-00855-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2024] [Accepted: 02/25/2025] [Indexed: 03/14/2025] Open
Abstract
Primary bone tumors (PBTs) present significant diagnostic challenges due to their heterogeneous nature and similarities with bone infections. This study aimed to develop an ensemble deep learning framework that integrates multicenter radiographs and extensive clinical features to accurately differentiate between PBTs and bone infections. We compared the performance of the ensemble model with four imaging models based solely on radiographs utilizing EfficientNet B3, EfficientNet B4, Vision Transformer, and Swin Transformers. The patients were split into external dataset (N = 423) and internal dataset [including training (N = 1044), test (N = 354), and validation set (N = 171)]. The ensemble model outperformed imaging models, achieving areas under the curve (AUCs) of 0.948 and 0.963 on internal and external sets, respectively, with accuracies of 0.881 and 0.895. Its performance surpassed junior and mid-level radiologists and was comparable to senior radiologists (accuracy: 83.6%). These findings underscore the potential of deep learning in enhancing diagnostic precision for PBTs and bone infections (Research Registration Unique Identifying Number (UIN): researchregistry10483 and with details are available at https://www.researchregistry.com/register-now#home/registrationdetails/6693845995ba110026aeb754/ ).
Collapse
Affiliation(s)
- Hua Wang
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Yu He
- Department of Radiology, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Lu Wan
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Chenbei Li
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Zhaoqi Li
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Zhihong Li
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China
- Shenzhen Research Institute of Central South University, Guangdong, China
| | - Haodong Xu
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China.
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China.
- Center for Precision Health, McWilliams School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA.
| | - Chao Tu
- Department of Orthopaedics, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China.
- Hunan Key Laboratory of Tumor Models and Individualized Medicine, The Second Xiangya Hospital of Central South University, Changsha, Hunan, China.
- Shenzhen Research Institute of Central South University, Guangdong, China.
- Changsha Medical University, Changsha, Hunan, China.
| |
Collapse
|
44
|
Morone G, De Angelis L, Martino Cinnera A, Carbonetti R, Bisirri A, Ciancarelli I, Iosa M, Negrini S, Kiekens C, Negrini F. Artificial intelligence in clinical medicine: a state-of-the-art overview of systematic reviews with methodological recommendations for improved reporting. Front Digit Health 2025; 7:1550731. [PMID: 40110115 PMCID: PMC11920125 DOI: 10.3389/fdgth.2025.1550731] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2024] [Accepted: 02/12/2025] [Indexed: 03/22/2025] Open
Abstract
Medicine has become increasingly receptive to the use of artificial intelligence (AI). This overview of systematic reviews (SRs) aims to categorise current evidence about it and identify the current methodological state of the art in the field proposing a classification of AI model (CLASMOD-AI) to improve future reporting. PubMed/MEDLINE, Scopus, Cochrane library, EMBASE and Epistemonikos databases were screened by four blinded reviewers and all SRs that investigated AI tools in clinical medicine were included. 1923 articles were found, and of these, 360 articles were examined via the full-text and 161 SRs met the inclusion criteria. The search strategy, methodological, medical and risk of bias information were extracted. The CLASMOD-AI was based on input, model, data training, and performance metric of AI tools. A considerable increase in the number of SRs was observed in the last five years. The most covered field was oncology accounting for 13.9% of the SRs, with diagnosis as the predominant objective in 44.4% of the cases). The risk of bias was assessed in 49.1% of included SRs, yet only 39.2% of these used tools with specific items to assess AI metrics. This overview highlights the need for improved reporting on AI metrics, particularly regarding the training of AI models and dataset quality, as both are essential for a comprehensive quality assessment and for mitigating the risk of bias using specialized evaluation tools.
Collapse
Affiliation(s)
- Giovanni Morone
- Department of Life, Health and Environmental Sciences, University of L'Aquila, L'Aquila, Italy
- San Raffaele Institute of Sulmona, Sulmona, Italy
| | - Luigi De Angelis
- Department of Translational Research and New Technologies in Medicine and Surgery, University of Pisa, Pisa, Italy
- Italian Society of Artificial Intelligence in Medicine (SIIAM, Società Italiana Intelligenza Artificiale in Medicina), Rome, Italy
| | - Alex Martino Cinnera
- Scientific Institute for Research, Hospitalisation and Health Care IRCCS Santa Lucia Foundation, Rome, Italy
| | - Riccardo Carbonetti
- Clinical Area of Neuroscience and Neurorehabilitation, Neurofunctional Rehabilitation Unit, IRCCS "Bambino Gesù" Children's Hospital, Rome, Italy
| | | | - Irene Ciancarelli
- Department of Life, Health and Environmental Sciences, University of L'Aquila, L'Aquila, Italy
| | - Marco Iosa
- Scientific Institute for Research, Hospitalisation and Health Care IRCCS Santa Lucia Foundation, Rome, Italy
- Department of Psychology, Sapienza University of Rome, Rome, Italy
| | - Stefano Negrini
- Department of Biomedical, Surgical and Dental Sciences, University 'La Statale', Milan, Italy
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
| | | | - Francesco Negrini
- Department of Biotechnology and Life Sciences, University of Insubria, Varese, Italy
- Istituti Clinici Scientifici Maugeri IRCCS, Tradate, Italy
| |
Collapse
|
45
|
Rusanov B, Ebert MA, Sabet M, Rowshanfarzad P, Barry N, Kendrick J, Alkhatib Z, Gill S, Dass J, Bucknell N, Croker J, Tang C, White R, Bydder S, Taylor M, Slama L, Mukwada G. Guidance on selecting and evaluating AI auto-segmentation systems in clinical radiotherapy: insights from a six-vendor analysis. Phys Eng Sci Med 2025; 48:301-316. [PMID: 39804550 PMCID: PMC11997002 DOI: 10.1007/s13246-024-01513-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Accepted: 12/20/2024] [Indexed: 04/15/2025]
Abstract
Artificial Intelligence (AI) based auto-segmentation has demonstrated numerous benefits to clinical radiotherapy workflows. However, the rapidly changing regulatory, research, and market environment presents challenges around selecting and evaluating the most suitable solution. To support the clinical adoption of AI auto-segmentation systems, Selection Criteria recommendations were developed to enable a holistic evaluation of vendors, considering not only raw performance but associated risks uniquely related to the clinical deployment of AI. In-house experience and key bodies of work on ethics, standards, and best practices for AI in Radiation Oncology were reviewed to inform selection criteria and evaluation strategies. A retrospective analysis using the criteria was performed across six vendors, including a quantitative assessment using five metrics (Dice, Hausdorff Distance, Average Surface Distance, Surface Dice, Added Path Length) across 20 head and neck, 20 thoracic, and 19 male pelvis patients for AI models as of March 2023. A total of 47 selection criteria were identified across seven categories. A retrospective analysis showed that overall no vendor performed exceedingly well, with systematically poor performance in Data Security & Responsibility, Vendor Support Tools, and Transparency & Ethics. In terms of raw performance, vendors varied widely from excellent to poor. As new regulations come into force and the scope of AI auto-segmentation systems adapt to clinical needs, continued interest in ensuring safe, fair, and transparent AI will persist. The selection and evaluation framework provided herein aims to promote user confidence by exploring the breadth of clinically relevant factors to support informed decision-making.
Collapse
Affiliation(s)
- Branimir Rusanov
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia.
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia.
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia.
| | - Martin A Ebert
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia
- Australian Centre for Quantitative Imaging, The University of Western Australia, Crawley, WA, Australia
- School of Medicine and Public Health, University of Wisconsin, Madison, WI, USA
| | - Mahsheed Sabet
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia
| | - Pejman Rowshanfarzad
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia
| | - Nathaniel Barry
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia
| | - Jake Kendrick
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Center for Advanced Technologies in Cancer Research, Perth, WA, Australia
| | - Zaid Alkhatib
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Suki Gill
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Joshua Dass
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Nicholas Bucknell
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Jeremy Croker
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Colin Tang
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Rohen White
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Sean Bydder
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Mandy Taylor
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Luke Slama
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| | - Godfrey Mukwada
- School of Physics, Mathematics and Computing, The University of Western Australia, Crawley, WA, Australia
- Department of Radiation Oncology, Sir Charles Gairdner Hospital, Nedlands, WA, Australia
| |
Collapse
|
46
|
Altom DS, Awad Taha AI, Mahmoud Hussein AAA, Ibrahim Elshiekh MA, Alata Abdelmajed AH, Abdalla Ibrahim FI, Abelgadir Mohammed SM, Elamin Eltain Tifoor MM. Artificial Intelligence-Based Chatbots in Chronic Disease Management: A Systematic Review of Applications and Challenges. Cureus 2025; 17:e81001. [PMID: 40260325 PMCID: PMC12011281 DOI: 10.7759/cureus.81001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/22/2025] [Indexed: 04/23/2025] Open
Abstract
Artificial intelligence (AI) is being used by an increasing number of conversational agents, sometimes known as chatbots. In applications related to health care, such as those that educate and assist patients with chronic illnesses, which are among the main causes of mortality in the 21st century, they are becoming more and more common. Chatbots powered by AI allows for more frequent and efficient engagement with these patients. This systematic review aimed to examine the traits, medical conditions, and AI architectures of conversational agents that are based on artificial intelligence and are specifically made for chronic illnesses. We searched four databases (Scopus, Web of Science, PubMed, and Cumulative Index to Nursing and Allied Health Literature [CINAHL]) to search for relevant studies using specific inclusion and exclusion criteria. Among these databases, we found 386 studies that were screened for duplicates and then assessed by inclusion and exclusion criteria. We included the 10 most relevant studies in this systemic review. There is a dearth of research on AI-based interactive agents for chronic illnesses, and what little is available is primarily quasi-experimental studies, including chatbots in prototype stages that employ natural language processing (NLP) and enable multimodal user engagement. Future studies could benefit from comparing and evaluating AI-based conversational bots within and between various chronic health disorders using evidence-based methodology. In addition to improving comparability, more structured development and standardized evaluation procedures could improve the caliber of chatbots created for certain chronic diseases and their subsequent effects on the target patients.
Collapse
Affiliation(s)
- Dalia Saad Altom
- Family Medicine, Najran Armed Forces Hospital, Ministry of Defense Health Services, Najran, SAU
| | | | | | | | | | | | | | | |
Collapse
|
47
|
Daga K, Agarwal S, Moti Z, Lee MBK, Din M, Wood D, Modat M, Booth TC. Machine Learning Algorithms to Predict the Risk of Rupture of Intracranial Aneurysms: a Systematic Review. Clin Neuroradiol 2025; 35:3-16. [PMID: 39546007 PMCID: PMC11832721 DOI: 10.1007/s00062-024-01474-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2024] [Accepted: 10/17/2024] [Indexed: 11/17/2024]
Abstract
PURPOSE Subarachnoid haemorrhage is a potentially fatal consequence of intracranial aneurysm rupture, however, it is difficult to predict if aneurysms will rupture. Prophylactic treatment of an intracranial aneurysm also involves risk, hence identifying rupture-prone aneurysms is of substantial clinical importance. This systematic review aims to evaluate the performance of machine learning algorithms for predicting intracranial aneurysm rupture risk. METHODS MEDLINE, Embase, Cochrane Library and Web of Science were searched until December 2023. Studies incorporating any machine learning algorithm to predict the risk of rupture of an intracranial aneurysm were included. Risk of bias was assessed using the Prediction Model Risk of Bias Assessment Tool (PROBAST). PROSPERO registration: CRD42023452509. RESULTS Out of 10,307 records screened, 20 studies met the eligibility criteria for this review incorporating a total of 20,286 aneurysm cases. The machine learning models gave a 0.66-0.90 range for performance accuracy. The models were compared to current clinical standards in six studies and gave mixed results. Most studies posed high or unclear risks of bias and concerns for applicability, limiting the inferences that can be drawn from them. There was insufficient homogenous data for a meta-analysis. CONCLUSIONS Machine learning can be applied to predict the risk of rupture for intracranial aneurysms. However, the evidence does not comprehensively demonstrate superiority to existing practice, limiting its role as a clinical adjunct. Further prospective multicentre studies of recent machine learning tools are needed to prove clinical validation before they are implemented in the clinic.
Collapse
Affiliation(s)
- Karan Daga
- School of Biomedical Engineering & Imaging Sciences, King's College London, BMEIS, King's College London. 1 Lambeth Palace Road, UK SE1 7EU, London, UK
- Guy's and St. Thomas' NHS Foundation Trust, Westminster Bridge Road, UK SE1 7EH, London, UK
| | - Siddharth Agarwal
- School of Biomedical Engineering & Imaging Sciences, King's College London, BMEIS, King's College London. 1 Lambeth Palace Road, UK SE1 7EU, London, UK
| | - Zaeem Moti
- Guy's and St. Thomas' NHS Foundation Trust, Westminster Bridge Road, UK SE1 7EH, London, UK
| | - Matthew B K Lee
- University College London Hospital NHS Foundation Trust, 235 Euston Rd, UK NW1 2BU, London, UK
| | - Munaib Din
- Guy's and St. Thomas' NHS Foundation Trust, Westminster Bridge Road, UK SE1 7EH, London, UK
| | - David Wood
- School of Biomedical Engineering & Imaging Sciences, King's College London, BMEIS, King's College London. 1 Lambeth Palace Road, UK SE1 7EU, London, UK
| | - Marc Modat
- School of Biomedical Engineering & Imaging Sciences, King's College London, BMEIS, King's College London. 1 Lambeth Palace Road, UK SE1 7EU, London, UK
| | - Thomas C Booth
- School of Biomedical Engineering & Imaging Sciences, King's College London, BMEIS, King's College London. 1 Lambeth Palace Road, UK SE1 7EU, London, UK.
- Department of Neuroradiology, King's College Hospital, Denmark Hill, UK SE5 9RS, London, UK.
| |
Collapse
|
48
|
Abbott LP, Saikia A, Anthonappa RP. ARTIFICIAL INTELLIGENCE PLATFORMS IN DENTAL CARIES DETECTION: A SYSTEMATIC REVIEW AND META-ANALYSIS. J Evid Based Dent Pract 2025; 25:102077. [PMID: 39947783 DOI: 10.1016/j.jebdp.2024.102077] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Revised: 11/25/2024] [Accepted: 11/29/2024] [Indexed: 05/09/2025]
Abstract
OBJECTIVES To assess Artificial Intelligence (AI) platforms, machine learning methodologies and associated accuracies used in detecting dental caries from clinical images and dental radiographs. METHODS A systematic search of 8 distinct electronic databases: Scopus, Web of Science, MEDLINE, Educational Resources Information Centre, Institute of Electrical and Electronics Engineers Explore, Science Direct, Directory of Open Access Journals and JSTOR, was conducted from January 2000 to March 2024. AI platforms, machine learning methodologies and associated accuracies of studies using AI for dental caries detection were extracted along with essential study characteristics. The quality of included studies was assessed using QUADAS-2 and the CLAIM checklist. Meta-analysis was performed to obtain a quantitative estimate of AI accuracy. RESULTS Of the 2538 studies identified, 45 met the inclusion criteria and underwent qualitative synthesis. Of the 45 included studies, 33 used dental radiographs, and 12 used clinical images as datasets. A total of 21 different AI platforms were reported. The accuracy ranged from 41.5% to 98.6% across reported AI platforms. A quantitative meta-analysis across 7 studies reported a mean sensitivity of 76% [95% CI (65% - 85%)] and specificity of 91% [(95% CI (86% - 95%)]. The area under the curve (AUC) was 92% [95% CI (89% - 94%)], with high heterogeneity across included studies. CONCLUSION Significant variability exists in AI performance for detecting dental caries across different AI platforms. Meta-analysis demonstrates that AI has superior sensitivity and equal specificity of detecting dental caries from clinical images as compared to bitewing radiography. Although AI is promising for dental caries detection, further refinement is necessary to achieve consistent and reliable performance across varying imaging modalities.
Collapse
Affiliation(s)
- Lyndon P Abbott
- Paediatric Dentistry, UWA Dental School, The University of Western Australia, Perth, Australia.
| | - Ankita Saikia
- Paediatric Dentistry, UWA Dental School, The University of Western Australia, Perth, Australia
| | - Robert P Anthonappa
- Professor Paediatric Dentistry, UWA Dental School, The University of Western Australia, Perth, Australia
| |
Collapse
|
49
|
van der Waerden RGA, Volleberg RHJA, Luttikholt TJ, Cancian P, van der Zande JL, Stone GW, Holm NR, Kedhi E, Escaned J, Pellegrini D, Guagliumi G, Mehta SR, Pinilla-Echeverri N, Moreno R, Räber L, Roleder T, van Ginneken B, Sánchez CI, Išgum I, van Royen N, Thannhauser J. Artificial intelligence for the analysis of intracoronary optical coherence tomography images: a systematic review. EUROPEAN HEART JOURNAL. DIGITAL HEALTH 2025; 6:270-284. [PMID: 40110224 PMCID: PMC11914731 DOI: 10.1093/ehjdh/ztaf005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/20/2024] [Revised: 10/14/2024] [Accepted: 11/26/2024] [Indexed: 03/22/2025]
Abstract
Intracoronary optical coherence tomography (OCT) is a valuable tool for, among others, periprocedural guidance of percutaneous coronary revascularization and the assessment of stent failure. However, manual OCT image interpretation is challenging and time-consuming, which limits widespread clinical adoption. Automated analysis of OCT frames using artificial intelligence (AI) offers a potential solution. For example, AI can be employed for automated OCT image interpretation, plaque quantification, and clinical event prediction. Many AI models for these purposes have been proposed in recent years. However, these models have not been systematically evaluated in terms of model characteristics, performances, and bias. We performed a systematic review of AI models developed for OCT analysis to evaluate the trends and performances, including a systematic evaluation of potential sources of bias in model development and evaluation.
Collapse
Affiliation(s)
- Ruben G A van der Waerden
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
| | - Rick H J A Volleberg
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
| | - Thijs J Luttikholt
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
| | - Pierandrea Cancian
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
- Quantitative Healthcare Analysis (qurAI) Group, Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands
| | - Joske L van der Zande
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
| | - Gregg W Stone
- The Zena and Michael A Wiener Cardiovascular Institute, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Niels R Holm
- Department of Cardiology, Aarhus University Hospital, Aarhus, Denmark
| | - Elvin Kedhi
- McGill University Health Center, Royal Victoria Hospital, Montreal, Canada
| | - Javier Escaned
- Hospital Clinico San Carlos IdISSC, Complutense University of Madrid, Madrid, Spain
| | - Dario Pellegrini
- U.O. Cardiologia Ospedaliera, IRCCS Ospedale Galeazzi Sant'Ambrogio, Milan, Italy
| | - Giulio Guagliumi
- U.O. Cardiologia Ospedaliera, IRCCS Ospedale Galeazzi Sant'Ambrogio, Milan, Italy
| | - Shamir R Mehta
- Population Health Research Institute, McMaster University and Hamilton Health Sciences, Hamilton, ON, Canada
| | - Natalia Pinilla-Echeverri
- Division of Cardiology, Hamilton General Hospital, Hamilton Health Sciences, McMaster University, Hamilton, ON, Canada
| | - Raúl Moreno
- Interventional Cardiology, University Hospital La Paz, Madrid, Spain
| | - Lorenz Räber
- Department of Cardiology, Bern University Hospital Inselspital, Bern, Switzerland
| | - Tomasz Roleder
- Faculty of Medicine, Wrocław University of Science and Technology, Wrocław, Poland
| | - Bram van Ginneken
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
| | - Clara I Sánchez
- Quantitative Healthcare Analysis (qurAI) Group, Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands
- Department of Biomedical Engineering and Physics, Amsterdam University Medical Center University of Amsterdam, Amsterdam, The Netherlands
| | - Ivana Išgum
- Quantitative Healthcare Analysis (qurAI) Group, Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands
- Department of Biomedical Engineering and Physics, Amsterdam University Medical Center University of Amsterdam, Amsterdam, The Netherlands
- Department of Radiology and Nuclear Medicine, Amsterdam University Medical Center University of Amsterdam, Amsterdam, The Netherlands
| | - Niels van Royen
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
| | - Jos Thannhauser
- Department of Cardiology, Radboud University Medical Center, Geert Grooteplein Zuid 10, Nijmegen 6525 GA, The Netherlands
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10, Nijmegen 6525 GA, The Netherlands
| |
Collapse
|
50
|
Cavallo AU, Stanzione A, Ponsiglione A, Trotta R, Fanni SC, Ghezzo S, Vernuccio F, Klontzas ME, Triantafyllou M, Ugga L, Kalarakis G, Cannella R, Cuocolo R. Prostate cancer MRI methodological radiomics score: a EuSoMII radiomics auditing group initiative. Eur Radiol 2025; 35:1157-1165. [PMID: 39739041 DOI: 10.1007/s00330-024-11299-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 09/05/2024] [Accepted: 10/10/2024] [Indexed: 01/02/2025]
Abstract
OBJECTIVES To evaluate the quality of radiomics research in prostate MRI for the evaluation of prostate cancer (PCa) through the assessment of METhodological RadiomICs (METRICS) score, a new scoring tool recently introduced with the goal of fostering further improvement in radiomics and machine learning methodology. MATERIALS AND METHODS A literature search was conducted from July 1st, 2019, to November 30th, 2023, to identify original investigations assessing MRI-based radiomics in the setting of PCa. Seven readers with varying expertise underwent a quality assessment using METRICS. Subgroup analyses were performed to assess whether the quality score varied according to papers' categories (diagnosis, staging, prognosis, technical) and quality ratings among these latter. RESULTS From a total of 1106 records, 185 manuscripts were available. Overall, the average METRICS total score was 52% ± 16%. ANOVA and chi-square tests revealed no statistically significant differences between subgroups. Items with the lowest positive scores were adherence to guidelines/checklists (4.9%), handling of confounding factors (14.1%), external testing (15.1%), and the availability of data (15.7%), code (4.3%), and models (1.6%). Conversely, most studies clearly defined patient selection criteria (86.5%), employed a high-quality reference standard (89.2%), and utilized a well-described (85.9%) and clinically applicable (87%) imaging protocol as a radiomics data source. CONCLUSION The quality of MRI-based radiomics research for PCa in recent studies demonstrated good homogeneity and overall moderate quality. KEY POINTS Question To evaluate the quality of MRI-based radiomics research for PCa, assessed through the METRICS score. Findings The average METRICS total score was 52%, reflecting moderate quality in MRI-based radiomics research for PCa, with no statistically significant differences between subgroups. Clinical relevance Enhancing the quality of radiomics research can improve diagnostic accuracy for PCa, leading to better patient outcomes and more informed clinical decision-making.
Collapse
Affiliation(s)
| | - Arnaldo Stanzione
- Department of Advanced Biomedical Sciences, University of Naples "Federico II", Naples, Italy
| | - Andrea Ponsiglione
- Department of Advanced Biomedical Sciences, University of Naples "Federico II", Naples, Italy.
| | - Romina Trotta
- Department of Radiology, Fatima Hospital, Seville, Spain
| | | | | | - Federica Vernuccio
- Section of Radiology, Department of Biomedicine, Neuroscience and Advanced Diagnostics (BIND), University of Palermo, Palermo, Italy
| | - Michail E Klontzas
- Artificial Intelligence and Translational Imaging (ATI) Lab, Department of Radiology, School of Medicine, University of Crete, Heraklion, Greece
- Department of Medical Imaging, University Hospital of Heraklion, Heraklion, Greece
- Division of Radiology, Department of Clinical Science, Intervention and Technology (CLINTEC), Karolinska Institutet, Stockholm, Sweden
| | - Matthaios Triantafyllou
- Artificial Intelligence and Translational Imaging (ATI) Lab, Department of Radiology, School of Medicine, University of Crete, Heraklion, Greece
- Department of Medical Imaging, University Hospital of Heraklion, Heraklion, Greece
| | - Lorenzo Ugga
- Department of Advanced Biomedical Sciences, University of Naples "Federico II", Naples, Italy
| | - Georgios Kalarakis
- Division of Radiology, Department of Clinical Science, Intervention and Technology (CLINTEC), Karolinska Institutet, Stockholm, Sweden
- Department of Neuroradiology, Karolinska University Hospital, Stockholm, Sweden
| | - Roberto Cannella
- Section of Radiology, Department of Biomedicine, Neuroscience and Advanced Diagnostics (BIND), University of Palermo, Palermo, Italy
| | - Renato Cuocolo
- Department of Medicine, Surgery, and Dentistry, University of Salerno, Baronissi, Italy
| |
Collapse
|