1
|
Reid JA. Building Clinical Simulations With ChatGPT in Nursing Education. J Nurs Educ 2025; 64:e6-e7. [PMID: 39073762 DOI: 10.3928/01484834-20240424-05] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/30/2024]
Abstract
BACKGROUND Competency-based nursing education necessitates effective instructional methods and assessment tools for evaluating students' knowledge, skills, and attitudes. Clinical simulation has emerged as a valuable approach, but creating well-crafted simulations traditionally requires substantial time and effort. The advent of artificial intelligence (AI), exemplified by ChatGPT (OpenAI), offers promising advancements in streamlining scenario creation. METHOD This article explores the application of ChatGPT-3, version GPT-3, created by OpenAI in generating clinical simulation scenarios for nursing education. The focus is on the convenience, speed, and creativity provided by ChatGPT, enabling nurse educators to save time while developing intricate and thought-provoking scenarios. RESULTS ChatGPT generates intricate scenarios that stimulate critical thinking, significantly reducing the time required for nurse educators to create simulations. This AI tool's ability to produce clinical simulations quickly demonstrates its potential to enhance educational experiences in nursing. CONCLUSION ChatGPT's convenience, speed, and innovative capabilities make it invaluable for constructing dynamic clinical simulations, opening new avenues for innovative instruction in nursing education. This article highlights the transformative role of AI in empowering educators and enhancing educational experiences, showcasing ChatGPT's potential to revolutionize nursing education despite ongoing discussions about its potential negative impacts. [J Nurs Educ. 2025;64(5):e6-e7.].
Collapse
|
2
|
Asal MGR, Alsenany SA, Elzohairy NW, El-Sayed AAI. The impact of digital competence on pedagogical innovation among nurse educators: The moderating role of artificial intelligence readiness. Nurse Educ Pract 2025; 85:104367. [PMID: 40209516 DOI: 10.1016/j.nepr.2025.104367] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 03/30/2025] [Accepted: 04/05/2025] [Indexed: 04/12/2025]
Abstract
AIM To investigate the relationships between digital competence, AI readiness and pedagogical innovation among nurse educators, with a specific focus on the moderating role of AI readiness. BACKGROUND Digital competence is vital for nurse educators, supporting technology integration and promoting pedagogical innovation. AI readiness further enhances this innovation, fostering dynamic learning environments. However, research on how digital competence and AI readiness together have an impact on pedagogical innovation among nurse educators remains limited. DESIGN Cross-sectional study. METHODS Data were collected from 600 nurse educators across various nursing faculties in Egypt. Validated scales measured digital competence, AI readiness and pedagogical innovation. Pearson correlation, multiple regression and moderation analyses were used to test study hypotheses. RESULTS Significant positive correlations were found between pedagogical innovation, digital competence (r = 0.546, p < 0.01) and AI readiness (r = 0.530, p < 0.01). Digital competence (B = 0.558, p < 0.001) and AI readiness (B = 0.580, p < 0.001) significantly predicted pedagogical innovation. AI readiness moderated this relationship (B = 0.199, p < 0.001, ΔR² = 0.0057), amplifying the effect at higher levels of AI readiness (B = 0.66, p < 0.001). CONCLUSION Digital competence and AI readiness play critical roles in promoting pedagogical innovation. Strengthening AI readiness through targeted training can enhance digital tools adoption in nursing education. It is crucial to revise academic standards for curricula and nurse educators to include AI competence, ensuring effective integration of AI and digital tools in nursing education through targeted training and infrastructure improvements.
Collapse
Affiliation(s)
- Maha Gamal Ramadan Asal
- Nursing Department, College of Pharmacy and Applied Medical Sciences, Dar Al Uloom University, Riyadh, Saudi Arabia.
| | - Samira Ahmed Alsenany
- Public Health Department, Faculty of Nursing, King Abdulaziz University, Jeddah, Saudi Arabia.
| | - Nadia Waheed Elzohairy
- Psychiatric and Mental Health Nursing Department, Faculty of Nursing, Damanhour University, Damanhour, Egypt.
| | | |
Collapse
|
3
|
Resnik DB, Hosseini M. The ethics of using artificial intelligence in scientific research: new guidance needed for a new tool. AI AND ETHICS 2025; 5:1499-1521. [PMID: 40337745 PMCID: PMC12057767 DOI: 10.1007/s43681-024-00493-8] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/14/2023] [Accepted: 05/07/2024] [Indexed: 05/09/2025]
Abstract
Using artificial intelligence (AI) in research offers many important benefits for science and society but also creates novel and complex ethical issues. While these ethical issues do not necessitate changing established ethical norms of science, they require the scientific community to develop new guidance for the appropriate use of AI. In this article, we briefly introduce AI and explain how it can be used in research, examine some of the ethical issues raised when using it, and offer nine recommendations for responsible use, including: (1) Researchers are responsible for identifying, describing, reducing, and controlling AI-related biases and random errors; (2) Researchers should disclose, describe, and explain their use of AI in research, including its limitations, in language that can be understood by non-experts; (3) Researchers should engage with impacted communities, populations, and other stakeholders concerning the use of AI in research to obtain their advice and assistance and address their interests and concerns, such as issues related to bias; (4) Researchers who use synthetic data should (a) indicate which parts of the data are synthetic; (b) clearly label the synthetic data; (c) describe how the data were generated; and (d) explain how and why the data were used; (5) AI systems should not be named as authors, inventors, or copyright holders but their contributions to research should be disclosed and described; (6) Education and mentoring in responsible conduct of research should include discussion of ethical use of AI.
Collapse
Affiliation(s)
- David B. Resnik
- National Institute of Environmental Health Sciences, Durham, USA
| | - Mohammad Hosseini
- Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL USA
- Galter Health Sciences Library and Learning Center, Northwestern University Feinberg School of Medicine, Chicago, IL USA
| |
Collapse
|
4
|
Raman R. Transparency in research: An analysis of ChatGPT usage acknowledgment by authors across disciplines and geographies. Account Res 2025; 32:277-298. [PMID: 37877216 DOI: 10.1080/08989621.2023.2273377] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 10/17/2023] [Indexed: 10/26/2023]
Abstract
This investigation systematically reviews the recognition of generative AI tools, particularly ChatGPT, in scholarly literature. Utilizing 1,226 publications from the Dimensions database, ranging from November 2022 to July 2023, the research scrutinizes temporal trends and distribution across disciplines and regions. U.S.-based authors lead in acknowledgments, with notable contributions from China and India. Predominantly, Biomedical and Clinical Sciences, as well as Information and Computing Sciences, are engaging with these AI tools. Publications like "The Lancet Digital Health" and platforms such as "bioRxiv" are recurrent venues for such acknowledgments, highlighting AI's growing impact on research dissemination. The analysis is confined to the Dimensions database, thus potentially overlooking other sources and grey literature. Additionally, the study abstains from examining the acknowledgments' quality or ethical considerations. Findings are beneficial for stakeholders, providing a basis for policy and scholarly discourse on ethical AI use in academia. This study represents the inaugural comprehensive empirical assessment of AI acknowledgment patterns in academic contexts, addressing a previously unexplored aspect of scholarly communication.
Collapse
Affiliation(s)
- Raghu Raman
- Amrita School of Business, Amrita Vishwa Vidyapeetham, Amritapuri, Kerala, India
| |
Collapse
|
5
|
Doru B, Maier C, Busse JS, Lücke T, Schönhoff J, Enax-Krumova E, Hessler S, Berger M, Tokic M. Detecting Artificial Intelligence-Generated Versus Human-Written Medical Student Essays: Semirandomized Controlled Study. JMIR MEDICAL EDUCATION 2025; 11:e62779. [PMID: 40053752 PMCID: PMC11914838 DOI: 10.2196/62779] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 11/28/2024] [Accepted: 01/16/2025] [Indexed: 03/09/2025]
Abstract
BACKGROUND Large language models, exemplified by ChatGPT, have reached a level of sophistication that makes distinguishing between human- and artificial intelligence (AI)-generated texts increasingly challenging. This has raised concerns in academia, particularly in medicine, where the accuracy and authenticity of written work are paramount. OBJECTIVE This semirandomized controlled study aims to examine the ability of 2 blinded expert groups with different levels of content familiarity-medical professionals and humanities scholars with expertise in textual analysis-to distinguish between longer scientific texts in German written by medical students and those generated by ChatGPT. Additionally, the study sought to analyze the reasoning behind their identification choices, particularly the role of content familiarity and linguistic features. METHODS Between May and August 2023, a total of 35 experts (medical: n=22; humanities: n=13) were each presented with 2 pairs of texts on different medical topics. Each pair had similar content and structure: 1 text was written by a medical student, and the other was generated by ChatGPT (version 3.5, March 2023). Experts were asked to identify the AI-generated text and justify their choice. These justifications were analyzed through a multistage, interdisciplinary qualitative analysis to identify relevant textual features. Before unblinding, experts rated each text on 6 characteristics: linguistic fluency and spelling/grammatical accuracy, scientific quality, logical coherence, expression of knowledge limitations, formulation of future research questions, and citation quality. Univariate tests and multivariate logistic regression analyses were used to examine associations between participants' characteristics, their stated reasons for author identification, and the likelihood of correctly determining a text's authorship. RESULTS Overall, in 48 out of 69 (70%) decision rounds, participants accurately identified the AI-generated texts, with minimal difference between groups (medical: 31/43, 72%; humanities: 17/26, 65%; odds ratio [OR] 1.37, 95% CI 0.5-3.9). While content errors had little impact on identification accuracy, stylistic features-particularly redundancy (OR 6.90, 95% CI 1.01-47.1), repetition (OR 8.05, 95% CI 1.25-51.7), and thread/coherence (OR 6.62, 95% CI 1.25-35.2)-played a crucial role in participants' decisions to identify a text as AI-generated. CONCLUSIONS The findings suggest that both medical and humanities experts were able to identify ChatGPT-generated texts in medical contexts, with their decisions largely based on linguistic attributes. The accuracy of identification appears to be independent of experts' familiarity with the text content. As the decision-making process primarily relies on linguistic attributes-such as stylistic features and text coherence-further quasi-experimental studies using texts from other academic disciplines should be conducted to determine whether instructions based on these features can enhance lecturers' ability to distinguish between student-authored and AI-generated work.
Collapse
Affiliation(s)
- Berin Doru
- University Hospital of Paediatrics and Adolescent Medicine, St. Josef-Hospital, Ruhr University Bochum, Bochum, Germany
| | - Christoph Maier
- University Hospital of Paediatrics and Adolescent Medicine, St. Josef-Hospital, Ruhr University Bochum, Bochum, Germany
| | - Johanna Sophie Busse
- University Hospital of Paediatrics and Adolescent Medicine, St. Josef-Hospital, Ruhr University Bochum, Bochum, Germany
| | - Thomas Lücke
- University Hospital of Paediatrics and Adolescent Medicine, St. Josef-Hospital, Ruhr University Bochum, Bochum, Germany
| | - Judith Schönhoff
- Departement of German Philology, General and Comparative Literary Studies, Ruhr University Bochum, Bochum, Germany
| | - Elena Enax-Krumova
- Department of Neurology, BG University Hospital Bergmannsheil gGmbH Bochum, Ruhr University Bochum, Bochum, Germany
| | - Steffen Hessler
- German Department, German Linguistics, Ruhr University Bochum, Bochum, Germany
| | - Maria Berger
- German Department, Digital Forensic Linguistics, Ruhr University Bochum, Bochum, Germany
| | - Marianne Tokic
- Department for Medical Informatics, Biometry and Epidemiology, Ruhr University Bochum, Bochum, Germany
| |
Collapse
|
6
|
Özçevik Subaşi D, Akça Sümengen A, Semerci R, Şimşek E, Çakır GN, Temizsoy E. Paediatric nurses' perspectives on artificial intelligence applications: A cross-sectional study of concerns, literacy levels and attitudes. J Adv Nurs 2025; 81:1353-1363. [PMID: 39003632 DOI: 10.1111/jan.16335] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2024] [Revised: 06/24/2024] [Accepted: 07/02/2024] [Indexed: 07/15/2024]
Abstract
AIMS This study aimed to explore the correlation between artificial intelligence (AI) literacy, AI anxiety and AI attitudes among paediatric nurses, as well as identify the influencing factors on paediatric nurses' AI attitudes. DESIGN A descriptive, correlational and cross-sectional research. METHODS This study was conducted between January and February 2024 with 170 nurses actively working in paediatric clinics in Turkey. The data collection tools included the Nurse Information Form, the General Attitudes Towards Artificial Intelligence Scale (GAAIS), the Artificial Intelligence Literacy Scale (AILS) and the Artificial Intelligence Anxiety Scale (AIAS). To determine the associations between the variables, the data was analysed using IBM SPSS 28, which included linear regression and Pearson correlation analysis. RESULTS The study indicated significant positive correlations between paediatric nurses' age and their AIAS scores (r = .226; p < .01) and significant negative correlations between paediatric nurses' age and their AILS (r = -.192; p < .05) and GAAIS scores (r = -.152; p < .05). The GAAIS was significantly predictive (p < .000) and accounted for 50% of the variation in AIAS and AILS scores. CONCLUSION Paediatric nurses' attitudes towards AI significantly predicted AI literacy and AI anxiety. The relationship between the age of the paediatric nurses and the anxiety, AI literacy and attitudes towards AI was demonstrated. Healthcare and educational institutions should create customized training programs and awareness-raising activities for older nurses, as there are noticeable variations in the attitudes of paediatric nurses towards AI based on their age. IMPLICATIONS FOR PROFESSION AND/OR PATIENT CARE Providing in-service AI training can help healthcare organizations improve paediatric nurses' attitudes towards AI, increase their AI literacy and reduce their anxiety. This training has the potential to impact their attitudes positively and reduce their anxiety. REPORTING METHOD The study results were critically reported using STROBE criteria. PATIENT OR PUBLIC CONTRIBUTION No patient or public contribution.
Collapse
Affiliation(s)
| | - Aylin Akça Sümengen
- Capstone College of Nursing, The University of Alabama, Tuscaloosa, Alabama, USA
| | - Remziye Semerci
- Department of Pediatric Nursing, School of Nursing, Koç University, Istanbul, Turkey
| | - Enes Şimşek
- Department of Pediatric Nursing, School of Nursing, Koç University, Istanbul, Turkey
| | - Gökçe Naz Çakır
- Department of Nursing, Faculty of Health Science, Yeditepe University, Istanbul, Turkey
| | - Ebru Temizsoy
- Department of Nursing, Faculty of Health Sciences, Istanbul Bilgi University, Istanbul, Turkey
| |
Collapse
|
7
|
Rodger D, O'Connor S. Using artificial intelligence in health research. Evid Based Nurs 2025:ebnurs-2025-104287. [PMID: 40015946 DOI: 10.1136/ebnurs-2025-104287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/01/2025]
Affiliation(s)
- Daniel Rodger
- School of Allied and Community Health, London South Bank University, London, England, UK
- School of Psychological Sciences, Birkbeck, University of London, London, England, UK
| | - Siobhan O'Connor
- Florence Nightingale Faculty of Nursing, Midwifery and Palliative Care, King's College London, London, England, UK
| |
Collapse
|
8
|
Abdelhafiz AS, Farghly MI, Sultan EA, Abouelmagd ME, Ashmawy Y, Elsebaie EH. Medical students and ChatGPT: analyzing attitudes, practices, and academic perceptions. BMC MEDICAL EDUCATION 2025; 25:187. [PMID: 39910593 PMCID: PMC11800517 DOI: 10.1186/s12909-025-06731-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 01/21/2025] [Indexed: 02/07/2025]
Abstract
BACKGROUND ChatGPT, a chatbot launched by OpenAI in November 2022, has generated both excitement and concern within the healthcare education, research, and practice communities. This study aimed to explore the knowledge, perceptions, attitudes, and practices of undergraduate medical students regarding the use of ChatGPT and similar chatbots in their academic work. METHODS An anonymous, structured questionnaire was developed using Google Forms and administered to medical students as part of a cross-sectional study. The survey targeted undergraduate medical students from four governorates in Egypt. The questionnaire link was distributed through social media platforms, including Facebook and WhatsApp. The survey comprised four sections: socio-demographic characteristics, perceptions, attitudes, and practices. RESULTS The survey achieved a response rate of 96%, with 614 out of 640 approached students participating. Prior to the study, most respondents (78.5%) had personal experience using it. Overall, respondents demonstrated positive perceptions, attitudes, and practices toward ChatGPT, with mean scores exceeding 3 for all three variables: 3.99 ± 0.60 for perceptions, 3.01 ± 0.46 for attitudes, and 3.55 ± 0.55 for practices. In general, the students exhibited a high degree of trust in the model, with approximately half trusting the accuracy and reliability of the information provided by ChatGPT. However, more than two-thirds expressed apprehension about its potential misuse in medical education, and around 60% were concerned about the accuracy of information ChatGPT might generate on complex medical topics. CONCLUSIONS Medical students show strong interest and trust in using ChatGPT and similar chatbots for academic purposes but have concerns about the reliability of the information and potential misuse in medical education. The use of AI tools should follow ethical guidelines set by academic institutions, with regular updates to keep pace with technological progress. Future research should focus on the impact of AI on education and personal development, especially among young people.
Collapse
Affiliation(s)
- Ahmed Samir Abdelhafiz
- Department of Clinical Pathology, National Cancer Institute, Cairo University, Kasr Al-Aini Street, Fom Elkhalig Square, Cairo, 11796, Egypt.
| | - Maysa I Farghly
- Department of Clinical and Chemical Pathology, Faculty of Medicine, Suez University, Suez, Egypt
| | - Eman Anwar Sultan
- Department of Community Medicine, Faculty of Medicine, Alexandria University, Alexandria, Egypt
| | | | | | - Eman Hany Elsebaie
- Department of Public Health and Community Medicine, Faculty of Medicine, Cairo University, Cairo, Egypt
| |
Collapse
|
9
|
Coşkun AB, Kenner C, Elmaoğlu E. Neonatal Intensive Care Nurses' Perceptions of Artificial Intelligence: A Qualitative Study on Discharge Education and Family Counseling. J Perinat Neonatal Nurs 2024:00005237-990000000-00076. [PMID: 39740134 DOI: 10.1097/jpn.0000000000000904] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/02/2025]
Abstract
OBJECTIVE This study aims to examine neonatal intensive care unit (NICU) nurses' perceptions of artificial intelligence (AI) technologies, particularly language models, and their impact on nursing practices. BACKGROUND AI is rapidly spreading in healthcare, transforming nursing practice. Understanding the role of AI in NICUs in the discharge process is crucial for understanding nurses' perceptions of these technologies. METHODS The qualitative, phenomenological study used semi-structured interviews. Data were collected in a public hospital in Gaziantep from January to June 2024. Fifteen NICU nurses participated. Data were analyzed using content analysis. RESULTS Most nurses found AI to be a valuable tool for saving time and simplifying information delivery in clinical processes. However, concerns were raised about AI potentially reducing human interaction and weakening the use of professional judgment. Serious concerns about AI's reliability and ethical implications were also expressed. CONCLUSIONS AI is considered a potentially supportive tool in nursing practice, but its integration must consider the ethical implications and impact on the use of professional judgment. Nursing is based on human interactions and AI should be considered an additive tool to enhance care. IMPLICATIONS FOR PRACTICE AND RESEARCH AI integration in nursing requires careful and balanced implementation. Future research should delve deeper into the ethical dimensions of AI and its long-term effects on nursing practices.
Collapse
Affiliation(s)
- Adnan Batuhan Coşkun
- Author Affiliations: Faculty of Health Sciences, Department of Nursing, Hasan Kalyoncu University, Gaziantep, Turkey (Dr Coşkun); School of Nursing and Health Sciences, The College of New Jersey, Council of International Neonatal Nurses, Inc, Augusta, Georgia (Dr Kenner); and Yusuf Şerefoğlu Faculty of Health Sciences, Department of Nursing, Kilis 7 Aralik University, Kilis, Turkey (Dr Elmaoğlu)
| | | | | |
Collapse
|
10
|
Tortora F, Guastaferro A, Barbato S, Febbraio F, Cimmino A. New Challenges in Bladder Cancer Diagnosis: How Biosensing Tools Can Lead to Population Screening Opportunities. SENSORS (BASEL, SWITZERLAND) 2024; 24:7873. [PMID: 39771612 PMCID: PMC11679013 DOI: 10.3390/s24247873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/15/2024] [Revised: 12/05/2024] [Accepted: 12/07/2024] [Indexed: 01/11/2025]
Abstract
Bladder cancer is one of the most common cancers worldwide. Despite its high incidence, cystoscopy remains the currently used diagnostic gold standard, although it is invasive, expensive and has low sensitivity. As a result, the cancer diagnosis is mostly late, as it occurs following the presence of hematuria in urine, and population screening is not allowed. It would therefore be desirable to be able to act promptly in the early stage of the disease with the aid of biosensing. The use of devices/tools based on genetic assessments would be of great help in this field. However, the genetic differences between populations do not allow accurate analysis in the context of population screening. Current research is directed towards the discovery of universal biomarkers present in urine with the aim of providing an approach based on a non-invasive, easy-to-perform, rapid, and accurate test that can be widely used in clinical practice for the early diagnosis and follow-up of bladder cancer. An efficient biosensing device may have a disruptive impact in terms of patient health and disease management, contributing to a decrease in mortality rate, as well as easing the social and economic burden on the national healthcare system. Considering the advantage of accessing population screening for early diagnosis of cancer, the main challenges and future perspectives are critically discussed to address the research towards the selection of suitable biomarkers for the development of a very sensitive biosensor for bladder cancer.
Collapse
Affiliation(s)
- Fabiana Tortora
- Institute of Genetics and Biophysics “A. Buzzati Traverso”, National Research Council (CNR), 80131 Naples, Italy; (F.T.); (A.G.); (S.B.); (A.C.)
| | - Antonella Guastaferro
- Institute of Genetics and Biophysics “A. Buzzati Traverso”, National Research Council (CNR), 80131 Naples, Italy; (F.T.); (A.G.); (S.B.); (A.C.)
| | - Simona Barbato
- Institute of Genetics and Biophysics “A. Buzzati Traverso”, National Research Council (CNR), 80131 Naples, Italy; (F.T.); (A.G.); (S.B.); (A.C.)
| | - Ferdinando Febbraio
- Institute of Biochemistry and Cell Biology, National Research Council (CNR), 80131 Naples, Italy
| | - Amelia Cimmino
- Institute of Genetics and Biophysics “A. Buzzati Traverso”, National Research Council (CNR), 80131 Naples, Italy; (F.T.); (A.G.); (S.B.); (A.C.)
| |
Collapse
|
11
|
Jin HK, Kim E. Performance of GPT-3.5 and GPT-4 on the Korean Pharmacist Licensing Examination: Comparison Study. JMIR MEDICAL EDUCATION 2024; 10:e57451. [PMID: 39630413 PMCID: PMC11633516 DOI: 10.2196/57451] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/17/2024] [Revised: 08/28/2024] [Accepted: 10/09/2024] [Indexed: 12/13/2024]
Abstract
Background ChatGPT, a recently developed artificial intelligence chatbot and a notable large language model, has demonstrated improved performance on medical field examinations. However, there is currently little research on its efficacy in languages other than English or in pharmacy-related examinations. Objective This study aimed to evaluate the performance of GPT models on the Korean Pharmacist Licensing Examination (KPLE). Methods We evaluated the percentage of correct answers provided by 2 different versions of ChatGPT (GPT-3.5 and GPT-4) for all multiple-choice single-answer KPLE questions, excluding image-based questions. In total, 320, 317, and 323 questions from the 2021, 2022, and 2023 KPLEs, respectively, were included in the final analysis, which consisted of 4 units: Biopharmacy, Industrial Pharmacy, Clinical and Practical Pharmacy, and Medical Health Legislation. Results The 3-year average percentage of correct answers was 86.5% (830/960) for GPT-4 and 60.7% (583/960) for GPT-3.5. GPT model accuracy was highest in Biopharmacy (GPT-3.5 77/96, 80.2% in 2022; GPT-4 87/90, 96.7% in 2021) and lowest in Medical Health Legislation (GPT-3.5 8/20, 40% in 2022; GPT-4 12/20, 60% in 2022). Additionally, when comparing the performance of artificial intelligence with that of human participants, pharmacy students outperformed GPT-3.5 but not GPT-4. Conclusions In the last 3 years, GPT models have performed very close to or exceeded the passing threshold for the KPLE. This study demonstrates the potential of large language models in the pharmacy domain; however, extensive research is needed to evaluate their reliability and ensure their secure application in pharmacy contexts due to several inherent challenges. Addressing these limitations could make GPT models more effective auxiliary tools for pharmacy education.
Collapse
Affiliation(s)
- Hye Kyung Jin
- Research Institute of Pharmaceutical Sciences, College of Pharmacy, Chung-Ang University, Seoul, Republic of Korea
- Data Science, Evidence-Based and Clinical Research Laboratory, Department of Health, Social, and Clinical Pharmacy, College of Pharmacy, Chung-Ang University, Seoul, Republic of Korea
| | - EunYoung Kim
- Research Institute of Pharmaceutical Sciences, College of Pharmacy, Chung-Ang University, Seoul, Republic of Korea
- Data Science, Evidence-Based and Clinical Research Laboratory, Department of Health, Social, and Clinical Pharmacy, College of Pharmacy, Chung-Ang University, Seoul, Republic of Korea
- Division of Licensing of Medicines and Regulatory Science, The Graduate School of Pharmaceutical Management and Regulatory Science Policy, The Graduate School of Pharmaceutical Regulatory Sciences, Chung-Ang University, 84 Heukseok-Ro, Dongjak-gu, Seoul, 06974, Republic of Korea, 82 2-820-5791, 82 2-816-7338
| |
Collapse
|
12
|
Bilika P, Stefanouli V, Strimpakos N, Kapreli EV. Clinical reasoning using ChatGPT: Is it beyond credibility for physiotherapists use? Physiother Theory Pract 2024; 40:2943-2962. [PMID: 38073539 DOI: 10.1080/09593985.2023.2291656] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Revised: 12/01/2023] [Accepted: 12/01/2023] [Indexed: 11/30/2024]
Abstract
BACKGROUND Artificial Intelligence (AI) tools are gaining popularity in healthcare. OpenAI released ChatGPT on November 30, 2022. ChatGPT is a language model that comprehends and generates human language, providing instant data analysis and recommendations. This is particularly significant in the dynamic field of physiotherapy, where its integration has the potential to enhance healthcare efficiency. OBJECTIVES This study aims to evaluate whether ChatGPT-3.5 (free version) provides consistent and accurate clinical responses, its ability to imitate human clinical reasoning in simple and complex scenarios, and its capability to produce a differential diagnosis. METHODS Two studies were conducted using the ChatGPT-3.5. Study 1 evaluated the consistency and accuracy of ChatGPT's responses in clinical assessment using ten user-participants who submitted the phrase "Which are the main steps for a completed physiotherapy assessment?" Study 2 assessed ChatGPT's differential diagnostic ability using published case studies by 2 independent participants. The case reports consisted of one simple and one complex scenario. RESULTS Study 1 underscored the variability in ChatGPT's responses, which ranged from comprehensive to concise. Notably, essential steps such as re-assessment and subjective examination were omitted in 30% and 40% of the responses, respectively. In Study 2, ChatGPT demonstrated its capability to develop evidence-based clinical reasoning, particularly evident in simple clinical scenarios. Question phrasing significantly impacted the generated answers. CONCLUSIONS This study highlights the potential benefits of using ChatGPT in healthcare. It also provides a balanced perspective on ChatGPT's strengths and limitations and emphasizes the importance of using AI tools in a responsible and informed manner.
Collapse
Affiliation(s)
- Paraskevi Bilika
- Physiotherapy Department, Faculty of Health Sciences, Clinical Exercise Physiology and Rehabilitation Research Laboratory, University of Thessaly, Lamia, Greece
| | - Vasiliki Stefanouli
- Physiotherapy Department, Faculty of Health Sciences, Health Assessment and Quality of Life Research Laboratory, University of Thessaly, Lamia, Greece
| | - Nikolaos Strimpakos
- Physiotherapy Department, Faculty of Health Sciences, Health Assessment and Quality of Life Research Laboratory, University of Thessaly, Lamia, Greece
- Division of Musculoskeletal and Dermatological Sciences, University of Manchester, Manchester, UK
| | - Eleni V Kapreli
- Physiotherapy Department, Faculty of Health Sciences, Clinical Exercise Physiology and Rehabilitation Research Laboratory, University of Thessaly, Lamia, Greece
| |
Collapse
|
13
|
Kleib M, Darko EM, Akingbade O, Kennedy M, Majekodunmi P, Nickel E, Vogelsang L. Current trends and future implications in the utilization of ChatGPT in nursing: A rapid review. INTERNATIONAL JOURNAL OF NURSING STUDIES ADVANCES 2024; 7:100252. [PMID: 39584012 PMCID: PMC11583729 DOI: 10.1016/j.ijnsa.2024.100252] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Revised: 09/16/2024] [Accepted: 10/21/2024] [Indexed: 11/26/2024] Open
Abstract
Background The past decade has witnessed a surge in the development of artificial intelligence (AI)-based technology systems for healthcare. Launched in November 2022, ChatGPT (Generative Pre-trained Transformer), an AI-based Chatbot, is being utilized in nursing education, research and practice. However, little is known about its pattern of usage, which prompted this study. Objective To provide a concise overview of the existing literature on the application of ChatGPT in nursing education, practice and research. Methods A rapid review based on the Cochrane methodology was applied to synthesize existing literature. We conducted systematic searches in several databases, including CINAHL, Ovid Medline, Embase, Web of Science, Scopus, Education Search Complete, ERIC, and Cochrane CENTRAL, to ensure no publications were missed. All types of primary and secondary research studies, including qualitative, quantitative, mixed methods, and literature reviews published in the English language focused on the use of ChatGPT in nursing education, research, and practice, were included. Dissertations or theses, conference proceedings, government and other organizational reports, white papers, discussion papers, opinion pieces, editorials, commentaries, and published review protocols were excluded. Studies involving other healthcare professionals and/or students without including nursing participants were excluded. Studies exploring other language models without comparison to ChatGPT and those examining the technical specifications of ChatGPT were excluded. Data screening was completed in two stages: titles and abstract and full-text review, followed by data extraction and quality appraisal. Descriptive analysis and narrative synthesis were applied to summarize and categorize the findings. Results Seventeen studies were included: 15 (88.2 %) focused on nursing education and one each on nursing practice and research. Of the 17 included studies, 5 (29.4 %) were evaluation studies, 3 (17.6 %) were narrative reviews, 3 (17.6 %) were cross-sectional studies, 2 (11.8 %) were descriptive studies, and 1 (5.9 %) was a randomized controlled trial, quasi-experimental study, case study, and qualitative study, respectively. Conclusion This study has provided a snapshot of ChatGPT usage in nursing education, research, and practice. Although evidence is inconclusive, integration of ChatGPT should consider addressing ethical concerns and ongoing education on ChatGPT usage. Further research, specifically interventional studies, is recommended to ascertain and track the impact of ChatGPT in different contexts.
Collapse
Affiliation(s)
- Manal Kleib
- Faculty of Nursing, University of Alberta, Edmonton, Alberta, Canada
| | | | | | - Megan Kennedy
- Library and Museums - Faculty Engagement (Health Sciences), University of Alberta, Edmonton, Alberta, Canada
| | | | - Emma Nickel
- Alberta Health Services, Calgary, Alberta, Canada
| | - Laura Vogelsang
- Faculty of Health Sciences, University of Lethbridge, Lethbridge, Alberta, Canada
| |
Collapse
|
14
|
Soulage CO, Van Coppenolle F, Guebre-Egziabher F. The conversational AI "ChatGPT" outperforms medical students on a physiology university examination. ADVANCES IN PHYSIOLOGY EDUCATION 2024; 48:677-684. [PMID: 38991037 DOI: 10.1152/advan.00181.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Revised: 06/13/2024] [Accepted: 07/09/2024] [Indexed: 07/13/2024]
Abstract
Artificial intelligence (AI) has gained massive interest with the public release of the conversational AI "ChatGPT," but it also has become a matter of concern for academia as it can easily be misused. We performed a quantitative evaluation of the performance of ChatGPT on a medical physiology university examination. Forty-one answers were obtained with ChatGPT and compared to the results of 24 students. The results of ChatGPT were significantly better than those of the students; the median (IQR) score was 75% (66-84%) for the AI compared to 56% (43-65%) for students (P < 0.001). The exam success rate was 100% for ChatGPT, whereas 29% (n = 7) of students failed. ChatGPT could promote plagiarism and intellectual laziness among students and could represent a new and easy way to cheat, especially when evaluations are performed online. Considering that these powerful AI tools are now freely available, scholars should take great care to construct assessments that really evaluate student reflection skills and prevent AI-assisted cheating.NEW & NOTEWORTHY The release of the conversational artificial intelligence (AI) ChatGPT has become a matter of concern for academia as it can easily be misused by students for cheating purposes. We performed a quantitative evaluation of the performance of ChatGPT on a medical physiology university examination and observed that ChatGPT outperforms medical students obtaining significantly better grades. Scholars should therefore take great care to construct assessments crafted to really evaluate the student reflection skills and prevent AI-assisted cheating.
Collapse
Affiliation(s)
- Christophe O Soulage
- CarMeN, INSERM U1060, INRAe U1397, Université Claude Bernard Lyon 1, Bron, France
| | | | - Fitsum Guebre-Egziabher
- CarMeN, INSERM U1060, INRAe U1397, Université Claude Bernard Lyon 1, Bron, France
- Department of Nephrology, Groupement Hospitalier Centre, Hospices Civils de Lyon, Hôpital E. Herriot, Lyon, France
| |
Collapse
|
15
|
Gauckler C, Werner MH. Artificial Intelligence: A Challenge to Scientific Communication. Klin Monbl Augenheilkd 2024; 241:1309-1321. [PMID: 39637910 DOI: 10.1055/a-2418-5238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2024]
Abstract
Recent years have seen formidable advances in artificial intelligence. Developments include a large number of specialised systems either existing or planned for use in scientific research, data analysis, translation, text production and design with grammar checking and stylistic revision, plagiarism detection, and scientific review in addition to general-purpose AI systems for searching the internet and generative AI systems for texts, images, videos, and musical compositions. These systems promise more ease and simplicity in many aspects of work. Blind trust in AI systems with uncritical, careless use of AI results is dangerous, as these systems do not have any inherent understanding of the content they process or generate, but only simulate this understanding by reproducing statistical patterns extracted from training data. This article discusses the potential and risk of using AI in scientific communication and explores potential systemic consequences of widespread AI implementation in this context.
Collapse
Affiliation(s)
| | - Micha H Werner
- Institut für Philosophie, Universität Greifswald, Deutschland
| |
Collapse
|
16
|
Bortoli M, Fiore M, Tedeschi S, Oliveira V, Sousa R, Bruschi A, Campanacci DA, Viale P, De Paolis M, Sambri A. GPT-based chatbot tools are still unreliable in the management of prosthetic joint infections. Musculoskelet Surg 2024; 108:459-466. [PMID: 38954323 PMCID: PMC11582126 DOI: 10.1007/s12306-024-00846-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Accepted: 06/21/2024] [Indexed: 07/04/2024]
Abstract
BACKGROUND Artificial intelligence chatbot tools responses might discern patterns and correlations that may elude human observation, leading to more accurate and timely interventions. However, their reliability to answer healthcare-related questions is still debated. This study aimed to assess the performance of the three versions of GPT-based chatbots about prosthetic joint infections (PJI). METHODS Thirty questions concerning the diagnosis and treatment of hip and knee PJIs, stratified by a priori established difficulty, were generated by a team of experts, and administered to ChatGPT 3.5, BingChat, and ChatGPT 4.0. Responses were rated by three orthopedic surgeons and two infectious diseases physicians using a five-point Likert-like scale with numerical values to quantify the quality of responses. Inter-rater reliability was assessed by interclass correlation statistics. RESULTS Responses averaged "good-to-very good" for all chatbots examined, both in diagnosis and treatment, with no significant differences according to the difficulty of the questions. However, BingChat ratings were significantly lower in the treatment setting (p = 0.025), particularly in terms of accuracy (p = 0.02) and completeness (p = 0.004). Agreement in ratings among examiners appeared to be very poor. CONCLUSIONS On average, the quality of responses is rated positively by experts, but with ratings that frequently may vary widely. This currently suggests that AI chatbot tools are still unreliable in the management of PJI.
Collapse
Affiliation(s)
- M Bortoli
- Orthopedic and Traumatology Unit, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| | - M Fiore
- Orthopedic and Traumatology Unit, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy.
- Department of Medical and Surgical Sciences, Alma Mater Studiorum University of Bologna, 40138, Bologna, Italy.
| | - S Tedeschi
- Department of Medical and Surgical Sciences, Alma Mater Studiorum University of Bologna, 40138, Bologna, Italy
- Infectious Disease Unit, Department for Integrated Infectious Risk Management, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| | - V Oliveira
- Department of Orthopedics, Centro Hospitalar Universitário de Santo António, 4099-001, Porto, Portugal
| | - R Sousa
- Department of Orthopedics, Centro Hospitalar Universitário de Santo António, 4099-001, Porto, Portugal
| | - A Bruschi
- Orthopedic and Traumatology Unit, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| | - D A Campanacci
- Orthopedic Oncology Unit, Azienda Ospedaliera Universitaria Careggi, 50134, Florence, Italy
| | - P Viale
- Department of Medical and Surgical Sciences, Alma Mater Studiorum University of Bologna, 40138, Bologna, Italy
- Infectious Disease Unit, Department for Integrated Infectious Risk Management, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| | - M De Paolis
- Orthopedic and Traumatology Unit, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| | - A Sambri
- Orthopedic and Traumatology Unit, IRCCS Azienda Ospedaliero-Universitaria Di Bologna, 40138, Bologna, Italy
| |
Collapse
|
17
|
Anthea P, Claudia C. Artificial intelligence: The researcher's assistant or sheep in wolf's clothing? United European Gastroenterol J 2024; 12:1354-1356. [PMID: 39508309 DOI: 10.1002/ueg2.12689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/15/2024] Open
Affiliation(s)
- Pisani Anthea
- Division of Gastroenterology, Department of Medicine, Mater Dei Hospital, Msida, Malta
| | - Campani Claudia
- Internal Medicine and Liver Unit, University Hospital Careggi, University of Florence, Florence, Italy
- Department of Experimental and Clinical Medicine, University of Florence, Florence, Italy
| |
Collapse
|
18
|
Khalifa AA, Ibrahim MA. Artificial intelligence (AI) and ChatGPT involvement in scientific and medical writing, a new concern for researchers. A scoping review. ARAB GULF JOURNAL OF SCIENTIFIC RESEARCH 2024; 42:1770-1787. [DOI: 10.1108/agjsr-09-2023-0423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/30/2024]
Abstract
PurposeThe study aims to evaluate PubMed publications on ChatGPT or artificial intelligence (AI) involvement in scientific or medical writing and investigate whether ChatGPT or AI was used to create these articles or listed as authors.Design/methodology/approachThis scoping review was conducted according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) guidelines. A PubMed database search was performed for articles published between January 1 and November 29, 2023, using appropriate search terms; both authors performed screening and selection independently.FindingsFrom the initial search results of 127 articles, 41 were eligible for final analysis. Articles were published in 34 journals. Editorials were the most common article type, with 15 (36.6%) articles. Authors originated from 27 countries, and authors from the USA contributed the most, with 14 (34.1%) articles. The most discussed topic was AI tools and writing capabilities in 19 (46.3%) articles. AI or ChatGPT was involved in manuscript preparation in 31 (75.6%) articles. None of the articles listed AI or ChatGPT as an author, and in 19 (46.3%) articles, the authors acknowledged utilizing AI or ChatGPT.Practical implicationsResearchers worldwide are concerned with AI or ChatGPT involvement in scientific research, specifically the writing process. The authors believe that precise and mature regulations will be developed soon by journals, publishers and editors, which will pave the way for the best usage of these tools.Originality/valueThis scoping review expressed data published on using AI or ChatGPT in various scientific research and writing aspects, besides alluding to the advantages, disadvantages and implications of their usage.
Collapse
|
19
|
Zhou Y, Li SJ, Tang XY, He YC, Ma HM, Wang AQ, Pei RY, Piao MH. Using ChatGPT in Nursing: Scoping Review of Current Opinions. JMIR MEDICAL EDUCATION 2024; 10:e54297. [PMID: 39622702 PMCID: PMC11611787 DOI: 10.2196/54297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/04/2023] [Revised: 07/25/2024] [Accepted: 08/19/2024] [Indexed: 12/06/2024]
Abstract
Background Since the release of ChatGPT in November 2022, this emerging technology has garnered a lot of attention in various fields, and nursing is no exception. However, to date, no study has comprehensively summarized the status and opinions of using ChatGPT across different nursing fields. Objective We aim to synthesize the status and opinions of using ChatGPT according to different nursing fields, as well as assess ChatGPT's strengths, weaknesses, and the potential impacts it may cause. Methods This scoping review was conducted following the framework of Arksey and O'Malley and guided by the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews). A comprehensive literature research was conducted in 4 web-based databases (PubMed, Embase, Web of Science, and CINHAL) to identify studies reporting the opinions of using ChatGPT in nursing fields from 2022 to September 3, 2023. The references of the included studies were screened manually to further identify relevant studies. Two authors conducted studies screening, eligibility assessments, and data extraction independently. Results A total of 30 studies were included. The United States (7 studies), Canada (5 studies), and China (4 studies) were countries with the most publications. In terms of fields of concern, studies mainly focused on "ChatGPT and nursing education" (20 studies), "ChatGPT and nursing practice" (10 studies), and "ChatGPT and nursing research, writing, and examination" (6 studies). Six studies addressed the use of ChatGPT in multiple nursing fields. Conclusions As an emerging artificial intelligence technology, ChatGPT has great potential to revolutionize nursing education, nursing practice, and nursing research. However, researchers, institutions, and administrations still need to critically examine its accuracy, safety, and privacy, as well as academic misconduct and potential ethical issues that it may lead to before applying ChatGPT to practice.
Collapse
Affiliation(s)
- You Zhou
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Si-Jia Li
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Xing-Yi Tang
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Yi-Chen He
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Hao-Ming Ma
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Ao-Qi Wang
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Run-Yuan Pei
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| | - Mei-Hua Piao
- School of Nursing, Chinese Academy of Medical Sciences, Peking Union Medical College, No. 33 Badachu Road, Shijingshan District, Beijing, 100433, China, 86 13522112889
| |
Collapse
|
20
|
Anderson HD, Kwon S, Linnebur LA, Valdez CA, Linnebur SA. Pharmacy student use of ChatGPT: A survey of students at a U.S. School of Pharmacy. CURRENTS IN PHARMACY TEACHING & LEARNING 2024; 16:102156. [PMID: 39029382 DOI: 10.1016/j.cptl.2024.102156] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/10/2024] [Revised: 07/01/2024] [Accepted: 07/02/2024] [Indexed: 07/21/2024]
Abstract
OBJECTIVE To learn how students in an accredited PharmD program in the United States are using ChatGPT for personal, academic, and clinical reasons, and whether students think ChatGPT training should be incorporated into their program's curriculum. METHODS In August 2023, an 18-item survey was developed, pilot tested, and sent to all students who were enrolled during the Spring 2023 semester in the entry-level PharmD program at the University of Colorado. E-mail addresses were separated from survey responses to maintain anonymity. Responses were described using descriptive statistics. RESULTS 206 pharmacy students responded to the survey for a 49% response rate. Nearly one-half (48.5%) indicated they had used ChatGPT for personal reasons; 30.2% had used it for academic reasons; and 7.5% had used it for clinical reasons. The most common personal use for ChatGPT was answering questions and looking-up information (67.0%). The top academic reason for using ChatGPT was summarizing information or a body of text (42.6%), while the top clinical reason was simplifying a complex topic (53.3%). Most respondents (61.8%) indicated they would be interested in learning about how ChatGPT could help them in pharmacy school, and 28.1% thought ChatGPT training should be incorporated into their pharmacy curriculum. CONCLUSION At the time of the survey, ChatGPT was being used by approximately one-half of our pharmacy student respondents for personal, academic, or clinical reasons. Overall, many students indicated they want to learn how to use ChatGPT to help them with their education and think ChatGPT training should be integrated into their curriculum.
Collapse
Affiliation(s)
- Heather D Anderson
- University of Colorado Anschutz Medical Campus, Skaggs School of Pharmacy and Pharmaceutical Sciences, Department of Clinical Pharmacy, 12850 E. Montview Blvd, Mail stop C238, Aurora, CO 80045, United States of America.
| | - Sue Kwon
- University of Colorado Anschutz Medical Campus, Skaggs School of Pharmacy and Pharmaceutical Sciences, Department of Clinical Pharmacy, 12850 E. Montview Blvd, Mail stop C238, Aurora, CO 80045, United States of America.
| | - Lauren A Linnebur
- University of Colorado Anschutz Medical Campus, School of Medicine, Division of Geriatric Medicine, 12631 East 17th Avenue, Suite 8111, Aurora, CO 80045, United States of America.
| | - Connie A Valdez
- University of Colorado Anschutz Medical Campus, Skaggs School of Pharmacy and Pharmaceutical Sciences, Department of Clinical Pharmacy, 12850 E. Montview Blvd, Mail stop C238, Aurora, CO 80045, United States of America.
| | - Sunny A Linnebur
- University of Colorado Anschutz Medical Campus, Skaggs School of Pharmacy and Pharmaceutical Sciences, Department of Clinical Pharmacy, 12850 E. Montview Blvd, Mail stop C238, Aurora, CO 80045, United States of America.
| |
Collapse
|
21
|
Kucukkaya A, Arikan E, Goktas P. Unlocking ChatGPT's potential and challenges in intensive care nursing education and practice: A systematic review with narrative synthesis. Nurs Outlook 2024; 72:102287. [PMID: 39413564 DOI: 10.1016/j.outlook.2024.102287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Revised: 07/13/2024] [Accepted: 09/15/2024] [Indexed: 10/18/2024]
Abstract
BACKGROUND The advancement of artificial intelligence (AI) in healthcare and nursing promises to enhance clinical outcomes and education. This review emphasizes integrating AI chatbots, specifically ChatGPT, into Intensive Care Units (ICU) to transform nursing education and practices, while addressing associated risks and ethical challenges. PURPOSE To evaluate ChatGPT's utility in ICU nursing education and practices, assessing its effectiveness and develop strategic recommendations for its future incorporation into critical care. METHODS This review employs systematic literature with narrative synthesis, adhering to PRISMA guidelines. DISCUSSION Five of 1,091 identified studies were eligible. These studies illustrate AI-driven applications' potential in clinical decision-making and educational efforts, emphasizing the need for improved AI accuracy, robust guidelines, and measures to address data privacy concerns to ensure reliable integration. CONCLUSION ChatGPT presents promising benefits for ICU applications but requires careful management. Ongoing research and adherence to ethical standards are essential to optimize its use in critical care. TWEETABLE ABSTRACT Explore how #ChatGPT revolutionizes ICU nursing & education, blending AI capabilities with ethical, human-centered care. #ICM #AIinHealthcare.
Collapse
Affiliation(s)
- Aycan Kucukkaya
- Istanbul University-Cerrahpasa, Florence Nightingale Faculty of Nursing, Department of Nursing Education, Istanbul, Turkey.
| | - Emine Arikan
- Akdeniz University Kumluca Health Science Faculty, Department of Internal Medicine Nursing, Antalya, Turkey
| | - Polat Goktas
- UCD School of Computer Science, University College Dublin, Dublin, Ireland.
| |
Collapse
|
22
|
Loconte R, Orrù G, Tribastone M, Pietrini P, Sartori G. Challenging large language models' " intelligence" with human tools: A neuropsychological investigation in Italian language on prefrontal functioning. Heliyon 2024; 10:e38911. [PMID: 39430451 PMCID: PMC11490853 DOI: 10.1016/j.heliyon.2024.e38911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Revised: 10/02/2024] [Accepted: 10/02/2024] [Indexed: 10/22/2024] Open
Abstract
The Artificial Intelligence (AI) research community has used ad-hoc benchmarks to measure the "intelligence" level of Large Language Models (LLMs). In humans, intelligence is closely linked to the functional integrity of the prefrontal lobes, which are essential for higher-order cognitive processes. Previous research has found that LLMs struggle with cognitive tasks that rely on these prefrontal functions, highlighting a significant challenge in replicating human-like intelligence. In December 2022, OpenAI released ChatGPT, a new chatbot based on the GPT-3.5 model that quickly gained popularity for its impressive ability to understand and respond to human instructions, suggesting a significant step towards intelligent behaviour in AI. Therefore, to rigorously investigate LLMs' level of "intelligence," we evaluated the GPT-3.5 and GPT-4 versions through a neuropsychological assessment using tests in the Italian language routinely employed to assess prefrontal functioning in humans. The same tests were also administered to Claude2 and Llama2 to verify whether similar language models perform similarly in prefrontal tests. When using human performance as a reference, GPT-3.5 showed inhomogeneous results on prefrontal tests, with some tests well above average, others in the lower range, and others frankly impaired. Specifically, we have identified poor planning abilities and difficulty in recognising semantic absurdities and understanding others' intentions and mental states. Claude2 exhibited a similar pattern to GPT-3.5, while Llama2 performed poorly in almost all tests. These inconsistent profiles highlight how LLMs' emergent abilities do not yet mimic human cognitive functioning. The sole exception was GPT-4, which performed within the normative range for all the tasks except planning. Furthermore, we showed how standardised neuropsychological batteries developed to assess human cognitive functions may be suitable for challenging LLMs' performance.
Collapse
Affiliation(s)
- Riccardo Loconte
- Molecular Mind Lab, IMT School of Advanced Studies Lucca, Lucca, Italy
| | | | - Mirco Tribastone
- Molecular Mind Lab, IMT School of Advanced Studies Lucca, Lucca, Italy
| | - Pietro Pietrini
- Molecular Mind Lab, IMT School of Advanced Studies Lucca, Lucca, Italy
| | - Giuseppe Sartori
- Department of General Psychology, University of Padova, Padova, Italy
| |
Collapse
|
23
|
Reading Turchioe M, Kisselev S, Van Bulck L, Bakken S. Increasing Generative Artificial Intelligence Competency among Students Enrolled in Doctoral Nursing Research Coursework. Appl Clin Inform 2024; 15:842-851. [PMID: 39053615 PMCID: PMC11483171 DOI: 10.1055/a-2373-3151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Accepted: 07/24/2024] [Indexed: 07/27/2024] Open
Abstract
BACKGROUND Generative artificial intelligence (AI) tools may soon be integrated into health care practice and research. Nurses in leadership roles, many of whom are doctorally prepared, will need to determine whether and how to integrate them in a safe and useful way. OBJECTIVE This study aimed to develop and evaluate a brief intervention to increase PhD nursing students' knowledge of appropriate applications for using generative AI tools in health care. METHODS We created didactic lectures and laboratory-based activities to introduce generative AI to students enrolled in a nursing PhD data science and visualization course. Students were provided with a subscription to Chat Generative Pretrained Transformer (ChatGPT) 4.0, a general-purpose generative AI tool, for use in and outside the class. During the didactic portion, we described generative AI and its current and potential future applications in health care, including examples of appropriate and inappropriate applications. In the laboratory sessions, students were given three tasks representing different use cases of generative AI in health care practice and research (clinical decision support, patient decision support, and scientific communication) and asked to engage with ChatGPT on each. Students (n = 10) independently wrote a brief reflection for each task evaluating safety (accuracy, hallucinations) and usability (ease of use, usefulness, and intention to use in the future). Reflections were analyzed using directed content analysis. RESULTS Students were able to identify the strengths and limitations of ChatGPT in completing all three tasks and developed opinions on whether they would feel comfortable using ChatGPT for similar tasks in the future. All of them reported increasing their self-rated competency in generative AI by one to two points on a five-point rating scale. CONCLUSION This brief educational intervention supported doctoral nursing students in understanding the appropriate uses of ChatGPT, which may support their ability to appraise and use these tools in their future work.
Collapse
Affiliation(s)
| | - Sergey Kisselev
- Columbia University School of Nursing, New York, New York, United States
| | - Liesbet Van Bulck
- Department of Public Health and Primary Care, KU Leuven - University of Leuven, Leuven, Belgium
| | - Suzanne Bakken
- Columbia University School of Nursing, New York, New York, United States
- Department of Biomedical Informatics, Columbia University, New York, New York, United States
- Data Science Institute, Columbia University, New York, New York, United States
| |
Collapse
|
24
|
Luo C, Mao B, Wu Y, He Y. The research hotspots and theme trends of artificial intelligence in nurse education: A bibliometric analysis from 1994 to 2023. NURSE EDUCATION TODAY 2024; 141:106321. [PMID: 39084073 DOI: 10.1016/j.nedt.2024.106321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 07/09/2024] [Accepted: 07/22/2024] [Indexed: 08/02/2024]
Abstract
OBJECTIVES To explore research hotspots and theme trends in artificial intelligence in nurse education using bibliometric analysis. DESIGN Bibliometric analysis. DATA SOURCES Literature from the Web of Science Core Collection from the time of construction to October 31, 2023 was searched. REVIEW METHODS Analyses of countries, authors, institutions, journals, and keywords were conducted using Bibliometrix (based on R language), CiteSpace, the online analysis platform (bibliometric), Vosviewer, and Pajek. RESULTS A total of 135 articles with a straight upward trend over the last three years were retrieved. By fitting the curve R2 = 0.6022 (R2 > 0.4), we predicted that the number of annual articles is projected to grow in the coming years. The United States (n = 38), the National University of Singapore (n = 16), Professor Jun Ota (n = 8), and Nurse Education Today (n = 14) are the countries, institutions, authors, and journals that contributed to the most publications, respectively. Collaborative network analysis revealed that 32 institutional and 64 author collaborative teams were established. We identified ten high-frequency keywords and nine clusters. We categorized the research hotspots of artificial intelligence in nurse education into three areas: (1) Artificial intelligence-enhanced simulation robots, (2) machine learning and data mining, and (3) large language models based on natural language processing and deep learning. By analyzing the temporal and spatial evolution of keywords and burst detection, we found that future research trends may include (1) expanding and deepening the application of AI technology, (2) assessment of behavioral intent and educational outcomes, and (3) moral and ethical considerations. CONCLUSIONS Future research should be conducted on technology applications, behavioral intent, ethical policy, international cooperation, interdisciplinary cooperation, and sustainability to promote the continued development and innovation of AI in nurse education.
Collapse
Affiliation(s)
- Chuhong Luo
- School of Medicine, Hunan Normal University, Changsha, Hunan, People's Republic of China
| | - Bin Mao
- School of Medicine, Hunan Normal University, Changsha, Hunan, People's Republic of China
| | - Ying Wu
- School of Medicine, Hunan Normal University, Changsha, Hunan, People's Republic of China
| | - Ying He
- School of Medicine, Hunan Normal University, Changsha, Hunan, People's Republic of China.
| |
Collapse
|
25
|
Shin H, De Gagne JC, Kim SS, Hong M. The Impact of Artificial Intelligence-Assisted Learning on Nursing Students' Ethical Decision-making and Clinical Reasoning in Pediatric Care: A Quasi-Experimental Study. Comput Inform Nurs 2024; 42:704-711. [PMID: 39152099 PMCID: PMC11458082 DOI: 10.1097/cin.0000000000001177] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/19/2024]
Abstract
The integration of artificial intelligence such as ChatGPT into educational frameworks marks a pivotal transformation in teaching. This quasi-experimental study, conducted in September 2023, aimed to evaluate the effects of artificial intelligence-assisted learning on nursing students' ethical decision-making and clinical reasoning. A total of 99 nursing students enrolled in a pediatric nursing course were randomly divided into two groups: an experimental group that utilized ChatGPT and a control group that used traditional textbooks. The Mann-Whitney U test was employed to assess differences between the groups in two primary outcomes: ( a ) ethical standards, focusing on the understanding and applying ethical principles, and ( b ) nursing processes, emphasizing critical thinking skills and integrating evidence-based knowledge. The control group outperformed the experimental group in ethical standards and demonstrated better clinical reasoning in nursing processes. Reflective essays revealed that the experimental group reported lower reliability but higher time efficiency. Despite artificial intelligence's ability to offer diverse perspectives, the findings highlight that educators must supplement artificial intelligence technology with strategies that enhance critical thinking, careful data selection, and source verification. This study suggests a hybrid educational approach combining artificial intelligence with traditional learning methods to bolster nursing students' decision-making processes and clinical reasoning skills.
Collapse
|
26
|
Prakash K, Prakash R. An artificial intelligence-based dental semantic search engine as a reliable tool for dental students and educators. J Dent Educ 2024; 88:1257-1266. [PMID: 38715215 DOI: 10.1002/jdd.13560] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 03/13/2024] [Accepted: 03/29/2024] [Indexed: 11/18/2024]
Abstract
PURPOSE/OBJECTIVES This study proposes the utilization of a Natural Language Processing tool to create a semantic search engine for dental education while addressing the increasing concerns of accuracy, bias, and hallucination in outputs generated by AI tools. The paper focuses on developing and evaluating DentQA, a specialized question-answering tool that makes it easy for students to seek information to access information located in handouts or study material distributed by an institution. METHODS DentQA is structured upon the GPT3.5 language model, utilizing prompt engineering to extract information from external dental documents that experts have verified. Evaluation involves non-human metrics (BLEU scores) and human metrics for the tool's performance, relevance, accuracy, and functionality. RESULTS Non-human metrics confirm DentQA's linguistic proficiency, achieving a Unigram BLEU score of 0.85. Human metrics reveal DentQA's superiority over GPT3.5 in terms of accuracy (p = 0.00004) and absence of hallucination (p = 0.026). Additional metrics confirmed consistent performance across different question types (X2 (4, N = 200) = 13.0378, p = 0.012). User satisfaction and performance metrics support DentQA's usability and effectiveness, with a response time of 3.5 s and over 70% satisfaction across all evaluated parameters. CONCLUSIONS The study advocates using a semantic search engine in dental education, mitigating concerns of misinformation and hallucination. By outlining the workflow and the utilization of open-source tools and methods, the study encourages the utilization of similar tools for dental education while underscoring the importance of customizing AI models for dentistry. Further optimizations, testing, and utilization of recent advances can contribute to dental education significantly.
Collapse
Affiliation(s)
- Krishna Prakash
- Anil Neerukonda Institute of Dental Sciences, Visakhapatnam, Andhra Pradesh, India
| | - Ram Prakash
- Anil Neerukonda Institute of Dental Sciences, Visakhapatnam, Andhra Pradesh, India
| |
Collapse
|
27
|
De Gagne JC, Hwang H, Jung D. Cyberethics in nursing education: Ethical implications of artificial intelligence. Nurs Ethics 2024; 31:1021-1030. [PMID: 37803810 DOI: 10.1177/09697330231201901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/08/2023]
Abstract
As the use of artificial intelligence (AI) technologies, particularly generative AI (Gen AI), becomes increasingly prevalent in nursing education, it is paramount to address the ethical implications of their implementation. This article explores the realm of cyberethics (a field of applied ethics that focuses on the ethical, legal, and social implications of cybertechnology), highlighting the ethical principles of autonomy, nonmaleficence, beneficence, justice, and explicability as a roadmap for facilitating AI integration into nursing education. Research findings suggest that ethical dilemmas that challenge these five principles can emerge within the context of nursing education; however, adherence to these very principles, which is essential to improving patient care, can offer solutions to these dilemmas. To ensure the ethical and responsible use of Gen AI in nursing education, these principles must be woven into the fabric of curricula, and appropriate guidelines must be developed. Nurse educators have a pivotal role in strategizing comprehensive approaches for ethical AI integration, establishing clear guidelines, and instilling critical thinking among students. Fostering lifelong learning and adaptability is key to ensuring that future nurses can successfully navigate the constantly evolving landscape of health care technology. Future research should investigate the long-term impacts of AI utilization on learning outcomes and ethical decision-making.
Collapse
|
28
|
Zhang J, Sun K, Jagadeesh A, Falakaflaki P, Kayayan E, Tao G, Haghighat Ghahfarokhi M, Gupta D, Gupta A, Gupta V, Guo Y. The potential and pitfalls of using a large language model such as ChatGPT, GPT-4, or LLaMA as a clinical assistant. J Am Med Inform Assoc 2024; 31:1884-1891. [PMID: 39018498 PMCID: PMC11339517 DOI: 10.1093/jamia/ocae184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 06/29/2024] [Accepted: 07/08/2024] [Indexed: 07/19/2024] Open
Abstract
OBJECTIVES This study aims to evaluate the utility of large language models (LLMs) in healthcare, focusing on their applications in enhancing patient care through improved diagnostic, decision-making processes, and as ancillary tools for healthcare professionals. MATERIALS AND METHODS We evaluated ChatGPT, GPT-4, and LLaMA in identifying patients with specific diseases using gold-labeled Electronic Health Records (EHRs) from the MIMIC-III database, covering three prevalent diseases-Chronic Obstructive Pulmonary Disease (COPD), Chronic Kidney Disease (CKD)-along with the rare condition, Primary Biliary Cirrhosis (PBC), and the hard-to-diagnose condition Cancer Cachexia. RESULTS In patient identification, GPT-4 had near similar or better performance compared to the corresponding disease-specific Machine Learning models (F1-score ≥ 85%) on COPD, CKD, and PBC. GPT-4 excelled in the PBC use case, achieving a 4.23% higher F1-score compared to disease-specific "Traditional Machine Learning" models. ChatGPT and LLaMA3 demonstrated lower performance than GPT-4 across all diseases and almost all metrics. Few-shot prompts also help ChatGPT, GPT-4, and LLaMA3 achieve higher precision and specificity but lower sensitivity and Negative Predictive Value. DISCUSSION The study highlights the potential and limitations of LLMs in healthcare. Issues with errors, explanatory limitations and ethical concerns like data privacy and model transparency suggest that these models would be supplementary tools in clinical settings. Future studies should improve training datasets and model designs for LLMs to gain better utility in healthcare. CONCLUSION The study shows that LLMs have the potential to assist clinicians for tasks such as patient identification but false positives and false negatives must be mitigated before LLMs are adequate for real-world clinical assistance.
Collapse
Affiliation(s)
- Jingqing Zhang
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
- Data Science Institute, Imperial College London, London, SW7 2AZ, United Kingdom
| | - Kai Sun
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
- Data Science Institute, Imperial College London, London, SW7 2AZ, United Kingdom
| | | | | | - Elena Kayayan
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
| | - Guanyu Tao
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
| | | | - Deepa Gupta
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
| | - Ashok Gupta
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
| | - Vibhor Gupta
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
| | - Yike Guo
- Pangaea Data Limited, London, SE1 7LY, United Kingdom
- Hong Kong University of Science and Technology, Hong Kong, China
| |
Collapse
|
29
|
Fatima A, Shafique MA, Alam K, Fadlalla Ahmed TK, Mustafa MS. ChatGPT in medicine: A cross-disciplinary systematic review of ChatGPT's (artificial intelligence) role in research, clinical practice, education, and patient interaction. Medicine (Baltimore) 2024; 103:e39250. [PMID: 39121303 PMCID: PMC11315549 DOI: 10.1097/md.0000000000039250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 07/19/2024] [Indexed: 08/11/2024] Open
Abstract
BACKGROUND ChatGPT, a powerful AI language model, has gained increasing prominence in medicine, offering potential applications in healthcare, clinical decision support, patient communication, and medical research. This systematic review aims to comprehensively assess the applications of ChatGPT in healthcare education, research, writing, patient communication, and practice while also delineating potential limitations and areas for improvement. METHOD Our comprehensive database search retrieved relevant papers from PubMed, Medline and Scopus. After the screening process, 83 studies met the inclusion criteria. This review includes original studies comprising case reports, analytical studies, and editorials with original findings. RESULT ChatGPT is useful for scientific research and academic writing, and assists with grammar, clarity, and coherence. This helps non-English speakers and improves accessibility by breaking down linguistic barriers. However, its limitations include probable inaccuracy and ethical issues, such as bias and plagiarism. ChatGPT streamlines workflows and offers diagnostic and educational potential in healthcare but exhibits biases and lacks emotional sensitivity. It is useful in inpatient communication, but requires up-to-date data and faces concerns about the accuracy of information and hallucinatory responses. CONCLUSION Given the potential for ChatGPT to transform healthcare education, research, and practice, it is essential to approach its adoption in these areas with caution due to its inherent limitations.
Collapse
Affiliation(s)
- Afia Fatima
- Department of Medicine, Jinnah Sindh Medical University, Karachi, Pakistan
| | | | - Khadija Alam
- Department of Medicine, Liaquat National Medical College, Karachi, Pakistan
| | | | | |
Collapse
|
30
|
Bumbach MD. The Use of AI Powered ChatGPT for Nursing Education. J Nurs Educ 2024; 63:564-567. [PMID: 38598788 DOI: 10.3928/01484834-20240318-04] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/12/2024]
Abstract
BACKGROUND In late 2022, an AI (artificial intelligence) application, ChatGPT (generative pre-trained transformer), was released free for public use. Although present use of AI applications are scant in nursing education, the easy access to ChatGPT will inevitably influence educational experiences for both educators and students. Nursing educators have an opportunity to leverage this new technology by understanding the functionality and limitations of ChatGPT. METHOD This article examines the framework and functionality of ChatGPT and considers a potential nursing education assignment using the AI powered ChatGPT. The AI application, ChatGPT, is reviewed within the context of health care and nursing education and a potential nursing assignment leveraging ChatGPT is considered. RESULTS Nursing educators will increase their knowledge about ChatGPT and consider a possible nursing curriculum assignment using ChatGPT. CONCLUSION Although not without limitations, nursing educators can leverage this new AI powered technology for an enhanced student experience. [J Nurs Educ. 2024;63(8):564-567.].
Collapse
|
31
|
Atwal K. Artificial intelligence in clinical nutrition and dietetics: A brief overview of current evidence. Nutr Clin Pract 2024; 39:736-742. [PMID: 38591653 DOI: 10.1002/ncp.11150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2024] [Revised: 03/07/2024] [Accepted: 03/13/2024] [Indexed: 04/10/2024] Open
Abstract
The rapid surge in artificial intelligence (AI) has dominated technological innovation in today's society. As experts begin to understand the potential, a spectrum of opportunities could yield a remarkable revolution. The upsurge in healthcare could transform clinical interventions and outcomes, but it risks dehumanization and increased unethical practices. The field of clinical nutrition and dietetics is no exception. This article finds a multitude of developments underway, which include the use of AI for malnutrition screening; predicting clinical outcomes, such as disease onset, and clinical risks, such as drug interactions; aiding interventions, such as estimating nutrient intake; applying precision nutrition, such as measuring postprandial glycemic response; and supporting workflow through chatbots trained on natural language models. Although the opportunity and scalability of AI is incalculably attractive, especially in the face of poor healthcare resources, the threat cannot be ignored. The risk of malpractice and lack of accountability are some of the main concerns. As such, the healthcare professional's responsibility remains paramount. The data used to train AI models could be biased, which could risk the quality of care to vulnerable or minority patient groups. Standardized AI-development protocols, benchmarked to care recommendations, with rigorous large-scale validation are required to maximize application among different settings. AI could overturn the healthcare landscape, and this article skims the surface of its potential in clinical nutrition and dietetics.
Collapse
Affiliation(s)
- Kiranjit Atwal
- Department of Nutritional Sciences, King's College London, London, UK
- School of Health Professions, University of Plymouth, Plymouth, UK
| |
Collapse
|
32
|
Zhui L, Yhap N, Liping L, Zhengjie W, Zhonghao X, Xiaoshu Y, Hong C, Xuexiu L, Wei R. Impact of Large Language Models on Medical Education and Teaching Adaptations. JMIR Med Inform 2024; 12:e55933. [PMID: 39087590 PMCID: PMC11294775 DOI: 10.2196/55933] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2023] [Revised: 04/25/2024] [Accepted: 06/08/2024] [Indexed: 08/02/2024] Open
Abstract
Unlabelled This viewpoint article explores the transformative role of large language models (LLMs) in the field of medical education, highlighting their potential to enhance teaching quality, promote personalized learning paths, strengthen clinical skills training, optimize teaching assessment processes, boost the efficiency of medical research, and support continuing medical education. However, the use of LLMs entails certain challenges, such as questions regarding the accuracy of information, the risk of overreliance on technology, a lack of emotional recognition capabilities, and concerns related to ethics, privacy, and data security. This article emphasizes that to maximize the potential of LLMs and overcome these challenges, educators must exhibit leadership in medical education, adjust their teaching strategies flexibly, cultivate students' critical thinking, and emphasize the importance of practical experience, thus ensuring that students can use LLMs correctly and effectively. By adopting such a comprehensive and balanced approach, educators can train health care professionals who are proficient in the use of advanced technologies and who exhibit solid professional ethics and practical skills, thus laying a strong foundation for these professionals to overcome future challenges in the health care sector.
Collapse
Affiliation(s)
- Li Zhui
- Department of Vascular Surgery, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| | - Nina Yhap
- Department of General Surgery, Queen Elizabeth Hospital, St Michael, Barbados
| | - Liu Liping
- Department of Ultrasound, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| | - Wang Zhengjie
- Department of Nuclear Medicine, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| | - Xiong Zhonghao
- Department of Acupuncture and Moxibustion, Chongqing Traditional Chinese Medicine Hospital, Chongqing, China
| | - Yuan Xiaoshu
- Department of Anesthesia, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| | - Cui Hong
- Department of Anesthesia, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| | - Liu Xuexiu
- Department of Neonatology, Children’s Hospital of Chongqing Medical University, Chongqing, China
| | - Ren Wei
- Department of Vascular Surgery, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
| |
Collapse
|
33
|
Yüce A, Yerli M, Misir A, Çakar M. Enhancing patient information texts in orthopaedics: How OpenAI's 'ChatGPT' can help. J Exp Orthop 2024; 11:e70019. [PMID: 39291057 PMCID: PMC11406043 DOI: 10.1002/jeo2.70019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 08/15/2024] [Accepted: 08/20/2024] [Indexed: 09/19/2024] Open
Abstract
Purpose The internet has become a primary source for patients seeking healthcare information, but the quality of online information, particularly in orthopaedics, often falls short. Orthopaedic surgeons now have the added responsibility of evaluating and guiding patients to credible online resources. This study aimed to assess ChatGPT's ability to identify deficiencies in patient information texts related to total hip arthroplasty websites and to evaluate its potential for enhancing the quality of these texts. Methods In August 2023, 25 websites related to total hip arthroplasty were assessed using a standardized search on Google. Peer-reviewed scientific articles, empty pages, dictionary definitions, and unrelated content were excluded. The remaining 10 websites were evaluated using the hip information scoring system (HISS). ChatGPT was then used to assess these texts, identify deficiencies and provide recommendations. Results The mean HISS score of the websites was 9.5, indicating low to moderate quality. However, after implementing ChatGPT's suggested improvements, the score increased to 21.5, signifying excellent quality. ChatGPT's recommendations included using simpler language, adding FAQs, incorporating patient experiences, addressing cost and insurance issues, detailing preoperative and postoperative phases, including references, and emphasizing emotional and psychological support. The study demonstrates that ChatGPT can significantly enhance patient information quality. Conclusion ChatGPT's role in elevating patient education regarding total hip arthroplasty is promising. This study sheds light on the potential of ChatGPT as an aid to orthopaedic surgeons in producing high-quality patient information materials. Although it cannot replace human expertise, it offers a valuable means of enhancing the quality of healthcare information available online. Level of Evidence Level IV.
Collapse
Affiliation(s)
- Ali Yüce
- Department of Orthopedic and Traumatology Prof. Dr. Cemil Taşcıoğlu City Hospital İstanbul Turkey
| | - Mustafa Yerli
- Department of Orthopedic and Traumatology Prof. Dr. Cemil Taşcıoğlu City Hospital İstanbul Turkey
| | - Abdulhamit Misir
- Department of Orthopedic and Traumatology Göztepe Medical Park Hospital İstanbul Turkey
| | - Murat Çakar
- Department of Orthopedic and Traumatology Prof. Dr. Cemil Taşcıoğlu City Hospital İstanbul Turkey
| |
Collapse
|
34
|
Rodrigues Alessi M, Gomes HA, Lopes de Castro M, Terumy Okamoto C. Performance of ChatGPT in Solving Questions From the Progress Test (Brazilian National Medical Exam): A Potential Artificial Intelligence Tool in Medical Practice. Cureus 2024; 16:e64924. [PMID: 39156244 PMCID: PMC11330648 DOI: 10.7759/cureus.64924] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/19/2024] [Indexed: 08/20/2024] Open
Abstract
Background The use of artificial intelligence (AI) is not a recent phenomenon, but the latest advancements in this technology are making a significant impact across various fields of human knowledge. In medicine, this trend is no different, although it has developed at a slower pace. ChatGPT is an example of an AI-based algorithm capable of answering questions, interpreting phrases, and synthesizing complex information, potentially aiding and even replacing humans in various areas of social interest. Some studies have compared its performance in solving medical knowledge exams with medical students and professionals to verify AI accuracy. This study aimed to measure the performance of ChatGPT in answering questions from the Progress Test from 2021 to 2023. Methodology An observational study was conducted in which questions from the 2021 Progress Test and the regional tests (Southern Institutional Pedagogical Support Center II) of 2022 and 2023 were presented to ChatGPT 3.5. The results obtained were compared with the scores of first- to sixth-year medical students from over 120 Brazilian universities. All questions were presented sequentially, without any modification to their structure. After each question was presented, the platform's history was cleared, and the site was restarted. Results The platform achieved an average accuracy rate in 2021, 2022, and 2023 of 69.7%, 68.3%, and 67.2%, respectively, surpassing students from all medical years in the three tests evaluated, reinforcing findings in the current literature. The subject with the best score for the AI was Public Health, with a mean grade of 77.8%. Conclusions ChatGPT demonstrated the ability to answer medical questions with higher accuracy than humans, including students from the last year of medical school.
Collapse
Affiliation(s)
| | - Heitor A Gomes
- School of Medicine, Universidade Positivo, Curitiba, BRA
| | | | | |
Collapse
|
35
|
Tessler I, Wolfovitz A, Alon EE, Gecel NA, Livneh N, Zimlichman E, Klang E. ChatGPT's adherence to otolaryngology clinical practice guidelines. Eur Arch Otorhinolaryngol 2024; 281:3829-3834. [PMID: 38647684 DOI: 10.1007/s00405-024-08634-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/01/2024] [Accepted: 03/22/2024] [Indexed: 04/25/2024]
Abstract
OBJECTIVES Large language models, including ChatGPT, has the potential to transform the way we approach medical knowledge, yet accuracy in clinical topics is critical. Here we assessed ChatGPT's performance in adhering to the American Academy of Otolaryngology-Head and Neck Surgery guidelines. METHODS We presented ChatGPT with 24 clinical otolaryngology questions based on the guidelines of the American Academy of Otolaryngology. This was done three times (N = 72) to test the model's consistency. Two otolaryngologists evaluated the responses for accuracy and relevance to the guidelines. Cohen's Kappa was used to measure evaluator agreement, and Cronbach's alpha assessed the consistency of ChatGPT's responses. RESULTS The study revealed mixed results; 59.7% (43/72) of ChatGPT's responses were highly accurate, while only 2.8% (2/72) directly contradicted the guidelines. The model showed 100% accuracy in Head and Neck, but lower accuracy in Rhinology and Otology/Neurotology (66%), Laryngology (50%), and Pediatrics (8%). The model's responses were consistent in 17/24 (70.8%), with a Cronbach's alpha value of 0.87, indicating a reasonable consistency across tests. CONCLUSIONS Using a guideline-based set of structured questions, ChatGPT demonstrates consistency but variable accuracy in otolaryngology. Its lower performance in some areas, especially Pediatrics, suggests that further rigorous evaluation is needed before considering real-world clinical use.
Collapse
Affiliation(s)
- Idit Tessler
- Department of Otolaryngology and Head and Neck Surgery, Sheba Medical Center, Ramat Gan, Israel.
- School of Medicine, Tel Aviv University, Tel Aviv, Israel.
- ARC Innovation Center, Sheba Medical Center, Ramat Gan, Israel.
| | - Amit Wolfovitz
- Department of Otolaryngology and Head and Neck Surgery, Sheba Medical Center, Ramat Gan, Israel
- School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Eran E Alon
- Department of Otolaryngology and Head and Neck Surgery, Sheba Medical Center, Ramat Gan, Israel
- School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Nir A Gecel
- School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Nir Livneh
- Department of Otolaryngology and Head and Neck Surgery, Sheba Medical Center, Ramat Gan, Israel
- School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Eyal Zimlichman
- School of Medicine, Tel Aviv University, Tel Aviv, Israel
- ARC Innovation Center, Sheba Medical Center, Ramat Gan, Israel
- The Sheba Talpiot Medical Leadership Program, Ramat Gan, Israel
- Hospital Management, Sheba Medical Center, Ramat Gan, Israel
| | - Eyal Klang
- The Division of Data-Driven and Digital Medicine (D3M), Icahn School of Medicine at Mount Sinai, New York, USA
| |
Collapse
|
36
|
Xu R, Wang Z. Generative artificial intelligence in healthcare from the perspective of digital media: Applications, opportunities and challenges. Heliyon 2024; 10:e32364. [PMID: 38975200 PMCID: PMC11225727 DOI: 10.1016/j.heliyon.2024.e32364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Accepted: 06/03/2024] [Indexed: 07/09/2024] Open
Abstract
Introduction The emergence and application of generative artificial intelligence/large language models (hereafter GenAI LLMs) have the potential for significant impact on the healthcare industry. However, there is currently a lack of systematic research on GenAI LLMs in healthcare based on reliable data. This article aims to conduct an exploratory study of the application of GenAI LLMs (i.e., ChatGPT) in healthcare from the perspective of digital media (i.e., online news), including the application scenarios, potential opportunities, and challenges. Methods This research used thematic qualitative text analysis in five steps: firstly, developing main topical categories based on relevant articles; secondly, encoding the search keywords using these categories; thirdly, conducting searches for news articles via Google ; fourthly, encoding the sub-categories using the elaborate category system; and finally, conducting category-based analysis and presenting the results. Natural language processing techniques, including the TermRaider and AntConc tool, were applied in the aforementioned steps to assist in text qualitative analysis. Additionally, this study built a framework, using for analyzing the above three topics, from the perspective of five different stakeholders, including healthcare demanders and providers. Results This study summarizes 26 applications (e.g., provide medical advice, provide diagnosis and triage recommendations, provide mental health support, etc.), 21 opportunities (e.g., make healthcare more accessible, reduce healthcare costs, improve patients care, etc.), and 17 challenges (e.g., generate inaccurate/misleading/wrong answers, raise privacy concerns, lack of transparency, etc.), and analyzes the reasons for the formation of these key items and the links between the three research topics. Conclusions The application of GenAI LLMs in healthcare is primarily focused on transforming the way healthcare demanders access medical services (i.e., making it more intelligent, refined, and humane) and optimizing the processes through which healthcare providers offer medical services (i.e., simplifying, ensuring timeliness, and reducing errors). As the application becomes more widespread and deepens, GenAI LLMs is expected to have a revolutionary impact on traditional healthcare service models, but it also inevitably raises ethical and security concerns. Furthermore, GenAI LLMs applied in healthcare is still in the initial stage, which can be accelerated from a specific healthcare field (e.g., mental health) or a specific mechanism (e.g., GenAI LLMs' economic benefits allocation mechanism applied to healthcare) with empirical or clinical research.
Collapse
Affiliation(s)
- Rui Xu
- School of Economics, Guangdong University of Technology, Guangzhou, China
| | - Zhong Wang
- School of Economics, Guangdong University of Technology, Guangzhou, China
- Key Laboratory of Digital Economy and Data Governance, Guangdong University of Technology, Guangzhou, China
| |
Collapse
|
37
|
Lukkahatai N, Han G. Perspectives on Artificial Intelligence in Nursing in Asia. Asian Pac Isl Nurs J 2024; 8:e55321. [PMID: 38896473 PMCID: PMC11222764 DOI: 10.2196/55321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 02/22/2024] [Accepted: 05/22/2024] [Indexed: 06/21/2024] Open
Abstract
Artificial intelligence (AI) is reshaping health care, including nursing, across Asia, presenting opportunities to improve patient care and outcomes. This viewpoint presents our perspective and interpretation of the current AI landscape, acknowledging its evolution driven by enhanced processing capabilities, extensive data sets, and refined algorithms. Notable applications in countries such as Singapore, South Korea, Japan, and China showcase the integration of AI-powered technologies such as chatbots, virtual assistants, data mining, and automated risk assessment systems. This paper further explores the transformative impact of AI on nursing education, emphasizing personalized learning, adaptive approaches, and AI-enriched simulation tools, and discusses the opportunities and challenges of these developments. We argue for the harmonious coexistence of traditional nursing values with AI innovations, marking a significant stride toward a promising health care future in Asia.
Collapse
Affiliation(s)
- Nada Lukkahatai
- School of Nursing, Johns Hopkins University, Baltimore, MD, United States
| | - Gyumin Han
- School of Nursing, Johns Hopkins University, Baltimore, MD, United States
- College of Nursing, Research Institute of Nursing Science, Pusan National University, Busan, Republic of Korea
| |
Collapse
|
38
|
Costa ICP, do Nascimento MC, Treviso P, Chini LT, Roza BDA, Barbosa SDFF, Mendes KDS. Using the Chat Generative Pre-trained Transformer in academic writing in health: a scoping review. Rev Lat Am Enfermagem 2024; 32:e4194. [PMID: 38922265 PMCID: PMC11182606 DOI: 10.1590/1518-8345.7133.4194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Accepted: 02/04/2024] [Indexed: 06/27/2024] Open
Abstract
OBJECTIVE to map the scientific literature regarding the use of the Chat Generative Pre-trained Transformer, ChatGPT, in academic writing in health. METHOD this was a scoping review, following the JBI methodology. Conventional databases and gray literature were included. The selection of studies was applied after removing duplicates and individual and paired evaluation. Data were extracted based on an elaborate script, and presented in a descriptive, tabular and graphical format. RESULTS the analysis of the 49 selected articles revealed that ChatGPT is a versatile tool, contributing to scientific production, description of medical procedures and preparation of summaries aligned with the standards of scientific journals. Its application has been shown to improve the clarity of writing and benefits areas such as innovation and automation. Risks were also observed, such as the possibility of lack of originality and ethical issues. Future perspectives highlight the need for adequate regulation, agile adaptation and the search for an ethical balance in incorporating ChatGPT into academic writing. CONCLUSION ChatGPT presents transformative potential in academic writing in health. However, its adoption requires rigorous human supervision, solid regulation, and transparent guidelines to ensure its responsible and beneficial use by the scientific community.
Collapse
Affiliation(s)
| | | | - Patrícia Treviso
- Universidade do Vale do Rio dos Sinos, Escola de Saúde, São Leopoldo, RS, Brazil
| | | | | | | | - Karina Dal Sasso Mendes
- Universidade de São Paulo, Escola de Enfermagem de Ribeirão Preto, PAHO/WHO Collaborating Centre for Nursing Research Development, Ribeirão Preto, SP, Brazil
| |
Collapse
|
39
|
Zhang F, Liu X, Wu W, Zhu S. Evolution of Chatbots in Nursing Education: Narrative Review. JMIR MEDICAL EDUCATION 2024; 10:e54987. [PMID: 38889074 PMCID: PMC11186796 DOI: 10.2196/54987] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Revised: 05/16/2024] [Accepted: 05/22/2024] [Indexed: 06/20/2024]
Abstract
Background The integration of chatbots in nursing education is a rapidly evolving area with potential transformative impacts. This narrative review aims to synthesize and analyze the existing literature on chatbots in nursing education. Objective This study aims to comprehensively examine the temporal trends, international distribution, study designs, and implications of chatbots in nursing education. Methods A comprehensive search was conducted across 3 databases (PubMed, Web of Science, and Embase) following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram. Results A total of 40 articles met the eligibility criteria, with a notable increase of publications in 2023 (n=28, 70%). Temporal analysis revealed a notable surge in publications from 2021 to 2023, emphasizing the growing scholarly interest. Geographically, Taiwan province made substantial contributions (n=8, 20%), followed by the United States (n=6, 15%) and South Korea (n=4, 10%). Study designs varied, with reviews (n=8, 20%) and editorials (n=7, 18%) being predominant, showcasing the richness of research in this domain. Conclusions Integrating chatbots into nursing education presents a promising yet relatively unexplored avenue. This review highlights the urgent need for original research, emphasizing the importance of ethical considerations.
Collapse
Affiliation(s)
- Fang Zhang
- Department of Science and Education, Shenzhen Baoan Women's and Children's Hospital, Shenzhen, China
| | - Xiaoliu Liu
- Medical Laboratory of Shenzhen Luohu People’s Hospital, Shenzhen, China
| | - Wenyan Wu
- Medical Laboratory of Shenzhen Luohu People’s Hospital, Shenzhen, China
| | - Shiben Zhu
- School of Nursing and Health Studies, Hong Kong Metropolitan University, Hong Kong, China
| |
Collapse
|
40
|
Srinivasan M, Venugopal A, Venkatesan L, Kumar R. Navigating the Pedagogical Landscape: Exploring the Implications of AI and Chatbots in Nursing Education. JMIR Nurs 2024; 7:e52105. [PMID: 38870516 PMCID: PMC11211702 DOI: 10.2196/52105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Revised: 11/30/2023] [Accepted: 12/12/2023] [Indexed: 06/15/2024] Open
Abstract
This viewpoint paper explores the pedagogical implications of artificial intelligence (AI) and AI-based chatbots such as ChatGPT in nursing education, examining their potential uses, benefits, challenges, and ethical considerations. AI and chatbots offer transformative opportunities for nursing education, such as personalized learning, simulation and practice, accessible learning, and improved efficiency. They have the potential to increase student engagement and motivation, enhance learning outcomes, and augment teacher support. However, the integration of these technologies also raises ethical considerations, such as privacy, confidentiality, and bias. The viewpoint paper provides a comprehensive overview of the current state of AI and chatbots in nursing education, offering insights into best practices and guidelines for their integration. By examining the impact of AI and ChatGPT on student learning, engagement, and teacher effectiveness and efficiency, this review aims to contribute to the ongoing discussion on the use of AI and chatbots in nursing education and provide recommendations for future research and development in the field.
Collapse
Affiliation(s)
| | - Ambili Venugopal
- College of Nursing, All India Institute of Medical Sciences, Mangalagiri, India
| | - Latha Venkatesan
- College of Nursing, All India Institute of Medical Sciences, New Delhi, India
| | - Rajesh Kumar
- College of Nursing, All India Institute of Medical Sciences, Rishikesh, India
| |
Collapse
|
41
|
Rao SJ, Isath A, Krishnan P, Tangsrivimol JA, Virk HUH, Wang Z, Glicksberg BS, Krittanawong C. ChatGPT: A Conceptual Review of Applications and Utility in the Field of Medicine. J Med Syst 2024; 48:59. [PMID: 38836893 DOI: 10.1007/s10916-024-02075-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Accepted: 05/07/2024] [Indexed: 06/06/2024]
Abstract
Artificial Intelligence, specifically advanced language models such as ChatGPT, have the potential to revolutionize various aspects of healthcare, medical education, and research. In this narrative review, we evaluate the myriad applications of ChatGPT in diverse healthcare domains. We discuss its potential role in clinical decision-making, exploring how it can assist physicians by providing rapid, data-driven insights for diagnosis and treatment. We review the benefits of ChatGPT in personalized patient care, particularly in geriatric care, medication management, weight loss and nutrition, and physical activity guidance. We further delve into its potential to enhance medical research, through the analysis of large datasets, and the development of novel methodologies. In the realm of medical education, we investigate the utility of ChatGPT as an information retrieval tool and personalized learning resource for medical students and professionals. There are numerous promising applications of ChatGPT that will likely induce paradigm shifts in healthcare practice, education, and research. The use of ChatGPT may come with several benefits in areas such as clinical decision making, geriatric care, medication management, weight loss and nutrition, physical fitness, scientific research, and medical education. Nevertheless, it is important to note that issues surrounding ethics, data privacy, transparency, inaccuracy, and inadequacy persist. Prior to widespread use in medicine, it is imperative to objectively evaluate the impact of ChatGPT in a real-world setting using a risk-based approach.
Collapse
Affiliation(s)
- Shiavax J Rao
- Department of Medicine, MedStar Union Memorial Hospital, Baltimore, MD, USA
| | - Ameesh Isath
- Department of Cardiology, Westchester Medical Center and New York Medical College, Valhalla, NY, USA
| | - Parvathy Krishnan
- Department of Pediatrics, Westchester Medical Center and New York Medical College, Valhalla, NY, USA
| | - Jonathan A Tangsrivimol
- Division of Neurosurgery, Department of Surgery, Chulabhorn Hospital, Chulabhorn Royal Academy, Bangkok, 10210, Thailand
- Department of Neurological Surgery, Weill Cornell Medicine Brain and Spine Center, New York, NY, 10022, USA
| | - Hafeez Ul Hassan Virk
- Harrington Heart & Vascular Institute, Case Western Reserve University, University Hospitals Cleveland Medical Center, Cleveland, OH, USA
| | - Zhen Wang
- Robert D. and Patricia E. Kern Center for the Science of Health Care Delivery, Mayo Clinic, Rochester, MN, USA
- Division of Health Care Policy and Research, Department of Health Sciences Research, Mayo Clinic, Rochester, MN, USA
| | - Benjamin S Glicksberg
- Hasso Plattner Institute for Digital Health, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Chayakrit Krittanawong
- Cardiology Division, NYU Langone Health and NYU School of Medicine, 550 First Avenue, New York, NY, 10016, USA.
| |
Collapse
|
42
|
Nachalon Y, Broer M, Nativ-Zeltzer N. Using ChatGPT to Generate Research Ideas in Dysphagia: A Pilot Study. Dysphagia 2024; 39:407-411. [PMID: 37907728 DOI: 10.1007/s00455-023-10623-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Accepted: 09/20/2023] [Indexed: 11/02/2023]
Abstract
Current research in dysphagia faces challenges due to the rapid growth of scientific literature and the interdisciplinary nature of the field. To address this, the study evaluates ChatGPT, an AI language model, as a supplementary resource to assist clinicians and researchers in generating research ideas for dysphagia, utilizing recent advancements in natural language processing and machine learning. The research ideas were generated through ChatGPT's command to explore diverse aspects of dysphagia. A web-based survey was conducted, 45 dysphagia experts were asked to rank each study on a scale of 1 to 5 according to feasibility, novelty, clinical implications, and relevance to current practice. A total of 26 experts (58%) completed the survey. The mean (± sd) rankings of research ideas were 4.03 (± 0.17) for feasibility, 3.5 (± 0.17) for potential impact on the field, 3.84 (± 0.12) for clinical relevance, and 3.08 (± 0.36) for novelty and innovation. Results of this study suggest that ChatGPT offers a promising approach to generating research ideas in dysphagia. While its current capability to generate innovative ideas appears limited, it can serve as a supplementary resource for researchers.
Collapse
Affiliation(s)
- Yuval Nachalon
- Department of Otolaryngology, Head and Neck Surgery and Maxillofacial Surgery, Tel-Aviv Sourasky Medical Center, 6 Weizman Street, 6423906, Tel-Aviv, Israel.
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel.
| | - Maya Broer
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | | |
Collapse
|
43
|
Tan S, Xin X, Wu D. ChatGPT in medicine: prospects and challenges: a review article. Int J Surg 2024; 110:3701-3706. [PMID: 38502861 PMCID: PMC11175750 DOI: 10.1097/js9.0000000000001312] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Accepted: 02/26/2024] [Indexed: 03/21/2024]
Abstract
It has been a year since the launch of Chat Generator Pre-Trained Transformer (ChatGPT), a generative artificial intelligence (AI) program. The introduction of this cross-generational product initially brought a huge shock to people with its incredible potential and then aroused increasing concerns among people. In the field of medicine, researchers have extensively explored the possible applications of ChatGPT and achieved numerous satisfactory results. However, opportunities and issues always come together. Problems have also been exposed during the applications of ChatGPT, requiring cautious handling, thorough consideration, and further guidelines for safe use. Here, the authors summarized the potential applications of ChatGPT in the medical field, including revolutionizing healthcare consultation, assisting patient management and treatment, transforming medical education, and facilitating clinical research. Meanwhile, the authors also enumerated researchers' concerns arising along with its broad and satisfactory applications. As it is irreversible that AI will gradually permeate every aspect of modern life, the authors hope that this review can not only promote people's understanding of the potential applications of ChatGPT in the future but also remind them to be more cautious about this "Pandora's Box" in the medical field. It is necessary to establish normative guidelines for its safe use in the medical field as soon as possible.
Collapse
Affiliation(s)
| | | | - Di Wu
- Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shijingshan, Beijing, China
| |
Collapse
|
44
|
Morya VK, Lee HW, Shahid H, Magar AG, Lee JH, Kim JH, Jun L, Noh KC. Application of ChatGPT for Orthopedic Surgeries and Patient Care. Clin Orthop Surg 2024; 16:347-356. [PMID: 38827766 PMCID: PMC11130626 DOI: 10.4055/cios23181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/02/2023] [Revised: 11/15/2023] [Accepted: 12/12/2023] [Indexed: 06/05/2024] Open
Abstract
Artificial intelligence (AI) has rapidly transformed various aspects of life, and the launch of the chatbot "ChatGPT" by OpenAI in November 2022 has garnered significant attention and user appreciation. ChatGPT utilizes natural language processing based on a "generative pre-trained transfer" (GPT) model, specifically the transformer architecture, to generate human-like responses to a wide range of questions and topics. Equipped with approximately 57 billion words and 175 billion parameters from online data, ChatGPT has potential applications in medicine and orthopedics. One of its key strengths is its personalized, easy-to-understand, and adaptive response, which allows it to learn continuously through user interaction. This article discusses how AI, especially ChatGPT, presents numerous opportunities in orthopedics, ranging from preoperative planning and surgical techniques to patient education and medical support. Although ChatGPT's user-friendly responses and adaptive capabilities are laudable, its limitations, including biased responses and ethical concerns, necessitate its cautious and responsible use. Surgeons and healthcare providers should leverage the strengths of the ChatGPT while recognizing its current limitations and verifying critical information through independent research and expert opinions. As AI technology continues to evolve, ChatGPT may become a valuable tool in orthopedic education and patient care, leading to improved outcomes and efficiency in healthcare delivery. The integration of AI into orthopedics offers substantial benefits but requires careful consideration and continuous improvement.
Collapse
Affiliation(s)
- Vivek Kumar Morya
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Ho-Won Lee
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Hamzah Shahid
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Anuja Gajanan Magar
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Ju-Hyung Lee
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Jae-Hyung Kim
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Lang Jun
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Kyu-Cheol Noh
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| |
Collapse
|
45
|
Zhang X, Ma L. Predictive Value of the Total Bilirubin and CA50 Screened Based on Machine Learning for Recurrence of Bladder Cancer Patients. Cancer Manag Res 2024; 16:537-546. [PMID: 38835478 PMCID: PMC11149634 DOI: 10.2147/cmar.s457269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2023] [Accepted: 05/27/2024] [Indexed: 06/06/2024] Open
Abstract
Purpose Recurrence is the main factor for poor prognosis of bladder cancer. Therefore, it is necessary to develop new biomarkers to predict the prognosis of bladder cancer. In this study, we used machine learning (ML) methods based on a variety of clinical variables to screen prognostic biomarkers of bladder cancer. Patients and Methods A total of 345 bladder cancer patients were participated in this retrospective study and randomly divided into training and testing group. We used five supervised clustering ML algorithms: decision tree (DT), random forest (RF), adaptive boosting (AdaBoost), gradient boosting machine (GBM), and extreme gradient boosting (XGBoost) to obtained prediction information through 34 clinical parameters. Results By comparing five ML algorithms, we found that total bilirubin (TBIL) and CA50 had the best performance in predicting the recurrence of bladder cancer. In addition, the combined predictive performance of the two is superior to the performance of any single indicator prediction. Conclusion ML technology can evaluate the recurrence of bladder cancer. This study shows that the combination of TBIL and CA50 can improve the prognosis prediction of bladder cancer recurrence, which can help clinicians make decisions and develop personalized treatment strategies.
Collapse
Affiliation(s)
- Xiaosong Zhang
- Department of Urology, Affiliated Hospital of Nantong University, Medical School of Nantong University, Nantong, 226001, People's Republic of China
- Department of Urology, Nantong Tongzhou District People's Hospital, Nantong, 226300, People's Republic of China
| | - Limin Ma
- Department of Urology, Affiliated Hospital of Nantong University, Medical School of Nantong University, Nantong, 226001, People's Republic of China
| |
Collapse
|
46
|
Xu Z, Fang Q, Huang Y, Xie M. The public attitude towards ChatGPT on reddit: A study based on unsupervised learning from sentiment analysis and topic modeling. PLoS One 2024; 19:e0302502. [PMID: 38743773 PMCID: PMC11093324 DOI: 10.1371/journal.pone.0302502] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 04/07/2024] [Indexed: 05/16/2024] Open
Abstract
ChatGPT has demonstrated impressive abilities and impacted various aspects of human society since its creation, gaining widespread attention from different social spheres. This study aims to comprehensively assess public perception of ChatGPT on Reddit. The dataset was collected via Reddit, a social media platform, and includes 23,733 posts and comments related to ChatGPT. Firstly, to examine public attitudes, this study conducts content analysis utilizing topic modeling with the Latent Dirichlet Allocation (LDA) algorithm to extract pertinent topics. Furthermore, sentiment analysis categorizes user posts and comments as positive, negative, or neutral using Textblob and Vader in natural language processing. The result of topic modeling shows that seven topics regarding ChatGPT are identified, which can be grouped into three themes: user perception, technical methods, and impacts on society. Results from the sentiment analysis show that 61.6% of the posts and comments hold favorable opinions on ChatGPT. They emphasize ChatGPT's ability to prompt and engage in natural conversations with users, without relying on complex natural language processing. It provides suggestions for ChatGPT developers to enhance its usability design and functionality. Meanwhile, stakeholders, including users, should comprehend the advantages and disadvantages of ChatGPT in human society to promote ethical and regulated implementation of the system.
Collapse
Affiliation(s)
- Zhaoxiang Xu
- Department of Data Science, School of Computer Science and Engineering, Guangzhou Institute of Science and Technology, Guangzhou, Guangdong, China
| | - Qingguo Fang
- Department of Management, School of Business, Macau University of Science and Technology, Macao, China
| | - Yanbo Huang
- Data Science Research Center, Faculty of Innovation Engineering, Macau University of Science and Technology, Macao, China
| | - Mingjian Xie
- Department of Decision Sciences, School of Business, Macau University of Science and Technology, Macao, China
| |
Collapse
|
47
|
Yalcinkaya T, Cinar Yucel S. Bibliometric and content analysis of ChatGPT research in nursing education: The rabbit hole in nursing education. Nurse Educ Pract 2024; 77:103956. [PMID: 38653086 DOI: 10.1016/j.nepr.2024.103956] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2023] [Revised: 03/16/2024] [Accepted: 03/31/2024] [Indexed: 04/25/2024]
Abstract
AIM This study was conducted to perform the bibliometric and content analysis of ChatGPT studies in nursing education. BACKGROUND ChatGPT is an artificial intelligence-based chatbot developed by OpenAI. The benefits and limitations of the use of ChatGPT in nursing education are still discussed; however, it is a tool having potential to be used in nursing education. DESIGN Bibliometric and content analysis. METHODS The study data were scanned through Scopus and Web of Science. Bibliometric analysis was carried out with VOSViewer and Bibliometrix software. In the bibliometric analysis, science mapping and performance analysis techniques were used. Various bibliometric data, including most cited publications, journals and countries, were analyzed and visualized. The synthetic knowledge synthesis method was used in content analysis. RESULTS We analyzed 53 publications to which 151 authors contributed. The publications had been published in 29 different journals. The average number of citations of publications is 8.2. It was determined that most of the articles were published in Nurse Education Today and Nurse Educator journals and that the leading countries were the USA and Canada. It was observed that international cooperation on the issue was weak. The most frequently mentioned keywords in the publications were "ChatGPT", "artificial intelligence" and "nursing". The following three themes emerged after the content analysis: (1) Integration of ChatGPT into nursing education; (2) Potential benefits and limitations of ChatGPT; and (3) Stepping down the rabbit hole. CONCLUSIONS We expect that the results of the study can give nursing faculties and academics ideas about the current status of ChatGPT in nursing education and enable them to make inferences for the future.
Collapse
Affiliation(s)
- Turgay Yalcinkaya
- Faculty of Health Sciences, Department of Nursing, Sinop University, Sinop, Turkey.
| | - Sebnem Cinar Yucel
- Nursing Faculty, Department of Fundamentals of Nursing, Ege University, Izmir, Turkey
| |
Collapse
|
48
|
Rosen S, Saban M. Evaluating the reliability of ChatGPT as a tool for imaging test referral: a comparative study with a clinical decision support system. Eur Radiol 2024; 34:2826-2837. [PMID: 37828297 DOI: 10.1007/s00330-023-10230-0] [Citation(s) in RCA: 14] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Revised: 07/28/2023] [Accepted: 08/01/2023] [Indexed: 10/14/2023]
Abstract
OBJECTIVES As the technology continues to evolve and advance, we can expect to see artificial intelligence (AI) being used in increasingly sophisticated ways to make a diagnosis and decisions such as suggesting the most appropriate imaging referrals. We aim to explore whether Chat Generative Pretrained Transformer (ChatGPT) can provide accurate imaging referrals for clinical use that are at least as good as the ESR iGuide. METHODS A comparative study was conducted in a tertiary hospital. Data was collected from 97 consecutive cases that were admitted to the emergency department with abdominal complaints. We compared the imaging test referral recommendations suggested by the ESR iGuide and the ChatGPT and analyzed cases of disagreement. In addition, we selected cases where ChatGPT recommended a chest abdominal pelvis (CAP) CT (n = 66), and asked four specialists to grade the appropriateness of the referral. RESULTS ChatGPT recommendations were consistent with the recommendations provided by the ESR iGuide. No statistical differences were found between the appropriateness of referrals by age or gender. Using a sub-analysis of CAP cases, a high agreement between ChatGPT and the specialists was found. Cases of disagreement (12.4%) were further analyzed and presented themes of vague recommendations such as "it would be advisable" and "this would help to rule out." CONCLUSIONS ChatGPT's ability to guide the selection of appropriate tests may be comparable to some degree with the ESR iGuide. Features such as the clinical, ethical, and regulatory implications are still warranted and need to be addressed prior to clinical implementation. Further studies are needed to confirm these findings. CLINICAL RELEVANCE STATEMENT The article explores the potential of using advanced language models, such as ChatGPT, in healthcare as a CDS for selecting appropriate imaging tests. Using ChatGPT can improve the efficiency of the decision-making process KEY POINTS: • ChatGPT recommendations were highly consistent with the recommendations provided by the ESR iGuide. • ChatGPT's ability in guiding the selection of appropriate tests may be comparable to some degree with ESR iGuide's.
Collapse
Affiliation(s)
- Shani Rosen
- Department of Health Technology and Policy Evaluation, Gertner Institute for Epidemiology and Health Policy, Institute of Epidemiology & Health Policy Research, Sheba Medical Center, Tel HaShomer, Ramat-Gan, Israel
- Nursing Department, School of Health Sciences, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Mor Saban
- Nursing Department, School of Health Sciences, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel.
| |
Collapse
|
49
|
Nazarovets S, Teixeira da Silva JA. ChatGPT as an "author": Bibliometric analysis to assess the validity of authorship. Account Res 2024:1-11. [PMID: 38693669 DOI: 10.1080/08989621.2024.2345713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2024] [Accepted: 04/08/2024] [Indexed: 05/03/2024]
Abstract
Background: Following the 2023 surge in popularity of large language models like ChatGPT, significant ethical discussions emerged regarding their role in academic authorship. Notable ethics organizations, including the ICMJE and COPE, alongside leading publishers, have instituted ethics clauses explicitly stating that such models do not meet the criteria for authorship due to accountability issues.Objective: This study aims to assess the prevalence and ethical implications of listing ChatGPT as an author on academic papers, in violation of existing ethical guidelines set by the ICMJE and COPE.Methods: We conducted a comprehensive review using databases such as Web of Science and Scopus to identify instances where ChatGPT was credited as an author, co-author, or group author.Results: Our search identified 14 papers featuring ChatGPT in such roles. In four of those papers, ChatGPT was listed as an "author" alongside the journal's editor or editor-in-chief. Several of the ChatGPT-authored papers have accrued dozens, even hundreds of citations according to Scopus, Web of Science, and Google Scholar.Discussion: The inclusion of ChatGPT as an author on these papers raises critical questions about the definition of authorship and the accountability mechanisms in place for content produced by artificial intelligence. Despite the ethical guidelines, the widespread citation of these papers suggests a disconnect between ethical policy and academic practice.Conclusion: The findings suggest a need for corrective measures to address these discrepancies. Immediate review and amendment of the listed papers is advised, highlighting a significant oversight in the enforcement of ethical standards in academic publishing.
Collapse
Affiliation(s)
- Serhii Nazarovets
- Library, Borys Grinchenko Kyiv Metropolitan University, Kyiv, Ukraine
| | | |
Collapse
|
50
|
Biswas S, Davies LN, Sheppard AL, Logan NS, Wolffsohn JS. Utility of artificial intelligence-based large language models in ophthalmic care. Ophthalmic Physiol Opt 2024; 44:641-671. [PMID: 38404172 DOI: 10.1111/opo.13284] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 01/23/2024] [Accepted: 01/25/2024] [Indexed: 02/27/2024]
Abstract
PURPOSE With the introduction of ChatGPT, artificial intelligence (AI)-based large language models (LLMs) are rapidly becoming popular within the scientific community. They use natural language processing to generate human-like responses to queries. However, the application of LLMs and comparison of the abilities among different LLMs with their human counterparts in ophthalmic care remain under-reported. RECENT FINDINGS Hitherto, studies in eye care have demonstrated the utility of ChatGPT in generating patient information, clinical diagnosis and passing ophthalmology question-based examinations, among others. LLMs' performance (median accuracy, %) is influenced by factors such as the iteration, prompts utilised and the domain. Human expert (86%) demonstrated the highest proficiency in disease diagnosis, while ChatGPT-4 outperformed others in ophthalmology examinations (75.9%), symptom triaging (98%) and providing information and answering questions (84.6%). LLMs exhibited superior performance in general ophthalmology but reduced accuracy in ophthalmic subspecialties. Although AI-based LLMs like ChatGPT are deemed more efficient than their human counterparts, these AIs are constrained by their nonspecific and outdated training, no access to current knowledge, generation of plausible-sounding 'fake' responses or hallucinations, inability to process images, lack of critical literature analysis and ethical and copyright issues. A comprehensive evaluation of recently published studies is crucial to deepen understanding of LLMs and the potential of these AI-based LLMs. SUMMARY Ophthalmic care professionals should undertake a conservative approach when using AI, as human judgement remains essential for clinical decision-making and monitoring the accuracy of information. This review identified the ophthalmic applications and potential usages which need further exploration. With the advancement of LLMs, setting standards for benchmarking and promoting best practices is crucial. Potential clinical deployment requires the evaluation of these LLMs to move away from artificial settings, delve into clinical trials and determine their usefulness in the real world.
Collapse
Affiliation(s)
- Sayantan Biswas
- School of Optometry, College of Health and Life Sciences, Aston University, Birmingham, UK
| | - Leon N Davies
- School of Optometry, College of Health and Life Sciences, Aston University, Birmingham, UK
| | - Amy L Sheppard
- School of Optometry, College of Health and Life Sciences, Aston University, Birmingham, UK
| | - Nicola S Logan
- School of Optometry, College of Health and Life Sciences, Aston University, Birmingham, UK
| | - James S Wolffsohn
- School of Optometry, College of Health and Life Sciences, Aston University, Birmingham, UK
| |
Collapse
|