Review ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns
综述ChatGPT在医疗教育、研究和实践中的应用:前景与隐忧的系统性评价
Malik Sallam $^{1,2(\oplus)}$
Malik Sallam $^{1,2(\oplus)}$
Citation: Sallam, M. ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns. Healthcare 2023, 11, 887. https://doi.org/10.3390/ healthcare 11060887
引用:Sallam, M. ChatGPT在医疗教育、研究和实践中的应用价值:关于前景与隐忧的系统性综述。Healthcare 2023, 11, 887. https://doi.org/10.3390/healthcare11060887
Academic Editor: Daniele Giansanti
学术编辑:Daniele Giansanti
Copyright: $\circledcirc$ 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license ( creative commons.org/licenses/by/ 4.0/).
版权:$\circledcirc$ 2023 作者所有。获许可方 MDPI,瑞士巴塞尔。本文是一篇开放获取文章,依据知识共享署名许可协议 (CC BY) (https://creativecommons.org/licenses/by/4.0/) 的条款和条件发布。
Abstract: ChatGPT is an artificial intelligence (AI)-based conversational large language model (LLM). The potential applications of LLMs in health care education, research, and practice could be promising if the associated valid concerns are pro actively examined and addressed. The current systematic review aimed to investigate the utility of ChatGPT in health care education, research, and practice and to highlight its potential limitations. Using the PRIMSA guidelines, a systematic search was conducted to retrieve English records in PubMed/MEDLINE and Google Scholar (published research or preprints) that examined ChatGPT in the context of health care education, research, or practice. A total of 60 records were eligible for inclusion. Benefits of ChatGPT were cited in 51/60 $(85.0%)$ ) records and included: (1) improved scientific writing and enhancing research equity and versatility; (2) utility in health care research (efficient analysis of datasets, code generation, literature reviews, saving time to focus on experimental design, and drug discovery and development); (3) benefits in health care practice (streamlining the workflow, cost saving, documentation, personalized medicine, and improved health literacy); and (4) benefits in health care education including improved personalized learning and the focus on critical thinking and problem-based learning. Concerns regarding ChatGPT use were stated in 58/60 $(96.7%)$ ) records including ethical, copyright, transparency, and legal issues, the risk of bias, plagiarism, lack of originality, inaccurate content with risk of hallucination, limited knowledge, incorrect citations, cyber security issues, and risk of infodemics. The promising applications of ChatGPT can induce paradigm shifts in health care education, research, and practice. However, the embrace of this AI chatbot should be conducted with extreme caution considering its potential limitations. As it currently stands, ChatGPT does not qualify to be listed as an author in scientific articles unless the ICMJE/COPE guidelines are revised or amended. An initiative involving all stakeholders in health care education, research, and practice is urgently needed. This will help to set a code of ethics to guide the responsible use of ChatGPT among other LLMs in health care and academia.
摘要:ChatGPT是一种基于人工智能(AI)的对话式大语言模型(LLM)。若能主动审视并解决相关合理关切,大语言模型在医疗教育、研究和实践中的潜在应用前景广阔。本系统综述旨在探讨ChatGPT在医疗教育、研究和实践中的效用,并强调其潜在局限性。根据PRIMSA指南,我们系统检索了PubMed/MEDLINE和Google Scholar中探讨ChatGPT在医疗教育、研究或实践背景下应用的英文文献(已发表研究或预印本)。共纳入60篇文献。51/60 $(85.0%)$ 篇文献指出ChatGPT的优势包括:(1)提升科学写作能力,促进科研公平性与多样性;(2)在医疗研究中的实用性(高效分析数据集、代码生成、文献综述、节省时间以专注实验设计、药物发现与开发);(3)在医疗实践中的益处(优化工作流程、节约成本、文档处理、个性化医疗及提升健康素养);(4)在医疗教育中的优势包括促进个性化学习、培养批判性思维和基于问题的学习。58/60 $(96.7%)$ 篇文献提出ChatGPT的使用顾虑,涉及伦理、版权、透明度、法律问题、偏见风险、剽窃、缺乏原创性、存在幻觉风险的不准确内容、知识有限、错误引用、网络安全问题及信息流行病风险。ChatGPT的广阔应用可能引发医疗教育、研究和实践的范式转变。然而,鉴于其潜在局限性,采用这款AI聊天机器人需极度谨慎。就目前而言,除非ICMJE/COPE指南修订,否则ChatGPT不具备被列为科学论文作者的资格。亟需发起一项涵盖医疗教育、研究和实践所有利益相关者的行动,以制定伦理准则,指导医疗界和学术界负责任地使用ChatGPT等大语言模型。
Keywords: machine learning; digital health; artificial intelligence; healthcare; ethics
关键词:机器学习;数字健康;人工智能;医疗保健;伦理
1. Introduction
1. 引言
Artificial intelligence (AI) can be defined as the multidisciplinary approach of computer science and linguistics that aspires to create machines capable of performing tasks that normally require human intelligence [1]. These tasks include the ability to learn, adapt, rationalize, understand, and to fathom abstract concepts as well as the reactivity to complex human attributes such as attention, emotion, creativity, etc. [2].
人工智能 (AI) 可定义为计算机科学与语言学的多学科交叉领域,旨在创造能够执行通常需要人类智能的任务的机器 [1]。这些任务包括学习、适应、推理、理解抽象概念的能力,以及对注意力、情感、创造力等复杂人类特征的响应能力 [2]。
The history of AI as a scientific discipline can be traced back to the mid-XX century at the Dartmouth Summer Research Project on AI [3]. This was followed by the development of machine learning (ML) algorithms that allow decision-making or predictions based on the patterns in large datasets [4]. Subsequently, the development of neural networks (brainmimicking algorithms), genetic algorithms (finding optimal solutions for complex problems by application of evolutionary principles), and other advanced techniques followed [5].
人工智能作为一门科学学科的历史可以追溯到20世纪中叶的达特茅斯夏季人工智能研究项目 [3]。随后出现了机器学习 (ML) 算法的发展,这些算法能够基于大数据集中的模式进行决策或预测 [4]。接着发展出了神经网络 (模拟大脑的算法)、遗传算法 (通过应用进化原理寻找复杂问题的最优解) 以及其他先进技术 [5]。
Launched in November 2022, “ChatGPT” is an AI-based large language model (LLM) trained on massive text datasets in multiple languages with the ability to generate humanlike responses to text input [6]. Developed by OpenAI (OpenAI, L.L.C., San Francisco, CA, USA), ChatGPT etymology is related to being a chatbot (a program able to understand and generate responses using a text-based interface) and is based on the generative pre-trained transformer (GPT) architecture [6,7]. The GPT architecture utilizes a neural network to process natural language, thus generating responses based on the context of input text [7]. The superiority of ChatGPT compared to its GPT-based predecessors can be linked to its ability to respond to multiple languages generating refined and highly sophisticated responses based on advanced modeling [6,7].
2022年11月推出的"ChatGPT"是一种基于AI的大语言模型(LLM),通过多语言海量文本数据集训练而成,能够针对文本输入生成类人响应[6]。由OpenAI(OpenAI, L.L.C., 美国旧金山)开发的ChatGPT,其名称源于chatbot(能够通过文本界面理解并生成响应的程序),并基于生成式预训练Transformer(GPT)架构[6,7]。GPT架构利用神经网络处理自然语言,从而根据输入文本的上下文生成响应[7]。ChatGPT相较于前代GPT模型的优势在于:通过先进建模技术,能够以多语言生成精炼且高度复杂的响应[6,7]。
In the scientific community and academia, ChatGPT has received mixed responses reflecting the history of controversy regarding the benefits vs. risks of advanced AI technologies [8–10]. On one hand, ChatGPT, among other LLMs, can be beneficial in conversational and writing tasks, assisting to increase the efficiency and accuracy of the required output [11]. On the other hand, concerns have been raised in relation to possible bias based on the datasets used in ChatGPT training, which can limit its capabilities and could result in factual inaccuracies, but alarmingly appear to be scientifically plausible (a phenomenon termed hallucination) [11]. Additionally, security concerns and the potential of cyberattacks with the spread of misinformation utilizing LLMs should also be considered [11].
在科学界和学术界,ChatGPT 引发的反响褒贬不一,这反映了关于先进 AI (Artificial Intelligence) 技术利弊的长期争议 [8–10]。一方面,ChatGPT 等大语言模型能在对话和写作任务中提升效率与输出准确性 [11];另一方面,人们担忧其训练数据可能导致的偏见——这些偏见不仅会限制模型能力,还可能产生看似科学合理实则失实的内容(这种现象称为幻觉)[11]。此外,大语言模型传播错误信息引发的安全问题及网络攻击风险也值得警惕 [11]。
The innate resistance of the human mind to any change is a well-described phenomenon and can be understandable from evolutionary and social psychology perspectives [12]. Therefore, the concerns and debate that arose instantaneously following the widespread release of ChatGPT appear to be understandable. The attention that ChatGPT received involved several disciplines. In education, for example, ChatGPT release could mark the end of essays as assignments [13]. In health care practice and academic writing, factual inaccuracies, ethical issues, and the fear of misuse including the spread of misinformation should be considered [14–16].
人类思维对任何变化的固有抗拒是一种已被充分描述的现象,从进化心理学和社会心理学角度可以理解 [12]。因此,ChatGPT广泛发布后立即引发的担忧和辩论似乎也在情理之中。ChatGPT受到的关注涉及多个学科领域。例如在教育界,ChatGPT的发布可能标志着论文作业时代的终结 [13];在医疗实践和学术写作领域,则需要考量事实性错误、伦理问题以及对滥用(包括虚假信息传播)的担忧 [14-16]。
The versatility of human intelligence (HI) compared to AI is related to its biological evolutionary history, adaptability, creativity, the ability of emotional intelligence, and the ability to understand complex abstract concepts [2]. However, HI-AI cooperation can be beneficial if an accurate and reliable output of AI is ensured. The promising utility of AI in health care has been outlined previously with possible benefits in personalized medicine, drug discovery, and the analysis of large datasets aside from the potential applications to improve diagnosis and clinical decisions [17,18]. Additionally, the utility of AI chatbots in health care education is an interesting area to probe. This is related to the massive information and various concepts that health care students are required to grasp [19]. However, all of these applications should be considered cautiously considering the valid concerns, risks, and categorical failures experienced and cited in the context of LLM applications. Specifically, Borji comprehensively highlighted the caveats of ChatGPT use that included, but were not limited to, the generation of inaccurate content, the risk of bias and discrimination, lack of transparency and reliability, cyber security concerns, ethical consequences, and societal implications [20].
人类智能 (HI) 相比人工智能的多样性源于其生物进化史、适应性、创造力、情商能力及理解复杂抽象概念的能力 [2]。然而,若能确保AI输出的准确性和可靠性,HI与AI的合作将大有裨益。此前已有研究概述了AI在医疗保健领域的潜在应用价值,包括个性化医疗、药物研发、大规模数据分析,以及改善诊断和临床决策的可能性 [17,18]。此外,AI聊天机器人在医疗教育中的应用也是一个值得探索的领域,这与医学生需要掌握海量信息和多样概念密切相关 [19]。但考虑到大语言模型应用中实际存在的隐患、风险及已被引证的典型失败案例,所有这些应用都应谨慎评估。Borji特别全面指出了ChatGPT的使用注意事项,包括但不限于:生成不准确内容、偏见与歧视风险、缺乏透明度和可靠性、网络安全问题、伦理影响及社会后果 [20]。
Therefore, the aim of the current review was to explore the future perspectives of ChatGPT as a prime example of LLMs in health care education, academic/scientific writing, health care research, and health care practice based on the existing evidence. Importantly, the current review objectives extended to involve the identification of potential limitations and concerns that could be associated with the application of ChatGPT in the aforementioned areas in health care settings.
因此,本次综述旨在基于现有证据,探讨以ChatGPT为代表的大语言模型(LLM)在医疗教育、学术/科研写作、医疗研究及临床实践中的未来前景。值得注意的是,本次综述目标还延伸至识别ChatGPT在上述医疗领域应用中可能存在的局限性与潜在问题。
2. Materials and Methods
2. 材料与方法
2.1. Search Strategy and Inclusion Criteria
2.1. 搜索策略与纳入标准
The current systematic review was conducted according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRIMSA) guidelines [21]. The information sources included PubMed/MEDLINE and Google Scholar.
本次系统综述按照系统综述和荟萃分析优先报告条目(PRIMSA)指南[21]开展。信息来源包括PubMed/MEDLINE和Google Scholar。
The eligibility criteria involved any type of published scientific research or preprints (article, review, communication, editorial, opinion, etc.) addressing ChatGPT that fell under the following categories: (1) health care practice/research; (2) health care education; and (3) academic writing.
入选标准包括涉及ChatGPT的任何类型的已发表科学研究或预印本(文章、综述、通讯、社论、观点等),且属于以下类别之一:(1) 医疗实践/研究;(2) 医学教育;(3) 学术写作。
The exclusion criteria included: (1) non-English records; (2) records addressing ChatGPT in subjects other than those mentioned in the eligibility criteria; and (3) articles from non-academic sources (e.g., newspapers, internet websites, magazines, etc.).
排除标准包括:(1) 非英文记录;(2) 记录涉及ChatGPT的主题不符合入选标准;(3) 非学术来源的文章(如报纸、网站、杂志等)。
The exact PubMed/MEDLINE search strategy, which concluded on 16 February 2023, was as follows: (ChatGPT) AND ((“2022/11/ $'30^{\prime\prime}$ [Date–Publication]: $\mathrm{'}3000^{\prime\prime}$ [Date–Publication])), which yielded 42 records.
截至2023年2月16日的PubMed/MEDLINE精确检索策略如下:(ChatGPT) AND (("2022/11/30"[Date-Publication] : "3000"[Date-Publication])),共获得42条记录。
The search on Google Scholar was conducted using Publish or Perish (Version 8) [22]. The search term was “ChatGPT” for the years 2022–2023, and the Google Scholar search yielded 238 records and concluded on 16 February 2023.
在Google Scholar上的搜索使用了Publish or Perish (版本8) [22]。搜索关键词为"ChatGPT",时间范围为2022年至2023年,Google Scholar搜索共获得238条记录,搜索结束于2023年2月16日。
2.2. Summary of the Record Screening Approach
2.2. 文献筛选方法概述
The records retrieved following the PubMed/MEDLINE and Google Scholar searches were imported to EndNote v.20 for Windows (Thomson Research Soft, Stanford, CA, USA), which yielded a total of 280 records.
通过PubMed/MEDLINE和Google Scholar检索获取的记录被导入至EndNote v.20 for Windows (Thomson Research Soft, Stanford, CA, USA),共获得280条记录。
Next, screening of the title/abstract was conducted for each record with the exclusion of duplicate records $(n=40)$ , followed by the exclusion of records published in languages other than English $(n=32)$ ). Additionally, the records that fell outside the scope of the review (records that examined ChatGPT in a context outside health care education, health care practice, or scientific research/academic writing) were excluded $(n=80)$ ). Moreover, the records published in non-academic sources (e.g., newspapers, magazines, Internet websites, blogs, etc.) were excluded $(n=18)$ .
接下来,对每条记录进行标题/摘要筛选,排除重复记录 $(n=40)$ ,随后剔除非英语发表的记录 $(n=32)$ 。此外,还排除了超出综述范围的记录(即在医疗教育、医疗实践或科学研究/学术写作之外研究ChatGPT的记录) $(n=80)$ 。同时,剔除非学术来源(如报纸、杂志、网站、博客等)发表的记录 $(n=18)$ 。
Afterward, full screening of the remaining records $(n=110)$ ) was carried out with the exclusion of an additional 41 records that fell outside the scope of the current review. An additional nine records were excluded due to the inability to access the full text of these records, being subscription-based. This yielded a total of 60 records eligible for inclusion in the current review.
随后对剩余记录 $(n=110)$ 进行了全面筛选,排除了另外41条不属于当前综述范围的记录。由于无法获取基于订阅的全文,又排除了9条记录。最终共有60条记录符合纳入当前综述的条件。
2.3. Summary of the Descriptive Search for ChatGPT Benefits and Risks in the Included Records
2.3. 纳入记录中ChatGPT益处与风险的描述性检索总结
Each of the included records was searched specifically for the following: (1) type of record (preprint, published research article, opinion, commentary, editorial, review, etc.); (2) the listed benefits/applications of ChatGPT in health care education, health care practice, or scientific research/academic writing; (3) the listed risks/concerns of ChatGPT in health care education, health care practice, or scientific research/academic writing; and (4) the main conclusions and recommendations regarding ChatGPT in health care education, health care practice, or scientific research/academic writing.
对每份纳入文献的检索具体包括以下方面:(1) 文献类型 (预印本、已发表研究论文、观点、评论、社论、综述等);(2) ChatGPT在医疗教育、医疗实践或科研/学术写作中列出的优势/应用;(3) ChatGPT在医疗教育、医疗实践或科研/学术写作中列出的风险/隐患;(4) 关于ChatGPT在医疗教育、医疗实践或科研/学术写作中的主要结论与建议。
Categorization of the benefits/applications of ChatGPT was as follows: (1) educational benefits in health care education (e.g., generation of realistic and variable clinical vignettes, customized clinical cases with immediate feedback based on the student’s needs, enhanced communications skills); (2) benefits in academic/scientific writing (e.g., text generation, sum mari z ation, translation, and literature review in scientific research); (3) benefits in scientific research (e.g., efficient analysis of large datasets, drug discovery, identification of potential drug targets, generation of codes in scientific research); (4) benefits in health care practice (e.g., improvements in personalized medicine, diagnosis, treatment, lifestyle recommendations based on personalized traits, documentation/generation of reports); and (5) being a freely available package.
ChatGPT的优势/应用分类如下:(1) 医疗教育领域优势(如生成真实多样的临床案例、根据学生需求提供定制化临床病例及即时反馈、提升沟通技巧);(2) 学术/科研写作优势(如文本生成、摘要总结、翻译及科研文献综述);(3) 科研领域优势(如高效分析大型数据集、药物发现、潜在药物靶点识别、科研代码生成);(4) 医疗实践优势(如改进个性化医疗、诊断治疗、基于个性化特征的生活方式建议、文档/报告生成);(5) 作为免费开源工具。
Categorization of the risks/concerns of ChatGPT was as follows: (1) ethical issues (e.g., risk of bias, discrimination based on the quality of training data, plagiarism); (2) hallucination (the generation of scientifically incorrect content that sounds plausible); (3) transparency issues (black box application); (4) risk of declining need for human expertise with subsequent psycho logic, economic and social issues; (5) over-detailed, redundant, excessive content; (6) concerns about data privacy for medical information; (7) risk of declining clinical skills, critical thinking and problem-solving abilities; (8) legal isues (e.g., copyright issues, authorship status); (9) interpret ability issues; (10) referencing issues; (11) risk of academic fraud in research; (12) incorrect content; and (13) infodemic risk.
ChatGPT的风险/担忧分类如下:(1) 伦理问题(如偏见风险、基于训练数据质量的歧视、剽窃);(2) 幻觉(生成听起来合理但科学上不正确的内容);(3) 透明度问题(黑箱应用);(4) 人类专业知识需求下降带来的心理、经济和社会问题风险;(5) 内容过度详细、冗余或过量;(6) 医疗信息数据隐私担忧;(7) 临床技能、批判性思维和解决问题能力下降风险;(8) 法律问题(如版权问题、作者身份认定);(9) 可解释性问题;(10) 参考文献问题;(11) 学术研究造假风险;(12) 内容错误;(13) 信息疫情风险。
3. Results
3. 结果
A total of 280 records were identified, and following the full screening process, a total of 60 records were eligible to be included in the review. The PRISMA flowchart of the record selection process is shown in Figure 1.
共识别出280条记录,经过全面筛选后,最终有60条记录符合纳入综述的条件。记录筛选过程的PRISMA流程图如图1所示。
Figure 1. Flowchart of the record selection process based on the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRIMSA) guidelines.
图 1: 基于系统评价与荟萃分析优先报告条目(PRIMSA)指南的记录筛选流程图。
3.1. Summary of the ChatGPT Benefits and Limitations/Concerns in Health Care
3.1. ChatGPT在医疗保健领域的优势与局限性/关注点总结
Summaries of the main conclusions of the included studies regarding ChatGPT utility in academic writing, health care education, and health care practice/research are provided in Table 1 for the records comprising editorials/letters to the editors, in Table 2 for the records comprising research articles, commentaries, news articles, perspectives, case studies, brief reports, communications, opinions, or recommendations, and in Table 3 for the records representing preprints
关于ChatGPT在学术写作、医疗教育和医疗实践/研究中效用的主要结论摘要如下:包含社论/致编辑信函的记录见表1,包含研究论文、评论、新闻文章、观点、案例研究、简报、通讯、意见或建议的记录见表2,预印本记录见表3
Table 1. A summary of the main conclusions of the included records comprising editorials/letters to the editors.
Author(s) [Record] | Design, Aims | Applications,Benefits | Risks,Concerns,Limitations | Suggested Action, Conclusions |
Chen [23] | Editorial on ChatGPT applications in scientific writing | ChatGPT helps to overcome language barriers promoting equity in research | Ethical concerns (ghostwriting); doubtful accuracy; citation problems | Embrace this innovation with an open mind; authors should have proper knowledge on how to exploit AI6 tools |
Thorp [24] | Editorial: "ChatGPT is not an author" | Content is not original; incorrect answers that sound plausible; issues of referencing; risk of plagiarism | Revise assignments in education In Science journals, the use of ChatGPT is considered as a scientific misconduct | |
Kitamura [25] | Editorial on ChatGPT and the future of medical writing | Improved efficiency in medical writing; translation purposes | Ethical concerns, plagiarism; lack of originality; inaccurate content; risk of bias | "AI in the Loop: Humans in Charge" |
Lubowitz [26] | Editorial, ChatGPT impact on medical literature | Inaccurate content; risk of bias; spread of misinformation and disinformation; lack of references; redundancy in text | Authors should notuse ChatGPT to compose any part of scientific submission; however, it can be used under careful human supervision to ensure the integrity and originality of the scientific work | |
Nature [27] | Nature editorial on the rules of ChatGPT use to ensure transparent science | ChatGPT can help to summarize research papers; to generate helpful computer codes | Ethical issues; transparency issues | LLM 7 tools will be accepted as authors;if LLM tools are to be used, it should be documented in the methods or acknowledgements; advocate for transparency in methods, and integrity and truth from researchers |
Moons and Van Bulck [28] | Editorial on ChatGPT potential in cardiovascular nursing practice and research | ChatGPT can help to summarize a large text; it can facilitate the work of researchers; it can help in | Information accuracy issues; the limited knowledge up to the year 2021; limited capacity | ChatGPT can be a valuable tool in health care practice |
Cahan and Treutlein [29] | Editorial reporting a conversation with ChatGPT on stem cell research | data collection ChatGPT helps to save time | Repetition; several ChatGPT responses lacked depth and insight; lack of references | ChatGPT helped to write an editorial saving valuable time |
Ahn [30] | A letter to the editor reporting a conversation of ChatGPT regarding CPR 1 | Personalized interaction; quick response time; it can help to provide easily accessible and understandable information | Inaccurate information might be generated with possible serious medical consequences | Explore the potential utility of ChatGPT to provide information and education on CPR |
Gunawan [31] | An editorial reporting a conversationwith ChatGPT regarding the future | regarding CPR to the general public ChatGPT can help to increase efficiency; it helps to reduce errors in | Lack of emotional and personal support | ChatGPT can provide valuable perspectives in health care |
D'Amico et al. [32] | of nursing An editorial reporting a conversation of ChatGPT regarding incorporating Chatbots into neurosurgical | care delivery ChatGPT can help to provide timely and accurate information for the patients about their treatment | Possibility of inaccurate information; privacy concerns; ethical issues; legal issues; | Neurosurgery practice can be leading in utilizing ChatGPT into patient care and research |
Fijacko et al. [33] | practice and research A letter to the editor to report the accuracy of ChatGPT responses with regards to life support exam | andcare ChatGPT provides relevant and accurate responses on occasions | risk of bias Referencing issues; over-detailed responses | ChatGPT did not pass any of the exams; however, it can be a powerful self-learning tool to prepare for the life |
Mbakwe et al. [34] | questions by the AHA 2 An editorial on ChatGPT ability to pass the USMLE 3 | Risk of bias; lack of thoughtful reasoning | support exams ChatGPT passing the USMLE revealed the deficiencies in medical education and assessment; there is a need to reevaluate the current medical students' training and educational tools |
表 1: 包含社论/致编辑信的主要结论摘要
作者 [文献] | 设计与目标 | 应用与优势 | 风险、担忧与局限 | 建议行动与结论 |
---|---|---|---|---|
Chen [23] | 关于ChatGPT在科学写作中应用的社论 | ChatGPT有助于克服语言障碍,促进研究公平性 | 伦理担忧(代笔);准确性存疑;引用问题 | 以开放心态拥抱创新;作者应掌握正确使用AI工具的知识 |
Thorp [24] | 社论:"ChatGPT不能成为作者" | 内容非原创;看似合理实则错误的答案;参考文献问题;抄袭风险 | 修订教育评估方式;《Science》期刊将使用ChatGPT视为学术不端行为 | |
Kitamura [25] | 关于ChatGPT与医学写作未来的社论 | 提高医学写作效率;翻译用途 | 伦理问题,抄袭;缺乏原创性;内容不准确;偏见风险 | "人在主导,AI辅助" |
Lubowitz [26] | 社论,ChatGPT对医学文献的影响 | 内容不准确;偏见风险;错误与虚假信息传播;缺乏参考文献;文本冗余 | 作者不应使用ChatGPT撰写科学投稿的任何部分;但在严格人工监督下可使用,以确保科研工作的完整性与原创性 | |
Nature [27] | 《自然》关于ChatGPT使用规则以确保科学透明性的社论 | ChatGPT可帮助总结研究论文;生成实用代码 | 伦理问题;透明度问题 | 大语言模型工具可作为作者被接受;若使用大语言模型工具,应在方法或致谢部分说明;倡导方法透明及研究者的诚信与真实 |
Moons和Van Bulck [28] | 关于ChatGPT在心血管护理实践与研究潜力的社论 | ChatGPT可帮助总结长文本;促进研究者工作;辅助 | 信息准确性问题;知识截止至2021年;能力有限 | ChatGPT可成为医疗实践中的有价值工具 |
Cahan和Treutlein [29] | 报告与ChatGPT关于干细胞研究对话的社论 | 数据收集 ChatGPT帮助节省时间 | 重复;部分回答缺乏深度与洞察力;缺少参考文献 | ChatGPT帮助撰写社论,节省宝贵时间 |
Ahn [30] | 致编辑信,报告与ChatGPT关于CPR的对话 | 个性化互动;响应迅速;可提供易获取且易懂的信息 | 可能生成不准确信息,导致严重医疗后果 | 探索ChatGPT在提供CPR信息与教育方面的潜在用途 |
Gunawan [31] | 报告与ChatGPT关于护理未来对话的社论 | 关于向公众普及CPR ChatGPT可提高效率;帮助减少 | 缺乏情感与个人支持 | ChatGPT可为医疗保健提供有价值视角 |
D'Amico等 [32] | 报告将Chatbot整合到神经外科实践与研究的社论 | 护理服务 ChatGPT可帮助为患者提供及时准确的治疗信息 | 信息不准确可能性;隐私担忧;伦理问题;法律问题 | 神经外科实践可引领将ChatGPT应用于患者护理与研究 |
Fijacko等 [33] | 关于ChatGPT对生命支持考试回答准确性的致编辑信 | 护理 ChatGPT有时能提供相关准确回答 | 偏见风险 参考文献问题;回答过于详细 | ChatGPT未通过任何考试;但可作为备考生命支持考试的有力自学工具 |
Mbakwe等 [34] | 关于ChatGPT通过USMLE能力的社论 | 偏见风险;缺乏深思熟虑的推理 | ChatGPT通过USMLE暴露了医学教育与评估的缺陷;需要重新评估现有医学生培训与教育工具 |
Table 1. Cont.
Author(s) [Record] | Design, Aims | Applications,Benefits | Risks, Concerns, Limitations | Suggested Action, Conclusions |
Huh [35] | An editorial of JEEHP 4 policy towards ChatGPT use | Reponses were not accurate in someareas | JEEHP will not accept ChatGPT as an author; however, ChatGPT content can be used if properly cited and documented | |
O'Connor * [36] | An editorial writtenwith ChatGPT assistance on ChatGPT in nursing education | ChatGPT can help to provide a personalized learning experience ChatGPT can help in the | Risk of plagiarism; biased or misleading results Risk of hallucination | Advocate ethical and responsible use of ChatGPT; improve assessment in nursing education |
Shen et al. [37] | An editorial on ChatGPT strengths and limitations | generation of medical reports; providing summaries ofmedical records; drafting letters to the insurance provider; improving the interpretability of CAD 5 systems | (inaccurateinformation that sounds plausible scientifically); the need to carefully construct ChatGPT prompts; possible inaccurate or incomplete results; dependence on the training data; risk of bias; risk of research fraud | Careful use of ChatGPT is needed to exploit its powerful applications |
Gordijn and Have [38] | Editorial on the revolutionary nature of ChatGPT | Risk of factual inaccuracies; risk of plagiarism; risk of fraud; copyright infringements possibility | In the near future, LLM can write papers with the ability to pass peer review; therefore, the scientific community should be prepared to address this serious issue | |
Mijwil et al. [39] | An editorial on the role of cybersecurity in the protection of medical information | Versatility; efficiency; high quality of text generated; cost saving; innovation potential; improved decision making; improved diagnostics; predictive modeling; improved personalized medicine; streamline clinical workflow increasing efficiency; remote monitoring | Data security issues | The role of cybersecurity to protect medical information should be emphasized |
TheLancet Digital Health [40] | An editorial on the strengths and limitations of ChatGPT | ChatGPT can help to improve language and readability | Over-detailed content;incorrect or biased content; potential to generate harmful errors; risk of spread of misinformation; risk of plagiarism; issues with | Widespread use of ChatGPT is inevitable; proper documentationofChatGPTuse is needed; ChatGPT should not belisted or cited as an author |
Aljanabi et al. [41] | An editorial on the possibilities provided by ChatGPT | ChatGPT can help in academic writing; helpful in code generation | integrity of scientific records Inaccurate content including inabilityhandlemathematical calculations to reliably | or co-author ChatGPT will receive a growing interest in the scientific community |
1 CPR: cardiopulmonary resuscitation; 2 AHA: American Heart Association; 3 USMLE: United States Medical Licensing Examination; 4 JEEHP: Journal of Educational Evaluation for Health Professions; 5 CAD: computeraided diagnosis; 6 AI: artificial intelligence; 7 LLM: large-scale language model. * ChatGPT generative pre-trained transformer was listed as an author.
表 1. (续)
作者 [记录] | 设计与目标 | 应用与优势 | 风险、问题与局限性 | 建议行动与结论 |
---|---|---|---|---|
Huh [35] | JEEHP 4 关于 ChatGPT 使用政策的社论 | 在某些领域回应不准确 | JEEHP 不接受 ChatGPT 作为作者;但如果正确引用和记录,可以使用 ChatGPT 生成的内容 | |
O'Connor * [36] | 一篇关于 ChatGPT 在护理教育中应用的社论,由 ChatGPT 辅助撰写 | ChatGPT 可提供个性化学习体验;有助于生成医疗报告、病历摘要、保险信函草拟,并提升 CAD 5 系统的可解释性 | 剽窃风险;结果可能存在偏见或误导;幻觉风险(看似科学但实际不准确的信息) | 倡导 ChatGPT 的伦理与负责任使用;改进护理教育评估 |
Shen 等 [37] | 关于 ChatGPT 优势与局限的社论 | 需谨慎构建提示词;结果可能不准确或不完整;依赖训练数据;存在偏见风险;研究欺诈风险 | 需谨慎使用 ChatGPT 以发挥其强大应用价值 | |
Gordijn 和 Have [38] | 关于 ChatGPT 革命性本质的社论 | 事实错误风险;剽窃风险;欺诈风险;可能侵犯版权 | 短期内大语言模型可撰写通过同行评审的论文,科学界应为此严峻问题做好准备 | |
Mijwil 等 [39] | 关于网络安全在医疗信息保护中作用的社论 | 多功能性;高效性;生成文本质量高;节约成本;创新潜力;改进决策;优化诊断;预测建模;提升个性化医疗;简化临床工作流程;远程监测 | 数据安全问题 | 应强调网络安全对医疗信息的保护作用 |
TheLancet Digital Health [40] | 关于 ChatGPT 优势与局限的社论 | ChatGPT 可提升语言表达与可读性 | 内容过度详细;存在错误或偏见内容;可能生成有害错误;传播错误信息风险;剽窃风险;科学记录完整性问题 | ChatGPT 的广泛使用不可避免;需规范使用记录;不应将 ChatGPT 列为作者或引用 |
Aljanabi 等 [41] | 关于 ChatGPT 可能性的社论 | ChatGPT 有助于学术写作;辅助代码生成 | 内容不准确(包括无法可靠处理数学计算) | ChatGPT 将日益受到科学界关注 |
1 CPR: 心肺复苏;2 AHA: 美国心脏协会;3 USMLE: 美国医师执照考试;4 JEEHP: 健康专业教育评价杂志;5 CAD: 计算机辅助诊断;6 AI: 人工智能;7 LLM: 大语言模型。* ChatGPT 生成式预训练 Transformer 被列为作者。
Table 2. A summary of the main conclusions of the included records comprising research articles, commentaries, news articles, perspectives, case studies, brief reports, communications, opinions, or recommendations.
Author(s) [Record] | Design, Aims | Applications,Benefits | Risks,Concerns,Limitations | Suggested Action,Conclusions |
Stokel- Walker [13] | News explainer | Well-organized content withdecentreferences;a free package Original, precise, and | Imminentendofconventional educationalassessment; concernsregardingthe effecton humanknowledge and ability Failuretofollow the | Theneedtoreviseeducational assessmenttoolstoprioritize criticalthinking orreasoning |
Kumar [42] | Briefreport;assessment of ChatGPTforacademic writinginbiomedicine | accurate responses with systematic approach; helpful for training and to improving topic clarity; efficiency in time; promotingmotivation towrite | instructionscorrectlyon occasions;failuretocite referencesin-text;inaccurate references;lack of practical examples;lackofpersonal experience highlights; superficialresponses | ChatGPT can help in improving academic writing skills; advocateforuniversaldesignfor learning; proper use of ChatGPT under academicmentoring |
表 2: 纳入记录(包括研究论文、评论、新闻文章、观点、案例研究、简报、通讯、意见或建议)的主要结论汇总
作者 [记录] | 设计与目标 | 应用与优势 | 风险/顾虑/局限性 | 建议措施与结论 |
---|---|---|---|---|
Stokel-Walker [13] | 新闻解读 | 结构清晰的内容与可靠参考文献;免费资源包;原创、精确且 | 传统教育评估即将终结;对人类知识与能力影响的担忧 | 需要改革教育评估工具以优先考察批判性思维与推理能力 |
Kumar [42] | 简报;评估ChatGPT在生物医学学术写作中的应用 | 系统方法生成准确回答;有助于培训与提升主题明确性;节省时间;提升写作动机 | 偶尔无法正确遵循指令;未标注文中引用;参考文献不准确;缺乏实际案例;缺少个人经验描述;回答流于表面 | ChatGPT可帮助提升学术写作能力;倡导通用学习设计;建议在学术指导下合理使用ChatGPT |
Table 2. Cont.
Author(s) [Record] | Design, Aims | Applications, Benefits | Risks, Concerns, Limitations | Suggested Action, Conclusions ChatGPT does not meet ICMJE 4 |
Zielinski et al. [43] | WAME1 recommendations on ChatGPT | ChatGPT can be a useful tool for researchers | Risk of incorrect or non-sensical answers; restricted knowledge to the period before 2021; lack of legal personality; risk of plagiarism | criteria and cannot belisted as an author; authors should be transparent regarding ChatGPT use and take responsibility for its content; editors need appropriate detection tools for ChatGPT-generated content |
Biswas [44] | A perspective record on the future of medical writing in light of ChatGPT | Improved efficiency in medical writing | Suboptimal understanding of the medicalfield;ethical concerns; risk of bias; legal issues;transparencyissues | A powerful tool in the medical field;however, several limitations of ChatGPTshould be considered |
Stokel- Walker [45] | A news article on the view of ChatGPT as an author | ChatGPT can help to | Risk of plagiarism; lack of accountability; concerns about misuse in the academia | ChatGPT should not be considered as an author ChatGPT ban will not work; develop rules for accountability, |
van Dis et al. [46] | A commentary on the priorities for ChatGPT research | accelerate innovation; to increase efficiency in publication time; it can make science more equitable and increase the diversity of scientific perspectives; more free time for experimental designs; it could optimize academic training | Compromised research quality; transparency issues; risk of spread of misinformation; inaccuracies in content, risk of bias and plagiarism; ethical concerns; possible future monopoly; lack of transparency | integrity, transparency and honesty; carefully consider which academic skills remain essential to researchers; widen the debateinthe academia; an initiative is needed to address the development and responsible use of LLM 5 for research |
Lund and Wang [47] | News article on ChatGPT impact in academia | Useful for literature review; can help in data analysis; can help in translation | Ethical concerns,issues about data privacy and security; risk of bias; transparency issues | ChatGPT has the potential to advance academia; consider how to use ChatGPT responsibly and ethically Robust AI authorship guidelines |
Liebrenz et al. [48] | A commentary on the ethical issues of ChatGPT use in medical publishing | ChatGPT can help to overcome language barriers | Ethical issues (copyright, attribution, plagiarism, and authorship); inequalities in scholarly publishing; risk of spread of misinformation; inaccurate content | are needed in scholarly publishing; COPE AI6 should be followed;AI cannot belisted as an author and it must be properly acknowledged upon its use |
Manohar and Prasad [49] | A case study written with ChatGPT assistance | ChatGPT helped to generate a clear, comprehensible text | Lack of scientific accuracy and reliability; citation inaccuracies | ChatGPT use is discouraged due to risk of false information and non-existent citations; it can be misleading in health care practice |
Akhter and Cooper [50] | A case study written with ChatGPT assistance | ChatGPT helped to provide a relevant general introductory summary | Inability to access relevant literature; the limited knowledge up to 2021; citation inaccuracies; limited ability to critically discuss results | Currently, ChatGPT cannot replaceindependentliterature reviews in scientific research |
Holzinger et al. [51] | An article on AI2/ ChatGPT use in biotechnology | Biomedical image analysis; diagnostics and disease prediction; personalized medicine; drug discovery and development | Ethical and legal issues; limited data availability to train the models; the issue of reproducibility of the runs | The scientists aspire for fairness, open science, and open data |
Mann [52] | A perspective on ChatGPT intranslationalresearch | Efficiency in writing; analysis oflarge datasets (e.g., electronic health records or genomic data); predict risk factors for disease; predict disease outcomes | Compromised quality of data available; inability to understand the complexity of biologic systems | ChatGPT use in scientific and medical journals is inevitable in near future |
Patel and Lam [53] | A commentary on ChatGPT utility in documentation of discharge summary | ChatGPT can help to reduce the burden of discharge summaries providing high-quality and efficient output | Data governance issues; risk of depersonalization of care; risk of incorrect or inadequate information | Proactive adoption of ChatGPT is needed to limit any possible future issues and limitations |
表 2. (续)
作者 [记录] | 设计与目标 | 应用与优势 | 风险、担忧与局限性 | 建议行动与结论 ChatGPT 不符合 ICMJE 4 项标准 |
---|---|---|---|---|
Zielinski 等 [43] | WAME1 关于 ChatGPT 的建议 | ChatGPT 可作为研究人员的实用工具 | 存在错误或无意义答案的风险;知识仅限于 2021 年前;缺乏法律人格;抄袭风险 | 不能列为作者;作者应透明披露 ChatGPT 使用情况并对其内容负责;编辑需配备检测 ChatGPT 生成内容的工具 |
Biswas [44] | 关于 ChatGPT 对医学写作未来影响的观点记录 | 提升医学写作效率 | 对医学领域理解不足;伦理问题;偏见风险;法律问题;透明度问题 | 医学领域的强大工具,但需考虑 ChatGPT 的多项局限性 |
Stokel-Walker [45] | 关于将 ChatGPT 视为作者的新闻文章 | ChatGPT 可协助 | 抄袭风险;缺乏问责;学术界滥用担忧 | ChatGPT 不应被视为作者;禁令无效,需制定问责规则 |
van Dis 等 [46] | 关于 ChatGPT 研究优先事项的评论 | 加速创新;缩短发表时间;促进科学公平性与视角多样性;为实验设计腾出时间;优化学术培训 | 研究质量受损;透明度问题;错误信息传播风险;内容不准确、偏见与抄袭;伦理问题;未来垄断可能;缺乏透明度 | 需确保诚信、透明与诚实;审慎评估研究者必备技能;扩大学术界讨论;需发起倡议推动大语言模型 (LLM) 5 的负责任研究应用 |
Lund 和 Wang [47] | 关于 ChatGPT 对学术界影响的新闻文章 | 助力文献综述;辅助数据分析;协助翻译 | 伦理问题;数据隐私与安全隐患;偏见风险;透明度问题 | ChatGPT 有推动学术进步的潜力;需考虑负责任与伦理的使用方式;需健全 AI 作者指南 |
Liebrenz 等 [48] | 关于医学出版中 ChatGPT 使用伦理问题的评论 | ChatGPT 可帮助克服语言障碍 | 伦理问题(版权、归属、抄袭与作者身份);学术出版不平等;错误信息传播风险;内容不准确 | 学术出版需遵循 COPE AI6 准则;AI 不能列为作者且使用时应明确标注 |
Manohar 和 Prasad [49] | 借助 ChatGPT 撰写的案例研究 | ChatGPT 帮助生成清晰易懂的文本 | 缺乏科学准确性与可靠性;引用不准确 | 因虚假信息与虚构引用风险不推荐使用;在医疗实践中可能产生误导 |
Akhter 和 Cooper [50] | 借助 ChatGPT 撰写的案例研究 | ChatGPT 协助提供相关通用性介绍摘要 | 无法获取相关文献;知识限于 2021 年前;引用不准确;结果批判性讨论能力有限 | 目前 ChatGPT 无法替代科研中的独立文献综述 |
Holzinger 等 [51] | 关于 AI2/ChatGPT 在生物技术中应用的文章 | 生物医学图像分析;诊断与疾病预测;个性化医疗;药物研发 | 伦理与法律问题;模型训练数据有限;实验可复现性问题 | 科学家追求公平性、开放科学与开放数据 |
Mann [52] | 关于 ChatGPT 在转化研究中应用的观点 | 提升写作效率;分析大型数据集(如电子健康记录或基因组数据);预测疾病风险因素;预测疾病结果 | 数据质量受损;无法理解生物系统复杂性 | ChatGPT 在科学与医学期刊中的应用在未来不可避免 |
Patel 和 Lam [53] | 关于 ChatGPT 在出院小结记录中效用的评论 | ChatGPT 可减轻出院小结负担,提供高效优质输出 | 数据治理问题;诊疗去个性化风险;信息错误或不完整风险 | 需主动采用 ChatGPT 以规避未来潜在问题与限制 |
Table 2. Cont.
Author(s) | |||||
[Record] Zhavoronkov * [54] | Design, Aims A perspective reporting a conversationwith ChatGPT about rapamycin use from | Applications, Benefits ChatGPT provided correct summary of rapamycin side effects;it referred to theneed to consult a health | Risks, Concerns, Limitations | Suggested Action, Conclusions Demonstration of ChatGPT's potential to generate complex philosophical arguments | |
Hallsworth et al. [55] | a philosophical perspective A comprehensive opinion article submitted before ChatGPT launching on the value of | care provider based on the specific situation It can help to circumvent language barriers; it can robustly help to process massive data in short time; it can stimulate creativity | Ethical issues; legal responsibility issues; lack of empathy and personal | Despite the AI potential in science, there is an intrinsic value of human engagement in the scientific process which | |
Stokel- Walker and Van | theory-based research Naturenewsfeature article on ChatGPT implications | by humans if "AI in the Loop: Humans in Charge" is applied More productivity among researchers | communication; lack oftransparency Problems in reliability and factual inaccuracies; misleading information that seem plausible | cannot be replaced by AI contribution "AI in the Loop: Humans in Charge" should be used; ChatGPT widespread use in the | |
Noorden [14] Huh [56] | in science A study to compare ChatGPT performance on a parasitology exam to the performance of Korean | ChatGPT performance will improve by deep learning | (hallucination); over-detailed content; risk of bias; ethical issues; copyright issues ChatGPT performance was lower compared to medical students; plausible explanations for incorrect | nearfuture would beinevitable ChatGPT performance will continue to improve, and health care educators/students are advised to incorporate this tool | |
Khan et al. [57] | medical students A communication on ChatGPT use in medical education and clinical management | ChatGPT can help in automated scoring; assistance in teaching; improved personalized learning; assistance in research; generation of clinical vignettes; rapid access to information; translation; documentation in clinical practice; support | answers (hallucination) Lack of human-like understanding; the limited knowledge up to 2021 | into the educational process ChatGPT is helpful in medical education, research, and in clinical practice; however, the human capabilities are still needed | |
Gilson et al. [58] | An article on the performance of ChatGPT on USMLE3 | in clinical decisions; personalized medicine Ability to understand context and to complete a coherent and relevant conversation in the medical field; can be used as an | The limited knowledge up to 2021 | ChatGPT passes the USMLE with performance at a 3rd year medical student level; can help to facilitate learning as a virtual medical tutor | |
Kung et al. [59] | An article showing the ChatGPT raised accuracy which enabled passing the USMLE | adjunct in group learning Accuracy withhigh concordance and insight; it can facilitate patient communication; improved personalized medicine | ChatGPT has a promising potential in medical education; future studies are recommended to consider non-biased approach withquantitativenatural language processing and text mining tools such as word | ||
Marchandot et al. [60] | A commentary on ChatGPT use in academic writing | ChatGPT can assist in literature review saving time; the ability to summarize papers; the ability to improve language | Risk of inaccurate content;risk of bias; ChatGPT may lead to decreased critical thinking and creativity in science; ethical concerns; risk of plagiarism | network analysis ChatGPT can be listed as an author based on its significant contribution |
1 WAME: World Association of Medical Editors; 2 AI: artificial intelligence; 3 USMLE: United States Medical Licensing Examination; 4 ICMJE: International Committee of Medical Journal Editors; 5 LLM: large-scale language model; 6 COPE AI in decision making: Committee on Publication Ethics, Artificial Intelligence (AI) in decision making, available from: https://publication ethics.org/node/50766, accessed on 18 February 2023. * ChatGPT generative pre-trained transformer was listed as an author.
表 2. 续
作者 | 设计与目标 | 应用与优势 | 风险、担忧与局限 | 建议行动与结论 |
---|---|---|---|---|
Zhavoronkov * [54] | 从哲学视角报道与ChatGPT关于雷帕霉素应用的对话 | ChatGPT正确总结了雷帕霉素副作用;建议根据具体情况咨询医疗专业人员 | 展示了ChatGPT生成复杂哲学论证的潜力 | |
Hallsworth等 [55] | 在ChatGPT发布前提交的关于理论研究的综合观点文章 | 可帮助克服语言障碍;快速处理海量数据;激发创造力 | 伦理问题;法律责任;缺乏同理心与个人沟通;透明度不足 | 尽管AI在科学中有潜力,但人类参与科研过程的内在价值无法被AI取代 |
Stokel-Walker与Van Noorden [14] | Nature新闻特写关于ChatGPT对科学的影响 | 若采用"AI在环:人类主导"模式可提升研究人员生产力 | 可靠性与事实准确性存疑;看似合理但误导的信息(幻觉);内容过度详细;偏见风险;伦理问题;版权问题 | "AI在环:人类主导"应被采用;ChatGPT在近未来的广泛使用不可避免 |
Huh [56] | 比较ChatGPT与韩国医学生在寄生虫学考试表现的研究 | ChatGPT表现将通过深度学习提升 | 表现低于医学生;错误答案的合理解释(幻觉) | ChatGPT表现将持续改进,建议医学教育者/学生将该工具纳入教学 |
Khan等 [57] | 关于ChatGPT在医学教育与临床管理中应用的通讯 | 可辅助自动评分、教学、个性化学习、研究、临床案例生成、快速获取信息、翻译、临床记录、临床决策支持、个性化医疗 | 缺乏类人理解;知识截止至2021年 | ChatGPT有助于医学教育、研究和临床实践,但仍需人类能力 |
Gilson等 [58] | 关于ChatGPT在美国医师执照考试(USMLE)表现的文章 | 能理解医学领域语境并完成连贯对话;可作为小组学习辅助工具 | 知识截止至2021年 | ChatGPT以三年级医学生水平通过USMLE;可作为虚拟医学导师促进学习 |
Kung等 [59] | 展示ChatGPT提高准确率从而通过USMLE的文章 | 高一致性准确率与洞察力;改善医患沟通;促进个性化医疗 | ChatGPT在医学教育中潜力巨大;建议未来研究采用无偏见的定量自然语言处理与文本挖掘工具 | |
Marchandot等 [60] | 关于ChatGPT在学术写作中应用的评论 | 可辅助文献综述节省时间;总结论文能力;改进语言能力 | 内容不准确风险;偏见风险;可能降低科研批判思维与创造力;伦理担忧;抄袭风险 | 基于显著贡献,ChatGPT可被列为作者 |
1 WAME:世界医学编辑协会;2 AI:人工智能;3 USMLE:美国医师执照考试;4 ICMJE:国际医学期刊编辑委员会;5 LLM:大语言模型;6 COPE AI决策:出版伦理委员会人工智能决策,来源:https://publicationethics.org/node/50766,访问于2023年2月18日。* ChatGPT生成式预训练Transformer被列为作者。
Table 3. A summary of the main conclusions of the included records representing preprints.
Author(s) | |||||
[Record] Wang et al. [61] | Design, Aims An arXiv preprint 1; investigating ChatGPT effectiveness to generate Boolean queries for | Applications, Benefits Higher precision compared to the current automatic | Risks, Concerns, Limitations Non-suitability for high-recall retrieval; many incorrect MeSH 11 terms; variability in query | Suggested Action, Conclusions A promising tool for research | |
systematic literature reviews An arXiv preprint; to | query formulation methods | effectiveness across multiple requests; a black-box application Problems in spatial, temporal, physical, psychological and logical reasoning; limited capability to calculate mathematical expressions; factual errors; risk of bias and discrimination; difficulty in using idioms; lack of real | Implementation ofresponsible use and precautions; proper | ||
Borji [20] | highlight the limitations of ChatGPT | Extremely helpful in scientific writing | emotions and thoughts; no perspective for the subject; over-detailed; lacks human-like divergences; lack of transparency and reliability; security concerns with vulnerability to data poisoning; violation of data privacy; plagiarism; impact on the environment and climate; ethical andsocial consequences | monitoring; transparent communication; regular inspection for biases, misinformation, among other harmful purposes (e.g., identity theft) | |
Cotton et al. [62] | An EdArXiv 2 preprint on the academic integrity in ChatGPT era A bioRxiv 3 preprint | Risk of plagiarism; academic dishonesty | Careful thinking of educational assessment tools | ||
Gao et al. [63] | comparing the scientific abstracts generated by ChatGPT to original abstracts | A tool to decrease the burden of writing and formatting; it can help to overcome language barriers ChatGPT can help to | Misuse to falsify research; risk of bias | The use of ChatGPTinscientific writing or assistance should be clearly disclosed and documented | |
Polonsky and Rotman [64] | An SSRN 4 preprint on listing ChatGPT as an author | accelerate the research process; it can help to increase accuracy and precision | Intellectual property issues if financial gains are expected | AI 12 can be listed as an author in some instances | |
Aczel and Wagenmak- ers [65] | A PsyArXiv 5 preprint as a guide of transparent ChatGPT use in scientific writing | Issues of originality, transparency issues | There is a need to provide sufficient information on ChatGPT use, with accreditation and verification of its use Carefully weigh ChatGPT | ||
De Angelis et al. [66] | An SSRN preprint discussing the concerns of an AI-driven infodemic | ChatGPT can support and expedite academic research | Generation ofmisinformation and the risk of subsequent infodemics; falsified or fake research; ethical concerns | possible benefits with its possible risks; there is a need to establish ethical guidelines for ChatGPT use; a science-driven debateis neededto address ChatGPT utility | |
Benoit [67] | A medRxiv 6 preprint on the generation, revision, and evaluation of clinical vignettes as a tool in health education using ChatGPT | Consistency, rapidity and flexibility of text and style; ability to generate plagiarism-free text | Clinical vignettes' ownership issues; inaccurate or non-existent references Risk of bias or inaccuracies; | ChatGPT can allow for improved medical education and better patient communication | |
Sharma and Thakur [68] | A ChemRxiv 7 preprint on ChatGPT possible use in drug discovery | ChatGPT can help to identify and validate new drug targets; to design new drugs; to optimize drug properties; to assess toxicity; and to generate drug-related reports | inability to understand the complexity of biologic systems; transparency issues;lack of experimental validation; limited interpretability; limited handling of uncertainty; ethical issues | ChatGPT can be a powerful and promising tool in drug discovery; however, its accompanying ethical issues should be addressed |
表 3: 收录预印本记录的主要结论摘要
作者 | 设计与目标 | 应用与优势 | 风险、问题与局限性 | 建议行动与结论 |
---|---|---|---|---|
Wang 等 [61] | 一项 arXiv 预印本;研究 ChatGPT 为系统文献综述生成布尔查询的有效性 | 相比当前自动查询生成方法具有更高精度 | 不适用于高召回检索;许多错误的 MeSH 术语;查询效果随请求变化;黑箱应用问题 | 研究中的潜力工具;需负责任使用并采取预防措施 |
Borji [20] | 强调 ChatGPT 的局限性 | 对科学写作极有帮助 | 空间、时间、物理、心理和逻辑推理问题;数学计算能力有限;事实错误;偏见与歧视风险;难以使用习语;缺乏真实情感与思考;透明性与可靠性不足;数据中毒漏洞的安全问题;数据隐私侵犯;剽窃;对环境和气候的影响;伦理与社会后果 | 需透明沟通;定期检查偏见、错误信息及其他有害用途(如身份盗用) |
Cotton 等 [62] | 一项 EdArXiv 预印本,探讨 ChatGPT 时代的学术诚信 | 剽窃风险;学术不端 | 需谨慎设计教育评估工具 | |
Gao 等 [63] | 一项 bioRxiv 预印本,比较 ChatGPT 生成的科学摘要与原始摘要 | 减轻写作与格式负担的工具;可帮助克服语言障碍 | 滥用以伪造研究;偏见风险 | 在科学写作或辅助中使用 ChatGPT 应明确披露并记录 |
Polonsky 和 Rotman [64] | 一项 SSRN 预印本,讨论将 ChatGPT 列为作者 | 加速研究进程;提高准确性与精确度 | 若涉及经济利益则存在知识产权问题 | 某些情况下可将 AI 列为作者 |
Aczel 和 Wagenmakers [65] | 一项 PsyArXiv 预印本,作为科学写作中透明使用 ChatGPT 的指南 | 原创性问题;透明度问题 | 需提供 ChatGPT 使用的充分信息,并验证其使用 | |
De Angelis 等 [66] | 一项 SSRN 预印本,讨论 AI 驱动信息疫情的担忧 | ChatGPT 可支持并加速学术研究 | 生成错误信息及后续信息疫情风险;伪造研究;伦理问题 | 需权衡利弊;制定 ChatGPT 使用的伦理指南;开展科学驱动的讨论 |
Benoit [67] | 一项 medRxiv 预印本,探讨 ChatGPT 作为健康教育工具生成、修订和评估临床案例 | 文本与风格的一致性、快速性和灵活性;可生成无剽窃文本 | 临床案例所有权问题;不准确或不存在的参考文献 | 可改善医学教育与患者沟通 |
Sharma 和 Thakur [68] | 一项 ChemRxiv 预印本,探讨 ChatGPT 在药物发现中的潜在用途 | 帮助识别和验证新药物靶点;设计新药;优化药物性质;评估毒性;生成药物相关报告 | 无法理解生物系统复杂性;透明度问题;缺乏实验验证;可解释性有限;不确定性处理能力有限;伦理问题 | ChatGPT 是药物发现中强大且有潜力的工具,但需解决其伦理问题 |
Table 3. Cont.
Author(s) [Record] | Design, Aims | Applications, Benefits | Risks,Concerns,Limitations Lack of references; alignment | Suggested Action, Conclusions Using ChatGPT for radiologic |
Rao et al. [69] | A medRxiv preprint on the usefulness of ChatGPT in radiologic decision making | ChatGPT showed moderate accuracy to determine appropriate imaging steps in breast cancer screening and evaluation of breast pain | with userintent;inaccurate information; over-detailed; recommending imaging in futile situations; providing rationale for incorrect imaging decisions; the black box nature with lack of transparency | decision making is feasible, potentially improving the clinical workflow and responsible use of radiology services |
Antaki et al. [70] | A medRxiv preprint assessing ChatGPT's ability to answer a diverse MCQ 8 exam in ophthalmology | ChatGPT currently performs at the level of an average first-year ophthalmology resident | Inability to process images; risk of bias; dependence on training dataset quality | There is a potential of ChatGPT use in ophthalmology; however, its applications should be carefully addressed Expression of knowledge can be |
Aydin and Karaarslan [71] | An SSRN preprint on the use of ChatGPT to conduct a literature review on digital twin in health care | Low risk of plagiarism; accelerated literature review; more free time for researchers ChatGPT can provide | Lack of originality | accelerated using ChatGPT; further work will use ChatGPT in citation analysis to assess the attitude towards the findings |
Sanmarchi et al. [72] | A medRxiv preprint evaluating ChatGPTvalue in an epidemiologic study following the STROBE 9 recommendations | appropriate responses if proper constructs are developed; more free time for researchers to focus on experimental phase Generation of rapid and accurate responses; easily | Risk of bias in the training data; risk of devaluation of human expertise; risk of scientific fraud; legal issues; reproducibility issues | Despite ChatGPT possible value, the research premise and originality will remain the function of human brain |
Duong and Solomon [73] | A medRxiv preprint evaluating ChatGPT versus human responses to questions on genetics | accessibleinformationfor the patients with genetic disease and theirfamilies;it can help can health professionals in the diagnosis and treatment of genetic diseases; it could make genetic information widely available and help non-experts to understand suchinformation | Plausible explanations for incorrect answers (hallucination); reproducibility issues | Thevalue of ChatGPTwill increase in research and clinical settings |
Yeo et al. [74] | A medRxiv preprint evaluating ChatGPT responses to questions on cirrhosis and hepatocellular carcinoma | Improved health literacy with better patient outcome; free availability; increased efficiency among health providers; emulation of empathetic responses | Non-comprehensive responses; the limited knowledge up to 2021; responses ca |