Published on in Vol 26 (2024)

This is a member publication of National University of Singapore

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/59439, first published .
Mitigating Cognitive Biases in Clinical Decision-Making Through Multi-Agent Conversations Using Large Language Models: Simulation Study

Mitigating Cognitive Biases in Clinical Decision-Making Through Multi-Agent Conversations Using Large Language Models: Simulation Study

Mitigating Cognitive Biases in Clinical Decision-Making Through Multi-Agent Conversations Using Large Language Models: Simulation Study

Original Paper

1Centre for Quantitative Medicine, Duke-NUS Medical School, Singapore, Singapore

2Department of Anesthesiology, Singapore General Hospital, Singapore, Singapore

3Information Technology Center, University of Tokyo, Tokyo, Japan

4Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore

5Institute of Data Science, National University of Singapore, Singapore, Singapore

Corresponding Author:

Nan Liu, PhD

Centre for Quantitative Medicine

Duke-NUS Medical School

8 College Road

Singapore, 169857

Singapore

Phone: 65 66016503

Email: liu.nan@duke-nus.edu.sg


Background: Cognitive biases in clinical decision-making significantly contribute to errors in diagnosis and suboptimal patient outcomes. Addressing these biases presents a formidable challenge in the medical field.

Objective: This study aimed to explore the role of large language models (LLMs) in mitigating these biases through the use of the multi-agent framework. We simulate the clinical decision-making processes through multi-agent conversation and evaluate its efficacy in improving diagnostic accuracy compared with humans.

Methods: A total of 16 published and unpublished case reports where cognitive biases have resulted in misdiagnoses were identified from the literature. In the multi-agent framework, we leveraged GPT-4 (OpenAI) to facilitate interactions among different simulated agents to replicate clinical team dynamics. Each agent was assigned a distinct role: (1) making the final diagnosis after considering the discussions, (2) acting as a devil’s advocate to correct confirmation and anchoring biases, (3) serving as a field expert in the required medical subspecialty, (4) facilitating discussions to mitigate premature closure bias, and (5) recording and summarizing findings. We tested varying combinations of these agents within the framework to determine which configuration yielded the highest rate of correct final diagnoses. Each scenario was repeated 5 times for consistency. The accuracy of the initial diagnoses and the final differential diagnoses were evaluated, and comparisons with human-generated answers were made using the Fisher exact test.

Results: A total of 240 responses were evaluated (3 different multi-agent frameworks). The initial diagnosis had an accuracy of 0% (0/80). However, following multi-agent discussions, the accuracy for the top 2 differential diagnoses increased to 76% (61/80) for the best-performing multi-agent framework (Framework 4-C). This was significantly higher compared with the accuracy achieved by human evaluators (odds ratio 3.49; P=.002).

Conclusions: The multi-agent framework demonstrated an ability to re-evaluate and correct misconceptions, even in scenarios with misleading initial investigations. In addition, the LLM-driven, multi-agent conversation framework shows promise in enhancing diagnostic accuracy in diagnostically challenging medical scenarios.

J Med Internet Res 2024;26:e59439

doi:10.2196/59439

Keywords



Human cognitive biases in clinical decision-making are increasingly recognized as a crucial factor in health care errors and suboptimal patient outcomes [1]. These biases stem from innate psychological tendencies and can potentially lead to misjudgments and suboptimal outcomes in patient care [2-4]. Despite concerted efforts involving educational strategies, optimal work environments, and a culture promoting bias awareness and correction [5,6], the eradication of these biases remains an elusive goal.

The integration of artificial intelligence (AI), and in particular, large language models (LLMs), into clinical medicine is on the horizon [7]. LLMs have advanced text generation capability and extensive domain-specific knowledge [8,9]. Notably, these models have demonstrated their proficiency by successfully passing advanced medical examinations [10] and scoring clinical risk gradings on par with experienced physicians [11].

However, the deployment of LLMs in actual clinical diagnosis and decision-making processes has been mired in controversy, primarily due to the high stakes involved. The use of AI in medical settings is not just a technological issue; it intersects with complex ethical, legal, and medical considerations [11-16]. The accuracy of ChatGPT making the correct emergency medicine diagnosis is still limited to 77% to 83% [17]. Thus, concerns centering around the legal implications and accountability in cases where AI-driven diagnostics might lead to errors or misjudgments are a major hurdle.

This debate is rooted in the fundamental difference between human and machine intelligence. While LLMs can process and analyze vast quantities of data far beyond human capacity [18], they lack the nuanced understanding, empathy, and ethical reasoning inherent to human practitioners [19]. Human cognitive biases can be mitigated through a combination of awareness, education, and structured approaches [20]. Simulations of such discussions through LLM agents could provide a new solution to increase the accuracy of diagnosis [21]. The multi-agent framework features dialogue agents with near-human performance and could introduce an innovative paradigm in health care [22-24]. By simulating scenarios that mirror real-life clinical decision-making processes, and through reading the multi-agent conversations, clinicians can be made aware of potential cognitive biases and how to correct them. This facilitates learning in a controlled, educational environment [25,26].

Despite significant advancements in LLM technology, especially within multi-agent systems, its application to identify and mitigate human cognitive biases in clinical settings remains largely unexplored in current research. This study seeks to evaluate the efficacy of the multi-agent framework in achieving accurate final diagnoses following discussions on cognitive biases that may be present within the initial diagnosis. In addition, it aims to compare these outcomes with the differential diagnoses provided by experienced clinicians after reviewing the same scenarios. By doing so, the research aims to shed light on the potential of the multi-agent system to support clinical decision-making processes and enhance diagnostic accuracy in health care settings.


Overview

We accessed GPT-4 [27] through an application programming interface call to the OpenAI server. The specific variant used was GPT-4 Turbo.

We implemented a comprehensive search strategy aimed at including all relevant reports on misdiagnoses attributed to cognitive biases. A selection of 15 case reports was identified after a full review of the published literature as a representative sampling.

Ethical Considerations

Due to the nature of the study, institutional review board approval was not required, as the research did not involve patient data and did not constitute human participants.

Search Strategy

In this study, we focused on case reports highlighting instances of misdiagnoses resulting from cognitive biases. Our research involved a comprehensive search of the PubMed database using the terms “case reports [Publication Type]” AND “cognitive bias[All Fields]”. PubMed was chosen for its comprehensive coverage of case studies across diverse medical disciplines, ensuring access to a broad spectrum of peer-reviewed case studies for our analysis.

Eligibility for inclusion requires case reports to meet four key criteria: (1) they must provide detailed case information sufficient for making the initial diagnosis; (2) they must include a final, accurate diagnosis for the patient; (3) the incorrect diagnosis must be linked to cognitive bias by the authors; and (4) the final diagnosis should not be a rare disease or unclear. A rare disease is a disease or condition that affects fewer than 200,000 patients per year. The list of exclusions for rare diseases was based on the National Organization for Rare Disorders [28]. We set no limits on the publication year of these reports.

Screening of abstracts for eligible studies was conducted by 2 independent clinically trained reviewers, YK and TXYL. Each reviewer separately assessed whether a case should be included or excluded based on predefined criteria. In instances of disagreement, SAL reviewed the justifications for exclusion and inclusion and made the final decision. The full texts were reviewed to obtain the case summary, the initial wrong diagnosis by the medical team, and the final diagnosis of the case reported.

For the studies included in the analysis, full-text extraction, including patient demographics, past medical history, initial presenting complaints, and results of the preliminary investigations, was conducted. For cases involving imaging data, such as chest x-rays, we did not incorporate the actual images into the query. Instead, we opted to include the legends or descriptions accompanying these images. In defining the boundaries of the clinical scenarios for our study, we restricted the information to that available up to the point of and before the initial diagnosis. This meant deliberately excluding any subsequent investigations, treatments, or management strategies that followed.

As GPT-4 Turbo has a knowledge base trained up to April 2023 [29], there is potential bias stemming from the inclusion of case reports that might have been part of the LLM’s pre-training data. Hence a personal clinical scenario that was not published on the internet was included. This complex case was derived from the critical care attending’s personal experience where cognitive biases had resulted in wrong and delayed diagnosis. A concise summary of these clinical scenarios, including the unique case, is provided in Multimedia Appendix 1.

Multi-Agent Conversation Framework

In this study, we used the multi-agent conversation framework provided by AutoGen [22] to assess its efficacy in mitigating cognitive biases in clinical decision-making. Within the system, each agent interacts based on their predefined role prompts, thereby simulating the collaborative decision-making process typically observed among health care professionals.

The suggested optimal group size to facilitate group discussion and performance has been proposed to be between 3 and 5 [30]. In the absence of established literature recommending an optimal team size for mitigating cognitive biases in medical settings, we constructed a simulation using 3 to 4 different agents, representing a typical clinical team composition [31]. These configurations aim to realistically emulate the dynamics of clinical teams and their potential to reduce cognitive biases.

As shown in Table 1, we tested 3 different frameworks to identify the most effective configuration. Framework 3 consisted of 3 agents (Junior Resident I, Junior Resident II, and Recorder), while Frameworks 4 (Junior Resident I, Junior Resident II, Professional Expert, and Recorder) and 4-C (Junior Resident I, Junior Resident II, Senior Doctor, and Recorder) each utilized 4 agents, with the fourth agent playing different roles. The distinguishing feature of Framework 4-C, in contrast to other frameworks, lies in its explicit directive for the Senior Doctor role to engage in discussions specifically focused on cognitive biases alongside the initial diagnosis. In addition, we experimented with combinations involving 5 or more agents, but the fifth agent did not effectively participate in the conversations despite modifications to the prompts. Consequently, the frameworks were limited to a maximum of 4 agents. All prompts for agent roles can be found in Multimedia Appendix 2.

Table 1. Different roles in the multi-agent conversation framework. Framework 3 consists of 3 agents, and Frameworks 4 and 4-C consist of 4 agents each.
Agents presentRole descriptionsMulti-agent framework


344-C
Junior Resident ITo make the final diagnosis after considering the discussions
Junior Resident IIThe devil’s advocate and correct confirmation and anchoring bias
Professional ExpertThe field expert in any subspecialization required (eg, radiologists and cardiologists)

Senior DoctorThe tutor and facilitator of the discussion to reduce premature closure bias

RecorderTo record and summarize the findings

The diagnostic process was orchestrated through the collaborative efforts of simulated medical professionals (agents) with varying levels of expertise, as shown in Figure 1. Junior Resident I, as the primary physician, was tasked with presenting the initial diagnosis. Junior Resident I was given the personality of making swift assumptions but is willing to embrace feedback and consider alternative diagnostic possibilities. After the group discussion, Junior Resident I is then allowed to reconsider the most probable differential diagnosis along with an alternative. Junior Resident II, a colleague of Junior Resident I, critically appraised the initial diagnosis, pinpointing inconsistencies and advocating for alternative differential diagnoses. This role was instrumental in addressing potential confirmation and anchoring biases in the diagnostic process. Complementing the juniors, the Senior Doctor brought in-depth experience to the table, crucially identifying cognitive biases in the initial diagnosis and steering the junior residents toward a more nuanced and accurate diagnosis, while the Professional Expert aims to provide any specialist knowledge required to help with the diagnosis without further encouraging discussions of cognitive biases. This guidance was vital in circumventing premature diagnostic closure and knowledge bias. In addition, the role of the Recorder was to distill the outcomes of the discussion, compiling a definitive list of differential diagnoses and extracting key learning points from the collaborative effort, thereby enriching the diagnostic process with a comprehensive and multifaceted approach.

Figure 1. Different roles in the multi-agent conversation framework.

Diagnostic Accuracy Assessment

The final accurate diagnoses for the clinical scenarios were extracted directly from the published cases. Summarized answers from both the multi-agent framework and human evaluators were marked as “Correct” if they matched the final accurate diagnoses. Vague answers, such as diagnosing “septic shock” when the accurate diagnosis was “endometriosis,” were marked as “Incorrect.” A total of 2 physicians graded the answers. In cases where there were discrepancies in their assessments, discussions were held to reach a consensus.

The overall performance of the framework was evaluated based on the accuracy of (1) the “initial diagnosis” made without any multi-agent discussions and (2) the “final diagnosis” after the discussions. Each clinical scenario within each multi-agent framework was simulated 5 times to assess the consistency of diagnoses across multiple iterations.

A total of 3 doctors, each with at least 5 years of clinical experience, were asked to provide their top differential diagnosis along with 2 other differentials based on the clinical scenarios. If the top differential was correct, both the “initial diagnosis” and “final diagnosis” were marked as correct. The diagnoses generated by the multi-agent framework were compared to those provided by human doctors using Fisher exact test.

Bias Identification and Mitigation

An integral part of the evaluation involved documenting the specific cognitive biases identified and addressed during the agents’ discussions. This aspect focused on understanding how effectively the multi-agent system could recognize and mitigate cognitive biases, which are crucial factors in diagnostic accuracy. Hallucinations are characterized by the dissemination of false medical information during multi-agent conversations or responses that fail to directly address the queries posed. This determination was made following an independent review by 2 doctors who thoroughly evaluated the provided answers. The interaction and decision-making process among the agents are illustrated in (Figure 2). This representation aids in visualizing the dynamics of the simulation and the interplay between different agents in reaching a diagnosis.

Figure 2. Schematic illustration of multi-agent (Framework 4-C) discussion dynamics leading to accurate differential diagnosis. ED: emergency department; LLM: large language model.

Overview

A comprehensive search of the PubMed database yielded 162 case reports, of which 37 were determined to be eligible for inclusion in the study. From this subset, 15 cases were selected for evaluation as a representative sampling. In addition, a 16th scenario, derived from critical care attending personal clinical experience, was included. The flow diagram can be viewed in Figure 3.

Figure 3. Flow diagram for identification of case reports with cognitive bias.

Overall Performance of Multi-Agent Conversation Framework

A total of 240 responses were generated by the multi-agent conversation framework, encompassing both initial and final diagnoses following discussions with the agents. The initial diagnosis made by the first-responder agent had a correctness rate of 0% (0/80) across all 3 multi-agent frameworks, whereas the human-generated answers had an accuracy rate of 27% (13/48). After the multi-agent discussions, the final diagnosis was correct in 76% (61/80) of cases in Framework 4-C, which was significantly better than the human answers (odds ratio 3.49; P=.002), as shown in Table 2.

Table 2. Number of correct responses across different multi-agent frameworks and humans.
Multi-agent frameworkInitial diagnosis, n (%)Final diagnosis, n (%)Odds ratioaP valuea
3 (n=80)0 (0)51 (64)1.91.10
4 (n=80)0 (0)55 (69)2.39.03
4-C (n=80)0 (0)61 (76)3.49.002
Human (n=48)13 (27)23 (48)b

aFisher exact test.

bNot applicable.

Clinical Scenarios

The clinical cases covered a broad spectrum of medical fields, ranging from pediatric to malignancy diagnosis. Specifically, 6 cases were centered on infectious disease diagnosis, 3 pertained to critical care, and 2 involved vascular-related diagnoses. The rest were diverse, spanning neurology, gynecology, cancer, urology, and endocrinology (Table 3). A notable aspect of these cases was the cross-disciplinary nature of most initial and final diagnoses, observed in 12 (75%) out of the 16 cases. For example, in one illustrative case (case 1 [32]), an older adult patient presented with symptoms of shortness of breath on exertion and cough, leading to an initial misdiagnosis of heart failure. However, further investigation, considering her ongoing treatment with infliximab for rheumatoid arthritis, revealed the actual diagnosis of miliary tuberculosis. Another case (case 12 [33]) involved a young woman presenting with sudden, left-sided sharp pleuritic chest pain, which lessened when sitting forward. Despite the initial chest radiograph being interpreted as showing no acute abnormalities, the AutoGen system, provided with this potentially misleading information, initially diagnosed the condition as a pulmonary embolism. Yet, after a thorough discussion and re-evaluation, the correct diagnosis of pneumothorax was established, indicating a missed finding in the chest radiograph.

Table 3. Clinical scenarios with the initial wrong and the final correct diagnosis are given in the scenarios.
ReferencePublication yearInitial wrong diagnosisFinal correct diagnosis
[32]2015Heart failureMiliary tuberculosis
[34]2017Bronchial asthma triggered by bacterial pneumoniaHeart failure secondary to dilated cardiomyopathy
[35]2018Syndrome of inappropriate secretion of antidiuretic hormoneNonfunctioning macropituitary adenoma causing adrenal insufficiency
[36]2019Headache and neck pain—migraines and muscle strainCryptococcal meningitis
[37]2020Complex regional pain syndrome flareLeft common and external iliac artery occlusion
[38]2022Pelvic inflammatory diseaseAtypical ectopic pregnancy
[39]2022COVID-19 pneumoniaeBacterial pneumonia (legionella pneumoniae)
[40]2022Urinary tract infection with complicated pyelonephritisVesicointestinal fistula due to Crohn disease
[41]2022Bone (sternum) tuberculosisSyphilitic gumma and osteomyelitis
[42]2022Urinary tract infectionVertebral osteomyelitis and bilateral psoas and retroperitoneal abscesses
[43]2022Anaphylaxis secondary to hennaSuperior vena cava syndrome
[33]2023Pulmonary embolismPneumothorax
[44]2023Diabetic ketoacidosis with infectionsThiamine deficiency
[45]2023EndometritisIschemic bowel
[46]2023Acute myocarditis likely due to MIS-CaAcute myocarditis caused by invasive bacterial infection
bCongestive cardiac failureMalignancy

aMIS-C: multisystem inflammatory syndrome in children.

bNot applicable.

Consistency of Multi-Agent Conversation Framework

There were variations observed in the repeated scenarios, particularly in the process of generating the top 2 differential diagnoses. For instance, in case 13 [44], a young patient presenting with lactic acidosis was initially diagnosed with diabetic ketoacidosis, and further discussions within the multi-agent environment led to the identification of thiamine deficiency. In 3 (60%) out of 5 simulations, Junior Resident I identified thiamine deficiency as the top differential diagnosis following the multi-agent discussions. In the remaining 2 simulations, gastrointestinal disorders were initially considered the most likely diagnosis, with thiamine deficiency being the second most likely differential.

The multi-agent discussions led to the correct final diagnosis in 13 (81%) out of the 16 scenarios across all 3 multi-agent frameworks. Furthermore, these discussions were effective in identifying various clinical biases, including anchoring bias, confirmation bias, availability bias, and premature closure. A detailed breakdown of the answers and the cognitive biases identified is provided in Multimedia Appendix 1. The detailed breakdown of correct answers for each scenario is available in Multimedia Appendix 3.


Principal Findings

This study assessed the effectiveness of the multi-agent conversation framework in improving diagnostic accuracy and mitigating cognitive biases in clinical decision-making. Our findings reveal that integrating multi-agent discussions substantially enhances diagnostic accuracy. The best-performing framework, which used 4 agents, included 1 agent specifically tasked with identifying cognitive biases. This 4-C multi-agent framework that includes discussions on cognitive biases performed significantly better compared with human-generated answers. However, it is important to note that while 4 agents performed best in our study setting, this may vary in general applications.

The increase in diagnostic accuracy demonstrates the value of multi-perspective analysis in medical diagnosis, a core feature of the multi-agent conversation environment. This is in line with previous research emphasizing the importance of collaborative decision-making in health care to mitigate individual cognitive biases [47]. The consistency of responses in the repeated scenarios further validates its reliability and potential applicability in real-world clinical settings.

The effectiveness of the multi-agent conversation system, particularly in scenarios involving misleading or misinterpreted initial investigations, is noteworthy. This was exemplified in case 12, where our multi-agent framework successfully identified a pneumothorax that had been initially overlooked by human clinicians, and improved the accuracy of the final diagnosis to 100% (5/5) in multi-agent framework 4-C. The multi-agent systems were able to critically examine and question potential misinterpretations. Such capabilities are crucial in refining the diagnostic process and enhancing accuracy. While Brown et al [33] discussed the role of AI in the identification of pneumothorax, there is a potential for pre-trained LLMs to replace the decision aid, rather than to develop new resource-intensive systems such as image deep learning. This strategy aligns with the current trajectory of AI development in health care, where the focus is on integrating and maximizing existing AI technologies to enhance clinical decision-making and focus on sustainable AI [48,49].

The integration of multi-agent frameworks in clinical practice holds promising implications for enhancing diagnostic accuracy and ultimately improving patient outcomes. By systematically addressing cognitive biases through collaborative discussions among agents, these frameworks offer a structured approach to refining diagnostic reasoning. This approach not only complements traditional diagnostic methods but also introduces a dynamic element that challenges and verifies initial clinical hypotheses. In practical terms, the ability of multi-agent systems to consistently identify and correct potential diagnostic errors, as demonstrated in our study, suggests a transformative potential in reducing patient morbidity and optimizing treatment strategies. Furthermore, the application of these frameworks within electronic medical records (EMRs) could revolutionize decision-making processes by providing real-time, data-driven insights that augment clinician judgment and ensure adherence to best practices. As health care systems evolve toward more integrated and technology-driven approaches, the strategic incorporation of multi-agent systems stands poised to contribute significantly to improving the quality and efficiency of patient care delivery.

The results of our study extend beyond the educational benefits of multi-agents, highlighting their potential for broader clinical integration. The reflective process fostered by engaging with LLMs in diagnosing and revising cases not only cultivates an educational atmosphere conducive to developing critical thinking skills but also suggests practical applications in clinical settings [50]. One notable avenue is the integration of multi-agent into EMR systems. This could enhance decision-making processes by providing real-time, data-driven insights and augmenting the cognitive capabilities of medical professionals. Such integration would not only streamline the diagnostic process but also aid in the identification of potential cognitive biases, thereby enhancing the quality of patient care. Furthermore, the incorporation of multi-agents in EMRs could facilitate continuous learning and improvement, ensuring that medical practitioners remain updated with the latest medical knowledge and best practices, crucial for maintaining high standards in patient treatment and care.

Limitations

The study, while providing valuable insights into the potential application of multi-agents in clinical diagnostics, is subject to several limitations. First, the reliance on published case reports limits the breadth of clinical scenarios, potentially affecting the generalizability of our findings to broader medical practice. Second, the exclusion of visual data, such as medical imaging, confines our model’s diagnostic capabilities to text-based information, omitting a critical component of clinical diagnosis. In addition, the inherent biases present in the LLMs, based on their pre-training data, could have influenced the diagnostic suggestions. Meanwhile, the technical limitations inherent in LLMs, including their understanding of complex medical terminologies and nuances [51], may not match the expertise of experienced clinicians, possibly limiting the scope of applicability.

Future studies could assess the effectiveness and adaptability of the multi-agent framework in evolving clinical scenarios. More importantly, while LLMs have demonstrated potential as a valuable clinical aid in correcting cognitive biases, the implementation of such technology in health care necessitates rigorous ethical and regulatory oversight [52] and should continue to augment rather than replace the human clinician’s expertise [14].

Conclusions

In conclusion, our study highlights the potential of multi-agents in enhancing diagnostic accuracy in clinical decision-making. The findings support the integration of advanced generative AI technology in educational and clinical settings as a tool for augmenting human decision-making. Future research should focus on the application of these systems in real-time clinical environments and their impact on patient outcomes. Moreover, ethical and legal considerations regarding the use of AI in health care need continued exploration to ensure patient safety and professional accountability.

Acknowledgments

This study was supported by the Duke-NUS Signature Research Program funded by the Ministry of Health, Singapore. Any opinions, findings conclusions, or recommendations expressed in this material are those of the author or authors and do not reflect the views of the Ministry of Health.

Authors' Contributions

YK and SAL conceived the study. YHK, RY, DSWT, and NL designed the study. YK and RY drafted the manuscript with critical appraisal and further development by DSWT, NL. YHK, TXYL, YN, RY, and IL conducted data analyses. NL supervised the study. All authors contributed to the revision of the manuscript and approval of the final version.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Summary of clinical scenarios and the initial and final correct diagnosis, as well as examples of cognitive biases present.

DOCX File , 25 KB

Multimedia Appendix 2

Prompts for different agent roles.

DOCX File , 16 KB

Multimedia Appendix 3

Detailed breakdown of the correct answers within each scenario stratified by each multi-agent framework and human.

DOCX File , 18 KB

  1. Saposnik G, Redelmeier D, Ruff CC, Tobler PN. Cognitive biases associated with medical decisions: a systematic review. BMC Med Inform Decis Mak. 2016;16(1):138. [FREE Full text] [CrossRef] [Medline]
  2. Korteling JEH, Paradies GL, Sassen-van Meer JP. Cognitive bias and how to improve sustainable decision making. Front Psychol. 2023;14:1129835. [CrossRef] [Medline]
  3. Berthet V. The impact of cognitive biases on professionals' decision-making: a review of four occupational areas. Front Psychol. 2021;12:802439. [CrossRef] [Medline]
  4. Beldhuis IE, Marapin RS, Jiang YY, Simões de Souza NF, Georgiou A, Kaufmann T, et al. Cognitive biases, environmental, patient and personal factors associated with critical care decision making: a scoping review. J Crit Care. 2021;64:144-153. [CrossRef] [Medline]
  5. Doherty TS, Carroll AE. Believing in overcoming cognitive biases. AMA J Ethics. 2020;22(9):E773-E778. [CrossRef] [Medline]
  6. Hershberger PJ, Markert RJ, Part HM, Cohen SM, Finger WW. Understanding and addressing cognitive bias in medical education. Adv Health Sci Educ Theory Pract. 1996;1(3):221-226. [CrossRef] [Medline]
  7. Yang R, Tan TF, Lu W, Thirunavukarasu AJ, Ting DSW, Liu N. Large language models in health care: development, applications, and challenges. Health Care Sci. 2023;2(4):255-263. [CrossRef] [Medline]
  8. Nori H, Lee YT, Zhang S, Carignan D, Edgar R, Fusi N, et al. Can generalist foundation models outcompete special-purpose tuning? Case study in medicine. arXiv. Preprint posted online on November 28, 2023. [FREE Full text] [CrossRef]
  9. Yang R, Liu H, Marrese-Taylor E, Zeng Q, Ke YH, Li W, et al. KG-Rank: enhancing large language models for medical QA with knowledge graphs and ranking techniques. arXiv. Preprint posted online on March 9, 2024. [FREE Full text] [CrossRef]
  10. Kung TH, Cheatham M, Medenilla A, Sillos C, de Leon L, Elepaño C, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. 2023;2(2):e0000198. [CrossRef] [Medline]
  11. Lim DYZ, Ke YH, Sng GGR, Tung JYM, Chai JX, Abdullah HR. Large language models in anaesthesiology: use of ChatGPT for American Society of Anesthesiologists Physical status classification. Br J Anaesth. 2023;131(3):e73-e75. [CrossRef] [Medline]
  12. Wang H, Fu T, Du Y, Gao W, Huang K, Liu Z, et al. Scientific discovery in the age of artificial intelligence. Nature. 2023;620(7972):47-60. [CrossRef] [Medline]
  13. Karabacak M, Margetis K. Embracing large language models for medical applications: opportunities and challenges. Cureus. 2023;15(5):e39305. [CrossRef] [Medline]
  14. Singhal K, Azizi S, Tu T, Mahdavi SS, Wei J, Chung HW, et al. Large language models encode clinical knowledge. Nature. 2023;620(7972):172-180. [CrossRef] [Medline]
  15. Harrer S. Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine. EBioMedicine. 2023;90:104512. [CrossRef] [Medline]
  16. Yang R, Ning Y, Keppo E, Liu M, Hong C, Bitterman DS, et al. Retrieval-augmented generation for generative artificial intelligence in medicine. arXiv. Preprint posted online on June 18, 2024. [FREE Full text] [CrossRef]
  17. Berg HT, van Bakel B, van de Wouw L, Jie KE, Schipper A, Jansen H, et al. ChatGPT and generating a differential diagnosis early in an emergency department presentation. Ann Emerg Med. 2024;83(1):83-86. [CrossRef] [Medline]
  18. Ke Y, Yang R, Liu N. Comparing open-access database and traditional intensive care studies using machine learning: bibliometric analysis study. J Med Internet Res. 2024;26:e48330. [CrossRef] [Medline]
  19. Clusmann J, Kolbinger FR, Muti HS, Carrero ZI, Eckardt JN, Laleh NG, et al. The future landscape of large language models in medicine. Commun Med (Lond). 2023;3(1):141. [CrossRef] [Medline]
  20. Satya-Murti S, Lockhart J. Recognizing and reducing cognitive bias in clinical and forensic neurology. Neurol Clin Pract. 2015;5(5):389-396. [CrossRef] [Medline]
  21. Nascimento N, Alencar P, Cowan D. Self-adaptive large language model (LLM)-based multiagent systems. 2023. Presented at: 2023 IEEE International Conference on Autonomic Computing and Self-Organizing Systems Companion (ACSOS-C); September 25-29, 2023:104-109; Toronto, ON. URL: http://arxiv.org/abs/2307.06187 [CrossRef]
  22. Wu Q, Bansal G, Zhang J, Wu Y, Li B, Zhu E, et al. AutoGen: enabling next-gen LLM applications via multi-agent conversation. arXiv. Preprint posted online on August 16, 2023. [FREE Full text] [CrossRef]
  23. Wang Z, Zhang G, Yang K, Shi N, Zhou W, Hao S, et al. Interactive natural language processing. arXiv. Preprint posted online on May 22, 2023. [FREE Full text] [CrossRef]
  24. Shanahan M, McDonell K, Reynolds L. Role play with large language models. Nature. Nov 2023;623(7987):493-498. [FREE Full text] [CrossRef] [Medline]
  25. Légaré F, Adekpedjou R, Stacey D, Turcotte S, Kryworuchko J, Graham ID, et al. Interventions for increasing the use of shared decision making by healthcare professionals. Cochrane Database Syst Rev. 2018;7(7):CD006732. [CrossRef] [Medline]
  26. Vela MB, Erondu AI, Smith NA, Peek ME, Woodruff JN, Chin MH. Eliminating explicit and implicit biases in health care: evidence and research needs. Annu Rev Public Health. 2022;43:477-501. [FREE Full text] [CrossRef] [Medline]
  27. ChatGPT. OpenAI. URL: https://chat.openai.com/ [accessed 2023-09-16]
  28. National Organization for Rare Disorders. URL: https://rarediseases.org/ [accessed 2023-11-16]
  29. GPT-4. OpenAI. URL: https://openai.com/research/gpt-4 [accessed 2024-01-01]
  30. Hackman JR, Vidmar N. Effects of size and task type on group performance and member reactions. Sociometry. 1970;33(1):37-54. [CrossRef]
  31. Rosen MA, DiazGranados D, Dietz AS, Benishek LE, Thompson D, Pronovost PJ, et al. Teamwork in healthcare: Key discoveries enabling safer, high-quality care. Am Psychol. 2018;73(4):433-450. [FREE Full text] [CrossRef] [Medline]
  32. Mull N, Reilly JB, Myers JS. An elderly woman with 'heart failure': cognitive biases and diagnostic error. Cleve Clin J Med. 2015;82(11):745-753. [FREE Full text] [CrossRef] [Medline]
  33. Brown C, Nazeer R, Gibbs A, Le Page P, Mitchell AR. Breaking bias: the role of artificial intelligence in improving clinical decision-making. Cureus. 2023;15(3):e36415. [FREE Full text] [CrossRef] [Medline]
  34. Tetsuhara K, Tsuji S, Nakano K, Kubota M. Case report: heart failure in dilated cardiomyopathy mimicking asthma triggered by pneumonia. BMJ Case Rep. 2017;2017:bcr2017222082. [FREE Full text] [CrossRef] [Medline]
  35. Ilaiwy A, Thompson NE, Waheed AA. Adenoma mimicking hyponatremia of SIAD. BMJ Case Rep. 2018;2018:bcr2018226458. [FREE Full text] [CrossRef] [Medline]
  36. Deming M, Mark A, Nyemba V, Heil EL, Palmeiro RM, Schmalzle SA. Cognitive biases and knowledge deficits leading to delayed recognition of cryptococcal meningitis. IDCases. 2019;18:e00588. [FREE Full text] [CrossRef] [Medline]
  37. Khawaja H, Font C. Common and external iliac artery occlusion in Behçet's disease: a case of anchoring bias. BMJ Case Rep. 2020;13(12):e236554. [FREE Full text] [CrossRef] [Medline]
  38. Birch EM, Torres Molina M, Oliver JJ. Not like the textbook: an atypical case of ectopic pregnancy. Cureus. 2022;14(10):e29881. [FREE Full text] [CrossRef] [Medline]
  39. Kyere K, Aremu TO, Ajibola OA. Availability bias and the COVID-19 pandemic: a case study of legionella pneumonia. Cureus. 2022;14(6):e25846. [FREE Full text] [CrossRef] [Medline]
  40. Miyagami T, Nakayama I, Naito T. What causes diagnostic errors? referred patients and our own cognitive biases: a case report. Am J Case Rep. 2022;23:e935163. [FREE Full text] [CrossRef] [Medline]
  41. Kamegai K, Yokoyama S, Takakura S, Takayama Y, Shiiki S, Koyama H, et al. Syphilitic osteomyelitis in a patient with HIV and cognitive biases in clinical reasoning: a case report. Medicine (Baltimore). 2022;101(40):e30733. [FREE Full text] [CrossRef] [Medline]
  42. Kawahigashi T, Harada Y, Watari T, Harada T, Miyagami T, Shikino K, et al. Missed opportunities for diagnosing vertebral osteomyelitis caused by influential cognitive biases. Am J Case Rep. 2022;23:e936058. [FREE Full text] [CrossRef] [Medline]
  43. Salama ME, Ukwade P, Khan AR, Qayyum H. Facial swelling mimicking anaphylaxis: a case of superior vena cava syndrome in the emergency department. Cureus. 2022;14(9):e29678. [FREE Full text] [CrossRef] [Medline]
  44. Chehayeb RJ, Ilagan-Ying YC, Sankey C. Addressing cognitive biases in interpreting an elevated lactate in a patient with type 1 diabetes and thiamine deficiency. J Gen Intern Med. 2023;38(6):1547-1551. [FREE Full text] [CrossRef] [Medline]
  45. Vittorelli J, Cacchillo J, McCool M, McCague A. Cognitive bias in the management of a critically Ill 29-year-old patient. Cureus. 2023;15(5):e39314. [CrossRef] [Medline]
  46. Stanzelova A, Debray A, Allali S, Belhadjer Z, Taha MK, Cohen JF, et al. Severe bacterial infection initially misdiagnosed as MIS-C: caution needed. Pediatr Infect Dis J. 2023;42(6):e201-e203. [CrossRef] [Medline]
  47. Lieder F, Griffiths TL, M Huys QJ, Goodman ND. The anchoring bias reflects rational use of cognitive resources. Psychon Bull Rev. 2018;25(1):322-349. [CrossRef] [Medline]
  48. Richie C. Environmentally sustainable development and use of artificial intelligence in health care. Bioethics. 2022;36(5):547-555. [FREE Full text] [CrossRef] [Medline]
  49. Liu M, Ning Y, Teixayavong S, Mertens M, Xu J, Ting DSW, et al. A translational perspective towards clinical AI fairness. NPJ Digit Med. 2023;6(1):172. [FREE Full text] [CrossRef] [Medline]
  50. Papathanasiou IV, Kleisiaris CF, Fradelos EC, Kakou K, Kourkouta L. Critical thinking: the development of an essential skill for nursing students. Acta Inform Med. 2014;22(4):283-286. [FREE Full text] [CrossRef] [Medline]
  51. Yang R, Zeng Q, You K, Qiao Y, Huang L, Hsieh CC, et al. Ascle-a Python natural language processing toolkit for medical text generation: development and evaluation study. J Med Internet Res. Oct 03, 2024;26:e60601. [FREE Full text] [CrossRef] [Medline]
  52. Meskó B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digit Med. 2023;6(1):120. [FREE Full text] [CrossRef] [Medline]


AI: artificial intelligence
EMR: electronic medical record
LLM: large language model


Edited by A Schwartz; submitted 12.04.24; peer-reviewed by WHK Chiu, E Amini-Salehi; comments to author 12.05.24; revised version received 21.06.24; accepted 12.09.24; published 19.11.24.

Copyright

©Yuhe Ke, Rui Yang, Sui An Lie, Taylor Xin Yi Lim, Yilin Ning, Irene Li, Hairil Rizal Abdullah, Daniel Shu Wei Ting, Nan Liu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 19.11.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.