Skip to main content

MINI REVIEW article

Front. Cardiovasc. Med., 04 July 2024
Sec. Pediatric Cardiology
This article is part of the Research Topic Advances in Understanding and Managing Congenital Heart Disease: From Bench to Bedside View all 5 articles

Cognitive biases in pediatric cardiac care

\r\nPaul Padovani,
Paul Padovani1,2*Arnaud Roy,Arnaud Roy3,4Amanda GuerraAmanda Guerra3Olivier CadeauOlivier Cadeau3Mohamed LyMohamed Ly1Corina M. VasileCorina M. Vasile5Robert H. PassRobert H. Pass6Alban-Elouen Baruteau,,,\r\nAlban-Elouen Baruteau1,2,7,8
  • 1CHU Nantes, Department of Pediatric Cardiology and Pediatric Cardiac Surgery, FHU PRECICARE, Nantes Université, Nantes, France
  • 2CHU Nantes, INSERM, CIC FEA 1413, Nantes Université, Nantes, France
  • 3LPPL, SFR Confluences, Nantes Université, Université d’Angers, Angers, France
  • 4CHU Nantes, Centre Référent des Troubles d’Apprentissage, Nantes Université, Nantes, France
  • 5Pediatrics Department at Filantropia Municipal Hospital of Craiova, Craiova, Romania
  • 6Department of Pediatric Cardiology, Mount Sinai Kravis Children’s Hospital, New York, NY, United States
  • 7CHU Nantes, CNRS, INSERM, l’institut du Thorax, Nantes Université, Nantes, France
  • 8INRAE, UMR 1280, PhAN, Nantes Université, Nantes, France

Medical practitioners are entrusted with the pivotal task of making optimal decisions in healthcare delivery. Despite rigorous training, our confidence in reasoning can fail when faced with pressures, uncertainties, urgencies, difficulties, and occasional errors. Day-to-day decisions rely on swift, intuitive cognitive processes known as heuristic or type 1 decision-making, which, while efficient in most scenarios, harbor inherent vulnerabilities leading to systematic errors. Cognitive biases receive limited explicit discussion during our training as junior doctors in the domain of paediatric cardiology. As pediatric cardiologists, we frequently confront emergencies necessitating rapid decision-making, while contending with the pressures of stress, fatigue, an earnest interest in “doing the right thing” and the impact of parental involvement. This article aims to describe cognitive biases in pediatric cardiology, highlighting their influence on therapeutic interventions for congenital heart disease. Whether future pediatric cardiologists or experienced professionals, understanding and actively combating cognitive biases are essential components of our ongoing medical education. Furthermore, it is our responsibility to thoroughly examine our own practices in our unwavering commitment to providing high-quality care.

1 Introduction

Type 1 processing, also known as intuitive decision-making or heuristic reasoning, refers to a rapid and automatic thinking process. It is characterized by the utilization of mental heuristics, or simple rules, to swiftly reach conclusions and make decisions without requiring in-depth analysis of available information (1). In addition to Type 1 processing, there exists Type 2 processing, characterized by slower, more deliberate analysis of information and consideration of multiple factors (2). Despite the essential nature of Type 1 processing (3), it can lead to systematic thinking errors, called cognitive biases. A cognitive bias is a systematic deviation from rationality or objectivity in judgment or decision-making, often stemming from mental shortcuts, perceptual distortions, or subjective influences. Importantly, these biases do not correlate with intelligence or cognitive ability (4, 5). They arise from a variety of sources including acquired or inherent predispositions, societal and cultural influences, deficits in statistical understanding and mathematical reasoning and, in particular, and environmental stimuli that demand our attention (6).

Cognitive biases can impact various facets of our existence, but their significance becomes pronounced when these errors manifest within the context of medical practice. Prevalence of diagnostic error has been estimated to be as high as 10%–15% in daily clinical practice (7). Cognitive factors are the main contributor to diagnostic errors, which are associated with a proportionately higher morbidity than is the case with other types of medical error (8, 9). Contrary to physician interviews that often identify system-related factors (organizational flaws, inadequate policies, staffing or equipment) as the main contributors to diagnostic errors, cognitive factors are more likely the primary driver of such errors (10).

Among medical specialties, paediatrics stands out as one where decisions are emotionally demanding, given the significant weight they carry for both parents and children (11, 12). Given the high stakes involved in pediatric care, understanding, and mitigating cognitive biases is paramount.

This work aims to examine the medical reasoning and practice by highlighting several cognitive biases specifically within the field of pediatric cardiology and congenital heart disease (CHD) care.

2 Subsections relevant for the subject

Cognitive biases are prevalent in pediatric cardiology practice and can significantly impact diagnostic decision-making, multidisciplinary collaboration, and technical procedures.

2.1 Diagnostic decision-making

This section delves into cognitive biases that influence diagnostic decision-making such as the availability bias, the anchoring bias, the attrition bias, the confirmation bias, overconfidence, and search satisfying.

Cognitive biases have been identified in all steps of decision making (1315), including information gathering, association triggering, context formulation, processing and verification (16). The diagnostic enterprise, construed as the distinctive characterisation of a specific disease or condition, hinges on factors including etiopathogenesis, parental interrogation, children signs, symptoms, physical examination results, diagnostic tests, and health history. Clinical expertise accrues through domain-based practice, augmented experience, enhanced knowledge and skills, and the development of domain-specific intuitive capacities. In addition to the advanced cognitive functions such as problem-solving, judgment, and decision-making mentioned earlier, formulating a diagnosis also relies on the utilization of social and emotional resources. Contemporary understanding of cognition underscores the dynamic interaction between cognitive processes and socioemotional factors in decision-making contexts. This includes the impact of emotional states, interpersonal dynamics, and cultural influences on cognitive functioning (17, 18). Expert clinicians find it easier to effectively focus attention on and evaluate details of the infant's clinical problem. This enables them to intuitively and deliberately generate several relevant differential diagnoses and potential strategies to address the identified clinical situation.

Nonetheless, experience can be a source of bias, such as availability bias (1921) or base rate neglect (22) (see Figure 1). Physicians often tend to perceive things as more likely if they readily come to mind: “common things are common (availability)”. They may form diagnostic hypotheses based on recent exposure to another patient with the same illness, particularly if the case was particularly notable or emotionally charged. Conversely, prolonged absence in encountering a disease tends to decrease the likelihood of considering it during diagnosis: “out of sight out of mind (non-availability)”. The bias of restricting representativeness is ingrained in the traditions of medical education (23). Our diagnostic approach often leads us to search for prototypical presentations of a disease: adhering to the adage “When you hear hoofbeats, think of horses rather than zebras”. These entrenched dogmas and mindsets tend to exclude rare diseases from consideration in the diagnostic process. This can be particularly detrimental in the field of CHD, which may present infrequently in emergency room consultations despite being relatively common, affecting 8 out of 1,000 births (24). Base rate neglect on the other hand may result in overestimates of unlikely diagnoses. This bias occurs when individuals fail to consider statistical or base rates in decision-making, instead prioritizing specific information or individual cases.

Figure 1
www.frontiersin.org

Figure 1 Clinical examples of cognitive biases during pediatric cardiac care.

Another potential bias, notably observed in emergency departments (ED), is anchoring bias whereby physicians may strongly adhere to their initial impressions, even when confronted with a substantial amount of conflicting data, particularly if they harbor certain preconceptions (23, 25). Anchoring bias is linked to attrition bias which is attempts to discover reason for observations (26). Stereotyping and gender bias are both good examples of attrition bias. In the ED, there is often significant time pressure to make swift decisions. This urgency is particularly concerning in the context of undiagnosed CHD, where achieving accurate diagnoses may necessitate thorough and careful consideration in addition to multiple examinations such as electrocardiography, chest radiographs, echocardiography, and blood tests.

Compared with an expert, novice specialists will likely make more diagnostic errors (18). Formulating a specific diagnosis largely relies on pattern recognition, the clinician's familiarity with similar clinical scenarios, and the ease with which relevant past cases come to mind. Confirmation Bias, Overconfidence and Search Satisfying are common bias for novice specialists (27). Confirmation bias, the selective acceptance of clinical data supporting a desired hypothesis while disregarding contradictory data, significantly heightens the risk of diagnostic errors (28). Overconfidence, also known as the Dunning-Kruger effect, describes the tendency for individuals lacking expertise in a given field to overestimate their proficiency, leading them to believe they possess more knowledge than they actually do (29, 30). This can lead to noncompliance with evidence-based guidelines. Additionally, search satisfying occurs when clinicians cease seeking additional information once they find a solution that appears satisfactory, even if it may not be the most appropriate or accurate (14). Unlike confirmation bias, where clinicians tend to seek evidence to confirm their initial hypotheses while ignoring contradictory evidence, search satisfying focuses more on the satisfaction of finding a plausible solution, even if it is not validated by additional evidence or thorough consideration. Paradoxically, qualified individuals tend to underestimate their own abilities. In other words and restated, inexperience may prevent people from recognising their own limitations and those who are more highly qualified and experienced may paradoxically underestimate the extent of their knowledge.

Another well-known bias is premature closure which is a form of bias in which we tend to be satisfied with a plausible hypothesis derived from our own experience and dismiss the possibility of uncertainty (31). Moreover, when contemplating the dynamics of the doctor-parent relationship, particularly within the nuanced context of paediatrics, a salient cognitive bias that warrants consideration is the outcome bias. This cognitive inclination manifests as a predisposition towards diagnostic decisions that are anticipated to yield positive outcomes, consciously avoiding choices associated with unfavourable consequences (32). Such a cognitive predisposition serves as a protective mechanism, strategically sidestepping the emotional distress that may be entwined with less favourable clinical outcomes. This can lead to minimizing serious diagnoses or hoping for a favourable outcome and may manifest, in pediatric cardiovascular care as the avoidance of surgery or a catheterization procedure, as two examples. Also, countertransference (positive or negative) can lead to an under-assessment of the severity of the child or adolescent's condition, to avoid breaking bad news (33, 34). In an ideal scenario, every clinical decision would be made impartially and consistently across different patients. However, this isn't always the case. Our interactions with patients and families can evoke both favorable and unfavorable sentiments, which might influence the quality of our decisions. Within the context of the ED, instances may arise where a physician experiences positive countertransference towards a patient, potentially exerting an influence on clinical decision-making processes. This phenomenon, driven by outcome bias, could lead to underinvestigation, as decisions favoring positive outcomes may take precedence over those indicating negative outcomes. Consequently, this bias may result in the omission of diagnostic tests critical for identifying unfavorable prognostic implications for the patient.

Added to and closely linked to cognitive biases is the notion of noise which reflects errors in judgement and measurement. Some situations are more prone to biased-reasoning due to this form of error (17). The profound influence of background noise on decision-making, occasionally impeding the activation of System 2 cognition—the reflective and analytical mode of thought—is highlighted by factors including workload overload, concurrent multitasking, task interruptions, fatigue, time pressure, the broader work environment, dysfunctional team dynamics, hyperconnectivity, and various forms of distraction. These distractions may arise from external sources or be influenced by temporal factors such as the day of the week, holidays, or the time of day or night. A constant self-inquiry that warrants consideration is whether the current situation is conducive to the prevalence of biases.

2.2 Multidisciplinary collaboration

Exploring cognitive biases during multidisciplinary decision-making highlights challenges and opportunities for enhancing teamwork. These biases include memory shifting/reconstruction bias, diagnosis momentum bias, framing effect bias, order effect bias, bias of omission, commission bias, conformity bias, authority gradient effect, and hindsight bias

Multidisciplinary consultations and collaborative thinking represent effective strategies for mitigating decisions influenced by rapid reasoning with cognitive biases (3538) (see Figure 1). However, certain inter-human factors may still be biased and affect such decision-making. Memory shifting, also called reconstruction bias, involves the inaccurate recall of information due to variations in coding of meaning and textual information, resulting in the “filling in” of details, sometimes with incorrect information, during memory recall. Diagnosis momentum bias is another concern, whereby diagnostic labels attached to patients tend to become increasingly ingrained over time, leading to the exclusion of alternative possibilities. This bias is closely related to the framing effect bias, which demonstrates how the formulation of a problem can strongly influence perception (39). Additionally, the order effect bias must be considered when presenting cases, as individuals tend to remember the beginning (primacy effect) or the end (recency effect) of a story being told, highlighting the importance of how cases are presented in such conferences (13, 17). This holds particularly true when presenting a case of a child who has undergone multiple surgeries for a complex cardiac disease. The order in which the varied problems and interventions have been presented can have a meaningful and unintentional impact on how the case is discussed and considered.

In medical decision-making, there is a tendency toward inaction, prioritizing the principle of non-maleficence, which leads to the omission bias (32, 36, 40). This bias, influenced by the perceived safety of inaction, can have severe consequences despite maintaining the status quo. Conversely, commission bias, stemming from the obligation toward beneficence, involves a propensity toward action. This bias is more prevalent in overconfident physicians. Discussing these two biases in the context of pediatrics is particularly important, given the unique ethical considerations and potential consequences of either form of bias on the care of pediatric patients.

Moreover, within the intricate dynamics of medical staff discussions, the susceptibility to groupthink or false consensus effect poses a significant challenge (36). The desire for consensus may lead to a conformity bias, wherein dissenting opinions are subdued in favour of achieving unanimous agreement. This can be particularly pronounced in hierarchical medical teams, where junior members may hesitate to challenge prevailing views. Specifically, the phenomenon referred to as the authority gradient effect comes into play in such situations. The authority gradient effect describes the reluctance of junior members to challenge the opinions of senior members within hierarchical structures. Furthermore, the phenomenon of hindsight bias whereby there may be a tendency to perceive something as having been more predictable (e.g., “I knew it all along”) than it truly is while making decisions (41). This retrospective distortion may impact how post-case analyses unfold and potentially impede the recognition of avoidable errors. These behaviours may be conscious and induced by social norms, but often, they are unconscious and characteristic of cognitive biases.

2.3 Technical procedure

Exploring cognitive biases during technical procedures is essential for identifying potential errors and implementing risk-mitigation strategies. Anesthesiologists, surgeons, and cardiac interventionists, encounter various cognitive biases during medical procedures such as anchoring bias, sunk cost fallacy, social desirability bias and the tunnel effect.

Anesthesiologists confront various cognitive biases that can significantly impact decision-making (6, 39). Anchoring bias—the tendency to fixate on specific features -, for instance, becomes pronounced when faced with unexpected challenges during surgery, including hemodynamic variation. For example, anchoring bias becomes prominent in such situations. The initial medication dosage administered acts as a cognitive anchor, disproportionately influencing subsequent decisions. This predisposition may lead to either inadequate or excessive adjustments.

Confronted with unexpected challenges in the operating room or the catheterisation laboratory, physicians (cardiac interventionists, surgeons, or anaesthesiologists) may grapple with the sunk cost fallacy, fostering a reluctance to deviate from initially planned courses of action (42). This psychological bias arises from a perceived investment of time and effort, impeding objective reassessment. Additionally, the dynamics of communication within the anesthesia/nurse/surgery team may exhibit a conformity bias, wherein team members may withhold concerns to maintain a positive atmosphere or avoid contradicting the opinions of others. Of course, physicians during technical care are also vulnerable to many of the previously described biases such as confirmation bias or overconfidence bias amongst others.

Surgeons and cardiac interventionists are highly sensitive to the tunnel effect (42, 43). This refers to a situation in which the physician's attention is so focused on an objective (expected outcome, management of a complication) that he neither hears nor sees the warning signals that should lead him to modify his approach, or even stop it before an undesirable effect occurs. In addition, proceduralists must deal with a variable range of emotions that may influence their decision-making during procedures. For example, anger can influence decisions made by oneself or the team, regret describes the tendency to let regret about past decisions influence future decisions and anticipatory regret is the desire to avoid regret about future consequences or outcomes of decision choices. All of these may negatively impact the proceduralist.

Finally, and somewhat paradoxically, the well-known blind spot bias, whereby individuals tend to recognize biases in others but not in themselves, is observed (15, 17, 36). A blind spot bias may occur when a physician, while evaluating a child with cardiac symptoms, readily identifies and corrects reasoning errors in their colleagues but fails to question their own diagnostic decisions or treatments, even if they are affected by their own similar cognitive biases.

3 Discussion

Identifying physicians’ cognitive biases at an early stage is essential for optimising medical decisions, preventing errors and creating realistic expectations for patients, ultimately reducing the rising costs of healthcare (44). Most cognitive biases probably arise from overuse of System 1 or when System 1 dominates over System 2. Techniques that enhance System 2 could counteract these biases, thereby improving diagnostic accuracy and decreasing management errors (45).

Overconfidence, anchoring bias, and availability bias were prevalent, impacting diagnostic accuracy in 36.5%–77% of case scenarios (46). For example, Mamede et al. found that availability bias increased with years of training, and reflective reasoning improved diagnostic accuracy among internal medicine residents (21). Additionally, biases such as information bias, representativeness bias, and premature closure were associated with diagnostic errors in over half of the evaluated scenarios. These findings highlight the widespread presence of cognitive biases across various medical tasks, from diagnosis to treatment and management (46).

Cognitive biases not only affect diagnostic accuracy but also influence therapeutic and management decisions. For instance, Yee et al. found that better-coping strategies and higher tolerance to ambiguity among obstetricians were associated with lower rates of instrumental vaginal deliveries and fewer management errors (47). They also indicated that higher tolerance to ambiguity among physicians was associated with increased medical complications, such as postpartum hemorrhage. These studies demonstrate how cognitive biases can lead to both over-treatment and under-treatment, illustrating the necessity for balanced decision-making frameworks. This also underscores the potential severity of cognitive biases, emphasizing the need for further research to establish clear links between these biases and patient outcomes (48).

Addressing cognitive biases requires a multi-faceted approach. Increasing awareness among physicians and medical students is crucial, as is incorporating training on cognitive biases into medical education programs. Effective strategies include reflective reasoning, the use of cognitive checklists, and heuristic approaches to simplify decision-making processes. Collaborative efforts from academic institutions, healthcare organizations, and policymakers are needed to implement these strategies and improve healthcare delivery (4951).

4 Conclusion

A significant portion of today's medical decision-making research evolved from studies conducted in the field of cognitive psychology in the late 20th century (52). Regrettably, despite this wealth of accumulated knowledge, medical practitioners still struggle to fully comprehend and integrate these findings into their practice. Grounded in real-world examples and informed by personal experiences, this perspective endeavors to elucidate the intricate relationship between cognitive biases and the landscape of pediatric cardiology. However, this work represents only an initial step toward grasping the complexities of human decision-making dynamics. There is a notable lack of comprehensive research in this arena in pediatrics, prompting critical inquiry into the interaction between cognitive biases and the powerful stresses and emotions experienced in pediatric care. To address these gaps, further research is imperative, along with the development of dedicated protocols and frameworks to help mitigate the impact of these biases on decision-making. Incorporating training or simulations in human factors, including the study of cognitive biases, into the curriculum of young pediatricians across all specialties should be considered as an essential component for enhancing clinical practice and patient care.

Author contributions

PP: Conceptualization, Investigation, Supervision, Writing – original draft. AR: Validation, Writing – review & editing. AG: Validation, Writing – review & editing. OC: Validation, Writing – review & editing. ML: Validation, Writing – review & editing. CMV: Validation, Writing – review & editing. RHP: Supervision, Validation, Writing – review & editing. A-EB: Conceptualization, Supervision, Writing – original draft.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article.

By a research grant from the French Federation of Cardiology (to PP); by a research grant from the Fondation Maladies Rares (to AE-B); by a research grant from the French Government as part of the “Investments of the future” program managed by the National Research Agency, grant reference ANR-16-IDEX-0007 (to AE-B).

Acknowledgments

We gratefully acknowledge GCS HUGO, AVIESAN and FHU PRECICARE.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Smallberg G. Bias is the nose for the story. In: Brockman J, editor. This Will Make you Smarter. New York: Harper Perennial (2012). 43–5.

Google Scholar

2. Mamede S, Schmidt HG. The structure of reflective practice in medicine. Med Educ. (2004) 38(12):1302–8. doi: 10.1111/j.1365-2929.2004.01917.x

PubMed Abstract | Crossref Full Text | Google Scholar

3. Croskerry P. Bias: a normal operating characteristic of the diagnosing brain. Diagnosis (Berl). (2014) 1(1):23–7. doi: 10.1515/dx-2013-0028

PubMed Abstract | Crossref Full Text | Google Scholar

4. Stanovich KE, West RF. On the relative independence of thinking biases and cognitive ability. J Pers Soc Psychol. (2008) 94(4):672–95. doi: 10.1037/0022-3514.94.4.672

PubMed Abstract | Crossref Full Text | Google Scholar

5. Detmer DE, Fryback DG, Gassner K. Heuristics and biases in medical decision-making. J Med Educ. (1978) 53(8):682–3. doi: 10.1097/00001888-197808000-00012

PubMed Abstract | Crossref Full Text | Google Scholar

6. Stiegler MP, Tung A. Cognitive processes in anesthesiology decision making. Anesthesiology. (2014) 120(1):204–17. doi: 10.1097/ALN.0000000000000073

PubMed Abstract | Crossref Full Text | Google Scholar

7. Elstein A. Clinical reasoning in medicine. In: Higgs J, editor. Clinical Reasoning in the Health Professions. Oxford, England: Butterworth-Heinemann Ltd (1995). p. 49–59.

Google Scholar

8. Brennan TA, Leape LL, Laird NM, Hebert L, Localio AR, Lawthers AG, et al. Incidence of adverse events and negligence in hospitalized patients. Results of the Harvard medical practice study I. N Engl J Med. (1991) 324(6):370–6. doi: 10.1056/NEJM199102073240604

PubMed Abstract | Crossref Full Text | Google Scholar

9. Thomas EJ, Studdert DM, Burstin HR, Orav EJ, Zeena T, Williams EJ, et al. Incidence and types of adverse events and negligent care in Utah and Colorado. Med Care. (2000) 38(3):261–71. doi: 10.1097/00005650-200003000-00003

PubMed Abstract | Crossref Full Text | Google Scholar

10. Warrick C, Patel P, Hyer W, Neale G, Sevdalis N, Inwald D. Diagnostic error in children presenting with acute medical illness to a community hospital. Int J Qual Health Care. (2014) 26(5):538–46. doi: 10.1093/intqhc/mzu066

PubMed Abstract | Crossref Full Text | Google Scholar

11. Feeks C, Chao J, Sinert R. Prevalence and risk factors for burnout in pediatric emergency medicine fellows. Pediatr Emerg Care. (2020) 36(5):236–9. doi: 10.1097/PEC.0000000000002093

PubMed Abstract | Crossref Full Text | Google Scholar

12. McClafferty HH, Hubbard DK, Foradori D, Brown ML, Profit J, Tawfik DS, et al. Physician health and wellness. Pediatrics. (2022) 150(5):e2022059665. doi: 10.1542/peds.2022-059665

PubMed Abstract | Crossref Full Text | Google Scholar

13. Croskerry P. The importance of cognitive errors in diagnosis and strategies to minimize them. Acad Med. (2003) 78(8):775–80. doi: 10.1097/00001888-200308000-00003

PubMed Abstract | Crossref Full Text | Google Scholar

14. Croskerry P. Achieving quality in clinical decision making: cognitive strategies and detection of bias. Acad Emerg Med. (2002) 9(11):1184–204. doi: 10.1197/aemj.9.11.1184

PubMed Abstract | Crossref Full Text | Google Scholar

15. Croskerry P. From mindless to mindful practice–cognitive bias and clinical decision making. N Engl J Med. (2013) 368(26):2445–8. doi: 10.1056/NEJMp1303712

PubMed Abstract | Crossref Full Text | Google Scholar

16. Kassirer JP, Kopelman RI. Cognitive errors in diagnosis: instantiation, classification, and consequences. Am J Med. (1989) 86(4):433–41. doi: 10.1016/0002-9343(89)90342-2

PubMed Abstract | Crossref Full Text | Google Scholar

17. Croskerry P, Singhal G, Mamede S. Cognitive debiasing 1: origins of bias and theory of debiasing. BMJ Qual Saf. (2013) 22(Suppl 2):ii58–64. doi: 10.1136/bmjqs-2012-001712

PubMed Abstract | Crossref Full Text | Google Scholar

18. Graber ML. The incidence of diagnostic error in medicine. BMJ Qual Saf. (2013) 22(Suppl 2):ii21–7. doi: 10.1136/bmjqs-2012-001615

PubMed Abstract | Crossref Full Text | Google Scholar

19. Kovacs G, Croskerry P. Clinical decision making: an emergency medicine perspective. Acad Emerg Med. (1999) 6(9):947–52. doi: 10.1111/j.1553-2712.1999.tb01246.x

PubMed Abstract | Crossref Full Text | Google Scholar

20. Tversky A, Kahneman D. Availability: a heuristic for judging frequency and probability. Cogn Psychol. (1973) 5(2):207–32. doi: 10.1016/0010-0285(73)90033-9

Crossref Full Text | Google Scholar

21. Mamede S, van Gog T, van den Berge K, Rikers RMJP, van Saase JLCM, van Guldener C, et al. Effect of availability bias and reflective reasoning on diagnostic accuracy among internal medicine residents. JAMA. (2010) 304(11):1198–203. doi: 10.1001/jama.2010.1276

PubMed Abstract | Crossref Full Text | Google Scholar

22. Elstein AS. Heuristics and biases: selected errors in clinical reasoning. Acad Med. (1999) 74(7):791–4. doi: 10.1097/00001888-199907000-00012

PubMed Abstract | Crossref Full Text | Google Scholar

23. Croskerry P. The cognitive imperative: thinking about how we think. Acad Emerg Med. (2000) 7(11):1223–31. doi: 10.1111/j.1553-2712.2000.tb00467.x

PubMed Abstract | Crossref Full Text | Google Scholar

24. Liu Y, Chen S, Zühlke L, Black GC, Choy MK, Li N, et al. Global birth prevalence of congenital heart defects 1970–2017: updated systematic review and meta-analysis of 260 studies. Int J Epidemiol. (2019) 48(2):455–63. doi: 10.1093/ije/dyz009

PubMed Abstract | Crossref Full Text | Google Scholar

25. Croskerry P, Sinclair D. Emergency medicine: a practice prone to error? CJEM. (2001) 3(4):271–6. doi: 10.1017/S1481803500005765

PubMed Abstract | Crossref Full Text | Google Scholar

26. Greenhalgh T. Narrative based medicine: narrative based medicine in an evidence based world. Br Med J. (1999) 318(7179):323–5. doi: 10.1136/bmj.318.7179.323

Crossref Full Text | Google Scholar

27. Scherer LD, de Vries M, Zikmund-Fisher BJ, Witteman HO, Fagerlin A. Trust in deliberation: the consequences of deliberative decision strategies for medical decisions. Health Psychol. (2015) 34(11):1090–9. doi: 10.1037/hea0000203

PubMed Abstract | Crossref Full Text | Google Scholar

28. Kern L, Doherty ME. “Pseudodiagnosticity” in an idealized medical problem-solving environment. J Med Educ. (1982) 57(2):100–4. doi: 10.1097/00001888-198202000-00004

PubMed Abstract | Crossref Full Text | Google Scholar

29. Leopold SS. Editorial: are we all better-than-average drivers, and better-than-average kissers? Outwitting the Kruger-Dunning effect in clinical practice and research. Clin Orthop Relat Res. (2019) 477(10):2183–5. doi: 10.1097/CORR.0000000000000948

PubMed Abstract | Crossref Full Text | Google Scholar

30. Howell WC. Uncertainty from internal and external sources: a clear case of overconfidence. J Exp Psychol. (1971) 89(2):240–3. doi: 10.1037/h0031206

PubMed Abstract | Crossref Full Text | Google Scholar

31. Voytovich AE, Rippey RM, Suffredini A. Premature conclusions in diagnostic reasoning. J Med Educ. (1985) 60(4):302–7. doi: 10.1097/00001888-198504000-00004

PubMed Abstract | Crossref Full Text | Google Scholar

32. Baron J, Hershey JC. Outcome bias in decision evaluation. J Pers Soc Psychol. (1988) 54(4):569–79. doi: 10.1037/0022-3514.54.4.569

PubMed Abstract | Crossref Full Text | Google Scholar

33. Andersen SM, Glassman NS, Chen S, Cole SW. Transference in social perception: the role of chronic accessibility in significant-other representations. J Pers Soc Psychol. (1995) 69(1):41–57. doi: 10.1037/0022-3514.69.1.41

PubMed Abstract | Crossref Full Text | Google Scholar

34. Scarfone RJ, Nagler J. Cognitive errors in pediatric emergency medicine. Pediatr Emerg Care. (2021) 37(2):96–103. doi: 10.1097/PEC.0000000000002342

PubMed Abstract | Crossref Full Text | Google Scholar

35. Croskerry P, Singhal G, Mamede S. Cognitive debiasing 2: impediments to and strategies for change. BMJ Qual Saf. (2013) 22(Suppl 2):ii65–72. doi: 10.1136/bmjqs-2012-001713

PubMed Abstract | Crossref Full Text | Google Scholar

36. Hammond MEH, Stehlik J, Drakos SG, Kfoury AG. Bias in medicine: lessons learned and mitigation strategies. JACC Basic Transl Sci. (2021) 6(1):78–85. doi: 10.1016/j.jacbts.2020.07.012

PubMed Abstract | Crossref Full Text | Google Scholar

37. Croskerry P. When I say… cognitive debiasing. Med Educ. (2015) 49(7):656–7. doi: 10.1111/medu.12670

PubMed Abstract | Crossref Full Text | Google Scholar

38. Eichbaum Q. Medical error, cognitive bias, and debiasing: the jury is still out. Acad Med. (2019) 94(8):1065–6. doi: 10.1097/ACM.0000000000002791

PubMed Abstract | Crossref Full Text | Google Scholar

39. Stiegler MP, Neelankavil JP, Canales C, Dhillon A. Cognitive errors detected in anaesthesiology: a literature review and pilot study. Br J Anaesth. (2012) 108(2):229–35. doi: 10.1093/bja/aer387

PubMed Abstract | Crossref Full Text | Google Scholar

40. Asch DA, Baron J, Hershey JC, Kunreuther H, Meszaros J, Ritov I, et al. Omission bias and Pertussis vaccination. Med Decis Making. (1994) 14(2):118–23. doi: 10.1177/0272989X9401400204

PubMed Abstract | Crossref Full Text | Google Scholar

41. Fischhoff B. Hindsight ^foresight: the effect of outcome knowledge on judgment under uncertainty. J Exp Psychol Hum Percept Perform. (1975) 1(3):288–99. doi: 10.1037/0096-1523.1.3.288

Crossref Full Text | Google Scholar

42. Antonacci AC, Dechario SP, Rindskopf D, Husk G, Jarrett M. Cognitive bias and severity of harm following surgery: plan for workflow debiasing strategy. Am J Surg. (2021) 222(6):1172–7. doi: 10.1016/j.amjsurg.2021.08.035

PubMed Abstract | Crossref Full Text | Google Scholar

43. Richburg CE, Dossett LA, Hughes TM. Cognitive bias and dissonance in surgical practice: a narrative review. Surg Clin North Am. (2023) 103(2):271–85. doi: 10.1016/j.suc.2022.11.003

PubMed Abstract | Crossref Full Text | Google Scholar

44. Saposnik G, Redelmeier D, Ruff CC, Tobler PN. Cognitive biases associated with medical decisions: a systematic review. BMC Med Inform Decis Mak. (2016) 16(1):138. doi: 10.1186/s12911-016-0377-1

PubMed Abstract | Crossref Full Text | Google Scholar

45. Ely JW, Graber ML, Croskerry P. Checklists to reduce diagnostic errors. Acad Med. (2011) 86(3):307–13. doi: 10.1097/ACM.0b013e31820824cd

PubMed Abstract | Crossref Full Text | Google Scholar

46. Dhillon BS. Human errors: a review. Microelectron Reliab. (1989) 29(3):299–304. doi: 10.1016/0026-2714(89)90612-4

Crossref Full Text | Google Scholar

47. Yee LM, Liu LY, Grobman WA. The relationship between obstetricians’ cognitive and affective traits and their patients’ delivery outcomes. Am J Obstet Gynecol. (2014) 211(6):692.e1–6. doi: 10.1016/j.ajog.2014.06.003

PubMed Abstract | Crossref Full Text | Google Scholar

48. Baldwin RL, Green JW, Shaw JL, Simpson DD, Bird TM, Cleves MA, et al. Physician risk attitudes and hospitalization of infants with bronchiolitis. Acad Emerg Med. (2005) 12(2):142–6. doi: 10.1197/j.aem.2004.10.002

PubMed Abstract | Crossref Full Text | Google Scholar

49. Balla JI, Heneghan C, Glasziou P, Thompson M, Balla ME. A model for reflection for good clinical practice. J Eval Clin Pract. (2009) 15(6):964–9. doi: 10.1111/j.1365-2753.2009.01243.x

PubMed Abstract | Crossref Full Text | Google Scholar

50. Graber ML, Kissam S, Payne VL, Meyer AND, Sorensen A, Lenfestey N, et al. Cognitive interventions to reduce diagnostic error: a narrative review. BMJ Qual Saf. (2012) 21(7):535–57. doi: 10.1136/bmjqs-2011-000149

PubMed Abstract | Crossref Full Text | Google Scholar

51. Raab M, Gigerenzer G. The power of simplicity: a fast-and-frugal heuristics approach to performance science. Front Psychol. (2015) 6:1672. doi: 10.3389/fpsyg.2015.01672

PubMed Abstract | Crossref Full Text | Google Scholar

52. Tversky A, Kahneman D. Judgment under uncertainty: heuristics and biases. Science. (1974) 185(4157):1124–31. doi: 10.1126/science.185.4157.1124

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: congenital heart disease, cognitive biases, pediatric cardiology, diagnostic errors, medical decision-making, human factors

Citation: Padovani P, Roy A, Guerra A, Cadeau O, Ly M, Vasile CM, Pass RH and Baruteau A-E (2024) Cognitive biases in pediatric cardiac care. Front. Cardiovasc. Med. 11:1423680. doi: 10.3389/fcvm.2024.1423680

Received: 26 April 2024; Accepted: 21 June 2024;
Published: 4 July 2024.

Edited by:

Maria Joao Baptista, Centro Hospitalar Universitário de São João (CHUSJ), Portugal

Reviewed by:

Marta Silva, University of Porto, Portugal

© 2024 Padovani, Roy, Guerra, Cadeau, Ly, Vasile, Pass and Baruteau. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Paul Padovani, padovanipaul17@gmail.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.