- Research
- Open Access
- Published:
False certainty in the acquisition of anatomical and physiotherapeutic knowledge
BMC Medical Education volume 22, Article number: 765 (2022)
Abstract
Background
Efficient metacognitive monitoring, that is the monitoring of one’s own thought processes and specifically one’s state of knowledge, is essential for effective clinical reasoning. Knowing what one does and does not know is a competency that students of health professions need to develop. Students often tend to develop false certainty in their own knowledge in the longer course of their education, but the time frame that is required for this effect to occur has remained unclear. We investigated whether students developed false certainty already after one course unit.
Methods
This study analysed data from one sample of medical students and four samples of physiotherapy students in two formal educational settings (total N = 255) who took knowledge tests before and after a course unit. We examined changes in students’ confidence separately for correctly and incorrectly answered questions and analysed their ability to assign higher levels of confidence to correct answers than to incorrect answers (discrimination ability).
Results
Students’ knowledge as well as confidence in their correct answers in knowledge tests increased after learning. However, consistently for all samples, confidence in incorrect answers increased as well. Students’ discrimination ability improved only in two out of the five samples.
Conclusions
Our results are in line with recent research on confidence increase of health professions students during education. Extending those findings, our study demonstrated that learning in two different formal educational settings increased confidence not only in correct but also in incorrect answers to knowledge questions already after just one learning session. Our findings highlight the importance of improving metacognition in the education of health professionals—especially their ability to know what they do not know.
Background
Metacognitive monitoring
Efficient metacognition, that is, a range of cognitive processes that are involved in self-assessment, cognitive control, and monitoring, is paramount for effective clinical reasoning processes [1]. One major part of metacognition is metacognitive monitoring of one’s own knowledge, which allows an individual to gain insight into what they know and what they do not know [2]. However, practicing medical professionals can lack awareness of what they do and do not know [3,4,5,6].
The ability of metacognitive monitoring can be expressed by metacognitive calibration, that is, the association between knowledge and confidence in that knowledge [7]. In general, if individuals display more confidence than would be appropriate given their knowledge, they are poorly calibrated, and overconfidence occurs. Even experienced medical practitioners might not always be aware whether their diagnoses are correct or not [8] and physicians’ levels of confidence in their diagnoses sometimes show little correlation with their diagnostic accuracy [9]. Overconfidence of medical professionals is widespread: They overestimate what they know in areas like dermatology [10], pathology [11], radiology [12], internal medicine [13], and dementia [14]. Dermatologists, for example, change their minds about the malignancy of lesion images less often if they are more confident in their decisions [10], yet they follow this tendency regardless of whether their initial decision was correct or incorrect. Apart from its importance in professional life, adequate monitoring of what is known is also critical in medical education since it allows a learner to control and regulate one’s own learning processes [15, 16].
Metacognitive judgments of confidence
One common method of measuring monitoring is collecting metacognitive judgments by asking students to rate confidence in the correctness of their answers to knowledge test items [17]. Adequate monitoring ability should be reflected by high confidence ratings to correctly and low confidence ratings to incorrectly answered questions [18]. Especially knowing what one does not know, is a competency that needs to be taught in medical and health professional education, since a health professional should sense when to ask for another opinion in making decisions [19].
The literature suggests that medical students show persistent overconfidence in diverse medical topics like delirium [15], surgery [20], or acne [21]. It should be expected that over a student’s years of study, overconfidence would give way to better monitoring ability. Yet the data of a recent longitudinal study by Kämmer et al. [17] points to the opposite direction. The authors measured students’ medical knowledge and confidence in knowledge questions asked over the course of 10 semesters. As a measure of metacognitive calibration, they analysed students’ discrimination ability, that is, the ability to assign higher levels of confidence to correct answers than to incorrect answers. The authors report an increase of medical knowledge over the course of the semesters but no change in the discrimination ability. Instead, results showed a general increase in the absolute level of confidence over time, regardless of knowledge.
The results of another recent longitudinal study by Cecilio-Fernandes and colleagues [22] found that students in their final year answered significantly more clinical knowledge questions correctly than first-year students. However, first-year students on average marked 71% of their incorrectly answered questions as “don’t know”, whereas final-year students did so only for 37%. Those results suggest that knowledge gain over the course of one’s medical education can be accompanied by an emerging reluctance to admit one’s lack of knowledge. Those results indicate that over the course of medical education, while students acquire medical knowledge and increase confidence in this knowledge, they also develop certainty in supposed knowledge they mistakenly think is correct. Due to the longitudinal nature of these studies, it is unclear what time frame it will take for this effect to emerge in health professions education. Thus, the goal of the study presented here was to explore how medical and physiotherapy students’ confidence in their knowledge as well as their metacognitive calibration changes after only one course unit. Specifically, we were interested in addressing the research question of whether students would develop false certainty under these conditions.
Materials and method
We analysed five datasets containing five student samples to explore medical and physiotherapy students’ change in confidence before and after a course unit in two different formal medical educational settings. The design of all of the studies was similar in that it featured a learning phase and the pre-/post-measurement of both knowledge and confidence. Courses varied in content and didactic methods. Samples 1 and 2 were taken from a study published by Grosser and colleagues [23] and included medical and physiotherapy students who watched an anatomy lecture broadcasted live from the dissection hall of a university clinic’s anatomical institute. This live dissection was moderated by an anatomy professor. The authors did not report changes in confidence or metacognitive indices in the previous publication, which are the dependent variables in our analyses. Samples 3 to 5 were taken from three different cohorts of a physiotherapy course about gait analysis and included physiotherapy students. The data of the students in Samples 3 and 4 have not been published; the data of Sample 5 was partly published by Bientzle and colleagues [24].
Participants
Data of initially N = 294 participants was collected in four separate studies (Samples 1 and 2 were collected together; Samples 3–5 were collected individually in each case but using the same setting). Participants for Samples 1 and 2 were recruited among medical and physiotherapy students who participated in an educational live event about anatomy. Participants of Samples 3–5 were students of physiotherapy who participated in a course about gait analysis. Participants’ knowledge about the course content was tested before and immediately after the course (Samples 1, 2) or one week later (Samples 3–5). Participants who responded to fewer than 75% of knowledge test items or confidence judgments at either time 1 or time 2 were excluded from our analyses. See Table 1 for the samples’ characteristics and Fig. 1 for a flow diagram.
Measures
Knowledge tests
All knowledge tests were forced choice tests in which participants were given one statement at a time and had to determine whether the statement was true or false. Samples 1 and 2 responded to a 24-item knowledge test about the anatomy of the shoulder. Samples 3–5 responded to a 20-item knowledge test on gait analysis. Presenting knowledge tests to students that require them to select the correct answer from multiple statements is a straightforward and standardized approach for assessing knowledge in medical education [25].
Confidence
Following each knowledge test item, students’ confidence was measured by asking them to indicate how confident they were that their given answer was true on a 6–point rating scale, with labelled endpoints of 1 = “not sure at all” and 6 = “absolutely sure” for every item of the knowledge tests.
Statistical analyses
For answering the research question of whether false certainty would already develop after only one course, we analysed the change of confidence over time for each sample. To compare effect sizes between samples we first z-transformed the confidence variable. Then we conducted mixed effects model regressions separately for correct and incorrect answers for each sample. This method is able to handle the unbalanced cell size of our data [26]. We specified confidence as the dependent variable and time as a fixed effect. As random effects, we specified random intercepts for participant and item, as well as a by-participant random slope for time. We report standardized coefficients as well as confidence intervals for all models.
For analysing knowledge gain, for each participant we calculated the mean number of knowledge test items answered correctly at t1 and t2. We then used a paired sample t-test to analyse if there was an increase in knowledge at t2.
As measure of metacognitive calibration, we examined participants’ discrimination ability. Specifically, we used the index of relative metacognitive accuracy by calculating the Goodman–Kruskal’s gamma correlation [27] between the correctness of an answer and the respective confidence judgment. This is a widely endorsed method in educational psychology and metacognition research [28,29,30,31,32]. Individuals’ relative accuracy can range between -1 and 1, with 1 indicating perfect discrimination ability between correct and incorrect answers. For each participant we calculated the mean relative accuracy at both measurement points and determined changes in those scores by paired sample t-tests.
An alpha level of 0.05 was used for all statistical tests. Effect sizes for all mean differences are indicated as Cohen’s d; confidence intervals are also given. We used statistical software R version 4.0.3 (2020–10-10) for all analyses. Data and code can be found at https://osf.io/6v8za/?view_only=4a8fd507cceb42c1b8a8e9070d247728.
Results
Knowledge gain
Students in Samples 1, 2, 3, and 5 increased their knowledge significantly as indicated by paired sample t-tests, whereas in Sample 4 we could not detect an increase in knowledge. The first section of Table 2 shows statistical parameters for those analyses including p-values. Figure 2 shows a bar graph depicting the mean percentage of correctly answered knowledge test items for all samples and both measurement points.
Confidence distributions
Figure 3 shows distributions of confidence at t1 and t2 for all samples separated for correct and incorrect answers. After course units, distributions for correct answers were steeper and had a longer left tail at t2, indicating an increase in high confidence responses or an increase in knowing what is known. For incorrectly answered questions, this pattern was similar, although curves were flatter overall. The distributions shifted to the right at t2, the opposite of what would be expected if individuals improved upon knowing what they do not know.
Change in confidence over time
For answering our main research question of whether students would develop false certainty after one course unit, we fitted several mixed effects models to data containing confidence responses to the knowledge test questions. Figure 4 shows the standardized mixed effects regression coefficients of the predictor time for all samples, separated for correct/incorrect answers. All coefficients showed a similar pattern: Confidence intervals did not cross zero, indicating that confidence increased significantly in all samples. The mean effect size of confidence increase over time aggregated over all samples was M = 0.50 (SD = 0.11) for correct questions and M = 0.41 (SD = 0.12) for incorrect ones. The latter finding supports the hypothesis that false certainty already developed after one course unit. See supplementary material for statistical parameters of all analyses (Supplement A).
Relative metacognitive accuracy
To answer the research question of how metacognitive calibration changes after learning, we analysed mean relative metacognitive accuracy scores by conducting paired sample t-tests for scores at t1 and t2 for each sample. The results showed a different pattern for Samples 1 and 2 than for Samples 3, 4, and 5. Sample 1 showed low relative metacognitive accuracy at t1 (M = 0.36, SD = 0.31) but participants were able to improve to a moderate degree at t2 (M = 0.52, SD = 0.24; d = 0.46; p = 0.005). Sample 2 displayed higher relative metacognitive accuracy at t1 (M = 0.45, SD = 0.28) but did not improve significantly (p = 0.403). Students of Samples 3–5 had low relative metacognitive accuracy at t1 and showed improvement at t2. However, only the increase of Sample 5 from M = 0.28 (SD = 0.31) to M = 0.40 (SD = 0.35) reached statistical significance (p = 0.003). Table 3 shows the statistical parameters of analyses conducted including confidence intervals for effect sizes.
Discussion
This study was conducted to examine the changes in medical and physiotherapy students’ confidence in their knowledge and metacognitive calibration after only one course unit in different educational contexts. The data showed substantial differences in mean confidence scores and discrimination ability, illustrating a complex picture of students’ metacognition. In general, results revealed that students’ metacognitive calibration could reach a moderate level after learning, although it was not often the case and far from optimal: Only students in two out of five samples were able to improve their discrimination ability. Most importantly, we demonstrated a robust substantial increase in false certainty, that is, increased confidence in incorrect answers to knowledge test items across all samples. Since measures of calibration are derived from confidence ratings and their association, increased false certainty might be responsible for the missed opportunity to improve discrimination ability after learning.
Before and after course units, all students did on average report higher confidence in correctly than in incorrectly answered questions, which is a typical finding [33] and mirrors the above-zero discrimination ability captured with our metacognitive calibration measure. Taken together, the results show that students were capable of moderate discrimination between their correct and incorrect answers. But there was substantial variation among different cohorts and across different course units regarding pre-existing discrimination ability and improvement after learning.
The results for general confidence measured separately for correct and incorrect answers were robust across samples: Confidence in correct answers increased after learning, but also confidence in incorrect ones. This effect of increased false certainty had a medium effect size and was nearly as large as the effect of the increase of confidence in correct answers. This could mean that after a learning experience, although students are able to better judge what they know, they also display false certainty—hence are worse in knowing what they do not know. Similar findings have been reported for short online information search scenarios [34]. Also, another recent study reports that more knowledgeable students are able to give a high confidence response more adequately than low performing students when they are actually correct [35]. However, they are worse in applying the low confidence response to wrong questions. Our findings complement findings of longitudinal studies that showed increased confidence over the longer course of academic education [17, 22]. To our knowledge this study is the first to illustrate that in health professional education one course unit alone can elicit false certainty. Ongoing learning experiences, where students acquire more knowledge, accompanied by increased false certainty, could accumulate over the course of a student’s medical education and lead to the effects observed in longitudinal studies.
Knowing what one knows and what one does not know is a highly important learning goal for health professionals because it is part of critically reflecting upon one’s own clinical decisions [36]. Having high confidence in wrong knowledge is potentially hazardous in clinical practice [37]. This means seeing increasing false certainty developing after only one course unit is troubling for educators in health professional education. Our results also support what has already been pointed out by other researchers: Medical professionals can be reluctant to admit their uncertainty in medical diagnoses [38], and students need to be familiarized with experiencing uncertainty [39, 40] and the feeling of not knowing something [1, 13] in their future work life. One limitation of our study is that it cannot be ruled out that initially emerging overall high confidence could fade away after subsequent learning experiences in additional course units. For probabilistic learning tasks, studies showed that learners can find themselves in a “beginner’s bubble” of overconfidence marked by quickly developing overconfidence after a little learning, which gives way to better metacognitive calibration after more learning [41, 42]. It has, however, also been shown that confidence in one’s ability after a short learning period may be more stable than knowledge itself [43], which points to a potential resilience of false certainty once it has emerged. Future research is needed to directly test stability of false certainty in health professional education. Another limitation concerns the generalizability of our results to different topics in health professional education. Our study investigated confidence change after learning only for knowledge in anatomy and gait analysis. Although similar confidence effects have been demonstrated for topics like meteorology [44], scuba diving [34], and overall clinical knowledge [17], to infer about its generalizability, the causes of this effect have yet to be investigated.
Conclusions
Our results alert educators that students of health professions are potentially capable of moderate metacognitive calibration but at the same increase their confidence in incorrect answers after only one course unit. This developing false certainty is troubling as high confidence in wrong medical knowledge, if persistent in professional life, can ultimately threaten patient safety. Medical and physiotherapy educators should be aware of this effect and think of means to counter it. Students need to be made familiar with uncertainty and not knowing, since these are part of most clinical routines. This could help to reduce diagnostic error in future professional practice and improve overall metacognition of health professionals.
Availability of data and materials
Data and code for analysis is provided at https://osf.io/6v8za/?view_only=4a8fd507cceb42c1b8a8e9070d247728.
References
Colbert CY, Graham L, West C, White BA, Arroliga AC, Myers JD, Ogden PE, Archer J, Mohammad ZTA, Clark J. Teaching metacognitive skills: Helping your physician trainees in the quest to “know what they don’t know.” Am J Med. 2015;128(3):318–24.
Koriat A. Metacognition and consciousness. In: The Cambridge Handbook of Consciousness. 2012:289–326. https://doi.org/10.1017/cbo9780511816789.012
Hodges B, Regehr G, Martin D. Difficulties in recognizing one’s own incompetence: Novice physicians who are unskilled and unaware of it. Acad Med. 2001;76(10):S87–9.
Barnsley L, Lyon PM, Ralston SJ, Hibbert EJ, Cunningham I, Gordon FC, Field MJ. Clinical skills in junior medical officers: A comparison of self-reported confidence and observed competence. Med Educ. 2004;38(4):358–67.
Burman NJ, Boscardin CK, Van Schaik SM. Career-long learning: Relationship between cognitive and metacognitive skills. Med Teach. 2014;36(8):715–23.
Brezis M, Orkin-Bedolach Y, Fink D, Kiderman A. Does Physician’s Training Induce Overconfidence That Hampers Disclosing Errors? J Patient Saf. 2019;15(4):296–8.
Mamede S, Hautz WE, Berendonk C, Hautz SC, Sauter TC, Rotgans J, Zwaan L, Schmidt HG. Think twice: Effects on diagnostic accuracy of returning to the case to reflect upon the initial diagnosis. Acad Med. 2020;95(8):1223–9.
Friedman CP, Gatti GG, Franz TM, Murphy GC, Wolf FM, Heckerling PS, Fine PL, Miller TM, Elstein AS. Do physicians know when their diagnoses are correct? Implications for decision support and error reduction. J Gen Intern Med. 2005;20(4):334–9.
Meyer AND, Payne VL, Meeks DW, Rao R, Singh H. Physicians’ diagnostic accuracy, confidence, and resource requests: A vignette study. JAMA Intern Med. 2013;173(21):1952–61.
Dreiseitl S, Binder M. Do physicians value decision support? A look at the effect of decision support systems on physician opinion. Artif Intell Med. 2005;33(1):25–30.
Podbregar M, Voga G, Krivec B, Skale R, Parežnik R, Gabršček L. Should we confirm our clinical diagnostic certainty by autopsies? Intensive Care Med. 2001;27(11):1750–5.
Potchen EJ. Measuring Observer Performance in Chest Radiology: Some Experiences. J Am Coll Radiol. 2006;3(6):423–32.
Friedman CP, Donaldson KM, Vantsevich AV. Educating medical students in the era of ubiquitous information. Med Teach. 2016;38(5):504–9.
Lerner BS, Kalish V, Ledford CJW. Exploring residents’ skills in diagnosing dementia: The unexpected dissonance between ability and confidence. Fam Med. 2017;49(6):460–3.
Artino AR, Dong T, Dezee KJ, Gilliland WR, Waechter DM, Cruess D, Durning SJ. Achievement goal structures and self-regulated learning: Relationships and changes in medical school. Acad Med. 2012;87(10):1375–81.
Cleary TJ, Durning SJ, Artino AR. Microanalytic Assessment of Self-Regulated Learning during Clinical Reasoning Tasks: Recent Developments and Next Steps. Acad Med. 2016;91(11):1516–21.
Kämmer JE, Hautz WE, März M. Self-monitoring accuracy does not increase throughout undergraduate medical education. Med Educ. 2020;54(4):320–7.
Hunt DP. The concept of knowledge and how to measure it. J Intellect Cap. 2003;4(1):100–13. https://doi.org/10.1108/14691930310455414.
Moulton CAE, Regehr G, Mylopoulos M, MacRae HM. Slowing down when you should: A new model of expert judgment. Acad Med. 2007;82(10 SUPPL.):109–16. https://doi.org/10.1097/ACM.0b013e3181405a76.
Borracci RA, Arribalzaga EB. The Incidence of Overconfidence and Underconfidence Effects in Medical Student Examinations. J Surg Educ. 2018;75(5):1223–9.
Alajmi RS, Aljefri YE, Naaman NK, Alraddadi AA, Alamri AM. Acne Self-medication among Pre-clinical and Clinical Years Medical Students. I J Surg Med. 2020;6(5):5–10.
Cecilio-Fernandes D, Kerdijk W, Jaarsma ADDC, Tio RA. Development of cognitive processing and judgments of knowledge in medical students: Analysis of progress test results. Med Teach. 2016;38(11):1125–9.
Grosser J, Bientzle M, Shiozawa T, Hirt B, Kimmerle J. Observing Interprofessional Collaboration: Impact on Attitude and Knowledge Acquisition. Anat Sci Educ. 2020;Published online 2020. https://doi.org/10.1002/ase.2007
Bientzle M, Lechner C, Cress U, Kimmerle J. Online peer consulting for health professionals. Clin Teach. 2018;16(5):463–7. https://doi.org/10.1111/tct.12950.
Epstein RM. Assessment in Medical Education. N Engl J Med. 2007;356(4):387–96.
Brauer M, Curtin JJ. Linear mixed-effects models and the analysis of nonindependent data: A unified framework to analyze categorical and continuous independent variables that vary within-subjects and/or within-items. Psychol Methods. 2018;23(3):389–411.
Nelson TO. A comparison of current measures of the accuracy of feeling-of-knowing predictions. Psychol Bull. 1984;95(1):109–33.
Jemstedt A, Kubik V, Jönsson FU. What moderates the accuracy of ease of learning judgments? Metacognition Learn. 2017;12(3):337–55.
Hughes GI, Taylor HA, Thomas AK. Study techniques differentially influence the delayed judgment-of-learning accuracy of adolescent children and college-aged adults. Metacognition Learn. 2018;13(2):109–26.
Canfield CI, Fischhoff B, Davis A. Correction to: Better beware: comparing metacognition for phishing and legitimate emails. Metacognition Learn. 2019;14(3):363–363.
Ackerman R, Bernstein DM, Kumar R. Metacognitive hindsight bias. Mem Cogn. 2020;48(5):731–44.
Jang Y, Lee H, Kim Y, Min K. The Relationship between Metacognitive Ability and Metacognitive Accuracy. Metacognition Learn. 2020;15(3):411–34.
Hautz WE, Schubert S, Schauber SK, Kunina-Habenicht O, Hautz SC, Kämmer JE, Eva KW. Accuracy of self-monitoring: does experience, ability or case difficulty matter? Med Educ. 2019;53(7):735–44.
von Hoyer JF, Kimmerle J, Holtz P. Acquisition of false certainty: Learners increase their confidence in the correctness of incorrect answers after online information search. J Comput Assist Learn. 2022;38(3):833–44. https://doi.org/10.1111/jcal.12657.
Tabibzadeh N, Mullaert J, Zafrani L, Balagny P, Frija-Masson J, Marin S, Lefort A, Vidal-Petiot E, Flamant M. Knowledge self-monitoring, efficiency, and determinants of self-confidence statement in multiple choice questions in medical students. BMC Med Educ. 2020;20(1):445.
Croskerry P. A Model for Clinical Decision-Making in Medicine. Med Sci Educ. 2017;27(1):9–13.
Dory V, Degryse J, Roex A, Vanpee D. Usable knowledge, hazardous ignorance beyond the percentage correct score. Med Teach. 2010;32(5):375–80.
Croskerry P, Norman G. Overconfidence in Clinical Decision Making. Am J Med. 2008;121:24–9.
Ilgen JS, Eva KW, de Bruin A, Cook DA, Regehr G. Comfort with uncertainty: reframing our conceptions of how clinicians navigate complex clinical situations. Adv Heal Sci Educ. 2019;24(4):797–809.
Dunlop M, Schwartzstein RM. Reducing Diagnostic Error in the Intensive Care Unit. Engaging Uncertainty When Teaching Clinical Reasoning. ATS Sch. 2020;1(4):364–71.
Sanchez C, Dunning D. Overconfidence among beginners: Is a little learning a dangerous thing? J Pers Soc Psychol. 2017;114(1):10–28.
Sanchez C, Dunning D. Decision Fluency and Overconfidence Among Beginners. Decision. 2020;7(2):225–37.
Pusic MV, Chiaramonte R, Gladding S, Andrews JS, Pecaric MR, Boutis K. Accuracy of self-monitoring during learning of radiograph interpretation. Med Educ. 2015;49(8):838–46.
von Hoyer J, Pardi G, Kammerer Y, Holtz P. Metacognitive judgments in searching as learning (SAL) Tasks: Insights on (Mis-) calibration, multimedia usage, and confidence. In: SALMM 2019 - Proceedings of the 1st International Workshop on Search as Learning with Multimedia Information, Co-Located with MM 2019. 3‑10. https://doi.org/10.1145/3347451.3356730
Acknowledgements
We thank Penelope Burr Pinson for proofreading the manuscript and Marie Hornberger for helping with the formatting.
Funding
Open Access funding enabled and organized by Projekt DEAL. Budget Resources of the Leibniz-Institut für Wissensmedien.
Author information
Authors and Affiliations
Contributions
Planning and designing of the experiments as well as data collection were carried out by Johannes Großer (Samples 1, 2) and Martina Bientzle (Samples 1–5). Johannes von Hoyer and Peter Holtz conceptualized the study with feedback and support of Martina Bientzle. Analyses were carried out by Johannes von Hoyer with support by Martina Bientzle and Johannes Großer. The first draft was written by Johannes von Hoyer with feedback by and support of Martina Bientzle, Ulrike Cress, Johannes Großer, Peter Holtz, and Joachim Kimmerle. The manuscript was written by Johannes von Hoyer, Martina Bientzle and Joachim Kimmerle. Extensive critical revision was provided by Ulrike Cress and Peter Holtz. All Authors approved the final version of the submitted manuscript.
Corresponding author
Ethics declarations
Ethics approval and Consent to participate
Since this study is a meta-analysis ethics approval is not applicable.
The study in which data of Samples 1 and 2 was collected was approved by the Commission of Ethics of the Leibniz-Institut für Wissensmedien (Approval-Code LEK 2017/035).
Research procedures in which data of Samples 3–5 were collected was performed in accordance with the Declaration of Helsinki. The Ulmkolleg administration provided ethical approval regarding the participation of students (as a result of legal specifications, the school administration was responsible for checking and approving the participation of its students). Participants gave their consent to participate via informed consent.
Consent for publication
Not applicable.
Competing interests
The authors declare that there are no conflicts of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1:
Supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
von Hoyer, J., Bientzle, M., Cress, U. et al. False certainty in the acquisition of anatomical and physiotherapeutic knowledge. BMC Med Educ 22, 765 (2022). https://doi.org/10.1186/s12909-022-03820-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12909-022-03820-x
Keywords
- Medical Education
- Testing/Assessment
- Metacognition
- False certainty