Skip to main content

Assessing clinical reasoning ability in fourth-year medical students via an integrative group history-taking with an individual reasoning activity

Abstract

Background

The most important factor in evaluating a physician’s competence is strong clinical reasoning ability, leading to correct principal diagnoses. The process of clinical reasoning includes history taking, physical examinations, validating medical records, and determining a final diagnosis. In this study, we designed a teaching activity to evaluate the clinical reasoning competence of fourth-year medical students.

Methods

We created five patient scenarios for our standardized patients, including hemoptysis, abdominal pain, fever, anemia, and chest pain. A group history-taking with individual reasoning principles was implemented to teach and evaluate students’ abilities to take histories, document key information, and arrive at the most likely diagnosis. Residents were trained to act as teachers, and a post-study questionnaire was employed to evaluate the students’ satisfaction with the training activity.

Results

A total of 76 students, five teachers, and five standardized patients participated in this clinical reasoning training activity. The average history-taking score was 64%, the average key information number was 7, the average diagnosis number was 1.1, and the average correct diagnosis rate was 38%. Standardized patients presenting with abdominal pain (8.3%) and anemia (18.2%) had the lowest diagnosis rates. The scenario of anemia presented the most difficult challenge for students in history taking (3.5/5) and clinical reasoning (3.5/5). The abdominal pain scenario yielded even worse results (history taking: 2.9/5 and clinical reasoning 2.7/5). We found a correlation in the clinical reasoning process between the correct and incorrect most likely diagnosis groups (group history-taking score, p = 0.045; key information number, p = 0.009 and diagnosis number, p = 0.004). The post-study questionnaire results indicated significant satisfaction with the teaching program (4.7/5) and the quality of teacher feedback (4.9/5).

Conclusions

We concluded that the clinical reasoning skills of fourth-year medical students benefited from this training course, and the lower correction of the most likely diagnosis rate found with abdominal pain, anemia, and fever might be due to a system-based teaching modules in fourth-year medical students; cross-system remedial reasoning auxiliary training is recommended for fourth-year medical students in the future.

Peer Review reports

Introduction

Differential diagnosis is important for clinicians and involves a deeper higher order thinking process about the evaluation of history taking, physical examination, review of laboratory data, and diagnostic images that exceed memorization, facts, and concepts [1]. Koenemann et al. reported that clinical case discussions with peer-taught and physician-supervised collaborative learning formats could promote clinical reasoning in medical students [2]. However, we used problem-based learning (PBL) in schools to train the clinical reasoning process for fourth-year medical students using a module system base that could easily restrict students’ thinking in the same organ system.

Assessing a medical student’s ability concerning differential diagnosis is important; however, there is no consensus on the most effective approaches to evaluate these reasoning skills. Many efforts have been made to develop a valid and reliable measure of clinical reasoning ability, including key feature questions (KF), diagnostic thinking inventory (DTI) and script concordance (SC) tests. Charlin B et al. reported that the Script Concordance test is a simple and direct approach to testing organization and use of knowledge with machine-scorable [3]. Bordage G et al. developed an inventory of diagnostic thinking which measure two aspects of diagnostic thinking: the degree of flexibility in thinking and the degree of knowledge structure in memory [4]. Fischer MR et al. demonstrated a modified electronic key feature examination of clinical decision-making skills for undergraduate medical students [5].

Simulation-based testing methods have also been developed to meet the need for assessment procedures that are both authentic and well structured. Sutherland et al. allowed students to watch a video trigger and discuss their diagnostic reasoning with an examiner demonstrating that it could be assessed [6]. Fürstenberg et al. developed a clinical reasoning indicators history-taking scale to quantitatively assess clinical reasoning indicators during history taking in undergraduate medical students [7], which was deemed suitable for the assessment of fourth-year medical students’ clinical reasoning ability.

The objective structured clinical examination (OSCE) evaluation has proven to be a reliable and valid method for assessing the six competencies defined by the Accreditation Council for Graduate Medical Education (ACGME) in surgery. The competencies assessed by a well-constructed OSCE include patient care and medical knowledge as well as skills like data synthesis and the ability to list differential diagnoses [8]. This evaluation method was able to identify significant deficiencies in musculoskeletal examination skills and the diagnostic reasoning of fourth-year medical students based on principles of the hypothesis-driven physical examination [9]. Using a group OSCE format makes it possible to assess the individual ability of a large number of students without the usual time and expense needed [10]. As the fourth-year medical students in MacKay Medical College had no clinical reasoning curriculum except an integrated eight modules of divided organ systems PBL training and clinical diagnosis medicine and practice. In this study, we aimed to evaluate the M4 students’ clinical reasoning ability through group history taking with standardized patient (SP) in different clinical scenarios that mimic clinical conditions and then creating individual students’ most likely diagnosis with supporting data to assess their clinical diagnosis ability.

Material and methods

Problem identification and target needs of fourth-year medical students’ clinical reasoning ability before clinical practice

The medical curriculum of Taiwan medical schools including 2 years general and basic education of the department of medicine. The third-year curriculum is the integration of contents based on the general structure, functions, pathology and behaviors of a normal person. The fourth-year curriculum is the integration of contents based on the concise clinically relevant anatomical structure, functions, and behaviors of abnormal and diseased persons, as well as clinical knowledge and skills. The fifth- and sixth-year curriculum is the clinical medicine education. The curriculum at Mackay Medical College for fourth-year medical students is an integrated eight modules of divided organ systems including (1) introduction to clinical medicine and cardiovascular system, (2) pulmonary system and endocrine and metabolism, (3) gastrointestinal system, (4) brain and behaviors, (5) musculoskeletal system and integument system, (6) host defense and infection, (7) fluid, electrolytes, renal and genitourinary system, (8) hematology and oncology. In this M4 systemic module, PBL, clinical diagnosis medicine and practice is implanted and integrated into semesters for clinical reasoning training. This study conducted during the period from March 2019 to August 2019 in MacKay Medical College, the major question in this study is wanted to know the M4 students’ clinical reasoning ability through history taking with standardized patient in different clinical scenarios.

Educational objectives and assessment method

Accurate diagnostic reasoning is the fundamental basis for ensuring patient care and safety; thus, the development of diagnostic reasoning is a key component of UGY medical education. We created a group history taking with individual reasoning activity at Mackay Medical College which is an elective pre-clerkship education course for fourth-year medical students, with the aim of training and assessing students’ clinical reasoning ability. To evaluate clinical reasoning ability, Haring et al. developed an observation tool that consists of an 11-item observation rating form and a post encounter rating tool which are both feasible, valid, and reliable to assess students’ clinical reasoning skills in clinical practice [11]. They also reported that by observing and assessing clinical reasoning during medical students’ history taking, general and specific phenomena could be used as indicators, including taking control, recognizing, and responding to relevant information, specifying symptoms, asking specific questions that point to pathophysiological thinking, placing questions in a logical order, checking agreement with patients, summarizing, and body language [12]. We modified the methods that include history taking and clinical reasoning each time we let four to five students visit a standardized patient with a clinical case to collect enough data, and then individually write down key information and make a correct differential diagnosis. For assessment of clinical reasoning, and to train their principal diagnosis ability, students were asked to list 15 items of key information for differential diagnoses and list the three most likely diagnoses, including the first. We didn’t use a standard test like modified electronic key feature examination (KF) or diagnostic thinking inventory (DTI) is due to Findyartini A. reported that higher DTI scores only reflect a greater familiarity with the thinking process in diagnostic reasoning and do not necessarily represent better diagnostic performance [13], and the modified electronic KF examination lack SP simulated clinical setting scenario with history-taking interaction.

Study setting of the individual reasoning teaching program

Five R1 residents of internal medicine were recruited as teaching faculty in Mackay Memorial Hospital and requested to recognize the educational content and assess the trainees’ history taking performance and provide 30 minutes immediate feedback including the definition of this symptom, etiology, how to evaluate and collect key information, the initial diagnosis and treatment planning. We created five clinical scenarios: hemoptysis, abdominal pain, fever, anemia, and chest pain. The participants were divided into small groups consisting of four to five trainees per team. During the process of reasoning training, the trainees were expected taking turns to ask question while gathering clinical information from the standardized patients as well as acknowledge the symptoms and signs of the disease in each scenario. Documenting key words from history taking which requires as many as 15 clues, is also crucial to reflect students’ comprehension of a patient’s problems. From the patient’s information, the trainees were asked to differentiate the diagnosis with associated status and finally reach their most likely diagnosis along with two tentative diagnoses. After reasoning training, the residents who observed the trainees’ performance would be required to share their comments and experiences as well as resolve the trainee’s questions. The post-reasoning questionnaire, which was designed to focus on the satisfaction of the teaching program, provides an opportunity for the students to review and self-evaluate improvement in clinical reasoning and history taking through the reasoning program. The process and learning objectives of the reasoning training program for fourth-grade medical students are presented in Table 1.

Table 1 Group history taking with individual clinical reasoning activity for M4 students

Participants and educational content

Fourth-grade medical students from MacKay Medical College were enrolled to participate electively in a system-based teaching program. We also trained five residents from MacKay Memorial Hospital as teachers to evaluate the trainees’ performances based on the checklists and then provide immediate feedback. Additionally, five standardized patients (SPs) who simulated the symptoms and signs of the teaching scenarios were also able to offer post-study remarks, all methods were carried out in accordance with relevant guidelines and the ‘Declaration of Helsinki’.

The reasoning teaching program was designed to provide an opportunity for fourth-grade medical students to perform history taking while gathering clinical information from the SPs as well as acknowledge the symptoms and signs of the disease in each scenario. Five patient scenarios were created as follows: lung cancer presenting with hemoptysis, acute pancreatitis presenting with abdominal pain, acute pyelonephritis presenting with fever, uterine myoma bleeding presenting with anemia, and acute coronary syndrome presenting with chest pain. The case settings of hemoptysis and chest pain were regarded as system-based thinking logics, whereas those with abdominal pain, fever, and anemia were multisystem-related. Each scenario highlighted different key words in the patient’s history, which could be clues for approaching the final and tentative diagnosis. The educational strategies for clinical reasoning in each scenario are presented in Table 2.

Table 2 Educational strategies of clinical reasoning in each scenario

Assessment and feedback

In this study, we focused on four major components: (1) the trainee’s group history taking score, (2) number of key information points, (3) diagnosis numbers and a correct most likely diagnosis rate, and (4) feedback questionnaire. The residents were capable of rating the scores of trainees using the checklist, based on observation of the overall performance of each trainee. By collecting the record paper at the end of this teaching program, we were able to accumulate the number of key information notes documented by each trainee. Meanwhile, from the record paper, we could check the number of diagnoses and whether they were correct or not. Moreover, a post-study questionnaire consisting of the degree of satisfaction with the reasoning teaching program, self-evaluation of ability and difficulty in clinical reasoning, and history taking was issued. The rubrics and questionnaire were based on a 5-point Likert scale for level of satisfaction.

Statistical analysis

Data from the post-course feedback questionnaire and group history taking scores are shown as the mean ± standard deviation (SD). The correlations observed between the correct diagnosis group and the incorrect diagnosis group were analyzed using Student’s t-test. Statistical analyses were performed using SPSS 23.0 statistical package (SPSS, Chicago, IL, USA). All statistical analyses were based on two-sided hypothesis tests with a significance level of p < 0.05.

Results

A total of 76 fourth-year medical students participated in this study, and five MacKay Memorial Hospital R1 residents were recruited as teachers in this clinical reasoning training program. Regarding the trainees’ score in group history taking with individual reasoning, the best score in the group history-taking scenario was chest pain (76.9%) and the worst was hemoptysis (50.1%), with an average group history-taking score of 63.5%. The number of key words the medical students were permitted to write down was between five and eight, and the average number was seven; fever had the highest number of key words in the scenarios (8), with the lowest number for abdominal pain (5). The average diagnosis number in each scenario was one to two; and the correct “the most likely diagnosis” rate ranged from 8.3 to 87.5%, with chest pain as the best scenario and abdominal pain as the worst (Table 3).

Table 3 Trainees’ scores in group history taking with individual reasoning activity

Concerning the post-history taking with individual reasoning feedback questionnaire, the overall satisfaction degree of the students with group history taking with reasoning was about 4.6/5 to 4.8/5. The teacher’s feedback and teaching ability were rated from approximately 4.7/5 to 4.9/5. The anemia score of the participants’ self-evaluation in these five scenarios was the most difficult in history taking (3.4/5) and clinical reasoning (3.5/5). Self-evaluation ability in history taking was worse for the abdominal pain (2.9/5) and anemia scenarios (2.8/5); the self-evaluation ability in clinical reasoning was also worst in the abdominal pain (2.7/5) and anemia scenarios (2.8/5) (Table 4).

Table 4 Post training feedback questionnaire of group history taking with individual reasoning

Regarding ‘the most likely diagnosis’ after group history taking with individual reasoning, we divided the participants into ‘correct the most likely diagnosis group (n=29)’ and ‘incorrect the most likely diagnosis group (n=47)’ and found significant differences between these two groups, including the group history-taking score (p = 0.045), number of key words written down (p = 0.009), and diagnosis numbers (p = 0.004) (Table 5).

Table 5 The correlation between correct “the most likely diagnosis group” and incorrect “the most likely diagnosis group” in clinical reasoning (n = 76)

Discussion

Clinical reasoning skills may help students to better focus on the efficient history taking and physical examinations that are required for making a correct diagnosis [14]. Williams DM et al. reported that a simulated clinic model which allowed one student to perform history taking and physical examination, and then following a small group collaboratively develop a prioritized differential diagnosis could support medical students’ basic and clinical science integration [15]. In our elective individual clinical reasoning training activity for M4 medical students, a group history taking with individual reasoning was based on educational strategies including five key scenarios integrated with R1 residents as teachers enabled to facilitate the clinical reasoning learning before clinical practice. This clinical reasoning teaching and assessment activity demonstrated that the M4 students needed remedial reasoning training, the teaching activity earned the students’ good satisfaction, and we also recorded these training courses as interactive e-learning education material for other medical students.

According to previous literature, Rencic et al. reported that most students enter clerkship with only poor (29%) to fair (55%) knowledge of key clinical reasoning concepts at US medical schools, thus developing clinical reasoning curricula to improve diagnostic accuracy and reduce diagnostic error is important [16]. Bonifacino et al. reported that exposure to a clinical reasoning curriculum was associated with both superior reasoning knowledge and written demonstration of clinical reasoning skills by third-year medical students on an internal medicine clerkship [17]. Harendza S et al. said that a longitudinal implementation of clinical reasoning in the curriculum would appear to be worthwhile on students’ case presentation and differential diagnostic skills [18]. For our M4 clinical reasoning curriculum in the future, we may construct self e-learning education materials including cross-system symptoms and signs.

Does the group history taking score reflect the clinical reasoning ability of medical students? Park et al. demonstrated that the clinical reasoning score was significantly correlated with diagnostic accuracy and grade point average but not with group history taking score or clinical knowledge score, and that some students might attain a high history taking score simply by asking and checking memorized items without adequate reasoning [19]. In our study, we found that the ability to correct “the most likely diagnosis” is very important, due to its correlation with history taking, identifying key information, and reasoning ability (Table 5). The purpose of group history taking is to increase more data of key information from the training group which can improve individual clinical reasoning, however, reasoning also depends on the background of individual knowledge and critical thinking ability.

Clinical reasoning is currently receiving much attention as its theory is complex and under-taught in both undergraduate and postgraduate education. The complexities of diagnostic reasoning and therapeutic concepts, including hypothesis generation, pattern recognition, context formulation, diagnostic test interpretation, differential diagnosis, and diagnostic verification, provide both the language and the methods of clinical problem-solving [20]. A challenge of teaching clinical reasoning at an undergraduate level is the lack of formal educational sessions [16]. The teaching should not be delayed until students fully understand anatomy and pathophysiology. Weinstein et al. successfully implemented a faculty development workshop for diagnosing and remediating clinical reasoning difficulties to help clinical teachers improve their skills [21]. In our study, we created a teaching faculty including R1 residents, and the R1 teacher gave students 30 minutes immediate group feedback post training which earned high satisfaction. This study has several limitations. First, it was performed in a single institution with only 76 fourth-grade students. Therefore, these results may not be generalizable to other institutions, which may have different clinical clerkship programs and student evaluation systems. Second, the authors tested only five case scenarios, which may not be sufficient to generalize these conclusions across a wider population. Third, the group history taking score used for the present analysis was based only on the group of patient encounters, not personal encounters.

In conclusion, formal teaching of the principles of diagnosis and the causes of diagnostic uncertainty and error is a fundamental requirement in medical education for the safety of doctors and patients. This can be delivered as an educational strand through the course of the program, starting with the principles of diagnosis, diagnostic uncertainty, and misdiagnosis in the first year. It can conclude with symptom-based and patient safety-focused diagnostic reasoning, along with strategies and interventions to manage diagnostic uncertainty safely and reduce the risk of diagnostic errors, in the final year. In this study, we developed auxiliary teaching material to improve diagnostic accuracy, including R1-residents as clinical reasoning teaching faculty and using group history with individual reasoning training. We created e-learning scenario videos to improve other students’ cross-system reasoning ability, which demonstrated high satisfaction.

Availability of data and materials

The datasets generated and/or analyzed during the current study are not publicly available due to this is a pilot study in M4 clinical diagnosis medicine practice for reasoning but are available from the corresponding author on reasonable request.

Abbreviations

PBL:

Problem-based learning

OSCE:

Objective structured clinical examination

ACGME:

Accreditation Council for Graduate Medical Education

SP:

Standardized patient

References

  1. Cook CE, Décary S. Higher order thinking about differential diagnosis. Braz J Phys Ther. 2020;24(1):1–7.

    Article  Google Scholar 

  2. Koenemann N, Lenzer B, Zottmann JM, Fischer MR, Weidenbusch M. Clinical Case Discussions – a novel, supervised peer-teaching format to promote clinical reasoning in medical students. GMS J Med Educ. 2020;37(5):Doc48. https://doi.org/10.3205/zma001341.

    Article  Google Scholar 

  3. Charlin B, Roy L, Brailovsky C, Goulet F, Van der Vleuten C. The script concordance test: a tool to assess the reflective clinician. Teach Learn Med. 2000;12(4):189–95.

    Article  Google Scholar 

  4. Bordage G, Grant J, Marsden P. Quantitative assessment of diagnostic ability. Med Educ. 1990;24(5):413–25.

    Article  Google Scholar 

  5. Fischer MR, Kopp V, Holzer M, Ruderich F, Jünger J. A modified electronic key feature examination for undergraduate medical students: validation threats and opportunities. Med Teach. 2005;27(5):450–5.

    Article  Google Scholar 

  6. Sutherland RM, Reid KJ, Chiavaroli NG, Smallwood D, McColl GJ. Assessing diagnostic reasoning using a standardized case-based discussion. J Med Educ Curric Dev. 2019;6:2382120519849411.

    Article  Google Scholar 

  7. Fürstenberg S, Helm T, Prediger S, Kadmon M, Berberat PO, Harendza S. Assessing clinical reasoning in undergraduate medical students during history taking with an empirically derived scale for clinical reasoning indicators. BMC Med Educ. 2020;20(1):368.

    Article  Google Scholar 

  8. Sidhu RS, Grober ED, Musselman LJ, Reznick RK. Assessing competency in surgery: where to begin? Surgery. 2004;135:6–20.

    Article  Google Scholar 

  9. Stansfield RB, Diponio L, Craig C, Zeller J, Chadd E, Miller J, et al. Assessing musculoskeletal examination skills and diagnostic reasoning of 4th year medical students using a novel objective structured clinical exam. BMC Med Educ. 2016;16(1):268.

    Article  Google Scholar 

  10. Elliot DL, Fields SA, Keenen TL, Jaffe AC, Toffler WL. Use of a group objective structured clinical examination with first-year medical students. Acad Med. 1994;69(12):990–2.

    Article  Google Scholar 

  11. Haring CM, Klaarwater CCR, Bouwmans GA, Cools BM, van Gurp PJM, van der Meer JWM, et al. Validity, reliability feasibility of a new observation rating tool and a post encounter rating tool for the assessment of clinical reasoning skills of medical students during their internal medicine clerkship: a pilot study. BMC Med Educ. 2020;20(1):198.

    Article  Google Scholar 

  12. Haring CM, Cools BM, van Gurp PJM, van der Meer JWM, Postma CT. Observable phenomena that reveal medical students’ clinical reasoning ability during expert assessment of their history taking: a qualitative study. BMC Med Educ. 2017;17(1):147.

    Article  Google Scholar 

  13. Findyartini A, Hawthorne L, McColl G, Chiavaroli N. How clinical reasoning is taught and learned: cultural perspectives from the University of Melbourne and Universitas Indonesia. BMC Med Educ. 2016;16:185.

    Article  Google Scholar 

  14. Higgs J, Jones MA, Loftus S, Christensen N. Clinical reasoning in the health professions. 3rd ed. Edinburgh: Elsevier (Butterworth Heinemann); 2008.

    Google Scholar 

  15. Williams DM, Bruggen JT, Manthey DE, Korczyk SS, Jackson JM. The GI simulated clinic: a clinical reasoning exercise supporting medical students’ basic and clinical science integration. MedEdPORTAL. 2020;16:10926.

    Article  Google Scholar 

  16. Rencic J, Trowbridge RL Jr, Fagan M, Szauter K, Durning S. Clinical reasoning education at US medical schools: results from a national survey of internal medicine clerkship directors. J Gen Intern Med. 2017;32(11):1242–6.

    Article  Google Scholar 

  17. Bonifacino E, Follansbee WP, Farkas AH, Jeong K, McNeil MA, DiNardo DJ. Implementation of a clinical reasoning curriculum for clerkship-level medical students: a pseudo-randomized and controlled study. Diagnosis (Berl). 2019;6(2):165–72.

    Article  Google Scholar 

  18. Harendza S, Krenz I, Klinge A, Wendt U, Janneck M. Implementation of a clinical reasoning course in the internal medicine trimester of the final year of undergraduate medical training and its effect on students’ case presentation and differential diagnostic skills. GMS. J Med Educ. 2017;34(5):Doc66.

    Google Scholar 

  19. Park WB, Kang SH, Lee YS, Myung SJ. Does objective structured clinical examinations score reflect the clinical reasoning ability of medical students? Am J Med Sci. 2015;350(1):64–7.

    Article  Google Scholar 

  20. Kassirer JP. Teaching clinical reasoning: case-based and coached. Acad Med. 2010;85(7):1118–24. https://doi.org/10.1097/acm.0b013e3181d5dd0d.

    Article  Google Scholar 

  21. Weinstein A, Gupta S, Pinto-Powell R, Jackson J, Appel J, Roussel D, et al. Diagnosing and remediating clinical reasoning difficulties: a faculty development workshop. MedEdPORTAL. 2017;13:10650. https://doi.org/10.15766/mep_2374-8265.10650.

    Article  Google Scholar 

Download references

Acknowledgements

“Not applicable”.

Funding

There is no funding in this study.

Author information

Authors and Affiliations

Authors

Contributions

Dr. Lin substantial contributions to conception and design of the study, Dr. Lai and Dr. Cheng acquisition of data, or analysis and interpretation of data; Dr. Lin and Dr. Lai and Dr. Cheng drafting the article, Dr. Lin and Dr. Wu revising it critically for important intellectual content and all authors read and approved the final manuscript.

Corresponding author

Correspondence to Ching-Chung Lin.

Ethics declarations

Ethics approval and consent to participate

This study was approved, and the informed consent was waived by the MacKay Memorial Hospital Institutional Review Board (21MMHIS281e), Taipei, Taiwan,

Consent for publication

“Not applicable”.

Competing interests

“The authors declare that they have no competing interests”.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lai, JH., Cheng, KH., Wu, YJ. et al. Assessing clinical reasoning ability in fourth-year medical students via an integrative group history-taking with an individual reasoning activity. BMC Med Educ 22, 573 (2022). https://doi.org/10.1186/s12909-022-03649-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12909-022-03649-4

Keywords