- Research article
- Open Access
Teaching clinical reasoning to undergraduate medical students by illness script method: a randomized controlled trial
BMC Medical Education volume 21, Article number: 87 (2021)
The illness script method employs a theoretical outline (e.g., epidemiology, pathophysiology, signs and symptoms, diagnostic tests, interventions) to clarify how clinicians organized medical knowledge for clinical reasoning in the diagnosis domain. We hypothesized that an educational intervention based on the illness script method would improve medical students’ clinical reasoning skills in the diagnosis domain.
This study is a randomized controlled trial involving 100 fourth-year medical students in Shiraz Medical School, Iran. Fifty students were randomized to the intervention group, who were taught clinical reasoning skills based on the illness script method for three diseases during one clinical scenario. Another 50 students were randomized to the control group, who were taught the clinical presentation based on signs and symptoms of the same three diseases as the intervention group. The outcomes of interest were learner satisfaction with the intervention and posttest scores on both an internally developed knowledge test and a Script Concordance Test (SCT).
Of the hundred participating fourth-year medical students, 47 (47%) were male, and 53 (53%) were female. On the knowledge test, there was no difference in pretest scores between the intervention and control group, which suggested a similar baseline knowledge in both groups; however, posttest scores in the intervention group were (15.74 ± 2.47 out of 20) statistically significantly higher than the control group (14.38 ± 2.59 out of 20, P = 0.009). On the SCT, the mean score for the intervention group (6.12 ± 1.95 out of 10) was significantly higher than the control group (4.54 ± 1.56 out of 10; P = 0.0001). Learner satisfaction data indicated that the intervention was well-received by students.
Teaching with the illness script method was an effective way to improve students’ clinical reasoning skills in the diagnosis domain suggested by posttest and SCT scores for specific clinical scenarios. Whether this approach translates to improved generalized clinical reasoning skills in real clinical settings merits further study.
Physicians use clinical reasoning skills to gather patient data, combine it with their previous knowledge, and then make a clinical impression, diagnosis, and management plan [1, 2]. It is a critical skill in caring for patients effectively and efficiently and must be a part of every health professional education curriculum . However, according to a survey of 123 United States internal medicine clerkship directors, medical students lack clinical reasoning concepts. The authors recommended that all undergraduate medical education curriculum incorporate a structured curriculum in the clinical reasoning field .
The conceptual framework of script theory holds that human brains interpret the world by comparing the features of the mental models it makes with a real scene’s structures, checking for consistencies and inconsistencies, patterns, and irregularities. Based on the script theory, expert clinicians make a diagnosis by considering related differential diagnoses based on comparing and contrasting key features. These expert clinicians activate networks of prearranged knowledge, called “illness scripts” . The illness script method uses a theoretical outline to clarify how medical diagnostic knowledge will be organized into different categories. These categories include epidemiology, the pathophysiology of diseases, symptoms, clinical signs, and interventions, leading to an accurate diagnosis .
As novice learners, most medical students currently develop clinical reasoning skills informally in clinical wards with varying degrees of supervision. They generally organize their medical knowledge according to the components of the curriculum. When making a diagnosis, medical students often use a process of hypothesis generation and try to test one symptom at a time.
Teaching clinical reasoning by illness script model could help medical students acquire acceptable skills in generating differential diagnoses and clinical data interpretations .
Different assessment methods like clinical scenario-based multiple-choice questions, extended matching questions, and well-known clinical reasoning tests, such as the Script Concordance Test (SCT), may provide reliable evidence of medical students’ clinical reasoning skills in the diagnosis domain . SCT is one assessment of clinical reasoning skills that emphasizes data interpretation by asking learners to estimate the impact of new information on a suggested hypothesis. The SCT is based on the illness script method, which was developed in the field of cognitive psychology. The SCT measures the progress of illness scripts method in medical students as novice learners by comparing their performance on this test to a panel of expert physicians .
To better prepare the students for clinical rotations, we designed a brief educational intervention around clinical reasoning in the diagnosis domain. We hypothesized that an education intervention based on the illness script method would improve students’ clinical reasoning skills in the diagnosis domain.
Study design and participants
This Ranomized controlled trial was conducted at Shiraz Medical School, which was established in 1952 in south Iran. The medical school has a seven-year undergraduate medical education curriculum, including horizontal integration of basic science courses and 36 months of clinical rotations. Graduates are qualified for medical practice as general practitioners, but they may continue their education in specialties and subspecialties [10, 11].
The sample size had been calculated to be 32 in each group, assuming a power of 90%, the confidence interval of 95%, standard deviation 5.89, and 4.04 to find a 3.6 difference in the groups’ mean. Regarding at least 20% of the missing, this number had increased to 50 in each group.
A total of hundred fourth-year medical students were selected randomly from all students who entered the didactic classes in the internal medicine department to participate in the randomized controlled trial. Fifty students were randomized to either the control group or the intervention group. We used the CONSORT statement about randomized controlled trials . A diagram of the study design is shown in Fig. 1. For ethical purposes, after the initial intervention and measurements of outcomes, students crossed over to the other group to ultimately receive the educational content in both formats. Both groups attended teaching sessions (workshops) that lasted around 7 h, excluding breaks. Both workshops were guided by an internal medicine expert who was highly familiar with teaching clinical reasoning skills in the diagnosis domain (third author). Several tutors also helped during small group sessions during both workshops. Descriptions, timings, and agendas of each group’s workshops are described in Table 1.
The workshop’s goal in the intervention group was to help students develop a correct problem representation from the patient’s clinical problem and organize data into three illness scripts; this was based on a clinical scenario adapted from a study by Levin et al. published in MedEdPortal . .Details on the case and facilitator guide are provided in Supplementary Appendix 1. To help students compare and contrast the findings, the ‘think aloud’ method was also used in the intervention group . An external observer, familiar with the clinical reasoning, observed both intervention and control group workshops, and confirmed using the facilitator’s guide by the facilitator and tutors.
Before conducting the lecture and illness script sessions, the researchers designed ten multiple choice-questions for pretest and posttest. Each question measured a specific teaching point. In both groups, the pretest was done before, and the posttest was done after the teaching sessions. Each question had 2 points, and the total score for each of the pretest and posttest was 20. Both the pretest and posttest were completed on paper in a proctored and closed-book setting. The scores were measured using an answer key that was developed before the administration of the tests. The scorers of the pretests and posttests were blinded to the intervention.
Script concordance test
Using SCT development guidelines developed by Bernard Charlin, internal medicine experts at Shiraz medical school developed ten Script Concordance Tests (SCTs) based on the illness scripts of these three important diseases (nephrotic syndrome, cirrhosis, and congestive heart failure) . Our SCT case-based vignettes and questions were designed to evaluate clinical reasoning ability in the diagnostic domain for early medical learners. Each describes a short scenario followed by questions presented in three parts: (1) an appropriate diagnosis option; (2) a new clinical finding; and (3) a five-point Likert scale from - 2 to + 2 that should be chosen by examinees . A sample of SCT is shown in Table 2. We invited ten internal medicine expert faculties to answer the SCT.
The correct answer for an SCT was weighted based on expert response. The credit for the best answer was 100%, and credit for other answers was calculated based on the expert panel’s percentage who chose that answer . Each SCT had 1 point, and the total score of all of the SCT was 10. The students in the intervention and control group participated in this SCT 4 weeks after the intervention.
After the SCT administration, students crossed over from the intervention group to the control group and vice versa. All medical students in both groups completed a satisfaction questionnaire/survey about the illness script method after participating in the teaching by illness script workshop. The questionnaire had 14 items/questions and was designed based on our previous questionnaires about educational workshops’ satisfaction and one other study in the illness script teaching method [10, 18, 19]. Students rated each item using a Likert scale (1 = strongly disagree to 5 = strongly agree). Medical education experts determined the validity of the questionnaire using the modified Kappa variation coefficient . The modified Kappa coefficient was 0.75, and the reliability was established after a pilot study (r = 0.87).
Data were analyzed with descriptive and analytic statistics such as paired t-test using SPSS, version 16. The alpha level was at 0.05. The effect size was measure by Cohen’s d method .
The Ethics Committee approved of Shiraz University of Medical Sciences approved our study by ethical code number IR.SUMS.REC.1397.470 and did not need to register with the Iranian Registry of Clinical Trials. Informed written consent to participate was obtained from all participants. Participants joined the study voluntarily, and their scores remain confidential.
Of the hundred participating students, 47 (47%) were male, and 53 (53%) were female. The effect size was 2.04 for pretests and posttests in the intervention group, and 1.99 for the control group. There was no difference in pretest scores between the intervention and control group by student t-test (10.87 + 2.49 vs. 9.84 + 2.85, p = 0.083) which was suggestive of similar baseline knowledge in both groups. On the knowledge test, the mean posttest scores (14.38 + 2.59 in control group & 15.74 + 2.47 in intervention group) were higher than the pretest scores (9.84 + 2.85 in control group & 10.87 + 2.49 in intervention group) in both groups, and this difference by paired t-test was statistically significant (p = 0.0001 in both groups). The comparison of posttest scores by student t-test in the intervention group was significantly higher than the control group (p = 0.009). The effect size for the difference between posttests was 0.54 that is a moderate effect size. (Table 3). The intervention group’s mean score was significantly higher on the SCT than the control group (6.12 ± 1.95 vs. 4.54 ± 1.56, p = 0.0001).
Our satisfaction survey results indicated that the intervention was generally well-received by students (Table 4). Most students (82%) strongly agreed or agreed that the tutor gave them appropriate feedback. Most students (80%) also believed that the illness script method emphasized learning, and 78% of students reported that they were overall satisfied with the workshop. When students were asked how they would improve the workshop, the main suggestion was that the workshop should stress more on the “thinking aloud” approach.
This randomized controlled trial aimed to identify the effect of teaching clinical reasoning skills for the diagnosis domain based on the illness script method. Despite the illness script workshop’s briefness in the intervention group and the lecturing in the control group, both the illness script method and lecturing appeared effective. Both groups showed significant improvement in posttest scores in comparison with pretest scores. The findings also showed a high effect size between pretests and posttests in both groups. The results that students improved on the posttest are likely unsurprising as most success in clinical reasoning is attributable to knowledge gained .
Our study also showed that illness script teaching intervention helps medical students earn better scores in posttest in contrast to the control group. This better score maybe because teaching by illness scripts method helps students recognize the standard and discriminating features in the intervention group better than the control group with a medium effect size between posttests in the intervention and control groups. A study by Linsen et al. about teaching clinical reasoning to first-year medical students showed that this education would increase students’ participation in the learning process .
Like the knowledge test, the SCT result showed a higher intervention group score than the control group. However, medical students’ SCT average score in the intervention groups was around 60% of the total possible score, suggesting less proficiency even in the intervention group. These low scores might be due to their first learning experience with the illness script method and less clinical experience.
Other studies have used SCT for the assessment of clinical reasoning skills. The SCT assesses illness scripts’ formation in medical students by comparing their answers on this test to a panel of experts’ responses. In our previous studies, SCT has been described as a valid and reliable assessing tool in the clinical reasoning field [24,25,26,27]. Compared to other studies, we had performed SCT 4 weeks after original workshops, which helped to understand retention of knowledge for some time compared to the immediate posttest, which was better in the intervention group than the control group.
Like our study, several studies have shown the effectiveness of teaching clinical reasoning skills during the formal curriculum. Lee et al.’s study about teaching clinical reasoning to medical students showed that the intervention group’s students scored better than the control group on clinical reasoning tests named clinical reasoning problems . Another study by Delavari et al. about the educational strategies inspired by theory in developing illness scripts revealed that the medical students’ scores on clinical reasoning problem tests improved after the intervention . Another study by Keemink et al. about illness script development in medical students showed that case-based teaching would foster the illness script’s richness .
The satisfaction questionnaire results showed that the students were satisfied overall with the intervention. They were also satisfied with the tutors’ appropriate feedback and believed that this course would lead to real learning.
Our study’s most important strength was a randomized controlled nature, which can eliminate several confounding factors. Additionally, assessment of learning was done by multiple methods like knowledge test, SCT, and satisfaction survey showed positive results. Another study’s strength was assessing intermediate-term knowledge retention tested by SCT 4 weeks after the original workshop. There are some limitations to the present study. We might have prepared the students for tests during teaching, especially with the same instructor and tutors for both groups; however, we used the help of an external observer to monitor the educational sections to reduce this limitation. Another limitation is that SCT was not a part of baseline testing, and we only used this test for posttests.
Third, while some changes are statistically significant, we cannot ascertain whether these are educationally significant. Because the results are symptom and sign-specific, we cannot conclude that medical students develop better clinical reasoning skills in general, nor can we be sure of the clinical reasoning skills application during actual clinical practice. Third, this study is a single-center study for a specific group of learners with small sample sizes, so generalizability is limited.
Teaching with illness scripts method provided an effective way to improve students’ clinical reasoning scores for diagnosis domain posttest and SCT. Whether this approach translates to improved clinical reasoning skills in real clinical settings merits further study. Our findings can serve as a rationale for implementing clinical reasoning education modules for undergraduate medical education curricula to empower medical students in clinical reasoning skills.
Availability of data and materials
The datasets used and analyzed during the current study are available from the corresponding author on request.
Script Concordance Test
Eva KW. What every teacher needs to know about clinical reasoning. Med Educ. 2007;39:98–106.
Young M, Thomas A, Lubarsky S, Ballard T, Gordon D, Gruppen LD, et al. Drawing boundaries: the difficulty in defining clinical reasoning. Acad Med. 2018;93(7):990–5.
Van der Vulten CPM, Newble DI. How can we test clinical reasoning? Lancet. 1995;345:1032–4.
Rencic J, Trowbridge RL, Fagan M, Szauter K, Durning S. Clinical reasoning education at US medical schools: results from a national survey of internal medicine clerkship directors. J Gen Intern Med. 2017;32(11):1242–6.
Bowen JL. Educational strategies to promote clinical diagnostic reasoning. N Engl J Med. 2006;355:2217–25.
Custers EJFM. Thirty years of illness scripts: theoretical origins and practical applications. Med Teach. 2014;37(5):457–62.
Lubarsky S, Dory V, Audétat M, Custers E, Charlin B. Using script theory to cultivate illness script formation and diagnostic reasoning in health professions education. Can Med Educ J. 2016;6(2):e61–70.
Amini M, Moghadami M, Kojuri J, Abbasi H, Abadi AAD, Molaee NA, et al. An innovative method to assess clinical reasoning skills: clinical reasoning tests in the second national medical science olympiad in Iran. BMC Res Notes. 2011;4(1):418.
Lubarsky S, Dory V, Duggan P, Gagnon R, Charlin B. Script concordance testing: from theory to practice: AMEE guide no. 75. Med Teach. 2013;35(3):184–93.
Nasr K. Shiraz University School of Medicine: its foundation and development. Arch Iran Med. 2009;12:87–92.
Amini M, Kojuri J, Mahbudi A, Lotfi F, Seghatoleslam A, Karimian Z, Sham M. Implementation and evolution of the horizontal integration at shiraz medical school. J Adv Med Educ Prof. 2013;1(1):21–7.
Moher D, Schulz KF, Altman DG. The CONSORT statement: revised recommendations for improving the quality of reports of parallel-group randomised trials. Lancet. 2001;357(9263):1191–4.
Levin M, Cennimo D, Chen S, Lamba S. Teaching clinical reasoning to medical students: a case-based illness script worksheet approach. MedEdPORTAL. 2016;12:10445.
Pinnock R, Young L, Spence F, Henning M, Hazell W. Can think aloud be used to teach and assess clinical reasoning in graduate medical education? J Grad Med Educ. 2015;7(3):334–7.
Charlin B, Roy L, Brailovsky C, et al. The script concordance test: a tool to assess the reflective clinician. Teach Learn Med. 2000;12:189–95.
Fournier JP, Demeester A, Charlin B. Script concordance tests: guidelines for construction. BMC Med Inform Decis Mak. 2008;8(18):1–7.
Bland AC, Clarence D, Kreiter CD, Gordon JA. The psychometric properties of five scoring methods applied to the script concordance test. Acad Med. 2005;80(4):395–9.
Lee A, Joynt GM, Lee AKT, Ho A, Groves M, Vlantis AC, Ma RC, Fung CS, Aun CS. Using illness scripts to teach clinical reasoning skills to medical students. Fam Med. 2010;42(4):255–61.
Kojuri J, Amini M, Karimian Z, Dehghani MR, Saber M, Bazrafcan L, et al. Needs assessment and evaluation of a short course to improve faculties teaching skills at a former World Health Organization regional teacher training center. J Adv Med Educ Prof. 2015;3(1):1–8.
Gilbert GE, Prion S. Making sense of methods and measurement: Lawshe's content validity index. Clin Simul Nurs. 2016;12:530–1.
Cohen J. Statistical Power Analysis for the Behavioral Sciences. New York: Routledge Academic; 1998.
Monteiro SM, Norman G. Diagnostic reasoning: where we've been, where we’re going. Teach Learn Med. 2013;25(suppl 1):S26–32.
Linsen A, Elshout G, Pols D, Zwaan L, Mamede S. Education in clinical reasoning: an experimental study on strategies to Foster novice medical Students’ engagement in learning activities. Health Prof Educ. 2018;4(2):86–96.
Sadeghi AAA, Moulaei N, Mohammadkarimi V, Delavari S, Amini M, et al. Combination of different clinical reasoning tests in a national exam. J Adv Med Educ Prof. 2019;7(7):227–31.
Delavari S, Amini M, Sohrabi Z, Koohestani H, Delavari S, Rezaee R, et al. development and psychometrics of script concordance test (SCT) in midwifery. Med J Islam Repub Iran. 2018;32:75.
Amini M, Shahabi A, Moghadami M, Shams M, Anooshirvani A, Rostamipour H, et al. Psychometric characteristics of script concordance test (SCT) and its correlation with routine multiple choice question (MCQ) in internal medicine department. Biomed Res. 2017;28(19):8397–401.
Iravani K, Amini M, Doostkam A, Dehbozorgian M. The validity and reliability of script concordance test in otolaryngology residency training. J Adv Med Educ Prof. 2016;4(2):93–6.
Delavari S, Soltani-Arabshahi K, Monajemi A, Baradaran HR, Yaghmaei M, Myint PK. How to develop clinical reasoning in medical students and interns based on illness script theory: an experimental study. Med J Islam Repub Iran. 2020;34(9):1–5.
Keemink Y, Custers EJFM, van Dijk S, et al. Illness script development in pre-clinical education through case-based clinical reasoning training. Int J Med Educ. 2018;9:35–41.
This study is a part of a thesis by the first author, Mana Moghadami, for obtaining a Medical Doctor’s degree in Shiraz Medical School. This study was approved and financially supported by Shiraz University of Medical Sciences, Shiraz, Iran, with an ethical code: IR.SUMS.REC.1397.470. This study was also a Foundation for Advancement of International Medical Education and Research (FAIMER) project by the second author, Mitra Amini, a professor of Shiraz University of Medical Sciences and a FAIMER fellow. The fourth author, Bhavin Dalal, guided this project at the FAIMER Institute. This study was supported in part by the Foundation for Advancement of International Medical Education and Research. However, the findings and conclusions do not necessarily reflect the opinions of this organization. We thank the faculty members and medical students at Shiraz Medical School for their participation in this study. We give our special thanks to all FAIMER faculty members and fellows for their continuous support of this project. We also thank Jennifer Wilson at Jefferson (Philadelphia University + Thomas Jefferson University) for editorial assistance.
All of the funds of the present study (including the funds for performing teaching sections) were provided by the vice-chancellor of research at Shiraz University of Medical Sciences.
Ethics approval and consent to participate
The Ethics Committee approved this study of Shiraz University of Medical Sciences by ethical code number IR.SUMS.REC. 1397.470 and did not require registration with the Iranian Registry of Clinical Trials. Informed written consent to participate was obtained from all students. Students participated in the study voluntarily, and their scores remain confidential.
Consent for publication
The authors declare that they have no competing interests. Mitra Amini is the associate editor of the BMC Medical Education Journal, but there is no competing interests to declare.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Moghadami, M., Amini, M., Moghadami, M. et al. Teaching clinical reasoning to undergraduate medical students by illness script method: a randomized controlled trial. BMC Med Educ 21, 87 (2021). https://doi.org/10.1186/s12909-021-02522-0
- Clinical reasoning
- Script concordance test
- Illness script