- Research article
- Open Access
A meta-analysis of the effects of non-traditional teaching methods on the critical thinking abilities of nursing students
BMC Medical Education volume 16, Article number: 240 (2016)
Scientific framework is important in designing curricula and evaluating students in the field of education and clinical practice. The purpose of this study was to examine the effectiveness of non-traditional educational methods on critical thinking skills.
A systematic review approach was applied. Studies published in peer-reviewed journals from January 2001 to December 2014 were searched using electronic databases and major education journals. A meta-analysis was performed using Review Manager 5.2. Reviewing the included studies, the California Critical Thinking Dispositions Inventory (CCTDI) and California Critical Thinking Skills Test (CCTST) were used to assess the effectiveness of critical thinking in the meta-analysis.
The eight CCTDI datasets showed that non- traditional teaching methods (i.e., no lectures) were more effective compared to control groups (standardized mean difference [SMD]: 0.42, 95 % confidence interval [CI]: 0.26–0.57, p < .00001). And six CCTST datasets showed the teaching and learning methods in these studies were also had significantly more effects when compared to the control groups (SMD: 0.29, 95 % CI: 0.10–0.48, p = 0.003).
This research showed that new teaching and learning methods designed to improve critical thinking were generally effective at enhancing critical thinking dispositions.
The medical delivery system is changing rapidly due to developments in health technology. Aging populations, complicated changes in diseases, and increases in the number of patients with advanced diseases result in diverse and high-level health needs. To satisfy these needs in the context of such changes, healthcare providers must possess skills such as critical thinking, independence, and creativity so that they can identify solutions to problems based on quick and accurate analyses [1–3]. The Institute of Medicine  specified Evidence-Based Practice (EBP) as a core competence for all professional healthcare providers, which by 2020 aims to apply evidence that is accurate, timely, and supported by the latest clinical research to 90 % of all clinical decisions. In the EBP process, healthcare providers are not just simple agents, but thinkers with expertise who search for and evaluate evidence to solve problems that emerge in clinical practice, subsequently making decisions to provide optimum treatment and intervention. In this process, critical thinking is vital.
Critical thinking is intentional and self-regulatory judgment that leads to interpretation, analysis, evaluation, and inference. In parallel, it produces explanations concerning whether evidence for a specific judgment is appropriate, and whether it properly considers evidential, conceptual, methodological, referential, and contextual aspects. A non-linear and cyclical process enables individuals to make decisions about what to believe and do in a given context . Based on the above, Facione et al.  argued that to improve critical thinking, people must value the cognitive skills required for critical thinking and have the disposition to use them.
In accordance with these aforementioned changes, college education is also taking steps toward designing curricula that promote teaching and learning methods, as well as learning experiences, that use the latest technologies and information to nurture critical thinking amongst students . Since the 1990s, an increasing number of colleges have begun forming curricula based on problem-based learning (PBL) and self-directed learning (SDL); likewise, since the late 2000s, the popularity of simulations and concept mapping has increased. However, a consensus has not been reached concerning the most effective teaching method for improving critical thinking. Therefore, this study aims to establish a scientific framework that will be useful for designing curricula and evaluating students in the field of nursing education and clinical practice. It seeks to achieve this by systematically examining the effects of teaching and learning methods used to improve critical thinking skills.
This study is a meta-analysis conducted according to the systematic review guidelines established by the Cochrane Collaboration . A completed PRISMA checklist is included in Additional file 1. It utilizes a quantitative approach to analyze the effect and impact of teaching and learning methods used to improve nurses’ critical thinking abilities.
This study was a meta-analysis, therefore ethics committee approval was not applicable.
Studies were limited to those published from January 2001 to December 2014 in English and Korean peer-reviewed journals using the PubMed, Cochrane Library, CINAHL, Embase, and KoreaMed databases. Reference lists and major Korean academic journals were hand-searched, including the Journal of the Korean Academic Society of Nursing Education, Korean Journal of Medical Education, and the Journal of the Korean Academy of Nursing. The key search terms used included “critical thinking,” “medical,” “nursing,” “dentist,” “pharmacist,” “students,” “healthcare personnel,” “education,” and “program,” with single search terms or in combination with Boolean and wildcard.
First, this review considered research papers documenting randomized controlled trials or control group pre-post designs targeting healthcare providers such as doctors, dentists, nurses, and students.
Second, we selected research that used non-traditional teaching and learning methods (i.e., no lectures) for intervention. Third, we selected studies assessing critical thinking as the outcome. Finally, we selected studies using identical measurement (e.g., California Critical Thinking Dispositions Inventory [CCTDI]) and included means and standard deviations to verify effectiveness in the meta-analysis. We excluded studies in languages other than English and Korean. In addition, grey literature, such as papers that were not peer-reviewed (e.g., academic reports, dissertations), was also excluded.
Reviewing the outcome measurements in these inclusion criteria studies, the CCTDI and California Critical Thinking Skills Test (CCTST) were used. The CCTDI consists of 75 items and 6-point Likert scale. This tool is classified into seven subscales of truth-seeking, open-mindedness, analyticity, systematicity, critical thinking confidence, inquisitiveness and maturity . The target score was 350, while the cutoff score was 280 for overall disposition for CCTDI. In subscale analysis, each subscale score 30 or less represented weakness; 40, average; and 50 or above, strength .
The CCTST is a 34-item, multiple choice tests. This tool is classified into 5 subscales of analysis, evaluation, inference, deduction and induction . The range of score in this study is 0–34, higher scores indicating higher critical thinking ability [11, 12].
Meta-analysis was conducted using RevMan version 5.2 after related content was extracted (e.g., regarding the researcher, publication year, research design, subjects, control/experimental group teaching/learning methods, education content, education hours, measurements, and outcomes—including means and standard deviations). As studies included in the meta-analysis used some partial modification (e.g., subscale) of the CCTDI and CCTST, the standardized mean difference (SMD) using means and standard deviations was used to measure the effect size. Five studies measured the CCTDI. However, while Tiwari et al.  conducted three posttests, Kaveevivitchai et al.  conducted two; consequently, each test was analyzed separately. Thus, eight datasets were analyzed with the CCTDI.
Five studies measured the CCTST. Among these studies, Kaveevivitchai et al.  conducted two posttests, which were analyzed separately. Six datasets were analyzed with the CCTST.
Heterogeneity was examined by calculating using I2 statistics. Heterogeneity is assumed for I2 values of 0–73 % ; in such cases, a random effects model was used. The statistical meaning of the effect size was determined according to a test of overall effect and 95 % confidence intervals (CIs), based on a 5 % significance level. To verify publication bias, symmetry was examined using a funnel plot; publication bias was absent if an even distribution existed within the triangular shape.
For the final selection of literature and quality assessment, two independent researchers conducted an evaluation by applying risk of bias from Cochrane Library . These seven items included the selection bias, i.e., random sequence generation and allocation concealment, performance bias, detection bias, attrition bias, reporting bias and other sources of bias. The evaluation was conducted by classifying each item as high risk of bias, low risk of bias and unclear risk of bias. In cases where researchers disagreed, decisions were eventually made based on mutual consent.
The search findings
Following the primary search, 2534 studies were found by reviewing and hand searching the databases and references; 2309 studies remained after redundant literature was eliminated. Upon reviewing various titles and abstracts, 19 studies that satisfied each of the selection standards were identified. Of the nine studies [9, 11–14, 16–19] selected for systemic review, eight [9, 11, 13, 14, 16–19] were conducted within the realm of nursing education, while the remaining study  involved occupational therapy students.
However, Velde and colleagues  didn’t report measurement tool’s subscale data. Therefore, this study was excluded in this meta-analysis. Consequently, eight studies were selected for the final review (Fig. 1).
Overall, eight selected studies were assessed on risk of bias (Fig. 2). The results of the quality assessment revealed one study  satisfied six items of risk bias, six studies [11, 13, 16–19] satisfied five items, and one study  satisfied only three items. Three studies [11, 16, 18] were judged as having high risk of random sequence generation because these studies didn’t randomly assigned control and experimental group. Furthermore, only one study  had low risk on allocation concealment, while remaining seven studies didn’t reported the allocation sequence. Only one study  didn’t blind the intervention program to experimental group and investigator. Also during the program, participants were realized that they were observed by the researcher. Therefore, this study had a potential risk of Hawthorne effect that can produce an invalid result attributed to participants’ expectation. One study  might have attrition bias and reporting bias. Because the study reported selectively, i.e., mentioning effective experimental results only. Additionally, this study didn’t report missing data, which is an attrition bias.
Of the selected eight studies, one was published in 2006; four of them were published in each of 2003, 2004, 2007, and 2008; and three were published in 2012. The studies were conducted in a wide variety of countries including Korea , China , Thailand , Hong Kong , Taiwan , Turkey , Iran , and the United States .
Regarding the research design employed by the studies, four (50 %) used a randomized pretest-posttest control group design [9, 13, 14, 16], while four (50 %) used a quasi-experimental, nonequivalent pretest-posttest control group [11, 17–19]. Concerning the measurement used to measure critical thinking, three studies [13, 16, 17] used the CCTDI, three [11, 18, 19] were based upon the CCTST and two studies [9, 14] utilized both the CCTDI and CCTST.
The research subjects in most studies (6 studies; 75 %) included nursing students (midwifery students in one study), while staff nurses and nurse practitioner students were the participants in the remaining two. The range of the sample size was between 23 and 67, while the pooled sample size was 647 (experimental group = 327, control group = 320) and 452 (experimental group = 230, control group = 222) in studies that measured CCTDI and CCTST, respectively.
Characteristics of educational method
For the teaching and learning methods used to improve the subjects’ critical thinking skills, three used PBL [11, 13, 19], three used concept mapping [9, 16, 18], one used bioscientific multimedia , and one used a collaborative method .
The intervention period varied from 8 weeks to two semesters. Regarding the PBL, Yuan et al.’s  implementation lasted one semester, i.e., 2 h weekly for 18 weeks, totaling 36 h. Tiwari et al.’s  spanned two semesters, which took 3–6 h weekly for 28 weeks. On the other hand, lessons using concept mapping were conducted for 40 min on a biweekly basis for 16 weeks ; alternatively, as in Wheeler and Collins’  implementation, participants prepared concept maps for practical training each week during a 15-week training period following a simple orientation.
To verify the long-term effects of education, Tiwari et al.  measured subjects three times following intervention, while Kaveevivitchai et al.  measured subjects two times after intervention. The remaining seven studies measured subjects only once immediately after intervention. The characteristics of the included studies are summarized in Additional file 2.
Results of the meta-analysis
The following are the results of the meta-analysis on the overall and subscale scores using eight and six CCTDI and CCTST outcome datasets, respectively, from each study. The eight CCTDI datasets showed moderate differences (χ2 = 19.08, p = .008, I2 = 63 %). The random effects model analysis revealed that the teaching and learning methods used in these studies were significantly different than the control group (SMD: 0.42, 95 % CI: 0.26–0.57, p < .00001; Fig. 3). The CCTDI cutoff and target scores were 280 and 350, respectively . Scores of the experimental group in three studies [13, 14, 17] exhibited higher than 280 after the non-traditional educational intervention. However, each of experimental group did not reach the target score, i.e., 350. Analysis of the CCTDI subscale scores for truth-seeking (SMD: 0.32, 95 % CI: 0.01–0.47, p < .0001), open-mindedness (SMD: 0.37, 95 % CI: 0.22–0.53, p < .00001), analyticity (SMD: 0.28, 95 % CI: 0.09–0.46, p = .004), critical thinking confidence (SMD: 0.34, 95 % CI: 0.18–0.49, p < .0001), inquisitiveness (SMD: 0.36, 95 % CI: 0.21–0.52, p < .00001), and maturity (SMD: 0.16, 95 % CI: −0.01–0.32, p = 0.06) revealed a more effective increase as compared to the control group (Additional file 3). When the score of the CCTDI subscale should be higher than 50 to indicate strengthen critical thinking disposition, only one study showed a score of 50 or higher for ‘open-mindedness’ and ‘inquisitiveness’ . In the funnel plot, there was symmetric shape suggesting a lack of publication bias (Fig. 4).
The six datasets presenting the effects of teaching and learning methods on CCTST exhibited a high level of difference (χ2 = 23.32, p = .0003, I2 = 79 %). Consequently, the random effects model was used for analysis, teaching and learning methods used in these studies were significant effects on the overall CCTST score when compared to the control group (SMD: 0.29, 95 % CI: 0.10–0.48, p = 0.003; Fig. 3). Analysis of the subscale scores, however were not revealed a more effective increase as compared to the control group (Additional file 4). Publication bias was examined using the funnel plot that revealed a symmetrical shape suggesting a lack of bias (Fig. 5).
Analysis of the effects of teaching and learning methods revealed that concept mapping (SMD: 0.68, 95 % CI: 0.26–1.11, p = 0.002, I2 = 77 %) was effective in improving critical thinking (Fig. 6). However, PBL (SMD: 0.34, 95 % CI: −0.03–0.70, p = 0.07, I2 = 62 %) was not significantly effective in improving critical thinking.
This study was conducted to verify the effects of teaching and learning methods used to improve the critical thinking of healthcare providers. As nurses must make correct judgments and efficient decisions in diverse and complex clinical situations, critical thinking is important in professional nursing. Therefore, the findings of this study are especially meaningful.
The meta-analysis revealed that diverse teaching and learning methods (i.e., concept mapping, bioscientific multimedia) are more effective than are traditional approaches in improving dispositions towards critical thinking. This result is similar with previous studies [16, 20, 21]. According to Taylor and Wros’s study , concept mapping was an effective visualizing learning method, especially organizing and analyzing the patient data. Concept mapping can provide the important factors as well as inter-relational knowledge, therefore, students construct the basic concept. Overall, concept mapping might be positive effect to develop students’ critical thinking.
The overall CCTST score maintained an average level of 12.4–21.8 according to Huang et al.’s  standard. Although critical thinking disposition significantly increased post-intervention, it did not reach a level of excellence. This indicates that it is difficult to anticipate sufficient improvement, as the intervention was performed between the first and second semester, which is a short period of time for enhancing critical thinking. Moreover, Tiwari et al.  measured the results three times: immediately after intervention, 1 year after, and 2 years after. The results revealed that the critical thinking disposition score gradually decreased as time passed, showing no significant difference after 2 years. This implies that sufficient effects cannot be anticipated after merely conducting education for one or two semesters in a single course, and that continuous education is needed in a variety of courses.
However, the CCTDI and CCTST are commercialized measurements that have been used in various studies. Simpson and Courtney’s research  asserted that these measurement tools had limitation for nursing students or nurses. Critical thinking in nursing education is the ability to assess, analyze and understand the patients’ contextual clinical situations . However, these commercialized tools measured the limited aspect of critical thinking, such as analyzing and interpreting the suggested patient written data [24, 25]. These tools are not able to measure the students’ performance for example, patient specific situation driven critical thinking. Thus, these measurements were insufficient measuring the critical thinking abilities of nursing students or nurses. Furthermore, these measurements use self-report method, which may cause participants to respond in a manner that they believe society anticipates. Thus, it is necessary to interpret this study’s results with caution.
More than 20 years have passed since concepts concerning critical thinking and the education of healthcare providers have been reformed to meet the demands of outcome-based education. As the concept of critical thinking in nursing education is in constant discussion, it is necessary to consider the clinical context and the patient’s situation, not merely evaluate critical thinking skills and dispositions. Therefore, an objective measurement of critical thinking with a focus on empirically measuring student performance must be developed to determine how critical thinking should be applied in evidence-based nursing practice, and whether patients’ health problems are solved as a result.
Verifying the effectiveness of teaching and learning method showed that concept mapping was effective in improving critical thinking. This result is consistent with the previous systemic review findings in nurse education . According to this study, reflective writing, concept mapping and case studies are interventions that enhance critical thinking in the context of nursing education.
In this meta-analysis, PBL was not effective in improving nursing students’ critical thinking. This finding is inconsistent with previous reports [19, 26, 27]. Recent a systemic review  study explained that learners’ readiness, fluency or trait, educators’ belief or attitude in critical thinking, or learning environment can bring different educational effect culturally. Kong and colleagues  also described that different educational method or environment can influence PBL educational effects. Additionally, Kong et al.’s  meta-analysis selected more than one instrument in order to examine the CCTDI and CCTST (i.e., Watson-Glaser Critical Thinking Appraisal [WGCTA], Assessment Technologies Institute Critical Thinking Test [ATI], etc.), therefore, there can be different results. However, there were insufficient evidences to support this study’s result. Thus, further studies should be conducted to examine the effectiveness of PBL on critical thinking ability.
A moderate and high degree of heterogeneity was presented in this study. We included diverse educational method to examine the effectiveness of non-traditional teaching methods on the critical thinking in this meta-analysis. Aforementioned difference can significantly affect the heterogeneity.
The quality appraisals of the eight studies equally demonstrated that an insufficient amount of research applied concealment, double blinding, and multiple study sites. The reason might be that researchers primarily functioned as educators providing the intervention. The studies did not indicate whether certain actions were taken to reduce any potential bias that may have arisen, given the issues mentioned above. Thus, to establish a solid foundation for the validity and generalization of the results, randomized controlled trials must be conducted at multiple sites by applying strict research designs.
Compared to previous studies, this study had the advantage of securing generally high-quality research for meta-analysis; this is evident in its use of studies that applied randomized controlled trials and pretest-posttest control group designs in their verification of teaching and learning methods designed to improve critical thinking.
There were several limitations in this study. First, only eight studies were included for meta-analysis. While visual inspection of the funnel plots revealed a symmetrical shape suggesting a lack of publication bias, the limitations of funnel plots to detect publication bias are well known, especially when the number of studies included is less than 10 and a large degree of heterogeneity exists among studies [29, 30]. Secondly, the specific intervention methods, duration, contents of the teaching and learning methods, and study quality were varied considerably by moderate to high heterogeneity reported. Thirdly, all eight studies were retrieved from the nursing literature which limits the ability of our results to be generalized to other healthcare providers.
This research showed that new teaching and learning methods designed to improve critical thinking were generally effective in enhancing critical thinking dispositions. In particular, concept mapping was effective in increasing both critical thinking skills and dispositions. However, teaching and learning methods for the improvement of critical thinking must be implemented continuously throughout a curriculum. As critical thinking is an essential concept for integrated problem solving in clinical situations, it is necessary to focus on measuring capabilities in practice rather than by evaluating critical thinking by dividing it into cognitive and affective domains. Furthermore, greater effort is needed to improve research quality in order to generalize the results.
Assessment technologies institute critical thinking test
California critical thinking dispositions inventory
California critical thinking skills test
Guided reciprocal peer questioning
Standardized mean difference
Watson-glaser critical thinking appraisal
Finkelman AW. Problem-solving, decision-making, and critical thinking: how do they mix and why bother? Home Care Provid. 2001;6(6):194–7. quiz 198–199.
Shin KR, Hwang JW, Shin SJ. Concept analysis on the clinical critical thinking ability in nursing. Korean J Adult Nurs. 2008;20(5):707–18.
Huang L, Wang Z, Yao Y, Shan C, Wang H, Zhu M, Lu Y, Sun P, Zhao X. Exploring the association between parental rearing styles and medical students’ critical thinking disposition in China. BMC Med Educ. 2015;15:88.
Institute of Medicine. Evidence-based medicine and the changing nature of health care. https://www.nap.edu/catalog/12041. Accessed 7 Mar 2012.
American Philosophical Association. Critical Thinking: A statement of expert consensus for purposes of educational assessment and instruction. “The Delphi Report”. In: Committee on Pre-college Philosophy; 1990.
Facione NC, Facione PA, Sanchez CA. Critical thinking disposition as a measure of competent clinical judgment: the development of the California Critical Thinking Disposition Inventory. J Nurs Educ. 1994;33(8):345–50.
Seymour B, Kinn S, Sutherland N. Valuing both critical and creative thinking in clinical practice: narrowing the research-practice gap? J Adv Nurs. 2003;42(3):288–96.
The Cochrane Collaboration. Cochrane Handbook for Systematic Reviews of Interventions. Version 5.1.0. 2011.
Huang YC, Chen HH, Yeh ML, Chung C. Case studies combined with or without concept maps improve critical thinking in hospital-based nurses: A randomized-controlled trial. Int J Nurs Stud. 2012;49(6):747–54.
Facione PA, Facione NC. Test manual: The California Critical Thinking Skills Test, Form A and Form B. Milbrae: The California Academic Press; 1993.
Choi HJ. The Effects of PBL(Problem-Based Learning) on the Metacognition, Critical Thinking, and Problem Solving Process of Nursing Students. J Korean Acad Nurs. 2004;34(5):712–21.
Velde BP, Wittman PP, Vos P. Development of critical thinking in occupational therapy students. Occup Ther Int. 2006;13(1):49–60.
Tiwari A, Lai P, So M, Yuen K. A comparison of the effects of problem-based learning and lecturing on the development of students’ critical thinking. Med Educ. 2006;40(6):547–54.
Kaveevivitchai C, Piaseu N, Luptrawan S, Sirikoon N, Panijpan B. Supportive-educational program: Using bioscientific multimedia to enhance clinical problem solving skill in general nurse practitioner students. Thai J Nurs Res. 2007;11(4):295.
Higgins JP, Thompson SG, Deeks JJ, Altman DG. Measuring inconsistency in meta-analyses. BMJ. 2003;327(7414):557–60.
Atay S, Karabacak U. Care plans using concept maps and their effects on the critical thinking dispositions of nursing students. Int J Nurs Pract. 2012;18(3):233–9.
Iranfar K, Iranfar S, Mohammadi G. Developing critical thinking disposition in the students of nursing and midwifery through collaborative and individual methods of learning. HealthMED. 2012;6(12):4047–52.
Wheeler LA, Collins SK. The influence of concept mapping on critical thinking in baccalaureate nursing students. J Prof Nurs. 2003;19(6):339–46.
Yuan H, Kunaviktikul W, Klunklin A, Williams BA. Improvement of nursing students’ critical thinking skills through problem-based learning in the People’s Republic of China: a quasi-experimental study. Nurs Health Sci. 2008;10(1):70–6.
Taylor J, Wros P. Concept mapping: a nursing model for care planning. J Nurs Educ. 2007;46(5):211–6.
Zunkel GM, Cesarotti EL, Rosdahl D, McGrath JM. Enhancing diagnostic reasoning skills in nurse practitioner students: a teaching tool. Nurse Educ. 2004;29(4):161–5.
Simpson E, Courtney M. Critical thinking in nursing education: literature review. Int J Nurs Pract. 2002;8(2):89–98.
Schank MJ. Wanted: nurses with critical thinking skills. J Contin Educ Nurs. 1990;21(2):86–9.
Daley BJ, Shaw CR, Balistrieri T, Glasenapp K, Piacentine L. Concept maps: A strategyto teach and evaluate critical thinking. J Nurs Educ. 1999;38(1):42–7.
Magnussen L, Ishida D, Itano J. The impact of the use of inquiry-based learning as a teaching methodology on the development of critical thinking. J Nurs Educ. 2000;39(8):360–4.
Yu D, Zhang Y, Xu Y, Wu J, Wang C. Improvement in critical thinking dispositions of undergraduate nursing students through problem-based learning: a crossover-experimental study. J Nurs Educ. 2013;52(10):574–81.
Kong LN, Qin B, Zhou YQ, Mou SY, Gao H. The effectiveness of problem-based learning on development of nursing students’ critical thinking: a systematic review and meta-analysis. Int J Nurs Stud. 2014;51(3):458–69.
Chan ZC. A systematic review of critical thinking in nursing education. Nurs Educ Today. 2013;33(3):236–40.
Lau J, Ioannidis JP, Terrin N, Schmid CH, Olkin I. The case of the misleading funnel plot. BMJ. 2006;333(7568):597–600.
Sterne JA, Sutton AJ, Ioannidis JP, Terrin N, Jones DR, Lau J, Carpenter J, Rucker G, Harbord RM, Schmid CH, et al. Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials. BMJ. 2011;343:d4002.
This study was supported by College of Nursing, Yonsei University and Pusan National University Research Grant, 2014.
Availability of data and materials
Authors can confirm that all relevant data are included in the article and its supplementary information files.
JL designed this study, interpreted the data, and drafted the manuscript. YL performed data collection, interpretation, statistical analyses, and drafted the manuscript. SG performed data collection, interpretation. JB performed data collection, interpretation, and helped to draft the manuscript. MC performed data interpretation, and helped to draft the manuscript. All authors read and approved the final manuscript.
Competing of interests
The authors declare that they have no competing interests.
Consent for publication
Ethics approval and consent to participate
About this article
Cite this article
Lee, J., Lee, Y., Gong, S. et al. A meta-analysis of the effects of non-traditional teaching methods on the critical thinking abilities of nursing students. BMC Med Educ 16, 240 (2016). https://doi.org/10.1186/s12909-016-0761-7
- Critical thinking