Skip to content

Advertisement

  • Research article
  • Open Access
  • Open Peer Review

Flipped classroom improves student learning in health professions education: a meta-analysis

BMC Medical EducationBMC series – open, inclusive and trusted201818:38

https://doi.org/10.1186/s12909-018-1144-z

  • Received: 10 June 2017
  • Accepted: 8 March 2018
  • Published:
Open Peer Review reports

Abstract

Background

The use of flipped classroom approach has become increasingly popular in health professions education. However, no meta-analysis has been published that specifically examines the effect of flipped classroom versus traditional classroom on student learning. This study examined the findings of comparative articles through a meta-analysis in order to summarize the overall effects of teaching with the flipped classroom approach. We focused specifically on a set of flipped classroom studies in which pre-recorded videos were provided before face-to-face class meetings. These comparative articles focused on health care professionals including medical students, residents, doctors, nurses, or learners in other health care professions and disciplines (e.g., dental, pharmacy, environmental or occupational health).

Method

Using predefined study eligibility criteria, seven electronic databases were searched in mid-April 2017 for relevant articles. Methodological quality was graded using the Medical Education Research Study Quality Instrument (MERSQI). Effect sizes, heterogeneity estimates, analysis of possible moderators, and publication bias were computed using the Comprehensive Meta-Analysis software.

Results

A meta-analysis of 28 eligible comparative studies (between-subject design) showed an overall significant effect in favor of flipped classrooms over traditional classrooms for health professions education (standardized mean difference, SMD = 0.33, 95% confidence interval, CI = 0.21–0.46, p < 0.001), with no evidence of publication bias. In addition, the flipped classroom approach was more effective when instructors used quizzes at the start of each in-class session. More respondents reported they preferred flipped to traditional classrooms.

Conclusions

Current evidence suggests that the flipped classroom approach in health professions education yields a significant improvement in student learning compared with traditional teaching methods.

Keywords

  • Flipped classroom
  • Flipped learning
  • Health professions education
  • Meta-analysis

Background

Many classes in higher education institutes now employ blended learning; whereby students learn in part at a supervised face-to-face location on campus, and in part through the Internet with some elements of student choice over place and pace [1].Of the many different models of blended learning in practice, the use of flipped classroom approach has become increasingly widespread [24].

Initially popularized in the United States [5], flipped classrooms replace teacher-led in-class instructions with individual homework or group activities [6]. Recently, the flipped classroom approach has made inroads into health professions education, and has even been touted “a new paradigm” in medical education [7]. Various health professions have adopted this instructional approach into their curricula [8]. A recent review of learner perceptions of flipped classrooms in health professions education [8] found an overwhelming positive response from students who attended flipped courses. More specifically, students expressed high levels of satisfaction with pre-class video lectures because the videos can be accessed at any time and as often as they desire. Students also highly regarded the use of small group discussion-based activities in flipped classroom face-to-face sessions because these sessions help increase their motivation to learn, enhance their level of engagement, and interest in the subject matter [8].

But does using the flipped classroom approach in health professions education really improve student learning? It is important to note that positive student perception toward flipped classrooms does not necessarily imply that this instructional approach will significantly improve student learning [8]. For example, flipping the evidence-based medicine (EBM) course by Ilic et al. [9] did not improve scores on the Berlin objective assessment of EBM competencies compared to its traditional counterpart, despite students reporting a positive perception of the flipped course.

Up to now the effectiveness of flipped classroom approach compared with traditional learning has not been established. Although several literature review studies have been conducted in health care professionals such as nursing [10, 11] and medical education [8, 12], no meta-analysis has been published that specifically examines the effect of flipped classroom versus traditional classroom on student learning. By traditional classroom, we refer to the approach of having students come to class during which teachers use a range of pedagogical strategies (e.g., lecture, case discussion, student presentation), and then students complete most of their homework after school [13, 14]. The uncertainty about the effectiveness of flipped classroom approach over traditional instruction provided the impetus for the current study. We opted to contrast flipped classrooms with traditional classrooms because the latter is still widely used in health professions education [15].

Conceptual framework

For the purposes of this study, we adopted the conceptual framework of flipped classroom approach by educause [16], one of the leading associations that focus on instructional technology in higher education, as a pedagogical strategy in which the typical lecture and homework elements of a course are reversed. In a typical traditional classroom, students listen to lectures in class and complete most of their homework after class. In a flipped classroom, students listen or watch pre-recorded lectures before class and perform active learning activities such as exercises, projects, or discussions [16]. Problem-based learning may be one of the activities used in flipped classroom [17].

Although pre-recorded lectures could be a podcast or other audio format, the use of videos has become so ubiquitous that the flipped classroom approach has come to be identified with pre-recorded videos [16, 18, 23]. Therefore, in the present review, we focused specifically on flipped classroom studies in which pre-recorded videos were provided (rather than live lectures, or intelligent tutoring system without a video or instructor) prior to face-to-face class meetings.

It is important to stress that this definition excludes the sole use of pre-class reading materials as a form of flipped learning. Text-based materials cannot “closely mimic what students in a traditional setting would experience” [19] since it does not involve instructors’ explanation and elaboration of contents [20]. Muir and Geiger [21] reported that “a book doesn’t really walk through the steps on how to do something.” In contrast, the use of video lectures enables the instructors to elaborate the course contents as in a traditional lecture [20, 21].

Methods

Data sources and search strategies

This meta-analysis and review were carried out according to the PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analysis) guidelines [22]. Relevant online databases were searched from January 2012 through March 2017. January 2012 onwards was chosen because 2012 was the year of the first publication of an application of flipped classroom approach to health professions student teaching [23]. Altogether, seven electronic databases were searched in mid-April 2017, including Academic Search Complete, PubMed, PsycINFO, CINAHL Plus, TOC Premier, British Nursing Index, and ERIC. To capture a broader range of potentially eligible articles, we employed the following search terms with Boolean operators: “(flip* or invert*) and (class* or learn* or instruction* or course*) and (medic* or nurs* or pharmac* or physiotherap* or dental or dentist* or chiropract*)”. The asterisk was used as a wild card to include most of the common expressions of the flipped classroom approach (e.g., flipped learning, flipped class, flipping the classroom). The search term was entered as a string and searched in each of the seven databases.

Eligibility criteria

To examine the possible effects of flipped classroom on student achievement, a meta-analysis was done on eligible articles. The eligibility criteria are as follows:

Inclusion

  1. a)

    The studies must report at least one comparison of a flipped classroom condition versus a traditional classroom condition (i.e., between-subject design) focusing on health care professionals including medical students, residents, doctors, nurses, or learners in other health care professions and disciplines (e.g., dental, pharmacy, environmental or occupational health). The flipped classroom implementation must consist of both pre-class and in-class activities.

     
  2. b)

    The pre-class flipped classroom activities must at least include the use of instructor-recorded classroom lectures, PowerPoints with instructor talking head, YouTube videos, Khan Academy videos, TED (technology, entertainment, design) video talks, screencast, or PowerPoints with instructor’s voice over.

     
  3. c)

    In the present review, we include only flipped classroom studies that have a face-to-face meeting because face-to-face class meeting is typically used in many flipped classroom implementations [18, 24]. Comparing online courses (without face-to-face meeting) with flipped classrooms is outside the scope of our review.

     
  4. d)

    The traditional classroom involves students coming to class during which teachers typically give a lecture [25], and might use other presentation strategies (e.g., case discussion, group-work, student presentation).

     
  5. e)

    Only comparative studies such as randomized controlled trial, quasi-experiments, and historical cohort controlled research designs were included for review.

     
  6. f)

    Articles must measure student learning on similar course topics using some form of objective-based assessments such as post-tests or exams. These assessment instruments must be similar or identical.

     
  7. g)

    No geographical restrictions were imposed, however the articles must be written in English and published in peer-reviewed journals. Searching for peer-reviewed publications is a useful criterion for selecting studies of sufficient quality [26].

     

Exclusion

a) Published studies were excluded if their datasets or results were incomplete – such as if they lacked sufficient information to calculate effect sizes (e.g., sample sizes), or if effort to obtain data from corresponding authors was not successful.

Study selection and data extraction

The title and abstract of the studies from initial the search process were screened in order to derive a preliminary set of full articles for potential final review. This was followed by a reading of the preliminary full articles by the authors individually to confirm the relevance of the studies before a final decision (through discussion) was made to confirm the studies to be included in the final review. To extract the data, we looked for information including authors of the study, publication year, location in which the study was conducted, subject topic, participant sample, study design such as quasi-experiments, randomized control trials, or historical controls, and details of the flipped classroom implementation such as the types of pre-class and in-class activities used. The percent agreement between the coders concerning the data extraction was high – 95%. To reach consensus, the discrepancies between the extracted data of the two researchers were reviewed, discussed, and resolved prior to data entry and analysis.

For all the eligible studies, one effect size was calculated for each study to meet the assumption of the independence of the effect sizes based on independent samples of students. In cases where articles reported multiple assessments of a single course subject, we selected the assessment that was most summative, as suggested by Freeman et al. [27]. For example, we chose final exam over other assessments (e.g., mid-term exam, weekly quiz). We also chose assessment that is recognized as a widely-used inventory [27] – for example, the Objective Structured Clinical Examination over an instructor-written examination. In cases, where a single study had multiple outcomes from different course subjects from the same set of students, we computed a single combined effect size using the formulas proposed by Borenstein et al. [28]. In doing so, we assumed that the correlation between the different outcomes within a comparison was 1 as suggested by Freeman et al. [27] since the same students were sampled for each outcome. This is a conservative measure as the actual correlation between outcomes is likely lower than 1 [27].

Meta analyses

We computed effect sizes using the Comprehensive Meta-Analysis Version 3 software (Biostat, Inc., Englewood Cliffs, NJ, USA). All reported p values are two-tailed, unless otherwise reported. To compare the effect sizes, we used a random effects model, or random effects analysis because conditions that could affect student achievements differed among studies in the analysis, including the frequency of lessons flipped, student population, and course level. We computed effect sizes using standardized mean differences (SMDs) from the means and standard deviations of student achievement data (e.g., exam scores, post-test scores). If standard errors were used in the previous empirical studies but not the standard deviations, we used the following formula [29] to calculate the standard deviations:
$$ SE=\frac{SD}{\sqrt{\mathrm{sample}\ \mathrm{size}}}. $$

If the means and standard deviations were not reported in the previous empirical studies, the standardized mean difference was estimated using a variety of sources, including t-tests (or formulas, see Borenstein et al. [28]; Lipsey & Wilson [30]). The presence of heterogeneity (i.e. the degree of inconsistency in the studies’ results) was detected by the I2 test. Publication bias (or otherwise known as file-drawer problem) occurs when researchers publish only favorable results [63]. Therefore, in order to determine whether the present review suffers from publication bias, we conducted the following standard tests used to analyze publication bias including: (a) assessing the funnel plot, (b) computing Begg and Mazumdar rank correlation, (c) calculating Egger’s regression, (d) computing Duval and Tweedie’s trim and fill, and (e) calculating the classic fail-safe N test. Currently, sufficient information was available for all these tests.

We also conducted subgroup analyses across six major categorical moderator variables in order to identify the possible source of variation among the effect sizes and the differences among the subgroups. These moderator variables were: (a) student initial equivalence, (b) instructor equivalence, (c) research design, (d) types of students; (e) pre-class component of flipped classroom, and (f) in-class component of flipped classroom.

To determine student initial equivalence, we examined whether the study design was based on the following categories: (a) comparative studies where authors provided no data or no statistical control on student equivalence in terms of initial academic performance, or where authors merely claimed students were equivalent but did not provide any relevant statistical evidence (e.g., t-test results); or (b) comparative studies where data indicated no statistical difference on a pretest that directly relate to the topic, or on a metric of academic performance (e.g., college GPA). To evaluate instructor equivalence, we checked whether the study involved: (a) identical instructor for the flipped-and-non-flipped classes; or (b) different instructors for different classes; or (c) no data provided.

In addition, we stratified the analysis according to research design (randomized control, quasi-experiment, or historical control), types of students (e.g., medicine, pharmacy, public health, etc.), pre-class component (availability of pre-class assessment/exercise or not, availability of readings/notes or not), and in-class component (use of quiz at start of lesson or not).

Results

The literature search process is depicted in Fig. 1. Our initial search yielded a total of 2129 articles, which was reduced to 1655 after duplicates were removed. However, after reviewing their titles and abstracts, many articles (n = 1611) were found to be irrelevant, particularly those that did not report empirical research, or were not related to health professions education. For example, many irrelevant search outcomes came from articles that did not collect and analyze any form of data (i.e., non-empirical), and from basic science disciplines (e.g., inverted structural plasticity, invertebrates, and tissue classification). A total of 44 full text articles were read and assessed for eligibility. Out of these 44 articles, six were removed because it did not fulfill our criteria of a flipped classroom approach, eight were removed due to insufficient data on student performance or descriptions of the learning activities, and three were removed because the flipped course and traditional course offering were not comparable. Twenty-eight articles were included in the final analysis.
Fig. 1
Fig. 1

PRISMA flow diagram of article selection

Study demographics

Table 1 descriptively summarizes the main elements of the 28 studies included in the meta-analysis. Most of the studies were based on historical control designs. Only four studies were quasi-experiments. The remaining four were randomized controlled trials.
Table 1

Overview of studies included in the meta-analysis

Study

Country of origin

Subject matter

Sample

Study design

Duration of intervention

Bakr et al. (2016) [31]

Australia

Dental anatomy

1st year dental students

Historical control

13 weeks

Bossaer et al. (2016) [32]

USA

Pharmacy

3rd year pharmacy students

Historical control

1 semester (actual duration not specified)

Cheng et al. (2016) [33]

China

Histology

1st year medicine students

Quasi-experiment

1 semester (actual duration not specified)

Cotta et al. (2016) [13]

USA

Pharmaceutical calculations

1st year pharmacy students

Historical control

10 weeks

Galway et al. (2014) [34]

Canada

Environmental and occupational health

1st or 2nd year graduate public health students

Historical control

13 week

Gillispie (2016) [35]

USA

Obstetrics and gynecology

3rd/4th year medical students

Historical control

8 weeksa

Giuliano and Moser (2016) [36]

USA

Drug literature

1st year pharmacy students

Historical control

1 semester (actual duration not specified)

Harrington et al. (2015) [37]

USA

Nursing

Baccalaureate students

Randomized

4 months

Hsu et al. (2016) [38]

Taiwan

General medicine

Postgraduate year medical students

Historical control

12 months

Ilic et al. (2015) [9]

Australia

Evidence-based medicine

3rd year medical students

Randomized

10 two-hour teaching sessions

Kiviniemi (2014) [39]

USA

Public health

Graduate students

Historical control

1 semester (actual duration not specified)

Koo et al. (2016) [40]

USA

Pharmacotherapy

2nd year doctor of pharmacy students

Historical control

8 sessions

Liebert et al. (2016) [41]

USA

Surgery

3rd year medical student

Historical control

8 weeks

Lin et al. (2017) [14]

China

Ophthalmology

Medical undergraduates

Randomized

1 semester (actual duration not specified)

McLaughlin et al. (2013) [42]

USA

Pharmaceutics

1st year pharmacy students

Historical control

13 weeks

McLaughlin et al. (2014) [43]

USA

Pharmaceutics

1st year professional students

Historical control

13 weeks

Morton and Colbert-Getz (2016) [44]

USA

Anatomy

1st year medical students

Historical control

17 weeks

Munson and Pierce (2015) [45]

USA

Pharmacogenomics

Mixture of students with bachelor degree and students without first degree

Historical control

15 weeks

O’Connor et al. (2016) [15]

USA

Radiology

3rd or 4th year medical students

Quasi-experiment

2 weeks

Pierce and Fox (2012) [6]

USA

Pharmacotherapy

Pharmacy students

Historical control

8 weeks

Porcaro et al. (2016) [46]

Australia

Hematology

Mixture of undergraduate and postgraduate students in medical science

Historical control

1 semester (actual duration not specified)

Prescott et al. (2016) [47]

USA

Patient assessment

1st year pharmacy students

Historical control

2 semesters (actual duration not specified)

Rui et al. (2017) [48]

China

Medical diagnostics

Junior-year medical students

Randomized

3 weeks

Sajid et al. (2016) [49]

Saudi Arabia

Hematology

3rd year medical students

Historical control

5 lectures

Street et al. (2015) [17]

USA

Physiology

1st year medical students

Quasi-experiment

6 weeks

Tune et al. (2013) [50]

USA

Physiology

1st year graduate medical students

Quasi-experiment

1 semester (actual duration not specified)

Whillier and Lystad (2015) [51]

Australia

Neuroanatomy

2nd year undergraduate chiropractic students

Historical control

5 weeks

Wong et al. (2014) [52]

USA

Pharmacy

1st year pharmacy students

Historical control

1 week

anote: Gillispie (2016) [35] - 8 weeks information obtained from the website: https://medicine-program.uq.edu.au/current-students/placements/core-rotations

Methodological quality was graded using the Medical Education Research Study Quality Instrument (MERSQI) [53]. We summarize the study quality in Table 2. The mean MERSQI score was 12.5 on an 18-point scale.
Table 2

Quality of studies (N = 28) based on MERSQI

Scale item (max. points)

Subscale (points if present)

No. (%) present

Study design (max. 3)

Non-randomized 2-group (2)

24 (86)

Randomized 2-group (3)

4 (14)

Sampling: no. of institutions (max. 1.5)

1(0.5)

26 (93)

2 (1)

1 (3.5)

3 (1.5)

1 (3.5)

aSampling: response rate (max. 1.5)

<  50% or not reported (0.5)

5 (17)

50%–74% (1) (1)

1 (3.5)

≥75% (1.5)

22 (79)

Type of data: outcome assessment (max. 3)

Objective (3)

28 (100)

Validity evidence (max. 3)

Content (1)

28 (100)

Data analysis: appropriate (max. 1)

Appropriate (1)

28 (100)

Data analysis: sophistication (max. 2)

Beyond descriptive analysis (2)

28 (100)

Highest outcome type (max. 3)

Knowledge, skills (1.5)

28 (100)

aNote: Sampling (response rate) refers to the proportion of students enrolled who completed the flipped classroom approach

Meta-analysis

A meta-analysis of 28 eligible comparative studies involving 2295 subjects exposed to flipped classroom and 2420 subjects exposed to traditional classroom showed an overall significant effect in favor of the flipped classroom approach for health professions education (SMD = 0.33, 95% CI 0.21–0.46, p < 0.001) as shown in Fig. 2. A significant Q statistic (p < 0.001) indicated the presence of heterogeneity (I2 = 75.6%).
Fig. 2
Fig. 2

Forest plot of effect sizes (standardized mean difference) using random effect model. Note that data in Pierce and Fox [6] were provided by the corresponding author

Moderator analyses

In order to explore the possible causes for the significant heterogeneity, we conducted several moderator analyses using the random-effects mode. The results of the analyses are summarized in Table 2.

When we analyzed the data based on whether the studies controlled for student or instructor equivalence, we found no evidence of heterogeneity between studies that reported initial student equivalence, studies that employed randomized student assignment, and studies that did not provide any such data (Q = 0.43, df = 2, p = 0.81).

Analyzing variation with respect to instructor equivalence also suggested no evidence of heterogeneity (Q = 4.72, df = 2, p = 0.09). Thus, the overall effect size for student performance data appears to be robust to varying methodological rigor of published studies (e.g., poorly controlled studies with different instructors, or with no data provided on student or instructor equivalence).

Heterogeneity analyses also indicated no significant variation when comparing (a) studies with different research design such as historical control, quasi-experiment, or randomized experiment (Q = 0.52, df = 2, p = 0.77); (b) studies with different types of students such as medicine, pharmacy (Q = 2.44, df = 5, p = 0.78); (c) studies that employed pre-class assessment/exercise, or not (Q = 2.67, df = 1, p = 0.10); or (d) studies that provided pre-class readings/notes, or not (Q = 0.11, df = 1, p = 0.74).

However, a heterogeneity analysis in Table 3 indicated that the effect size is significantly higher when instructor(s) employed quizzes at the start of an in-class session to assess students’ learning of the pre-class video contents as opposed to instructor(s) who did not (Q = 5.34, df = 1, p = 0.02).
Table 3

Moderator analyses

Moderator

   

95% CI

n

SMD

SE

LL

UL

For student equivalence

 No data provided

13

0.31

0.10

0.11

0.50

 No statistical difference (pretest, or other scores)

11

0.39

0.10

0.18

0.59

 Randomized assignment

4

0.28

0.18

−0.08

0.63

For instructor equivalence

 No data

17

0.23

0.08

0.07

0.39

 Different instructors

3

0.39

0.19

0.01

0.76

 Identical instructors

8

0.55

0.12

0.31

0.79

Research design

 Historical control

20

0.32

0.08

0.17

0.48

 Quasi-experiment

4

0.45

0.18

0.09

0.81

 Random

4

0.28

0.18

−0.08

0.63

Types of students

 Medicine

13

0.26

0.10

0.08

0.45

 Pharmacy

10

0.45

0.10

0.24

0.65

 Public health

2

0.40

0.29

−0.18

0.98

 Nursing

1

0.08

0.36

−0.62

0.79

 Dental

1

0.19

0.30

−0.39

0.78

 Chiropractic

1

0.29

0.39

−0.47

1.04

Availability of pre-class assessment/ exercise?

 No

17

0.25

0.08

0.09

0.41

 Yes

11

0.46

0.10

0.26

0.65

Availability of pre-class readings/ notes?

 No

18

0.35

0.08

0.19

0.51

 Yes

10

0.30

0.11

0.08

0.53

Availability of quiz at start of in-class?

 No

20

0.26

0.07

0.12

0.38

 Yes

8

0.56

0.11

0.34

0.78*

n number of studies, SMD standardized mean difference, SE standard error, 95% CI 95% confidence interval, LL lower limit, UL upper limit, *p < 0.05

Publication bias

Visual inspection of Fig. 3 suggested no presence of publication bias. This is supported by two statistical analyses: Begg and Mazumdar rank correction (Kendall’s Tau with continuity correction) = 0.08, one-tailed p = 0.27; and Egger’s regression intercept 0.79, one-tailed p = 0.24. Computation of Duval and Tweedie’s trim and fill method using the random effects model revealed no studies were trimmed using the random effects model. We also conducted a classic fail-safe N test to determine the number of null effect studies needed to raise the p value associated with the mean effect above an arbitrary alpha level (α = 0.05). Results showed that 747 additional missing studies with zero mean effect size would be required to make the overall effect statistically insignificant. There would therefore have to be an unreasonably large number of undetected studies with zero effect to bring the effect sizes reported in this paper to values that might be statistically insignificant. Based on the visual inspection of funnel plot, statistical analyses, and class fail-safe N, we believe that the overall mean effect size is not inflated by publication bias.
Fig. 3
Fig. 3

Funnel plot assessing publication bias

Analysis of learners’ preference

In the course of the review, we found five articles explicitly compared student preference for flipped or traditional classroom via survey questionnaires (Table 4). Overall, among the studies listed in Table 4, preference for flipped classroom was reported by an average of 70% of total respondents.
Table 4

Summary of survey results on student preference for flipped or traditional classroom

Study

Prefer flipped classroom n (%)

Prefer traditional classroom n (%)

Total respondents

Cotta et al. (2016) [13]

87 (73)

33 (27)

120

Galway et al. (2014) [34]

9 (82)

2 (18)

11

Giuliano and Moser (2016) [36]

49 (60)

33 (40)

82

aKiviniemi (2014) [39]

33 (89)

4 (11)

37

Porcaro et al. (2016) [46]

50 (71)

20 (29)

70

Total

228 (71)

92 (29)

320

aThe remaining 3 participants in Kiviniemi [57] indicated a preference for a combination of flipped and traditional classrooms

Discussion

Overall, the data reported in this study indicate that more students favored the flipped classroom approach over traditional classroom. In addition, the flipped classroom approach was more effective than traditional classroom in increasing student learning performance. One explanation for the more positive student perception, as well as the greater effect of flipped classroom over traditional classroom, is that having unrestricted access to pre-recorded video lectures before class enables students to learn anywhere and at any time, at their own pace [36]. Students can also watch the videos multiple times to better understand a particular topic [13, 36]. Another explanation is the availability of more in-class active learning time to help increase students’ understanding of the subject material. Many of the in-class activities such as small-group discussion promoted students’ interactions with their peers [34]. Instructors also felt they had greater opportunity to provide more feedback during in-class sessions [34]. There were also greater opportunities for students to apply their knowledge in flipped classes [13, 34].

Further analyses suggest that the increase in performance holds across studies of different research designs or methodological quality. Experiments where students were randomly assigned to flipped classrooms produced results that were indistinguishable from quasi-experiments and historical controls. Analyzing variation with respect to controls over student or instructor equivalence also produced no evidence of heterogeneity. In addition, the availability of online assessment/exercises, or readings before face-to-face class did not appear to moderate performance gains, as no heterogeneity was detected between the subgroups.

However, we found that the use of quizzes at the start of a face-to-face class would make flipped classroom more effective. The quizzes consisted of specific questions that were developed by the instructor beforehand and were used to assess student learning of the pre-class learning materials. One explanation for this finding is that quizzes at the beginning of class helped students recall the knowledge learned prior to the class. Prior knowledge has long been considered an important factor influencing learning [54, 55]. Stimulating the recall of prior knowledge helps learners to make better sense of new information by connecting it to something they already know. In addition, the retrieval of information from memory makes the path to that information in memory stronger; this consequently enables the information to be more easily retrieved by the learner on the next occasion [56].

Having quizzes at the beginning of class also allows an instructor to identify students’ possible misconceptions of the pre-class materials. Students’ misconceptions can prevent further learning if not addressed. Based on student performance, instructors can provide remedial action if necessary such as reviewing the pre-class video lectures or making adjustments to the in-class teaching plans to specifically address the students’ misconceptions. The use of quizzes at the beginning of an in-class lesson can also serve as a strong motivator for students to watch the pre-class video lectures [34, 50]. This finding thus implies that instructors use quizzes as a regular part of the in-class activities to assess students’ mastery of the pre-class learning materials.

Students who preferred a traditional classroom reported that watching video lectures took a lot of additional time [13]. In a traditional class, students learn about the subject matter through a teacher-led lecture format during class time [13]; however in a flipped class students are now required to watch the video lectures before class. Students were unhappy being asked to do work at home that was traditionally done in a face-to-face class format, and considered watching the pre-class videos as burdensome in terms of time [13]. Studies from several non-health professions education flipped classrooms also supported this finding. For example, half of the students who would unwilling to take another flipped class cited the additional time required to complete the pre-class work as a reason [57]. This finding thus implies that instructors who wish to employ flipped classroom should first promote students’ understanding of this new instructional approach by explaining the rationale, and potential benefits of flipped classroom [20]. In addition, instructors may consider limiting total length of all combined video segments to about 20 min. Support for this comes from several non health professions education-related flipped classroom studies [58, 59] which reported that most students spent up to 20–25 min on viewing pre-class video lectures.

Strengths and limitations

Our meta-analysis has several strengths. The flipped classroom approach has grown rapidly and is now widely used in health professions education. To our knowledge, this is the first meta-analysis to summarize the evidence to date concerning the effectiveness of flipped classroom on student learning compared with traditional classroom. Evaluating the effectiveness of flipped classroom is therefore both timely and important for instructors and learners. We also intentionally kept our literature search very recent (up to April 15, 2017) and broad in terms of subjects from various health professions using multiple academic online databases.

However, there are some limitations that should be considered. First, this review focused on flipped classroom studies in which pre-class videos were provided, and class attendance was mandatory. To broaden the scope of review, future reviews can examine other flipped learning studies that do not restrain the instructors’ use of technological tools. Future reviews can also compare the use of online course without face-to-face meeting and flipped classes with face-to-face meeting. Second, the flipped classroom designs in the reviewed studies were not always clearly reported. For example, the specific types of video lecture used were not described. Video lectures can include many different styles including recorded classroom lecture, Khan-style freehand writing video, PowerPoint presentation with instructor talking head, PowerPoint presentation with more than one people in conversation, among others [6062]. Different video styles may affect student learning. We also could not identify information related to the actual time allocated to different instructional activities (e.g., small-group activities), and the actual details of small-group learning activities. Specifically, with regard to the small-group learning activities, there is no consensus about what various activities actually entail in practice [27]. As a result, we could not really differentiate the activities when the authors merely stated the use of small-group discussion, in-class collaboration, or group problem solving without providing specific details on the actual tasks involved in the group activities because group discussion, collaboration, problem-solving all involved discussions; and problem-solving can also be a form of collaboration. The absence of all this information prevented us from conducting further moderator analyses to discern additional key factors that could affect flipped classroom effectiveness. Additionally, the results are limited because no study included long-term follow-ups to assess learning retention. Also, the overall effect size for randomized controlled trials is usually bigger than for cohort studies.

Conclusions

Current evidence suggests that the flipped classroom approach in health professions education overall yields a statistically significant improvement in learner performance compared with traditional teaching methods. In addition, the flipped classroom would be more effective when instructors use quizzes at the start of each in-class session. Future research can be conducted to examine the possible effect of specific types of teaching method or presentation on student learning. Future research should also examine the possible impact of video styles. Despite the increasing popularity of using video-recorded lectures, we still understand little about how different video styles may impact student learning. Longitudinal studies should also be conducted to examine whether the flipped classroom approach can foster learning retention over a long period of time.

Abbreviations

CI: 

Confidence interval

EBM: 

Evidence-based medicine

LL: 

Lower limit

MERSQI: 

Medical education research study quality instrument

PRISMA: 

Preferred reporting items for systematic reviews and meta-analysis

SE: 

Standard error

SMD: 

Standardized mean differences

TED: 

Technology, entertainment, design

UL: 

Upper limit

Declarations

Acknowledgements

Not applicable.

Funding

No funding was received.

Availability of data and materials

The data used in this study can be found in previously published manuscripts.

Authors’ contributions

Both authors contributed to the design of the study. KFH performed the statistical analysis and wrote the first draft. Both authors carried out data collection and data coding. CKL helped to revise the manuscript. Both authors read and approved the final manuscript.

Ethics approval and consent to participate

Ethical approval and consent to participate are not applicable for this study.

Consent for publication

Consent for publication is not applicable for this work.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Authors’ Affiliations

(1)
The University of Hong Kong, Pok Fu Lam, Road, Hong Kong

References

  1. Horn MB, Staker HC. The rise of K-12 blended learning. San Mateo: Innosight Institute, Inc.; Retrieved from https://www.christenseninstitute.org/wp-content/uploads/2013/04/The-rise-of-K-12-blended-learning.pdf. 2011.
  2. Giannakos MN, Krogstie J, Chrisochoides N. Reviewing the flipped classroom research: reflections for computer science education. In: Proceedings of the computer science education research conference; 2014. p. 23–9.Google Scholar
  3. Karabulut-Ilgu A, Jaramillo Cherrez N, Jahren CT. A systematic review of research on the flipped learning method in engineering education. Br J Educ Technol. 2017; [cited 2017 Mar 29]. Available from: http://onlinelibrary.wiley.com/doi/10.1111/bjet.12548/full
  4. O’Flaherty J, Phillips C. The use of flipped classrooms in higher education: a scoping review. Internet High Educ. 2015;25:85–95.View ArticleGoogle Scholar
  5. Bergmann J, Sams A. Remixing chemistry class. Learn Lead Technol. 2008;36(4):24–7.Google Scholar
  6. Pierce R, Fox J. Vodcasts and active-learning exercises in a “flipped classroom” model of a renal pharmacotherapy module. Am J Pharm Educ. 2012;76(10):196.View ArticleGoogle Scholar
  7. Mehta NB, Hull AL, Young JB, Stoller JK. Just imagine: new paradigms for medical education. Acad Med. 2013;88(10):1418–23.View ArticleGoogle Scholar
  8. Ramnanan CJ, Pound LD. Advances in medical education and practice: student perceptions of the flipped classroom. Adv Med Educ Pract. 2017;8:63–73.View ArticleGoogle Scholar
  9. Ilic D, Nordin RB, Glasziou P, Tilson JK, Villanueva EA. Randomised controlled trial of a blended learning evaluation intervention for teaching evidence-based medicine. BMC Med Educ. 2015;15:39.View ArticleGoogle Scholar
  10. Betihavas V, Bridgman H, Kornhaber R, Cross M. (2016). The evidence for ‘flipping out’: a systematic review of the flipped classroom in nursing education. Nurse Educ Today. 2016;38:15–21.View ArticleGoogle Scholar
  11. Presti CR. The flipped learning approach in nursing education: a literature review. J Nurs Educ. 2016;55(5):252–7.View ArticleGoogle Scholar
  12. Chen F, Lui AM, Martinelli SM. A systematic review of the effectiveness of flipped classrooms in medical education. Med Educ. 2017;51(6):585–97.View ArticleGoogle Scholar
  13. Cotta KI, Shah S, Almgren MM, Macías-Moriarity LZ, Mody V. Effectiveness of flipped classroom instructional model in teaching pharmaceutical calculations. Curr Pharm Teach Learn. 2016;8(5):646–53.View ArticleGoogle Scholar
  14. Lin Y, Zhu Y, Chen C, Wang W, Chen T, Li T, Li Y, Liu B, Lian Y, Lu L, Zou Y, Liu Y. Facing the challenges in ophthalmology clerkship teaching: is flipped classroom the answer? PLoS One. 2017;12(4):e0174829.View ArticleGoogle Scholar
  15. O’Connor EE, Fried J, McNulty N, Shah P, Hogg JP, Lewis P, Zeffiro T, Agarwal V, Reddy S. Flipping radiology education right side up. Acad Radiol. 2016;23(7):810–22.View ArticleGoogle Scholar
  16. Educause. 7 things you should know about flipped classrooms. [cited 2017 Mar 29]. Available from https://www.rit.edu/academicaffairs/tls/sites/rit.edu.academicaffairs.tls/files/directory/ELI7081-1.pdf. 2012.
  17. Street SE, Gilliland KO, McNeil C, Royal K. The flipped classroom improved medical student performance and satisfaction in a pre-clinical physiology course. Med Sci Educ. 2015;25(1):35–43.View ArticleGoogle Scholar
  18. Bishop JL, Verleger MA. The flipped classroom: a survey of the research. In: Proceedings of the 120th ASEE National Conference and exposition, Atlanta, GA (paper ID 6219); 2013.Google Scholar
  19. Love B, Hodge A, Corritore C, Ernst DC. Inquiry-based learning and the flipped classroom model. Primus. 2015;25(8):745–62.View ArticleGoogle Scholar
  20. Lo CK, Hew KFA. Critical review of flipped classroom challenges in K-12 education: possible solutions and recommendations for future research. Res Pract Technol Enhanc Learn. 2017;12(1):4.View ArticleGoogle Scholar
  21. Muir T, Geiger V. The affordances of using a flipped classroom approach in the teaching of mathematics: a case study of a grade 10 mathematics class. Math Educ Res J. 2016;28(1):149–71.View ArticleGoogle Scholar
  22. Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann Intern Med. 2009;151(4):264–9.View ArticleGoogle Scholar
  23. Prober CG, Heath C. Lecture halls without lectures – a proposal for medical education. N Engl J Med. 2012;366(18):1657–9.View ArticleGoogle Scholar
  24. Abeysekera L, Dawson P. Motivation and cognitive load in the flipped classroom: definition, rationale and a call for research. High Educ Res Dev. 2015;34(1):1–14.View ArticleGoogle Scholar
  25. Williams DE. The future of medical education: flipping the classroom and education technology. Ochsner J. 2016;16(1):14–5.Google Scholar
  26. Korpershoek H, Harms T, de Boer H, van Kuijk M, Doolaard S. A meta-analysis of the effects of classroom management strategies and classroom management programs on students’ academic, behavioral, emotional, and motivational outcomes. Rev Educ Res. 2016;86(3):643–80.View ArticleGoogle Scholar
  27. Freeman S, Eddy SL, McDonough M, Smith MK, Okoroafor N, Jordt H, Wenderoth MP. Active learning increases student performance in science, engineering, and mathematics. Proc Natl Acad Sci U S A. 2014;111(23):8410–5.View ArticleGoogle Scholar
  28. Borenstein M, Hedges LV, Higgins JPT, Rothstein HR. Introduction to meta-analysis. Chichester, UK: Wiley; 2009.View ArticleGoogle Scholar
  29. Altman DG, Bland JM. Standard deviations and standard errors. BMJ. 2005;331(7521):903.View ArticleGoogle Scholar
  30. Lipsey MW, Wilson DB. Practical meta-analysis. Thousand Oaks, CA: Sage; 2001.Google Scholar
  31. Bakr MM, Massey WL, Massa HM. Flipping a dental anatomy course: a retrospective study over four years. Educ Res Int. 2016;2016:7097398.Google Scholar
  32. Bossaer JB, Panus P, Stewart DW, Hagemeier NE, George J. Student performance in a pharmacotherapy oncology module before and after flipping the classroom. Am J Pharm Educ. 2016;80(2):31.View ArticleGoogle Scholar
  33. Cheng X, Lee KKH, Chang EY, Yang X. The “flipped classroom” approach: stimulating positive learning attitudes and improving mastery of histology among medical students. Anat Sci Educ. 2016; https://doi.org/10.1002/ase.1664.
  34. Galway LP, Corbett KK, Takaro TK, Tairyan K, Frank EA. Novel integration of online and flipped classroom instructional models in public health higher education. BMC Med Educ. 2014;14(1):181.View ArticleGoogle Scholar
  35. Gillispie V. Using the flipped classroom to bridge the gap to generation Y. Ochsner J. 2016;16(1):32–6.Google Scholar
  36. Giuliano CA, Moser LR. Evaluation of a flipped drug literature evaluation course. Am J Pharm Educ. 2016;80(4):66.View ArticleGoogle Scholar
  37. Harrington SA, Bosch MV, Schoofs N, Beel-Bates C, Anderson K. Quantitative outcomes for nursing students in a flipped classroom. Nurs Educ Perspect. 2015;36(3):179–81.View ArticleGoogle Scholar
  38. Hsu SD, Chen CJ, Chang WK, An HYJ. Investigation of the outcomes of PGY students’ cognition of and persistent behavior in learning through the intervention of the flipped classroom in Taiwan. PLoS One. 2016;11(12):e0167598.View ArticleGoogle Scholar
  39. Kiviniemi MT. Effects of a blended learning approach on student outcomes in a graduate-level public health course. BMC Med Educ. 2014;14:181.View ArticleGoogle Scholar
  40. Koo CL, Demps EL, Farris C, Bowman JD, Panahi L, Boyle P. Impact of flipped classroom design on student performance and perceptions in a pharmacotherapy course. Am J Pharm Educ. 2016;80(2):33.View ArticleGoogle Scholar
  41. Liebert CA, Lin DT, Mazer LM, Bereknyei S, Lau JN. Effectiveness of the surgery core clerkship flipped classroom: a prospective cohort trial. Am J Surg. 2016;211(2):451–7.View ArticleGoogle Scholar
  42. McLaughlin JE, Griffin LM, Esserman DA, Davidson CA, Glatt DM, Roth MT, Gharkholonarehe N, Mumper RJ. Pharmacy student engagement, performance, and perception in a flipped satellite classroom. Am J Pharm Educ. 2013;77(9):196.View ArticleGoogle Scholar
  43. McLaughlin JE, Roth MT, Glatt DM, Gharkholonarehe N, Davidson CA, Griffin LM, Esserman DA, Mumper RJ. The flipped classroom: a course redesign to foster learning and engagement in a health professions school. Acad Med. 2014;89(2):236–43.View ArticleGoogle Scholar
  44. Morton DA, Colbert-Getz JM. Measuring the impact of the flipped anatomy classroom: the importance of categorizing an assessment by Bloom’s taxonomy. Anat Sci Educ. 2016;10(2):170–5.View ArticleGoogle Scholar
  45. Munson A, Pierce R. Flipping content to improve student examination performance in a pharmacogenomics course. Am J Pharm Educ. 2015;79(7):103.View ArticleGoogle Scholar
  46. Porcaro PA, Jackson DE, McLaughlin PM, O’Malley CJ. Curriculum design of a flipped classroom to enhance haematology learning. J Sci Educ Technol. 2016;25(3):345–57.View ArticleGoogle Scholar
  47. Prescott Jr WA, Woodruff A, Prescott GM, Albanese N, Bernhardi C, Doloresco F. Introduction and assessment of a blended-learning model to teach patient assessment in a doctor of pharmacy program. Am J Pharm Educ. 2016;80(10):176.Google Scholar
  48. Rui Z, Lian-rui X, Rong-zheng Y, Jing Z, Xue-hong W, Chuan Z. Friend or foe? Flipped classroom for undergraduate electrocardiogram learning: a randomized controlled study. BMC Med Educ. 2017;17:53.View ArticleGoogle Scholar
  49. Sajid MR, Laheji AF, Abothenain F, Salam Y, AlJayar D, Obeidat A. Can blended learning and the flipped classroom improve student learning and satisfaction in Saudi Arabia? Int J Med Educ. 2016;7:281–5.View ArticleGoogle Scholar
  50. Tune JD, Sturek M, Basile DP. Flipped classroom model improves graduate student performance in cardiovascular, respiratory, and renal physiology. AJP Adv Physiol Educ. 2013;37(4):316–20.View ArticleGoogle Scholar
  51. Whillier S, Lystad RP. No differences in grades or level of satisfaction in a flipped classroom for neuroanatomy. J Chiropr Educ. 2015;29(2):127–33.View ArticleGoogle Scholar
  52. Wong TH, Ip EJ, Lopes I, Rajagopalan V. Pharmacy students’ performance and perceptions in a flipped teaching pilot on cardiac arrhythmias. Am J Pharm Educ. 2014;78(10):185.View ArticleGoogle Scholar
  53. Reed DA, Cook DA, Beckman TJ, Levine RB, Kern DE, Wright SM. Association between funding and quality of published medical education research. JAMA. 2007;298:1002–9.View ArticleGoogle Scholar
  54. Hailikari T, Katajavuori N, Lindblom-Ylanne S. The relevance of prior knowledge in learning and instructional design. Am J Pharm Educ. 2008;72(5):113.View ArticleGoogle Scholar
  55. Merrill MD. First principles of instruction. Educ Technol Res Dev. 2002;50(3):43–59.View ArticleGoogle Scholar
  56. Dirkx KJH, Kester L, Kirschner PA. The testing effect for learning principles and procedures from texts. J Educ Res. 2014;107:357–64.View ArticleGoogle Scholar
  57. Murphy J, Chang JM, Suaray K. Student performance and attitudes in a collaborative and flipped linear algebra course. Int J Math Educ Sci Technol. 2016;47(5):653–73.View ArticleGoogle Scholar
  58. McGivney-Burelle J, Xue F. Flipping calculus. Primus. 2013;23(5):477–86.View ArticleGoogle Scholar
  59. Vazquez JJ, Chiang EP. Flipping out! A case study on how to flip the principles of economics classroom. Int Adv Econ Res. 2015;21:379–90.View ArticleGoogle Scholar
  60. Chen CM, Wu CH. Effects of different video lecture types on sustained attention, emotion, cognitive load, and learning performance. Comput Educ. 2015;80:108–21.View ArticleGoogle Scholar
  61. Chorianopoulos K, Giannakos MN. Usability design for video lectures. In: Proceedings of the 11th European conference on interactive TV and video; 2013. p. 163–4.View ArticleGoogle Scholar
  62. Guo PJ, Kim J, Rubin R. How video production affects student engagement: an empirical study of MOOC videos. In: Proceedings of the first ACM conference on learning@ scale conference; 2014. p. 41–50.View ArticleGoogle Scholar
  63. Peplow M. Social Sciences suffer from severe publication bias. Nature. Retrieved on August 25, 2017 from http://www.nature.com/news/social-sciences-suffer-from-severe-publication-bias-1.15787.

Copyright

Advertisement