Internet-based medical education: a realist review of what works, for whom and in what circumstances
© Wong et al; licensee BioMed Central Ltd. 2010
Received: 13 May 2009
Accepted: 2 February 2010
Published: 2 February 2010
Educational courses for doctors and medical students are increasingly offered via the Internet. Despite much research, course developers remain unsure about what (if anything) to offer online and how. Prospective learners lack evidence-based guidance on how to choose between the options on offer. We aimed to produce theory driven criteria to guide the development and evaluation of Internet-based medical courses.
Realist review - a qualitative systematic review method whose goal is to identify and explain the interaction between context, mechanism and outcome. We searched 15 electronic databases and references of included articles, seeking to identify theoretical models of how the Internet might support learning from empirical studies which (a) used the Internet to support learning, (b) involved doctors or medical students; and (c) reported a formal evaluation. All study designs and outcomes were considered. Using immersion and interpretation, we tested theories by considering how well they explained the different outcomes achieved in different educational contexts.
249 papers met our inclusion criteria. We identified two main theories of the course-in-context that explained variation in learners' satisfaction and outcomes: Davis's Technology Acceptance Model and Laurillard's model of interactive dialogue. Learners were more likely to accept a course if it offered a perceived advantage over available non-Internet alternatives, was easy to use technically, and compatible with their values and norms. 'Interactivity' led to effective learning only if learners were able to enter into a dialogue - with a tutor, fellow students or virtual tutorials - and gain formative feedback.
Different modes of course delivery suit different learners in different contexts. When designing or choosing an Internet-based course, attention must be given to the fit between its technical attributes and learners' needs and priorities; and to ways of providing meaningful interaction. We offer a preliminary set of questions to aid course developers and learners consider these issues.
The Internet is widely used in medical education . Several previous systematic reviews and two meta-analyses have compared the efficacy and utility of Internet-based education with conventional teaching methods or no teaching [2–8]. Two main questions face researchers in this field: efficacy (can Internet-based medical education work, and if so what is the 'effect size' compared to conventional teaching?) and effectiveness (under what real-world circumstances does it actually work, and how might its impact and cost-effectiveness be maximised?).
Cook et al.'s 2008 meta-analysis addressed efficacy, and concluded that, on average, Internet formats were equivalent to non-Internet formats in terms of learner satisfaction and changes in knowledge, skills and behaviour . Their findings indicated that substantial heterogeneity existed and their meta-analysis was unable to account for the complexity of the interactions within their included studies.
In trying to make sense of this heterogeneity we conceptualised educational courses as complex interventions and used the realist review method. Complex interventions consist of multiple human components (teachers, learners etc.) that interact in a non-linear fashion to produce outcomes which are highly context dependent [9–11]. Outcomes in such interventions depend on humans making decisions in a semi-predictable (demi-regular) manner about how to use the resources available to them in the context they find themselves in. Our rationale for using the realist review method is explained in the Methods section below.
In this realist review we set out to supplement and extend previous systematic reviews and meta-analyses. In particular we sought initially to [a] explain what sort of Internet-based medical education 'works', for whom and in what circumstances; [b] produce pragmatic guidance that could be used by developers to optimise the design of their courses and by potential learners to evaluate whether a particular course is right for them; and [c] extend the methodological knowledge base in relation to secondary research in medical education.
The realist review method
The realist approach to reviewing the evidence from complex interventions assumes that no deterministic theories can always explain nor predict outcomes in every context . Instead it is based on the principle that, though human agency and interaction is involved, in certain contexts or situations, individuals are likely, though not always certain, to make similar choices about which resources they will use . In other words, particular contexts influence human choice such that semi-predictable reoccurring patterns of behaviour ('demi-regularities') occur. Realist review seeks to uncover the underlying theories that explain these demi-regularities by critically scrutinising the interaction between context, mechanism and outcome in a sample of primary studies. Mechanisms are processes operating within an intervention that describes how the 'human components' use the resources available to them [14, 15]. In particular middle-range theory (that is theory that "involves abstraction... but [is] close enough to observed data to be incorporated in propositions that permit empirical testing.") is specifically sought as their level of abstraction provides a more generalisable explanation of demi-regularities. More than one middle-range theory may explain the influences of context on a mechanism to produce an outcome .
Importantly, realist review methodology acknowledges that within complex interventions there are many dimensions and layers of explanation that warrant exploration. For example, there are human behaviours as well as multiple interactions between the numerous components of the intervention. A realist review does not seek to explain all these layers; it is specifically focused on the demi-regularities in the social (and socio-technical) world which create preconditions for particular human behaviours . To that end, we sought to extract theories from our dataset of primary studies which would explain whether or not an Internet based course was considered a 'success' and especially whether it produced effective learning. We sought to try to gain insights and explanations that would be generalisable across a whole range of different types of Internet based courses and so theories that focused on specific aspects of such courses (for example only computer mediated conferencing) were not central to our inquiry.
Inclusion and exclusion criteria
Studies were included if they had any medical students or doctors as learners; used the Internet to support learning; and contained at least one level of evaluation as described by Kirkpatrick . Studies were excluded if the Internet was used for purposes other than learning (e.g. tracking website use, examinations only, course administration).
Identifying primary studies
We searched 15 electronic databases relevant to medical education from their inception dates to April 2006 using guidance provided by Haig and Dozier [18, 19]. No language restrictions were applied (non-English language papers were translated), and publications of any type were included. Details of the databases and search strategy are available in Additional file 1.
In the first stage of searching, GW screened the title, abstract and subject headings (where available) against inclusion and exclusion criteria. Potentially eligible studies were obtained in full text and re-screened in the second stage. A random subset (200/12586 and 50/514 citations respectively) at each stage was screened independently by TG and disagreements resolved by discussion.
Identifying candidate theories
The initial identification of candidate (middle-range) theories in realist reviews is necessarily an iterative and speculative process. Whilst a review team may initially have theories that they believe to be in operation to explain why certain outcomes occur, a key element in realist review is to explore the presence of these 'educated guess' theories and where applicable, test their explanatory value. Candidate theories are not considered definitive until they have been tested. Much of the work in realist review involves not only repeatedly questioning the validity of any candidate theory and refining it but also seeking out new candidate theories from included studies if existing ones are found wanting.
We used a variety of methods to derive our list of candidate theories. This included; brain-storming within the review team, browsing through specialist educational library collections, discussions with fellow educators and pursuing references of references . We did not specifically consult individual experts in the field. We iteratively [re-]checked all the included studies against the candidate theories so as to establish which (if any) explained differences in outcomes. In each paper, we sought data to test (affirm, refute or refine) the candidate theory by assessing their relevance and rigour . Throughout our data extraction and synthesis phases, we continually sought out further candidate theories that might better explain the data in the included studies.
Data management, analysis and synthesis
In a first phase, study characteristics (e.g. sample type and size, setting, course objectives) and theoretical contribution (e.g. 'how', 'why, 'in what circumstances') were tabulated on an Excel spreadsheet using data domains informed by previous systematic reviews in this field [2, 5, 6, 21]. In a second phase, the NVivo qualitative software was used to index and link relevant sections of text of included articles to our emerging analytic framework . As each included article was read and re-read, we created and iteratively revised codes to capture themes or concepts that might contribute to theory testing . In particular, we sought to identify prominent demi-regularities that might help us to understand Internet-based interventions better. We classified 'interaction' in the online environment according to the criteria of Vrasidas and Glass (in sum, learner-tutor, learner-learner, learner-content and learner-software, the last of these being technical feedback such as automated replies to multiple choice questions) .
Data synthesis involved both individual reflection and team discussions that considered the ability of the candidate theories to explain the data reported in empirical studies (especially in relation to any prominent demi-regularities we encountered). The sections of texts from our included studies, which we had coded and captured within NVivo formed the raw materials for our interpretations. We used these sections of texts to see if they were able to confirm, refute or refine our candidate theories. Specifically, we attempted to identify recurrent demi-regularities which might act as barriers or enablers to Internet-based learning and tested the explanatory powers of our initial candidate theories against these. Where candidate theories failed to explain the data we sought new ones, either from the included studies or wider educational or sociological literature. Throughout this process, we deliberately sought out disconfirming data - i.e. data that might refute our provisional candidate theories. In line with realist review methodology, we also used the information we gleaned from our immersion in our included studies to refine our initial review goals .
Search results and study characteristics
We initially selected four candidate theories for further testing: Laurillard's conversational framework , Schon's reflective practitioner , Slotnick's how doctors learn  and Reeves' effective dimensions of interactive learning .
These theories provided only a starting point in our attempt to explain what sort of Internet-based medical education 'works', for whom and in what circumstances. As we extracted our data, we noted further candidate theories and proceeded to test these as well. Additional candidate theories that we attempted to tested included: Vygotski , Danchak , Schon , Garrison [32, 33], Dewey and Brookfield , Kolb , Moshman , Eraut , Boettcher , Wenger , Koschmann , Nahapiet and Ghoshal , Socrates  Problem Based Learning [31, 42–48], Constructivism [29–31, 33–35, 37, 44, 45, 49–57] and adult learning theory/principles [31, 32, 47, 50, 53, 54, 58–70].
As no previous realist review had been undertaken in this field, we were initially unclear as to how suitable the data reported in our included studies would be for answering the broad research question goal we had set ourselves. As the review progressed we became aware of various data suitability limitations (see Discussion) and the emergence of two prominent demi-regularities prompted us to narrow our review focus to the two candidate theories discussed below. This is an example of progressive focusing - a well-established technique in qualitative research in which the focus of the inquiry is iteratively sharpened by reflection on emerging data .
Technology acceptance: getting learners to log on
At an early stage in this review, our reading and interpretation of the reported data in our included studies showed that educators often faced a substantial barrier of getting learners to use their Internet-based course. This demi-regularity of getting learners to log onto - or engage with - a course was clearly an important factor in explaining the fortunes of such courses. We noted that learners needed to have to have good reasons to engage and that unless they did, the outcomes reported were less favourable. Examples of the texts we used to support our interpretation may be found in Additional file 2: Table s1.
Engagement and acceptance was not explained by any of our initial candidate theories, but we noted that one of our included papers  mentioned the value of conceptualising Internet based courses as innovations and specifically Rogers' diffusion of innovations theory . We found that Davis's Technology Acceptance Model , which is derived from Rogers' theory, was a more precise articulation of innovation acceptance when the innovation involved was a technology. Drawing on both Rogers' and Davis's theories, the attribute of an Internet-based course that provided the most coherent and complete explanation of technology acceptance was the perceived usefulness of the technological medium (in the eyes of potential learners) over an alternative delivery format. From our included studies, we identified that perceived usefulness - or in Rogers' original terminology 'relative advantage', included 7 sub-components, representing the contexts that influence whether learners choose to engage with an Internet-based course: access to learning; access to consistent content; links with assessment; convenience; cost saving; interactivity; and time saving.
Overall, 38% (95/249) of our included studies provided some data to support the central importance of perceived usefulness and none provided data to refute it. Two other attributes - perceived ease of use (from Davis's Technology Acceptance Model) and compatibility with the learner's norms and values (from Rogers' original diffusion of innovations theory), also explained some of the variability in acceptance of the Internet medium, and evidence to support these attributes was found in 13% (32/249) and 3% (7/249) of studies respectively. Again, we found no disconfirming studies.
Five questions for developers and prospective learners to ask of an Internet-based course
How useful will the prospective learners perceive the Internet technology to be?
For example, in any particular context and compared to what is currently available to them, to what extent will this technology
a. Increase their access to learning?
b. Provide consistent, high-quality content?
c. Be a convenient format in which to receive their education?
d. Save them money?
e. Save them time?
f. Link to course assessment?
How easy will the prospective learners find this technology to use?
How well does this format fit in with what learners are used to and expect?
Achieving interactive dialogue
How will high-quality human-human (learner-tutor and learner-learner) interaction and feedback be achieved? For example what use will be made of
a. Structured virtual seminars?
b. Email, bulletin boards?
c. Real-time chat?
d. Supplementary media e.g. video, audio, phone calls, videoconferencing?
e. Course assessment and feedback on performance?
How will high-quality human-technical interaction and feedback be achieved? For example what use will be made of
f. Questions with automated feedback?
Interaction: building a learning dialogue
In our recommendations in Table 1, we have again converted our insights about the importance of interaction and feedback into two questions which remind course developers to think about this issue. The examples that we have provided of how the interaction and feedback might be enabled technically are drawn from our analysis of the methods used in our included studies.
An important finding of this review was that 'success features' did not seem to be intrinsic to any course but a function of the course-context interaction. One group of learners might perceive the a technologically based course as having very high 'usefulness' while a different group would find the same course much less useful. For example, in studies comparing virtual microscopy (where glass slides were digitised and the features of a traditional light microscope simulated by software) with conventional microscopy, medical student learners were reported as valuing the Internet-based materials much more highly and utilised these more. Features of perceived usefulness included assessment linkage (virtual material was used in exams) , consistent high quality content (whereas traditional slides may or may not show the feature concerned) ; convenience (they did not have to conform to laboratory opening times) ; cost saving (rental cost of a microscope) ; and time saving (journey times to the laboratory were cut to zero) . The course's ease of use (comments included "doesn't hurt my eyes" "stays in focus") was also highly rated compared to conventional alternatives . However, this same Internet-based application was reported as having little or no perceived usefulness for trainee pathologists, who must learn not merely to evaluate standardised slides in formal examinations but to deal with the inconsistencies and contextual complexities of real slides in the real world .
The above example also suggests that the construct 'ease of use' does not operate independently of other course features, especially its perceived usefulness. For example, we encountered studies utilising virtual textbooks (where text and/or images were digitised and placed online) where despite efforts to ensure the technology was easy to use, learner engagement remained low (e.g. because the learners perceived that they could access 'better' but similar content face to face or in other formats) [80–83]. Conversely, we found a 1996 paper describing a bio-computing course that had been set up to allow teaching expertise to be shared between the few geographically dispersed experts there were in this field . The tutors and highly computer literate students communicated using a very rudimentary and technically complicated email system. Despite these challenges, most students persisted with it and rated their learning experience as positive. It appears that the advantage of being able to learn with otherwise hard-to-reach experts ('improved access to learning') more than made up for the technical limitations of the learning technology.
Summary of main findings
This realist review of 249 primary studies has produced two key findings which are important if somewhat unsurprising. First, Internet-based courses must engage their target group of learners to use the technology. This is likely to occur only if the technology is perceived as 'useful' (e.g. increases access to learning or saves time) and 'easy to use', though benefits in the former can outweigh challenges in the latter. Second, 'interactivity' is highly valued by learners. Learners wanted to be able to enter into a dialogue with the course tutor, fellow students and/or a virtual tutorial and obtain ongoing feedback on their understanding and performance.
Course design is an important factor in Internet-based courses, but attention must also be paid to course-context interaction. A pedagogically sound course may prove technically acceptable and produce positive learning outcomes in one group of learners in one context but the same course may be technically unacceptable and/or fail to achieve effective learning in a different context. The skills of learners, course learning objectives and the availability, quality and cost of non-Internet alternatives are particularly important contextual factors.
Strengths and limitations of the review
To our knowledge, this review represents the first use of realist review in medical educational research. It contributes to an emerging field in systematic review, in which qualitative reviews are undertaken to supplement and extend the findings of meta-analyses and other quantitative reviews [85, 86]. The advantage of using both approaches is that the strengths and weaknesses of each method are complementary [87–89]. Realist reviews are a type of theory driven qualitative review and so differs in many respects to more quantitative (for example Cochrane) reviews. A discussion of the advantages and disadvantages between these review methods is beyond the scope of this paper and interested readers are directed to Chapter 3 of Pawson's Evidence-based Policy: A Realist Perspective .
The recent meta-analysis by Cook et al (see Background) provided much-needed evidence that the overall educational impact of Internet-based medical education can be equivalent to that of conventional formats. In their discussion, these authors raised two further questions which they acknowledged had not been addressed by their meta-analysis: "How can Internet-based learning be effectively implemented?" and "When should Internet-based learning be used?" . Cook has previously observed that "...the appropriateness of web-based learning as a learning tool will vary upon the instructional context..." - a comment which raises the question of what sort of course is 'appropriate' in what sort of context .
Our review has begun to extend the knowledge base by identifying and refining some of the middle-range theories that explain the 'how', 'why' and 'in what circumstances' questions about Internet-based medical education. We acknowledge that our progressive focus on two prominent demi-regularities has meant that we have not addressed all aspects of our initial review's goals. However, it is reassuring that the key findings of this review align with, and illuminate, the findings of previous systematic reviews. For example, the quantitative observation that the speed of downloading is associated with learner satisfaction  may be explained qualitatively by the 'ease of use' construct within the Technology Acceptance Model (and, more widely, diffusion of innovation theory). Similarly, the observation that 'dialogue'  and interaction  is associated with improved learner performance is explained qualitatively by the Conversational Framework.
Perhaps more significantly, theory-driven qualitative systematic reviews may also throw light on the reason why there is a lack of association between variables and outcomes seen in quantitative (Cochrane-type) reviews. We suggest that a paradigm shift may need to occur in how interventions that involve human agency should be viewed - namely as complex interventions [12, 13].
The pursuit of rigour in realist review follows similar principles to the pursuit of rigour in qualitative research more generally . The essence of such research is interpretation, hence key processes are immersion (reading and re-reading texts), reflection, discussion amongst team members, comparison and continuing to seek explanations and test theories until saturation of the data is reached. Our sample included a heterogeneous group of primary studies of different learner groups in diverse contexts, with no restrictions by study design or language of publication - in other words, we had what is known in qualitative research as a 'maximum variety sample'. This allowed us to explore a wide range of context-mechanism-outcome combinations and use the available qualitative data reported in the primary studies to build and refine theories of how Internet-based learning 'works'. Whilst we have followed the realist review method and documented the steps we took to arrive at the middle-range theories presented here, we are fully aware that (in common with other qualitative research) this method is subjective and interpretive. Therefore another team reviewing the same literature may arrive at a different set of middle-range theories with which to make sense of this vast field.
We did not consult individual experts in this field and acknowledge that had we done so, we may well have had a wider set of additional candidate theories to test. We did not set out to be all-inclusive in our review but have been able to uncover key middle-range theories that begin to help to explain the fortunes of Internet based courses. We are certain that other middle-range theories will be needed and are important in furthering understanding and believe that there is more work to be done in unravelling the multitude of theories that are in operation within Internet based courses. More specifically, we believe that more theory drive reviews, such as ours and that by Ruiz et al.  hold the greatest promise to understanding medical educational interventions.
Whatever review method is used in secondary research, the resulting synthesis is only as good as the primary data on which the synthesis is built. A major limitation we encountered in our review was that many primary studies included only cursory descriptions of their Internet-based educational intervention (e.g. educational setting, teaching practices and rationale of course design). The paucity of such data placed two important limitations on our review. Firstly, we were not able to test in detail all aspects of our candidate theories. If richer descriptions been reported in our included studies, we would have been able to undertake a more fine-grained analysis of both technology acceptance and interactivity. Secondly, we were aware that a large number of theories exist on how learners learn online and in more traditional settings. In our included studies alone, 17 specific theories were named in 58 articles. However, within the included studies, we could not find sufficient reported detail to enable comprehensive testing of these theories.
Limitations in the type of data, depth and quality of reporting of studies in medical education are well recognised . We strongly recommend that authors of primary studies in this field produce detailed descriptions of the intervention and context as well as quantitative data on satisfaction and impacts, and that journal editors make space for these rich descriptions, since the ability of future realist and other theory driven reviews to extend the knowledge base further will depend on the quality and completeness of the qualitative data gathered and reported.
Based on the findings of this review we suggest a set of questions that educators should address in order to maximise the chance that their Internet-based courses will be perceived as useful and provide an effective learning opportunity, and which prospective learners may use to evaluate whether a course is right for them (Table 1). Given our findings above about the importance of course-context interactions, it follows that the factors referred to in Table 1 cannot be 'built into' courses independently of a consideration of learners' needs and priorities or assessment of other courses available locally and indeed, on the Internet - in other words the course's context. Nor can our guidance be seen as a deterministic 'law of nature' which if slavishly followed will invariably lead to a successful course. The questions in Table 1 are designed to complement existing guidance on course design (such as for example by Grant  or McKendree ), and should be seen as part of the entire curriculum design process and not as a substitute for these.
We would like to thank the following: Marcia Rigby for her administrative support; UCL's Interlending and Document Supply Office for their assistance with obtaining manuscripts and the peer reviewers, Jarmila Potomkova and in particular David Cook for their constructive comments.
- Harden R: Trends and the future of postgraduate medical education. Emerg Med J. 2006, 23: 798-802. 10.1136/emj.2005.033738.View ArticleGoogle Scholar
- Adler MD, Johnson KB: Quantifying the literature on computer-aided instruction in medical education. Acad Med. 2000, 75: 1025-1028. 10.1097/00001888-200010000-00021.View ArticleGoogle Scholar
- Childs S, Blenkinsopp B, Hall A, Walton G: Effective e-learning for health professionals and students - barriers and their solutions. A systematic review of the literature - findings from the HeXL project. Health Info and Libr J. 2005, 22 (Suppl 2): 20-32. 10.1111/j.1470-3327.2005.00614.x.View ArticleGoogle Scholar
- Coomey M, Stephenson J: Online learning: it is all about dialogue, involvement, support and control - according to the research. Teaching and Learning Online: Pedagogies for New Technologies. Edited by: Stephenson J. 2001, London: Kogan Page, 37-52.Google Scholar
- Curran V, Fleet L: A review of evaluation outcomes of web-based continuing medical education. Med Educ. 2005, 39: 561-567. 10.1111/j.1365-2929.2005.02173.x.View ArticleGoogle Scholar
- Wutoh R, Boren S, Balas A: eLearning: A Review of Internet-Based Continuing Medical Education. J Contin Educ Health Prof. 2004, 24: 20-30. 10.1002/chp.1340240105.View ArticleGoogle Scholar
- Bernard R, Abrami P, Lou Y, Borokhovski E, Wade A, Wozney L, et al: How Does Distance Education Compare With Classroom Instruction? A Meta-Analysis of the Empirical Literature. Rev Educ Res. 2004, 74: 379-439. 10.3102/00346543074003379.View ArticleGoogle Scholar
- Cook D, Levinson A, Garside S, Dupras D, Erwin P, Montori V: Internet-Based Learning in Health Professionals: A Meta-analysis. JAMA. 2008, 300: 1181-1196. 10.1001/jama.300.10.1181.View ArticleGoogle Scholar
- Anderson R: New MRC guidance on evaluating complex interventions. BMJ. 2008, 337: a1937-10.1136/bmj.a1937.View ArticleGoogle Scholar
- Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M: Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ. 2008, 337: a1655-10.1136/bmj.a1655.View ArticleGoogle Scholar
- Pawson R: Nothing as Practical as a Good Theory. Evaluation. 2003, 9: 471-490. 10.1177/1356389003094007.View ArticleGoogle Scholar
- Shiell A, Hawe P, Gold G: Complex interventions or complex systems? Implications for health economic evaluation. BMJ. 2008, 336: 1281-1283. 10.1136/bmj.39569.510521.AD.View ArticleGoogle Scholar
- Shepperd S, Lewin L, Straus S, Clarke M, Eccles M, Fitzpatrick R, et al: Can We Systematically Review Studies That Evaluate Complex Interventions?. PLoS Med. 2009, 6: e1000086-10.1371/journal.pmed.1000086.View ArticleGoogle Scholar
- Pawson R: Evidence-based Policy. A Realist Perspective. 2006, London: SageView ArticleGoogle Scholar
- Pawson R, Greenhalgh T, Harvey G, Walshe K: Realist review - a new method of systematic review designed for complex policy interventions. J Health Serv Res Policy. 2005, 10: 21-34. 10.1258/1355819054308530.View ArticleGoogle Scholar
- Merton R: On Theoretical Sociology. Five Essays, Old and New. 1967, New York: The Free PressGoogle Scholar
- Kirkpatrick D: The Four Levels of Evaluation: Measurement and Evaluation. 2007, Alexandria: American Society for Training and DevelopmentGoogle Scholar
- Haig A, Dozier M: BEME Guide No 3: Systematic searching for evidence in medical education - Part 1: Sources of information. Med Teach. 2003, 25: 352-363.Google Scholar
- Haig A, Dozier M: BEME Guide No 3: Systematic searching for evidence in medical education - Part 2: Constructing searches. Med Teach. 2003, 25: 463-484. 10.1080/01421590310001608667.View ArticleGoogle Scholar
- Greenhalgh T, Peacock R: Effectiveness and efficiency of search methods in systematic reviews of complex evidence: audit of primary sources. BMJ. 2005, 331: 1064-1065. 10.1136/bmj.38636.593461.68.View ArticleGoogle Scholar
- Chumley-Jones H, Dobbie A, Alford C: Web-based learning: Sound Educational Method or Hype? A review of the evaluation Literature. Acad Med. 2005, 77: S86-S93. 10.1097/00001888-200210001-00028.View ArticleGoogle Scholar
- Bazeley P: Qualitative data analysis with NVIVO. 2007, London: SageGoogle Scholar
- Glaser B, Strauss A: The constant comparative method of qualitative analysis. The Discovery of Grounded Theory. Edited by: Glaser B, Strauss A. 1967, Chicago: AdlineGoogle Scholar
- Vrasidas C, Glass G: A conceptual framework for studying distance education. Distance Education And Distributed Learning. Edited by: Vrasidas C, Glass G. 2002, Greenwich, Connecticut: Information Age Publishing, 31-55.Google Scholar
- Laurillard D: Rethinking University Teaching: a conversational framework for the effective use of learning technologies. 2002, London: RoutledgeFalmer, SecondView ArticleGoogle Scholar
- Schon D: Educating the Reflective Practitioner. 1987, San Francisco, California: Jossey-Bass Inc, FirstGoogle Scholar
- Slotnick H: How Doctors Learn: The Role of Clinical Problems across the Medical School-to-Practice Continuum. Acad Med. 1996, 71: 28-34. 10.1097/00001888-199601000-00014.View ArticleGoogle Scholar
- Reeves T: Effective Dimensions of Interactive Learning on the Wolrd Wide Web. Web-Based Instruction. Edited by: Khan B. 2005, Englewood Cliffs, New Jersey: Educational Technology Publications, Inc, 59-66.Google Scholar
- Alant E, Dada S: Group learning on the web. Int J Educ Dev. 2005, 25: 305-316. 10.1016/j.ijedudev.2004.11.010.View ArticleGoogle Scholar
- Boulos MN, Taylor AD, Breton A: A synchronous communication experiment within an online distance learning program: a case study. Telemed J E Health. 2005, 11: 583-593. 10.1089/tmj.2005.11.583.View ArticleGoogle Scholar
- Casebeer LL, Strasser SM, Spettell CM, Wall TC, Weissman N, Ray M, et al: Designing tailored Web-based instruction to improve practicing physicians' preventive practices. J Med Internet Res. 2003, 5: e20-10.2196/jmir.5.3.e20.View ArticleGoogle Scholar
- Curran VR, Lockyer J, Kirby F, Sargeant J, Fleet L, Wright D: The nature of the interaction between participants and facilitators in online asynchronous continuing medical education learning environments. Teach Learn Med. 2005, 17: 240-245. 10.1207/s15328015tlm1703_7.View ArticleGoogle Scholar
- Kamin C, O'Sullivan P, Deterding R, Younger M: A comparison of critical thinking in groups of third-year medical students in text, video, and virtual PBL case modalities. Acad Med. 2003, 78: 204-211. 10.1097/00001888-200302000-00018.View ArticleGoogle Scholar
- Lang EV, Sood A, Anderson B, Kettenmann E, Armstrong E: Interpersonal and communication skills training for radiology trainees using a rotating peer supervision model (microteaching). Acad Radiol. 2005, 12: 901-908. 10.1016/j.acra.2005.03.064.View ArticleGoogle Scholar
- Liaw ST, Pearce C, Keppell M: Developing a Web-based Learning Network for Continuing Medical Education. J Workplace Learn. 2002, 14: 98-108. 10.1108/13665620210421911.View ArticleGoogle Scholar
- Maier P, Armstrong R, Hall W, Ng M: JointZone: users' views of an adaptive online learning resource for rheumatology. Learn Media Technol. 2005, 30: 281-297.View ArticleGoogle Scholar
- Mash B, Marais D, Walt Van Der S, Van Deventer I, Steyn M, Labadarios D: Assessment of the quality of interaction in distance learning programmes utilizing the Internet or interactive television: perceptions of students and lecturers. Med Teach. 2006, 28: e1-e9. 10.1080/01421590600568439.View ArticleGoogle Scholar
- Nathoo AN, Goldhoff P, Quattrochi JJ: Evaluation of an Interactive Case-based Online Network (ICON) in a problem based learning environment. Adv Health Sci Educ Theory Pract. 2005, 10: 215-230. 10.1007/s10459-005-7851-3.View ArticleGoogle Scholar
- Raffety B, Allendoerfer C, Minstrell J, Chabal C, Dunbar P, Nakamura Y: A facet-based system for computer-assisted instruction in pain management for elderly patients. Proc AMIA Symp. 2000, 670: 1-5.Google Scholar
- Sandars J, Langlois M: Online learning networks for general practitioners: Evaluation of a pilot project. Educ Prim Care. 2005, 16: 688-696.Google Scholar
- Turchin A, Lehmann CU: Active Learning Centre: utilization patterns of an interactive educational World Wide Web site. Proc AMIA Symp. 1999, 627-31.Google Scholar
- Allen M, Sargeant J, Mann K, Fleming M, Premi J: Videoconferencing for practice-based small-group continuing medical education: feasibility, acceptability, effectiveness, and cost. J Contin Educ Health Prof. 2003, 23: 38-47. 10.1002/chp.1340230107.View ArticleGoogle Scholar
- Alverson DC, Saiki SMJ, Jacobs J, Saland L, Keep MF, Norenberg J, et al: Distributed interactive virtual environments for collaborative experiential learning and training independent of distance over Internet2. Stud Health Technol Inform. 2004, 98: 7-12.Google Scholar
- Fieschi M, Soula G, Giorgi R, Gouvernet J, Fieschi D, Botti G, et al: Experimenting with new paradigms for medical education and the emergence of a distance learning degree using the internet: teaching evidence-based medicine. Med Inform Internet Med. 2002, 27: 1-11. 10.1080/14639230110105301.View ArticleGoogle Scholar
- Guerandel A, Felle P, Malone K: Computer-assisted learning in undergraduate psychiatry (CAL-PSYCH): Evaluation of a pilot programme. Ir J Psychol Med. 2003, 20: 84-87.View ArticleGoogle Scholar
- Kim S, Kolko BE, Greer TH: Web-based problem solving learning: Third-year medical students' participation in end-of-life care virtual clinic. Comput Hum Behav. 2002, 18: 761-772. 10.1016/S0747-5632(02)00029-8.View ArticleGoogle Scholar
- O'Rourke A, Dolman E, Fox N, Lane P, Roberts C: The Wisdom Project: virtual education in primary care. Health Libr Rev. 1999, 16: 73-81. 10.1046/j.1365-2532.1999.00214.x.View ArticleGoogle Scholar
- Seabra D, Srougi M, Baptista R, Nesrallah LJ, Ortiz V, Sigulem D: Computer aided learning versus standard lecture for undergraduate education in urology. J Urol. 2004, 171: 1220-1222. 10.1097/01.ju.0000114303.17198.37.View ArticleGoogle Scholar
- Booth A, Levy P, Bath PA, Lacey T, Sanderson M, Diercks O'Brien G: Studying health information from a distance: refining an e-learning case study in the crucible of student evaluation. Health Info Libr J. 2005, 22: 8-19. 10.1111/j.1470-3327.2005.00610.x.View ArticleGoogle Scholar
- Bowdish BE, Chauvin SW, Kreisman N, Britt M: Travels towards Problem Based Learning in Medical Education (VPBL). Instr Sci. 2003, 31: 231-253. 10.1023/A:1024625707592.View ArticleGoogle Scholar
- Bryant SL, Ringrose T: Evaluating the Doctors.net.uk model of electronic continuing medical education. Work Based Learn Prim Care. 2005, 129-142.Google Scholar
- Carr MM, Hewitt J, Scardamalia M, Reznick RK: Internet-based otolaryngology case discussions for medical students. J Otolaryngol. 2002, 31: 197-201. 10.2310/7070.2002.21057.View ArticleGoogle Scholar
- Curran V, Kirby F, Parsons E, Lockyer J: Discourse analysis of computer-mediated conferencing in World Wide Web -based continuing medical education. J Contin Educ Health Prof. 2003, 23: 229-238. 10.1002/chp.1340230506.View ArticleGoogle Scholar
- Fordis M, King JE, Ballantyne CM, Jones PH, Schneider KH, Spann S, et al: Comparison of the instructional efficacy of Internet-based CME with live interactive CME workshops: a randomized controlled trial. JAMA. 2005, 294: 1043-1051. 10.1001/jama.294.9.1043.View ArticleGoogle Scholar
- Friedl R, Hoppler H, Ecard K, Scholz W, Hannekum A, Ochsner W, et al: Multimedia-driven teaching significantly improves students' performance when compared with a print medium. Ann Thorac Surg. 2006, 81: 1760-1766. 10.1016/j.athoracsur.2005.09.048.View ArticleGoogle Scholar
- Kemp M, Davis H, Roche W, Hall W: From classroom tutor to hypertext adviser: an evaluation. Alt-J. 2002, 10: 41-53. 10.1080/0968776020100304.View ArticleGoogle Scholar
- Mash RJ, Marais D, Walt Van Der S, Van Deventer I, Steyn M, Labadarios D: Assessment of the quality of interaction in distance learning programmes utilising the Internet (WebCT) or interactive television (ITV). Med Educ. 2005, 39: 1093-1100. 10.1111/j.1365-2929.2005.02315.x.View ArticleGoogle Scholar
- Barnes K, Itzkowitz S, Brown K: Teaching clinical management skills for genetic testing of hereditary nonpolyposis colorectal cancer using a Web-based tutorial. Genet Med. 2003, 5: 43-48. 10.1097/00125817-200301000-00007.View ArticleGoogle Scholar
- Cook DA, Dupras DM: Teaching on the web: automated online instruction and assessment of residents in an acute care clinic. Med Teach. 2004, 26: 599-603. 10.1080/01421590400004932.View ArticleGoogle Scholar
- Curran V, Kirby F, Parsons E, Lockyer J: Short report: satisfaction with on-line CME. Evaluation of the ruralMDcme website. Can Fam Physician. 2004, 50: 271-274.Google Scholar
- Curran VR, Hoekman T, Gulliver W, Landells I, Hatcher L: Web-based continuing medical education (I): field test of a hybrid computer -mediated instructional delivery system. J Contin Educ Health Prof. 2000, 20: 97-105. 10.1002/chp.1340200206.View ArticleGoogle Scholar
- Fox NJ, Dolman EA, Lane P, Rourke O, Roberts C: The WISDOM project: training primary care professionals in informatics in a collaborative 'virtual classroom'. Med Educ. 1999, 33: 365-370. 10.1046/j.1365-2923.1999.00309.x.View ArticleGoogle Scholar
- Macrae HM, Regehr G, McKenzie M, Henteleff H, Taylor M, Barkun J, et al: Teaching practicing surgeons critical appraisal skills with an Internet -based journal club: A randomized, controlled trial. Surgery. 2004, 136: 641-646. 10.1016/j.surg.2004.02.003.View ArticleGoogle Scholar
- Mandayam S: Enhancing clinical epidemiology concepts of internal medicine residents: A Web-based approach. 2004, University of Texas Graduate School of Biomedical Sciences at Galveston, MPH; 4Google Scholar
- Peterson MW, Galvin JR, Dayton C, D'Alessandro MP: Realizing the promise: delivering pulmonary continuing medical education over the Internet. Chest. 1999, 115: 1429-1436. 10.1378/chest.115.5.1429.View ArticleGoogle Scholar
- Robinson L, Cruickshank N: Improving primary care nutrition skills. Asia Pac J Clin Nutr. 2005, 14: S92-S96.Google Scholar
- Schilling K, Wiecha J, Polineni D, Khalil S: An Interactive Web-based Curriculum on Evidence-based Medicine: Design and Effectiveness. Fam Med. 2006, 38: 126-132.Google Scholar
- Shaffer K, Small JE: Blended learning in medical education: use of an integrated approach with web-based small group modules and didactic instruction for teaching radiologic anatomy. Acad Radiol. 2004, 11: 1059-1070. 10.1016/j.acra.2004.05.018.View ArticleGoogle Scholar
- Wiecha JM, Gramling R, Joachim P, Vanderschmidt H: Collaborative e-learning using streaming video and asynchronous discussion boards to teach the cognitive foundation of medical interviewing: a case study. J Med Internet Res. 2003, 5: e13-10.2196/jmir.5.2.e13.View ArticleGoogle Scholar
- Zebrack JR, Mitchell JL, Davids SL, Simpson DE: Web-based curriculum. A practical and effective strategy for teaching women's health. J Gen Intern Med. 2005, 20: 68-74. 10.1111/j.1525-1497.2005.40062.x.View ArticleGoogle Scholar
- Britten N, Jones R, Murphy M, Stacy R: Qualitative research methods in general practice and primary care. Fam Pract. 1995, 12: 104-114. 10.1093/fampra/12.1.104.View ArticleGoogle Scholar
- Sargeant J, Curran V, Jarvis SS, Ferrier S, Allen M, Kirby F, et al: Interactive on-line continuing medical education: physicians' perceptions. J Contin Educ Health Prof. 2004, 24: 227-236. 10.1002/chp.1340240406.View ArticleGoogle Scholar
- Rogers EM: Diffusion of innovations. 2003, New York: Free Press, 5Google Scholar
- Davis FD: Perceived usefulness, perceived ease of use, and user acceptance of information technology. Mis Quarterly. 1989, 13: 319-340. 10.2307/249008.View ArticleGoogle Scholar
- Klatt E: Web-based teaching in pathology. JAMA. 1997, 278: 1787-10.1001/jama.278.21.1787.View ArticleGoogle Scholar
- Blake CA, Lavoie HA, Millette CF: Teaching medical histology at the University of South Carolina School of Medicine: Transition to virtual slides and virtual microscopes. Anat Rec B New Anat. 2003, 275: 196-206. 10.1002/ar.b.10037.View ArticleGoogle Scholar
- Harris T, Leaven T, Heidger P, Kreiter C, Duncan J, Dick F: Comparison of a virtual microscope laboratory to a regular microscope laboratory for teaching histology. Anat Rec. 2001, 265: 10-14. 10.1002/ar.1036.View ArticleGoogle Scholar
- Kumar RK, Velan GM, Korell SO, Kandara M, Dee FR, Wakefield D: Virtual microscopy for learning and assessment in pathology. J Pathol. 2004, 204: 613-618. 10.1002/path.1658.View ArticleGoogle Scholar
- Marchevsky AM, Relan A, Baillie S: Self-instructional "virtual pathology" laboratories using web-based technology enhance medical school teaching of pathology. Hum Pathol. 2003, 34: 423-429. 10.1016/S0046-8177(03)00089-3.View ArticleGoogle Scholar
- Chou MT, McGinnis P, Tello R: A web based video tool for MR arthrography. Comput Biol Med. 2003, 33: 113-117. 10.1016/S0010-4825(02)00061-6.View ArticleGoogle Scholar
- Gomez Arbones X, Ferreira A, Pique M, Roca J, Tomas J, Frutos JL, et al: A cardiological web as an adjunct to medical teaching: prospective analysis. Med Teach. 2004, 26: 187-189. 10.1080/01421590310001653991.View ArticleGoogle Scholar
- Horsch A, Balbach T, Hogg M, Sturm F, Minov C: The case-based Internet textbook ODITEB for multi-modal diagnosis of tumors --development, features and first experiences. Stud Health Technol Inform. 1999, 68: 513-516.Google Scholar
- Horsch A, Balbach T, Melnitzki S, Knauth J: Learning tumor diagnostics and medical image processing via the WWW--the case-based radiological textbook ODITEB. Int J Med Inform. 2000, 58-59: 39-50. 10.1016/S1386-5056(00)00074-5.View ArticleGoogle Scholar
- De La Vega F, Giegerich R, Fuellen G: Distance Education Through The Internet: The GNA-VSNS Biocomputing Course. Pac Symp Biocomput. 1996, 203-215.Google Scholar
- Dixon-Woods M, Agarwal S, Jones D, Young B, Sutton A: Synthesising qualitative and quantitative evidence: a review of possible methods. J Health Serv Res Policy. 2005, 10: 45-53. 10.1258/1355819052801804.View ArticleGoogle Scholar
- Thomas J, Harden A, Oakley A, Oliver S, Sutcliffe K, Rees R, et al: Integrating qualitative research with trials in systematic reviews. BMJ. 2004, 328: 1010-1012. 10.1136/bmj.328.7446.1010.View ArticleGoogle Scholar
- Kristjansson E, Robinson V, Petticrew M, Macdonald B, Krasevec J, Janzen L, et al: School feeding for improving the physical and psychosocial health of disadvantaged elementary school children. Cochrane Database Syst Rev. 2007, CD004676.Google Scholar
- Greenhalgh T, Kristjansson E, Robinson V: Realist review to understand the efficacy of school feeding programmes. BMJ. 2007, 335: 858-861. 10.1136/bmj.39359.525174.AD.View ArticleGoogle Scholar
- Cook D: Narrowing the focus and broadening horizons: complementary roles for nonsystematic and systematic reviews. Adv in Health Sci Educ. 2008, 13: 391-395. 10.1007/s10459-008-9140-4.View ArticleGoogle Scholar
- Cook D: Web-based learning: pros, cons and controversies. Clinical Medicine. 2007, 7: 737-742.View ArticleGoogle Scholar
- Bernard R, Abrami P, Borokhovski E, Wade A, Tamin R, Surkes M, et al: A Meta-Analysis of Three Types of Interaction Treatments in Distance Education. Rev Educ Res. 2009, 79: 1243-1289. 10.3102/0034654309333844.View ArticleGoogle Scholar
- Malterud K: Qualitative research: standards, challenges, and guidelines. Lancet. 2001, 358: 483-488. 10.1016/S0140-6736(01)05627-6.View ArticleGoogle Scholar
- Ruiz J, Cook D, Levinson A: Computer animations in medical education: a critical literature review. Med Educ. 2009, 43: 838-846. 10.1111/j.1365-2923.2009.03429.x.View ArticleGoogle Scholar
- Cook D, Beckman T, Bordage G: Quality of reporting of experimental studies in medical education: a systematic review. Med Educ. 2007, 41: 737-745. 10.1111/j.1365-2923.2007.02777.x.View ArticleGoogle Scholar
- Grant J: Principles of curriculum design. 2006, Edinburgh: Association for the Study of Medical EducationGoogle Scholar
- McKendree J: eLearning. 2006, Edinburgh: Association for the Study of Medical EducationGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6920/10/12/prepub