- Research article
- Open Access
- Open Peer Review
Eight years’ experience with a Medical Education Journal Club in Mexico: a quasi-experimental one-group study
BMC Medical Education volume 15, Article number: 222 (2015)
A time-honored strategy for keeping up to date in medicine and improving critical appraisal skills is the Journal Club (JC). There are several reports of its use in medicine and allied health sciences but almost no reports of JC focused on medical education. The purpose of the study is to describe and evaluate an eight years’ experience with a medical education Journal Club (MEJC).
We started a monthly medical education JC in 2006 at UNAM Faculty of Medicine in Mexico City. Its goal is to provide faculty with continuing professional development in medical education. A discussion guide and a published paper were sent 2 weeks before sessions. We reviewed the themes and publication types of the papers used in the sessions, and in June-July 2014 administered a retrospective post-then-pre evaluation questionnaire to current participants that had been regular attendees to the JC for more than 2 years. The retrospective post-then-pre comparisons were analyzed with Wilcoxon signed-rank test. Effect sizes were calculated for the pre-post comparisons with Cohen’s r.
There have been 94 MEJC sessions until July 2014. Average attendance is 20 persons, a mix of clinicians, educators, psychologists and a sociologist. The articles were published in 32 different journals, and covered several medical education themes (curriculum, faculty development, educational research methodology, learning methods, assessment, residency education). 22 Attendees answered the evaluation instrument. The MEJC had a positive evaluation from good to excellent, and there was an improvement in self-reported competencies in medical education literature critical appraisal and behaviors related to the use of evidence in educational practice, with a median effect size higher than 0.5. The evaluation instrument had a Cronbach’s alpha of 0.96.
A periodic Medical Education Journal Club can improve critical appraisal of the literature, and be maintained long-term using evidence-based strategies. This activity is a useful adjunct to the scholarship of teaching.
Journal Clubs (JC) are a time-honored method of reviewing and discussing the scientific literature in medicine [1, 2]. There is no universally accepted definition of JC, even though this educational modality has been around for more than a century and has evolved substantially during this time. Several published papers about JC in the health sciences define it as “a group of individuals who meet regularly to discuss critically the clinical applicability of articles in the current medical journals” . Some authors have described the origin and history of JC in medicine, apparently Sir William Osler was the first to establish a formal JC at McGill University in Montreal, Canada, in 1875 [1, 2].
Initially the main goal of the JC was to help participants keep abreast of the growing body of medical literature. In the era of paper-only medical journals it was time-consuming and expensive to access the original papers needed to maintain competence in medicine. Over time the focus of JC has moved to teaching critical appraisal skills and evidence-based medicine, while preserving the original objective of helping students, residents and practicing physicians keep up-to-date in the research literature of their field [2, 3]. The appropriate methodology to implement an effective JC in medicine, as well as the factors associated with a successful one, have been extensively reviewed and discussed [3–6].
Even though JC are popular and used frequently in academic health centers and medical schools, there is little published information about JC dedicated to the subject of medical education. There are only three brief case reports with scant data [7–9], and a review paper in the “12 tips” format . This limited amount of information is surprising, since the field of medical education has grown enormously in the last decades, as shown by the large global number of master in health professions education programs, the variety of related research papers and scientific journals, and the rising number of medical education organizations and academic venues [11–13]. There is a need to formally explore the use of JC in medical education settings, to promote the use of educational evidence by clinicians and basic science teachers. Our research question was: does a medical education Journal Club (MEJC) increase self-reported educational research knowledge and skills in our University medical educators? We describe the design, long-term implementation and evaluation of a medical education Journal Club (MEJC) at UNAM Faculty of Medicine in Mexico City.
The National Autonomous University of Mexico (UNAM) Faculty of Medicine in Mexico City is the largest medical school in the country and one of the largest in Latin America, with more than 7,000 undergraduate students and about 9,000 medical residents. It is a public institution and the largest generator of basic and clinical medical research in Mexico, through its affiliations with major national academic medical centers. Its Postgraduate Studies Division (PSD) is the largest graduate program in Mexico. The PSD offers 78 medical specialty courses, and currently has 8,739 registered medical residents and almost fifteen hundred clinical teachers . The Division has academic staff responsible for faculty development activities, medical education service and research.
Origin and development of UNAM MEJC
In January 2006, we designed a Journal Club focused on medical education, specifically for the Postgraduate Studies Division. The goals of the MEJC were: to provide continuous professional development to the Division scholars; to discuss the use of research in addressing medical residency educational challenges; to keep up to date in the medical education field; to identify sources of medical education publications and to improve knowledge of medical education research methodology. The MEJC was targeted to PSD academic staff and clinicians responsible for residents’ education from the neighboring academic health centers affiliated with the university.
The MEJC is a monthly one-hour activity, in a regular time and day. It is organized by the PSD academic staff, and has been led for the last eight years by a pediatrician with a Masters’ degree in Health Professions Education. The PSD office sends a formal invitation and the documents to be discussed to the regular attendees.
The MEJC coordinator selects papers from the medical education literature, relevant to graduate medical education. The selection is informed by requests from regular session attendees and the current educational challenges of the Division. The selected paper and a list of questions designed as discussion guide to promote reflection, analysis and application of the concepts described in the study, are sent by e-mail to the group of potential attendees, 1 to 2 weeks before the session. A core regular population of participants has developed, as well as a smaller, less-regular attendee group. The design, planning and implementation of the MEJC has used several principles and strategies reported in the literature, to increase its chance of success and sustainability [3–6, 15].
The authors reviewed and classified the papers discussed in the sessions, using a schema that identified papers by research design (quantitative, qualitative, mixed methods), type of article (case report, essay, observational or experimental study, systematic or narrative review), content area (curriculum, teaching, learning, faculty development, assessment, program evaluation, education research), purpose of research , type of journal (medical education, clinical or specialty journal, general education journal) and journal title.
We designed a questionnaire to evaluate this academic activity, that included attendees’ demographic data, sessions’ characteristics, and a “retrospective pre-post” section that assessed their self-reported change in medical education literature critical appraisal knowledge and competencies, as well as behavior modification related to their use of the sessions’ concepts in their educational practice. The respondents were asked to provide their best overall estimate for the sessions of each specific aspect explored.
We used a quasi-experimental “retrospective pre-post” methodology (also called “retrospective post-then-pre” design) to control the response-shift bias that frequently occurs in traditional pre-post evaluations, minimizing pre-test over or underestimation [17–19].
“Retrospective pre-post” is a term that is not widely used in clinical research. This method refers to the assessment of learners’ self-reported changes in knowledge, skills, confidence, attitudes or behaviors where both before and after information is collected at the same time. After the educational experience learners are asked to rate their current knowledge, skill, attitude, behavior as a result of the program, and then, to reflect back and rate that same knowledge, skill, attitude, behavior before participating in the program [17–19]. This methodology was described in the 1970’s as a way to minimize the “response shift bias” that occurs in traditional pre-post design. This bias happens when a participant uses a different frame of understanding about an item between the pre and post periods, since learners may not accurately assess their pre-program knowledge or behaviors. At the end of the program, their new understanding of content may affect their response on the post self-assessment, so they are actually responding based on two different frames of reference [20–23].
Research has shown that response shift can mask program effectiveness. The retrospective pre-post design reduces or eliminates response shift bias; compared with the traditional pre- and post design, results from the retrospective post-then-pre design are more consistent with interview data collected from program participants. The retrospective post-then-pre method has been used in many settings and appears to reduce response shift bias across contexts. Responding to both measures at the same time is less burdensome and intrusive for learners, and provides before and after data for each learner [20–23].
The questionnaire also included open items related to suggestions about the MEJC. The instrument was administered to the regular attendees of the MEJC in a voluntary and anonymous fashion. We distributed the questionnaires at the end of the June and July 2014 sessions, but only to the participants that had attended the Journal Club for more than 2 years. The response rate of the current 2014 regular attendees (those that attend more than 80 % of the sessions per year) was 100 %, two participants that are authors of this study did not respond the questionnaire. During the eight-year period, there are some participants that have moved to other institutions, we included only the current participants.
Data were analyzed with PRISM Version 6 for the Macintosh (http://www.graphpad.com/scientific-software/prism/), obtaining descriptive and inferential statistics. The retrospective pre-post comparisons were analyzed with the non-parametric Wilcoxon signed-rank test. P values less than 0.05 were considered statistically significant. Effects sizes for non-parametric Wilcoxon-signed rank test results were calculated for the pre-post comparisons with Cohen’s r .
The study was in compliance with the Declaration of Helsinki of ethical principles for research involving human subjects. It was approved by the Research Department of the Postgraduate Studies Division, UNAM Faculty of Medicine. Verbal informed consent was obtained. The data were managed anonymously, and analyzed in the aggregate.
94 MEJC sessions have taken place between 2006 and 2014, about 11 per year. The average attendance is 20, a mix of clinicians, educators, psychologists and one sociologist. Attendance is voluntary, and a certificate of attendance with curricular value in our university is given to the participants every year. The type of articles reviewed cover the breadth of medical education research, mostly related to postgraduate medical education (curriculum development, hidden curriculum, faculty development, educational research methodology –quantitative, qualitative and mixed methods-, learning methods, assessment, systematic reviews, e-learning, among others). The articles’ scope and relevant methodological data are presented in Tables 1, 2, 3, 4.
The papers discussed in the MEJC were published in 32 different journals, the more frequent sources were: Medical Education = 14 (15 %), Academic Medicine = 13 (14 %), Medical Teacher = 12 (13 %), and Advances in Health Sciences Education = 7 (7 %). The evaluation questionnaire was responded by 22 participants, excluding the MEJC coordinator. The data were captured by a person not involved in the implementation of the MEJC, and analyzed in aggregate form. The demographic data of the participants are outlined in Table 5.
Eight aspects of the MEJC sessions were assessed with a 1 to 9 scale, where the respondents were asked to provide their best overall estimate for each specific aspect of the sessions. The results are shown in Table 6. The majority of these aspects had a positive evaluation.
The “retrospective pre-post” portion of the questionnaire showed statistically significant increases in knowledge and skills related to critical appraisal of the medical education literature, as well as positive changes in behaviors related to the use of published evidence in their educational practice (Tables 7 and 8). The overall median increase in knowledge score of medical education research concepts and critical appraisal skills was 3 (in a scale of 0 to 10), with a median effect size of 0.80. The instrument’s reliability was 0.96, measured with Cronbach’s alpha.
This study is the first report that provides a detailed description of the implementation of a journal club dedicated to medical education. It suggests that this scholarly activity may improve knowledge and behaviors related to the teaching activities of medical educators, and promote evidence-based medical education in a medical school context. Traditional JC are ubiquitous in medical schools and health care institutions, this educational modality has a long and rich history in medicine and allied health sciences all over the world [1–3, 5, 15, 25–29]. The flexibility and intellectual challenge provided by JC have stimulated its use not only in general medicine and several medical specialties, but also in nursing, dentistry and other allied health professions [15, 27, 30–35]. JC have diverse goals, like promoting self-directed learning, keeping up-to-date in the published literature, linking published science with practical clinical problems, and recently they are frequently used to teach critical appraisal skills and evidence-based medicine [36–40].
It is interesting and paradoxical that with such extensive dissemination of the JC methodology, there are almost no published papers related to its use in the field of medical education. There are many reports of its use as an educational tool and strategy, but almost none about its utilization for discussing medical education published papers. We did a wide-ranging search in the most important biomedical and educational databases for published paper descriptions and research studies related to medical education journal clubs, and found many papers (171 references) describing JC in medicine and other health professions, but only four related to medical education JC [7–10]. Three are short reports: the first describes the use of JC as a continuing medical education activity for teachers at Universidad Austral in Buenos Aires, Argentina ; the second is a description of an advanced education faculty development program for primary care physicians at the Medical College of Wisconsin, USA ; and the third is a case report of a faculty development initiative for a major curricular change using the JC as a culture changing strategy, at the University of Virginia School of Medicine, USA . These papers are brief descriptions of their MEJC, and do not include data to evaluate their educational effectiveness. The fourth paper is a set of recommendations published in the Medical Teacher “12 tips” format, where after surveys and interviews with the McGill University Centre for Medical Education scholars, they propose guidelines for a successful medical education journal club .
We searched Google with the term “medical education journal club”, and we found many web sites that announce MEJC activities in many medical schools in several countries. We can only speculate on the reasons for this asymmetry in the number of scholarly publications about the JC method in medicine and its limited number in medical education JC, but we think it is important to approach MEJC with a scholarly lens, and report positive experiences that could be relevant to clinicians and medical educators in a variety of settings.
Our study describes a specific medical education journal club methods and strategies, and presents data on self-reported educational effectiveness. There are guidelines for organizing and maintaining a JC, most from clinical JC papers, and one of the consistent recommendations is that these educational events need specific goals and a designated leader/coordinator (which can be one person or a rotating position) with disciplinary expertise, in this case a medical education background. There are no follow-up reports of the published MEJC experiences, so we don’t know if they are currently active, but those papers emphasize that faculty with educational background are important to start and maintain the project, preferably from an office or department of medical education, and that clinicians with educational responsibilities should be involved [7–9]. In our setting the group coordinating the MEJC has been the same for more than 8 years, led by a pediatrician with clinical experience and formal training in health professions education. McLeod and colleagues suggest that the leadership of a MEJC should vary, so every regular participant can have the opportunity of leading the discussion . In any case, the decision of who coordinates the MEJC should be an academic one obtained by group consensus.
Our MEJC has several similarities with the published case reports, like its monthly periodicity (Simpson reported a bimonthly, Centeno monthly and Pollart a biweekly JC), short duration of the sessions, broad variety of articles and content, and high levels of satisfaction in the participants. There are some differences in our approach: we focus on papers that are relevant for residency education; and we choose only one paper for discussion in each session, unlike Simpson’s approach where they review 12 to 15 journals in an hour and choose about eight papers for critical appraisal and follow-up. One of our main goals is to learn medical education critical appraisal skills and research methodologies, and a one-hour session is just enough to discuss one paper appropriately. The variety of topics and types of papers discussed in our MEJC provides a reasonable overview of medical education in residency training. We have emphasized themes like assessment, research education, duty hours and fatigue, faculty development, among others.
We have followed several of McLeod’s recommendations for a successful MEJC: broad representation of participants; visible leadership support (the Director and high-level staff from the Division usually attend); provision of coffee and pastries; consistency and punctuality, every year we have had about 11 sessions, rain or shine; a pleasant and modern physical facility; participants’ enthusiasm and loyalty (more than 70 % of the current attendees have participated since 2006); an stress-free open environment for discussion; focus on methodology and practical implications of educational research in our setting; use of a semi-structured discussion guide with open-ended questions; it is not mandatory but we register attendance, and at the end of the year we provide a valid certificate that is useful in our university recognition system.
Our findings show a high level of satisfaction with the MEJC activity, from the physical setting to more complex aspects like the quality of group discussions (Table 6). Our data show a substantial and statistically significant increase in self-reported knowledge of several domains related to medical education research and evidence-based education (Tables 7 and 8). Cohen’s guidelines for the interpretation of effect sizes expressed as “r” are that a large effect is 0.5, a medium effect 0.3, and a small effect 0.1 . The effect size in all our items for knowledge and behavior change is large (>0.5), which suggests an effect that can be educationally relevant . This is apparently the first study that provides educational impact data with a MEJC, and paves the way for other studies. Evidence-based recommendations for JC effectiveness, McLeod and colleagues’ 12 tips for a successful MEJC, and our findings are a reasonable starting point for planning, implementing and developing a MEJC in a medical education setting.
There are some limitations to the study: it is the experience of a single medical school in a developing country, with its inherent external validity and ecological generalizability implications; the data are self-reports, which rises the question of the shortcomings of self-assessment in medicine and the frequent overestimation of one’s abilities. We are aware that there is a need for objective external instruments that measure knowledge and behavior modification, but that was not feasible in our study settings. Nonetheless self-efficacy may be associated with future performance and behavior in some contexts [42, 43], and when individuals transit from the pre-contemplation phase to the contemplation and preparation stages of Prochaska’s transtheoretical model of behavior change, there is a possibility they move to the action phase . Our results suggest that MEJC participants are aware of the need and potential for behavior change and maybe these self-reported increases in knowledge and behavior could be educationally relevant.
We used a quasi-experimental one-group retrospective pre-post research design methodology (also called “post-then-pre”) instead of the traditional one-group pretest-posttest design, because the retrospective strategy is less intrusive, takes less time, and minimizes pretest sensitivity and the response shift bias that can occur as a consequence of over or underestimation of the individuals’ self-perceived changes in knowledge and abilities [17, 18]. On the other hand, in our case several years had passed since we started the JC, so an actual pretest was not feasible.
There are also some limitations to the JC methodology itself, which can explain its limited capacity to produce behavior change in residents and physicians, as has been shown in several studies [5, 27, 38, 45]. It is important to be cognizant of the educational limitations of the JC to improve clinical and educationally relevant outcomes. Some of these factors are: the artificial setting of the JC which occurs usually in a classroom or conference room, away from the real clinical arena; the risk of heterogeneous participation of JC attendees, promoting passivity in some participants; the lack of motivation to read a complex research paper that does not seem immediately relevant to clinical practice; the difficulty of some of the concepts related to critical appraisal skills, research methods, statistics and evidence-based medicine, which can be hard to grasp and difficult to understand in the JC scenario, where most of the interaction is verbal.
It is noteworthy that the published evidence about JC educational effectiveness in clinically relevant outcomes is limited. Deenadayalan et al. performed an extensive systematic review and found 101 papers, of which 21 fulfilled the inclusion criteria , and only 12 described JC effectiveness. Their methodological quality was moderate. More than 80 % of the papers reported that their intervention improved knowledge and critical appraisal skills. Just a few described the psychometric properties of their instrument, and no paper reported evidence of JC impact in clinical practice . There are several initiatives to improve the educational significance and impact of the JC, like the use of e-learning modalities, team-based-learning, problem-based learning, positive-deviant strategies, which could be considered in the implementation of a MEJC [45–55].
We have shown that a continuous professional development activity in the form of a periodic face-to-face Medical Education Journal Club can be maintained in the long-term using evidence-based strategies, and that the experience can have positive effects. The MEJC is an academic activity that can improve the use and appraisal of the medical education literature and promote socialization and community building in health professions’ educational settings.
A MEJC is an educational activity that can be included in the routine scholarly activities of a medical school department, be a useful adjunct to improve the scholarship of teaching and promote evidence-based medical education practice.
Linzer M. The journal club and medical education: over one hundred years of unrecorded history. Postgrad Med J. 1987;63(740):475–8.
Valentini RP, Daniels SR. The journal club. Postgrad Med J. 1997;73(856):81–5.
Alguire PC. A review of journal clubs in postgraduate medical education. J Gen Intern Med. 1998;13(5):347–53.
Deenadayalan Y, Grimmer-Somers K, Prior M, Kumar S. How to run an effective journal club: a systematic review. J Eval Clin Pract. 2008;14(5):898–911.
Ebbert JO, Montori VM, Schultz HJ. The journal club in postgraduate medical education: a systematic review. Med Teach. 2001;23(5):455–61.
Forsen JW, Hartman JM, Neely JG. Tutorials in clinical research, part VIII: Creating a journal club. Laryngoscope. 2003;113(3):475–83.
Centeno AM, Blanco A, Arce M. Journal club devoted to educational issues. Acad Med. 1999;74(5):464.
Pollart SM, Caelleigh AS. Changing conversations, changing culture: a medical education journal club. Med Educ. 2011;45(11):1134.
Simpson D, Flynn C, Wendelberger K. An evidence-based education journal club. Acad Med. 1997;72(5):464.
McLeod P, Steinert Y, Boudreau D, Snell L, Wiseman J. Twelve tips for conducting a medical education journal club. Med Teach. 2010;32(5):368–70.
Lee K, Whelan JS, Tannery NH, Kanter SL, Peters AS. 50 Years of Publication in the Field of Medical Education. Med Teach. 2013;35(7):591–8.
Sampson M, Horsley T, Doja A. A bibliometric analysis of evaluative medical education studies: characteristics and indexing accuracy. Acad Med. 2013;88(3):421–7.
Tekian A, Harris I. Preparing health professions education leaders worldwide: A description of masters-level programs. Med Teach. 2012;34(1):52–8.
Sánchez-Mendiola M. El Seminario de Educación del Plan Único de Especialidades Médicas de la Facultad de Medicina UNAM: una Reflexión Crítica. Los Retos de la Educación Médica. 2012;1(1):135–62.
Sidorov J. How are internal medicine residency journal clubs organized, and what makes them successful? Arch Intern Med. 1995;155(11):1193–7.
Cook DA, Bordage G, Schmidt HG. Description, justification and clarification: a framework for classifying the purposes of research in medical education. Med Educ. 2008;42(2):128–33.
Bhanji F, Gottesman R, de Grave W, Steinert Y, Winer LR. The retrospective pre-post: a practical method to evaluate learning from an educational program. Acad Emerg Med. 2012;19(2):189–94.
Hill LG. Revisiting the Retrospective Pretest. Am J Eval. 2005;26(4):501–17.
Skeff KM, Stratos GA, Bergen MR. Evaluation of a medical faculty development program. A comparison of traditional pre-post and retrospective pre-post self-assesment ratings. Eval Health Prof. 1992;15(3):350–66.
Howard GS. Response-shift bias a problem in evaluating interventions with pre/post self-reports. Eval Rev. 1980;4(1):93–106.
Howard GS, Ralph KM, Gulanick NA, Maxwell SE, Nance SW, Gerber SK. Internal invalidity in pre-test-post-test self-report evaluations and a re-evaluation of retrospective pre-tests. Appl Psychol Meas. 1979;3:1–23.
Lam TC, Bengo P. A comparison of three retrospective self-reporting methods of measuring change in instructional practice. Am J Eval. 2003;24(1):65–80.
Pratt CC, McGuigan WM, Katzev AR. Measuring program outcomes: Using retrospective pretest methodology. Am J Eval. 2000;21(3):341–9.
Fritz CO, Morris PE, Richler JJ. Effect size estimates: Current use, calculations, and interpretation. J Exp Psychol Gen. 2012;141(1):2–18.
Cave MT, Clandinin DJ. Revisiting the journal club. Med Teach. 2007;29(4):365–70.
Fowler L, Gottschlich MM, Kagan RJ. Burn center journal club promotes clinical research, continuing education, and evidence-based practice. J Burn Care Res. 2013;34(2):e92–8.
Honey CP, Baker JA. Exploring the impact of journal clubs: a systematic review. Nurse Educ Today. 2011;31(8):825–31.
Pato MT, Cobb RT, Lusskin SI, Schardt C. Journal club for faculty or residents: a model for lifelong learning and maintenance of certification. Int Rev Psychiatry. 2013;25(3):276–83.
Ruíz García V, Cabello JB. Journal clubs at XXIth century. Med Clín. 2010;135(12):556–60.
Akhund S, Kadir MM. Do community medicine residency trainees learn through journal club? An experience from a developing country. BMC Med Educ. 2006;6:43. http://www.biomedcentral.com/1472-6920/6/43.
Dirschl DR, Tornetta P, Bhandari M. Designing, conducting, and evaluating journal clubs in orthopaedic surgery. Clin Orthop Relat Res. 2003;413:146–57.
Moro JK, Bhandari M. Planning and executing orthopedic journal clubs. Indian J Orthop. 2007;41(1):47–54.
Lee AG, Arnold AC, Olson RJ, Carter K. Using the Journal Club to teach and assess competence in practice-based learning and improvement: a literature review and recommendation for implementation. Surv Ophthalmol. 2005;50(6):542–8.
Phitayakorn R, Gelula MH, Malangoni MA. Surgical journal clubs: A bridge connecting experiential learning theory to clinical practice. J Am Coll Surg. 2007;204(1):158–63.
Shifflette V, Mitchell C, Mangram A, Dunn E. Current approaches to journal club by general surgery programs within the Southwestern surgical congress. J Surg Educ. 2012;69(2):162–6.
Ahmadi N, McKenzie ME, Maclean A, Brown CJ, Mastracci T, McLeod RS. Teaching evidence based medicine to surgery residents-is journal club the best format? A systematic review of the literature. J Surg Educ. 2012;69(1):91–100.
Chávez-Romero SA, Sánchez-Mendiola M. Evaluación del Journal Club en residentes del Área de Pediatría del Hospital Central Militar. Rev Sanid Milit Méx. 2005;59(3):147–54.
Harris J, Kearley K, Heneghan C, Meats E, Roberts N, Perera R, et al. Are journal clubs effective in supporting evidence-based decision making? A systematic review. BEME Guide No. 16. Med Teach. 2011;33(1):9–23.
Wright J. Journal clubs--science as conversation. New Engl J Med. 2004;351(1):10–2.
Young T, Rohwer A, Volmink J, Clarke M. What are the effects of teaching evidence-based health care (EBHC)? Overview of systematic reviews. PLoS One. 2014;9(1), e86706.
Sullivan G, Feinn R. Using effect size-or why the P value is not enough. J Grad Med Educ. 2012;4(3):279–82.
Bandura A. Self-efficacy: toward a unifying theory of behavioral change. Psychol Rev. 1977;84(2):191–215.
Hughes A, Galbraith D, White D. Perceived competence: a common core for self-efficacy and self-concept? J Pers Assess. 2011;93(3):278–89.
Prochaska JO. Decision making in the transtheoretical model of behavior change. Med Decis Making. 2009;28(6):845–9.
Hatala R, Keitz SA, Wilson MC, Guyatt G. Beyond journal clubs. Moving toward an integrated evidence-based medicine curriculum. J Gen Intern Med. 2006;21(5):538–41.
Anzarut A, Martens B, Tredget E. Improving journal clubs through the use of positive deviance: A mixed-methods study. Can J Plast Surg. 2011;19(3):82–4.
Baird JS. Journal clubs: what not to do. Acad Med. 2012;87(3):257.
Chakraborti C. Teaching evidence-based medicine using team-based learning in journal clubs. Med Educ. 2011;45(5):516–7.
Currier RL, Schneider MR, Heubi JE. Taking Journal Clubs off Autopilot: A Case Study of Teaching Literature Evaluation Skills to Preclinical MD/PhD Students. Med Sci Educ. 2013;23(4):572–7.
Freshwater MF. The four C’s for a journal club--ingredients for success or failure. J Plast Reconstr Aesthet Surg. 2011;64(6):839–41.
Hartzell JD, Veerappan GR, Posley K, Shumway NM, Durning SJ. Resident run journal club: a model based on the adult learning theory. Med Teach. 2009;31(4):e156–61.
Lizarondo L, Kumar S, Grimmer-Somers K. Online journal clubs: an innovative approach to achieving evidence-based practice. J Allied Health. 2010;39(1):e17–22.
Touchet BK, Coon KA, Walker A. Journal Club 2.0: using team-based learning and online collaboration to engage learners. Acad Psychiatry. 2013;37(6):442–3.
Willett LR, Kim S, Gochfeld M. Enlivening journal clubs using a modified “jigsaw” method. Med Educ. 2013;47(11):1127–8.
Yang PR, Meals RA. How to establish an interactive eConference and eJournal Club. J Hand Surg Am. 2014;39(1):129–33.
We acknowledge the enthusiastic participation of UNAM’s Postgraduate Studies Division faculty and staff during all the MEJC sessions.
The authors declare that they have no competing interests.
MSM developed the concept of the paper. All the authors contributed to the development of the evaluation instrument. MSM, DMC and UTG did the data management and statistical analysis. MSM, DMC and MV did the literature search and wrote the first draft of the paper. All authors contributed to the final manuscript.
• Melchor Sánchez-Mendiola, MD, MHPE, PhD, is a pediatrician and medical educator at the National Autonomous University of Mexico (UNAM) Faculty of Medicine. He is Professor of Medical Education at UNAM Postgraduate Studies Division, where he is also the coordinator of the Masters in Health Professions Education program.
• Daniel Morales-Castillo, MEd, is a professional in pedagogy with a master in education. He works at the Medical Education Research Department, UNAM Faculty of Medicine in Mexico City, and is editorial assistant of the journal “Investigación en Educación Médica”, UNAM’s peer-reviewed journal dedicated to medical education research.
• Uri Torruco-García, MD, MEd, is a specialist in Internal Medicine and Infectious diseases. He works in the Medical Education Research Department at UNAM Faculty of Medicine.
• Margarita Varela-Ruiz, MSc, is a psychologist with Master’s degree in educational technology. She is the Chief of the Medical Education Research Department at UNAM Faculty of Medicine, and is heavily involved in the Master’s program in Health Professions Education at our institution.
About this article
- Journal club
- Continuing professional development
- Evidence based medical education
- Postgraduate medical education
- Faculty development