- Open Access
How to define core entrustable professional activities for entry into residency?
BMC Medical Education volume 18, Article number: 87 (2018)
Institutions considering to employ core Entrustable Professional Activities (EPAs) for entry into postgraduate training as outcomes for their undergraduate medical programs can partly build on published examples, but also have to undergo their own content validation process to take their specific context into consideration. This process involves several challenges and is not well-described in the literature. Here, we report in detail on a systematic, literature-based approach we recently utilised at our institution to define core EPAs for entry into residency.
Central to the process was a modified Delphi consent procedure. It involved a multistep interaction between a writing team and a multidisciplinary panel of experienced physicians. Panel members provided both quantitative ratings and qualitative feedback on the EPA categories title, specification/limitations, conditions and implications of entrustment decision, knowledge, skills, and attitude. Consent was achieved when a Content Validity Index (CVI) of ≥80% was reached. The writing team adjusted the EPA category descriptions on the basis of panel members´ ratings and comments, and specified the EPA categories’ link to competencies and assessment sources. This process produced a description and definition of a full set of core EPAs for entry into residency adapted to our context.
This process description for locally adapted core EPAs for entry into residency may support and guide other medical schools in the development and implementation of EPAs into their programs.
The definition of core Entrustable Professional Activities (EPAs) for entry into postgraduate training has become an active field of development. Many institutions are currently considering the use of EPAs as outcomes for their undergraduate medical programs . These institutions can build in part on EPAs which have been reported at a national level [2,3,4] and at a local level , but will be required to undertake their own content validation process to adapt these EPAs to their specific context. However, available reports do not include a fully detailed description of the EPA development process which could guide other institutions. In this article, we report in detail on a systematic, literature-based approach we employed to define core EPAs for entry into residency as outcomes for the undergraduate medical curriculum at Charité - Universitaetsmedizin Berlin, Germany (Charité).
We chose a modified Delphi study procedure, an established method for anonymised, non-hierarchical content validation, including EPA development in medical education [1, 6]. As a modification of the Delphi process, panel members received a predefined list of EPAs in the first round. Our goal was the definition of a full set of core EPAs with a seven-category description for each EPA according to current recommendations in the literature (1, 7, 8]. The definition of educational outcomes by EPAs is generally achieved in an iterative process, beginning with the identification of authentic professional tasks, followed by the elaboration of its characteristics, and finally validation of the content by a group of experts [1, 7]. Figure 1 provides an overview of our Delphi study process which involved a multistep interaction between a writing team of educationalists and a panel of experienced physicians.
Delphi study process
Panel selection and writing team
A total of 45 panel members were purposely selected from the Charité faculty body. All panel members had long-time supervision experience in both undergraduate and postgraduate medical training and were actively involved in the curriculum development process for the current undergraduate program. The EPA writing team, consisting of the authors of this article, were members of both the curriculum development group for the undergraduate medical program and educational researchers in the field of EPAs.
Guiding principles for EPA content definition
The following guiding principles were formulated for EPA content definition: 1) They should comply with the recommendations for EPA definition , i.e. represent independently executable tasks which are observable, measurable, confined to qualified personal and suitable for an entrustment decision. 2) The EPAs should consist of full, seven-category descriptions, including the following categories: ‘title’, ‘specification/limitations’, ‘knowledge, skills and attitudes’ (KSA), ‘conditions and implications of entrustment decision’, ‘most relevant domains of competence’, ‘assessment sources’, and the ‘expected supervision level at the stage of training’. 3) The EPA content elaboration should use clear language describing tasks and workplace context and avoid educational jargon. This includes a short brief title, succinct descriptions as well as an alignment of structure, language and wording within the set of EPAs. 4) The EPAs for entry into residency constitute the core, this is the full set of professional activities expected from a graduating physician. 5) The breadth and level of difficulty of the EPAs should be manageable for graduating physicians, align with the workflow and the supervision routines in the clinical setting, and 6) the supervision level is defined by the time it takes for the supervisor to be physical available as well as the degree of subsequent work verification.
Drafting of the initial EPA list and category description
In an iterative process, the writing team drafted the initial list of tasks to be considered as EPAs for entry into residency according to the specific context. This involved mapping these tasks to the Charité competency framework, a search and appraisal of the literature, along with continuous discussions and developments within the Charité curriculum development group. The AAMC core EPAs were used as a starting point . In addition, the writing team consulted articles on the EPAs concept in general [8,9,10,11,12] and articles covering the development of EPAs for postgraduate training [13,14,15,16]. The draft of the initial EPA list included tasks which graduating physicians should be able to perform under a granular operationalised level of supervision . The following categories were elaborated for each EPA: ‘title’, ‘specification/limitations’, and ‘expected supervision level at stage of training’. These categories are thought of as those representing the quintessence of an EPA description, upon which the other categories subsequently build.
The writing team developed the questionnaire for the Delphi process based on the literature on EPA development. For EPA identification and content validation, panel members were asked to rate the relevance of professional tasks for new residents, the clarity of each EPA title and the completeness of the EPA category descriptions on a 4-point scale. The questionnaire was administered online using EvaSys (Electric Paper Evaluationssysteme GmbH, Lüneburg, Germany), a software for survey based-research.
Establishing consensus among panellists’ ratings
Content validity indices (CVI) were calculated to establish consensus among the ratings of panel members .This included the relevance ratings of the EPAs, the ratings of the ‘clarity of the title’ and the completeness of the EPA categories ‘specification/limitations’, ‘conditions and implications of entrustment decision’ and ‘KSA’. The CVI describes the percentage of respondents who rated the relevance of the EPAs or the categories with ‘agree’ or ‘somewhat agree’. A CVI of at least 80% was set as the predefined consensus level. If this level was reached, consensus was assumed and no further validation was deemed necessary.
Panel member invitation and briefing
The panel members were invited to a formal meeting at the beginning of the Delphi study to prepare them for their participation. During the meeting, they were informed about the EPA concept and the aim and structure of the Delphi process. Similar panel meetings were held again before the second and third Delphi rounds. Here, panel members were provided with an anonymised summary of the previous round’s results, the refined EPA content descriptions, and information on subsequent tasks. The meetings were audio-recorded, screen-casted, and sent out to panel members as podcasts along with other material shown in the panel meeting.
Panel members received the initial draft of EPAs relevant for entering residency including titles and specification/limitations. Panel members provided ratings and could add narrative text for explanations or suggestions for refinement. They were also asked to propose relevant tasks which they felt were missing for entry into residency. The EPA writing team summarised the quantitative and qualitative information provided and refined the EPAs accordingly. The qualitative feedback was clustered inductively and allocated to the corresponding EPA text passages. The proposed changes were then discussed within the writing team until a consensus was reached on the EPA description refinement. The topics for additional EPAs were discussed within the writing team and reviewed on the basis of the above-described guiding principles for EPA content definition.
Panel members received the anonymised, summarised panel rating results of the first round along with the refined EPA titles and specification/limitations descriptions. Changes made following the first round were highlighted. The panel members received the same questions as in Delphi Round 1. In addition, they were asked to rate a description drafted by the writing team on the EPA category ‘conditions and implication of entrustment decision’ which specifies how the supervision level is operationalised into the workplace. Again, all quantitative ratings could be supplemented with narrative feedback. The EPA writing team summarised the quantitative and qualitative information and adjusted the EPA descriptions as described above for Round 1.
The panel members were given the anonymised, summarised panel rating results of Round 2. They also received the refined EPA titles and descriptions of the categories ‘specification/limitations’ and ‘conditions and implications of entrustment decision’ with an indication of changes made following feedback in the previous round. Panel members were asked to rate again on the content of the refined categories in those EPAs which had not received sufficient consensus on the relevance rating in the previous round. For the third Delphi round, the writing team drafted the EPA category ‘KSA’ for each EPA. The panel members rated the completeness of the categories ‘conditions and implication of entrustment decision’ and ‘KSA’ in all EPAs. The ratings could be supplemented by narrative comments. In the final round, a CVI of over 80% was reached in the panellists’ ratings on the EPA category descriptions.
Finalisation of EPA list and category descriptions
The writing team made final changes to the content of the EPA categories on the basis of panel member ratings and comments from Round 3. EPA categories ‘most relevant domains of competence’ and ‘assessment sources’ were defined in an iterative consensus process with the Charité curriculum development group. Furthermore, special attention was paid to harmonising structure, language and wording in the EPA descriptions.
This article reports in detail on the process of defining a full set of core EPAs for entry into residency. Our process description may provide support and guidance to other medical schools for the development and implementation of EPAs for their own programs according to their specific contexts.
Content validity indices
Entrustable Professional Activity
Knowledge, skills and attitudes
ten Cate O, Chen HC, Hoff RG, Peters H, Bok H, van der Schaaf M. Curriculum development for the workplace using Entrustable professional activities (EPAs): AMEE guide no. 99. Med Teach. 2015;37(11):983–1002.
Englander R, Flynn T, Call S, Carraccio C, Cleary L, Fulton TB, et al. Toward defining the foundation of the MD degree: Core Entrustable professional activities for entering residency. Acad Med. 2016;91(10):1352–8.
The Association of Faculties of Medicine of Canada. AFMC Entrustable professional activities for the transition from medical school to residency: The Association of Faculties of Medicine of Canada; 2016. https://afmc.ca/medical-education/entrustable-professional-activities-epas. Accessed 08 May 2017.
Michaud PA, Jucker-Kupper P. The profiles working G. The “profiles” document: a modern revision of the objectives of undergraduate medical studies in Switzerland. Swiss Med Wkly. 2016;146:w14270.
ten Cate O, Graafmans L, Posthumus I, Welink L, van Dijk M. The EPA-based Utrecht undergraduate clinical curriculum: Development and implementation. Med Teach. 2018;1-8.
Humphrey-Murto S, Varpio L, Gonsalves C, Wood TJ. Using consensus group methods such as Delphi and nominal group in medical education research. Med Teach. 2017;39(1):14–9.
Chen HC, McNamara M, Teherani A, Cate OT, O'Sullivan P. Developing Entrustable Professional Activities for Entry Into Clerkship. Acad Med. 2016;91(2):247–55.
ten Cate O. Nuts and bolts of entrustable professional activities. J Grad Med Educ. 2013;5(1):157–8.
ten Cate O, Scheele F. Competency-based postgraduate training: can we bridge the gap between theory and clinical practice? Acad Med. 2007;82(6):542–7.
ten Cate O, Snell L, Carraccio C. Medical competence: the interplay between individual ability and the health care environment. Med Teach. 2010;32(8):669–75.
ten Cate O. Competency-based education, entrustable professional activities, and the power of language. J Grad Med Educ. 2013;5(1):6–7.
Mulder H, Ten Cate O, Daalder R, Berkvens J. Building a competency-based workplace curriculum around entrustable professional activities: the case of physician assistant training. Med Teach. 2010;32(10):e453–9.
ten Cate O, Young JQ. The patient handover as an entrustable professional activity: adding meaning in teaching and practice. BMJ Qual Saf. 2012;21(Suppl 1):i9–12.
Shaughnessy AF, Sparks J, Cohen-Osher M, Goodell KH, Sawin GL, Gravel J Jr. Entrustable professional activities in family medicine. J Grad Med Educ. 2013;5(1):112–8.
Hauer KE, Kohlwes J, Cornett P, Hollander H, Ten Cate O, Ranji SR, et al. Identifying entrustable professional activities in internal medicine training. J Grad Med Educ. 2013;5(1):54–9.
Hauer KE, Soni K, Cornett P, Kohlwes J, Hollander H, Ranji SR, et al. Developing entrustable professional activities as the basis for assessment of competence in an internal medicine residency: a feasibility study. J Gen Intern Med. 2013;28(8):1110–4.
Peters H, Holzhausen Y, Boscardin C, Ten Cate O, Chen HC. Twelve tips for the implementation of EPAs for assessment and entrustment decisions. Med Teach. 2017:1–6. https://doi.org/10.1080/0142159X.2017.1331031.
Lynn MR. Determination and quantification of content validity. Nurs Res. 1986;35(6):382–5.
The study was funded as part of the initiative “Bologna – Zukunft der Lehre” by the foundation “Stiftung Mercator and VolkswagenStiftung”; the “European Union’s Seventh Framework Programme for research, technological development and demonstration”, under grant agreement 619349 (WATCHME Project) and the initiative “Modellstudiengang Medizin 2.0” (01PL16036) by the foundation “Bundesministerium für Bildung und Forschung”. We acknowledge support from the German Research Foundation (DFG) and the Open Access Publication Fund of Charité – Universitätsmedizin Berlin.
Availability of data and materials
The materials used in this study are available from the corresponding author on request where warranted.
Ethics approval and consent to participate
The data acquisition procedure was approved by the local data protection authorities and the local ethics board (No EA2/091/14, Ethics Board Charité, Campus Mitte).
Consent for publication
All participants in the Delphi Study gave their informed consent at the beginning of the study.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Holzhausen, Y., Maaz, A., Renz, A. et al. How to define core entrustable professional activities for entry into residency?. BMC Med Educ 18, 87 (2018). https://doi.org/10.1186/s12909-018-1159-5
- Entrustable Professional Activities
- Curriculum development
- Undergraduate medical education
- Consensus methods