The role of feedback in improving the effectiveness of workplace based assessments: a systematic review

  • Habiba Saedon1Email author,

    Affiliated with

    • Shizalia Salleh2,

      Affiliated with

      • Arun Balakrishnan3,

        Affiliated with

        • Christopher HE Imray4 and

          Affiliated with

          • Mahmud Saedon5

            Affiliated with

            BMC Medical Education201212:25

            DOI: 10.1186/1472-6920-12-25

            Received: 20 November 2011

            Accepted: 2 May 2012

            Published: 2 May 2012



            With recent emphasis placed on workplace based assessment (WBA) as a method of formative performance assessment, there is limited evidence in the current literature regarding the role of feedback in improving the effectiveness of WBA. The aim of this systematic review was to elucidate the impact of feedback on the effectiveness of WBA in postgraduate medical training.


            Searches were conducted using the following bibliographic databases to identify original published studies related to WBA and the role of feedback: Medline (1950-December 2010), Embase (1980-December 2010) and Journals@Ovid (English language only, 1996-December 2010). Studies which attempted to evaluate the role of feedback in WBA involving postgraduate doctors were included.


            15 identified studies met the inclusion criteria and minimum quality threshold. They were heterogeneous in methodological design. 7 studies focused on multi source feedback, 3 studies were based on mini-clinical evaluation exercise, 2 looked at procedural based assessment, one study looked at workplace based assessments in general and 2 studies looked at a combination of 3 to 6 workplace based assessments. 7 studies originated from the United Kingdom. Others were from Canada, the United States and New Zealand. Study populations were doctors in various grades of training from a wide range of specialties including general practice, general medicine, general surgery, dermatology, paediatrics and anaesthetics. All studies were prospective in design, and non-comparative descriptive or observational studies using a variety of methods including questionnaires, one to one interviews and focus groups.


            The evidence base contains few high quality conclusive studies and more studies are required to provide further evidence for the effect of feedback from workplace based assessment on subsequent performance. There is, however, good evidence that if well implemented, feedback from workplace based assessments, particularly multisource feedback, leads to a perceived positive effect on practice.


            Feedback in clinical education has been defined as “specific information about the comparison between a trainee’s observed performance and a standard, given with the intent to improve the trainee’s performance” [1]. It has been suggested that the provision of feedback from formative assessments leads to a positive impact on doctors’ learning and performance [2].

            Recent reforms in postgraduate medical education have brought about a greater emphasis on competency based training which focuses on outcomes rather than processes of learning. Workplace based assessment (WBA) is a system whereby doctors are assessed on clinical skills and other attributes in the context of his or her working environment. Various methods are used to provide this information including mini-clinical evaluation exercise (mini-CEX), case-based discussion (CBD), direct observation of procedural skills (DOPS), procedure-based assessment (PBA), objective structured assessment of technical skills (OSATS) and multi-source feedback (MSF). Feedback and scoring are given by the assessor and this information is compiled and fed back to educational supervisors.

            Although there is considerable emphasis placed on WBA as a method of formative performance assessment, there is limited evidence in the current literature regarding the effectiveness of WBA in changing the behaviour of doctors and improving their performance. A recent literature review set out to explore the impact of WBA on doctors' education and performance [3]. The authors found that multisource feedback can lead to performance improvement although other factors have a major impact upon the response. There is a dearth of evidence about the outcome and use of feedback for continued learning and improvement. Anecdotally, trainees perceive feedback as the most useful aspect of WBA and believe that greater emphasis on the feedback component of WBA will improve its effectiveness as a formative assessment tool, hence improving trainees’ performance. The aim of this systematic review was to elucidate the impact of feedback on the effectiveness of WBAs in postgraduate medical training.


            Search strategy

            Searches were conducted using the following bibliographic databases to identify original published studies related to WBA and the role of feedback: Medline (1950-December 2010), Embase (1980-December 2010) and Journals@Ovid (English language only, 1996-December 2010). The search terms used were “feedback”, “workplace based assessment”, “direct observation of procedural skills”, “mini clinical evaluation exercise”, “case based discussion”, “multisource feedback”, “procedure-based assessment,” “objective structured assessment of technical skills”, “training” and “medical education”. In addition, hand searches using reference lists and bibliographies of included studies and review articles were performed.

            Inclusion and exclusion criteria

            Studies which attempted to evaluate the role of feedback in WBA (multi-source feedback, direct observation of procedural skills, mini-clinical evaluation exercise, procedure based assessment, objective structured assessment of technical skills or case-based discussion) involving postgraduate doctors were included. Both quantitative and qualitative studies were included. Non-English literature, case reports, review articles, ‘grey literature’ (non-refereed journals and conference proceedings), commentaries, letters, editorials and studies which only included undergraduate medical students and other health care professionals were excluded. The methodological quality of the selected studies was assessed according criteria developed by Colthart et al (Table 1) [4]. Only studies where conclusions were supported by the evidence presented (grades 3 and above) were considered. All studied were initially reviewed by two reviewers (HS, MS) and summaries of the findings were derived after discussion among other reviewers (SS, AB and CHEI).
            Table 1

            Gradings of Strength of Findings of the Paper [[4]]


            Strength of Findings


            No clear conclusions can be drawn.


            Results ambiguous, but there appears to be a trend.


            Conclusions can probably be based on the results.


            Results are clear and very likely to be true.


            Results are unequivocal

            Data extraction

            Two members of the review team (HS, MS) independently assessed the titles and abstracts of all identified citations. Reviewers were not masked to journal or author name because previous work has shown that this does not make a significant difference to the results of data extraction [5]. Data were extracted using a proforma including details of the research question, number of subjects, study design, setting, findings and limitations. Decisions of the two reviewers were recorded and then compared. Any disagreements were resolved by consensus with close attention to the inclusion/exclusion criteria. Other members of the review team evaluated the full text of the remaining potentially eligible papers and made a decision whether to definitely include or exclude each study according to the inclusion and exclusion criteria specified above. Data were synthesized using Kirkpatrick's four levels of educational outcomes and strength of findings (Table 2). Findings were grouped by type of intervention and described according to levels of outcome.
            Table 2

            The Kirkpatrick (1967) model of education outcomes [[6]]




            Learners’ reactions


            Learning of skills and knowledge


            Changes in learner behaviour


            Wider changes in the delivery of care

            Evidence synthesis

            A statistical synthesis of the evidence was not conducted because no randomised trials involving feedback in formative assessments were identified and the prospective and retrospective studies included a variety of methods of assessment.


            Search results

            Our initial search using the Ovid database yielded 3486 citations. Of these, 75 were found to be relevant articles. Following further screening of the title and abstract 41 were excluded. The remaining 34 articles were scrutinised and 12 studies fulfilled the inclusion criteria. Further searches of the Medline and Embase databases led to an additional 3 studies being identified and included. Manual searching of reference lists did not identify any additional articles resulting in a total of 15 studies, summarized in Table 3.
            Table 3

            Summary of studies included in the review


            Study design

            Data collection methods



            of origin


            Aim of study (implied/stated)

            Type of WBA

            Grade of Strength of findings and main findings

            Archer et al [7]


            Analysis of MSF data


            United Kingdom


            To report the evidence for and challenges to the validity of Sheffield Peer Review Assessment Tool (SPRAT) with paediatric specialist trainees across the UK as part of Royal College of Paediatrics and Child Health workplace based assessment programme.

            SPRAT (MSF)

            Grade 3. Assessor seniority is important. Free text boxes allow feedback for personal development

            Bullock et al [8]


            Analysis of MSF data


            United Kingdom

            Junior doctors

            To address differences in staff groups in their assessment of junior doctors’ professional attitudes and behaviour.

            TAB (MSF)

            Grade 3. Peers and administrators were less likely to indicate concern compared to consultants and senior nurses.

            Burford et al [9]




            United Kingdom

            Junior doctors and trainers

            To compare perceptions of two tools for giving MSF to UK junior doctors, based on usability, usefulness and validity.


            Grade 3. Trainees were asked in detail whether they would change their behaviour. Attitudes towards MSF in principle were positive and tools felt to be usable. Text-oriented tool rated more useful for giving feedback on communication and attitude

            Canavan et al [10]




            United States

            Five medical and one surgical specialty

            To assess qualitatively written comments on multisource assessments based on psychological feedback theory for professional development


            Grade 3. Quality of written feedback varies; a substantial portion of comments were useless and at worst detrimental to progress

            Violato et al [11]

            Longitudinal comparative study

            Forms analysed on two occasions, 5 years apart



            General Practice

            Examining the validity and reliability of MSF for general practice and whether it has led to change in performance when reassessed in 5 years


            Grade 4. There is evidence for the construct validity of the instruments and stability over time

            Sargeant et al [12]

            Qualitative study




            General Practice

            To increase understanding of the consequential validity of MSF by exploring how doctors used their feedback and the conditions influencing this use.


            Grade 3.Feedback usefulness enhanced by increasing its specificity. Strong influence of direct patient feedback on doctors’ performance

            Sargeant et al [13]

            Observational study

            Focus group



            General Practice

            Exploration of physicians’ reactions to MSF, perceptions influencing these and the acceptance and use of feedback


            Grade 3. Physicians’ perceptions of the MSF process and feedback can influence how and if they use the feedback for practice improvement.

            Wellor et al [14]

            Observational study

            Questionnaire based ratings and written answers


            New Zealand


            To evaluate mini-CEX for both summative and formative assessment for anaesthetics training

            Mini CEX

            Grade 3. Factors that facilitated or hindered implementation or limited effective feedback were identified

            Wellor et al [15]


            Analysis of mini-CEX forms


            New Zealand


            Psychometric characteristics, logistics of application, and impact on the quality of supervision of the mini- CEX

            Mini CEX

            Grade 3. The positive effect of the mini CEX on feedback, its relative feasibility, and acceptance as a potential assessment tool was demonstrated

            Holmboe et al [16]

            Observational study

            Videotaping of feedback sessions


            United States

            Primary care and internal medicine

            To examine how often faculty provided recommendations and used interactive techniques when providing feedback as part of a mini CEX.

            Mini CEX

            Programs should consider both specific training in feedback and changes to the miniCEX form to facilitate interactive feedback.

            James et al [17]

            Observational study

            Times taken to complete the consenting and operative components of the forms were recorded.


            United Kingdom


            Assessing the time required to complete PBA forms and ease of use in the surgical workplace.


            Grade 3. PBAs are feasible in clinical practice and are valued by trainees as a means of enabling focused feedback and targeted training.

            Marriot et al [18]

            Prospective observational study

            Direct observation using the PBA.


            United Kingdom


            The aims were to evaluate the validity, reliability and acceptability of PBA.


            Grade 3. PBA demonstrated good overall validity and acceptability, and exceptionally high reliability.

            Murphy et al [19]

            Prospective study



            United Kingdom

            General Practice

            To investigate the reliability and feasibility of six potential workplace-based assessment methods

            MSF, criterion audit, patient feedback, referral letters, significant event analysis, and video analysis of consultations.

            Grade 3. Two WBA tools involving patient and colleague feedback have high reliability suitable for high stakes WBA in the general practice setting.

            Cohen et al [20]




            United Kingdom


            To collate the experience and views on three workplace assessments

            DOPS, mini- CEX, MSF

            Grade 3.Trainees appreciate the formative benefits which derive from the assessments, namely feedback and reassurance of

            satisfactory performance.

            Johnson et al [21]

            Observational study

            Questionnaires and focus groups


            United Kingdom


            To gain feedback from trainees and supervisors in relation to components of core medical training including workplace- based assessments,

            All WBA

            Grade 4.WPBA assessments were well received as means of evidencing achievement and for learning development The majority of trainees felt that in particular the feedback following WBA assessments had been useful.

            General findings

            The 15 identified studies which met the inclusion criteria and minimum quality threshold were heterogeneous in their methodological design. A narrative overview is therefore provided rather than a meta-analysis. A wide range of WBAs were covered in the included studies. 7 studies focused on MSF, 3 studies were based on mini-CEX, 2 looked at PBA, one study looked at WBAs in general and 2 studies looked at a combination of 3 to 6 WBAs. 7 studies originated from the United Kingdom. Others were from Canada, the United States and New Zealand. Study populations were doctors in various grades of training from a wide range of specialties including general practice, general medicine, general surgery, dermatology, paediatrics and anaesthetics. All studies were prospective in design, and non-comparative descriptive or observational studies using a variety of methods including questionnaires, one to one interviews and focus groups. They all showed a modification of skills and attitudes or behavioural or willingness of learners to apply new knowledge & skills (Kirkpatrick Levels 2 and 3) [6]. None of the studies showed an improvement in learning and performance as a direct result of WBA (Kirkpatrick Level 4).

            Specific findings

            Multisource feedback (MSF)

            MSF is believed to increase motivation among staff, translating into positive behaviour change, increased productivity and self awareness which are fundamental for the progress of any organisation [22]. A non-comparative action based study by Archer et al found that MSF in the form of the Sheffield Peer Review Assessment Tool (SPRAT) does not provide enough data on trainees about whom concerns are raised, and more assessments are required for these trainees [7]. They also felt that unregulated self-selection of assessors introduces leniency bias and that this should end. Although free-text boxes allowed comments for feedback, no clear evidence was presented to show a change in practice. In an analysis of MSF data, Bullock et al demonstrated a trend towards becoming more critical in assessing trainees as seniority increases [8]. Feedback was provided by a designated trainer after completed forms were returned unseen to a central point and they stated that remedial action is undertaken as appropriate.

            A postal questionnaire to trainees and trainers showed that the perceived effectiveness of multisource feedback was low [9]. There were small but significant preferences for textual feedback, shown by the team assessment of behaviour (TAB), which has large free-text boxes, being perceived as more useful than the mini-PAT, which has a numerical scale and only a small space for comments. Elements which were more likely to be changed as a result of feedback were medical knowledge and teaching and training skills. The aspect which was least likely to change was relationships with patients. TAB was felt to be more useful on items related to communication and professionalism. The expected influence of the feedback was low, with nearly a third of trainees not anticipating to change in response to feedback. The relationship between intention to change in any area and the perceived positivity or negativity of feedback was also extremely low. Assessors based their feedback on both direct and indirect observation, in conjunction with discussion with colleagues and comments from patients and other health care professionals.

            Canavan et al analysed phrases in feedback comments written by observers who completed surveys to provide developmental feedback to residents and fellows [10]. They looked at the valence of feedback (positive, negative, or neutral), its level of specificity, and whether it was behaviour based or directed toward the learner’s “self”. 74.5% of surveys contained at least one global judgement. Behaviour-oriented phrases occurred less frequently, and general behaviours were mentioned more often than specific behaviours. Negative feedback phrases were found in 10.3% surveys. Similar to the positive comments, many were self-oriented, which can lead to a decline in performance [23]. The desirable characteristics of feedback were found to be specificity, behavioural focus, and sufficient clarity to be of great potential value to trainees.

            A longitudinal study investigated changes in performance for 250 doctors who participated in MSF twice, 5 years apart [11]. All the ratings increased between times 1 and 2, although the increase for patient ratings was not significant. The change in ratings by co-workers and medical colleagues were in the small-to-moderate range. The reasons for relatively little change occurring between the two time-points include the scores being high initially or that the data were not sufficiently compelling. Also, when only a few aspects of behaviour are advised to change in a survey containing more than 100 items, its effect will not be great.

            A qualitative study by Sargeant et al found that doctors did not make changes if feedback from MSF was positive, and only 7 out of thirteen doctors who received negative feedback changed their behaviour [12]. The feedback most consistently used was specific, received from patients, and addressed communication skills. The feedback least frequently used addressed clinical competence and came from medical colleagues. Another qualitative study by Sargeant et al using focus group interviews found that family physicians generally agreed with their patients’ feedback [13]. However, responses to medical colleague and co-worker feedback ranged from positive to negative, and did not always result in a change in behaviour.

            Mini-clinical evaluation exercise (Mini-CEX)

            Studies on the mini-CEX in trainee anaesthetists in New Zealand showed a positive effect of feedback and a perceived very positive educational impact [14, 15]. In the written feedback fields of the Mini-CEX form, 95% of specialists wrote comments under ‘things that the trainee did well’, 70% recorded comments in ‘areas for improvement’, and 60% wrote down an ‘agreed action’ [15]. Trainees felt there was not a strong culture of feedback, but that the mini-CEX facilitated feedback. Holmboe et al recorded feedback from mini-CEX sessions in a prospective observational cohort study and showed that mini-CEX frequently leads to a recommendation for improvement, with the majority of the recommendations focused on the clinical skills of medical interviewing, physical examination, and counselling [16].

            Procedure based assessment (PBA)

            James et al looked at the PBA tool in a non-comparative observational study and found that completion of the PBAs resulted in focused feedback to the trainees about their practice [17]. As a result, the trainees in this study valued this structured approach because it enabled subsequent training to be targeted appropriately. Marriot et al also studied PBA and showed that trainees reported the feedback provided by the clinical supervisor as moderately useful to very useful. Clinical supervisors rated feedback similarly [18].

            Other assessments

            Murphy et al investigated 6 different instruments (criterion audit, multisource feedback, patient satisfaction ratings, assessment of referral letters, significant events analysis, and analysis of videotaped patient interactions) in General Practice registrars [19]. They highlighted the important role of feedback from patients and colleagues. A questionnaire survey of dermatology trainees collated the experience and views on MSF, DOPS and mini-CEX [20]. Trainees appreciated the formative aspects of the assessments, especially feedback, although not all trainees reported receiving useful feedback. Johnson et al’s questionnaire and focus group study of core medical trainees on their views of the curriculum and assessment found that the majority of them felt that in particular the feedback component of WBA assessments had been useful [21].


            This systematic review aimed to evaluate the effectiveness of feedback in WBAs. The studies were all observational and there were no randomised controlled trials. The majority of the studies were seeking perceptions and self-reported changes rather than measuring actual change in practice. This is because measuring changes in practice and attributing them to feedback from the WBA is extremely difficult due to confounding factors and problems with study design. Most of the evidence to support the use of feedback from WBAs comes from studies on MSF. This may be because, whereas in other assessments the emphasis may be upon performing a procedure correctly or the management of a particular patient, MSF has the sole purpose of providing feedback of doctors’ practice and behaviours. This opportunity is often missed, as found in the study by Canavan et al which analysed comments made on MSF forms [10]. Many forms contained no comments at all and, of those that did contain comments, a significant proportion were found to lack actionable information, thus limiting their usefulness. Global judgments were more frequently used and although these may build the confidence of the person being assessed, they do not give an indication of how they should behave in order to improve their practice and future actions. Most of the trainees in the study by Burford et al did not anticipate changing their behaviour as a result of feedback from the MSF tools used, but the perceived usefulness was consistently higher with the TAB compared to the mini-PAT [9]. The greater space for free text in the former tool allows valuable information to be transmitted back to the trainee which they can use to inform a change in practice, rather than simply a numerical score.

            MSF has the potential to be a useful tool but the current evidence suggests that in order for this to occur, the way in which it is used must be improved. Comments should be provided and these should be specific and action-based. Reasons why it is currently under-utilised include time constraints of an already busy clinical workload, regarding WBA as cumbersome, a lack of training on how to provide feedback and a lack of trust in the formative nature of the assessment, as learners may feel that the feedback may have a negative impact on their training [10].

            Other WBAs methods such as the mini-CEX, and DOPS did not show any clear evidence of leading to a change in behaviour. The use of the mini-CEX was strongly advocated to improve feedback, but pointed out that feedback is offered less frequently than is desirable [14]. Cohen et al found that half of the dermatology trainees surveyed reported that learning points had been identified from the mini-CEX, and that feedback and learning were identified most frequently as positive aspects of the process [20]. This implies that feedback is valued and a change in behaviour may occur, but does not show this. A fifth of respondents on the mini-CEX expressed reservations about the quality of feedback; for DOPS, 14% reported that insufficient time was allowed for feedback and only 45% identified learning points arising from the process. There were no studies looking at case based discussion so the effect of this assessment on doctors’ performance is undeterminable. Further research in this area is therefore warranted.

            The highest Kirkpatrick level reached by any of the studies was level 3 which indicates a change in behaviour and documents the transfer of learning to the workplace or willingness of learners to apply new knowledge and skills. Others were level 2, showing changes in the attitudes or perceptions among participant groups towards teaching and learning.

            Feedback may not produce intended outcomes and may even have detrimental consequences, such as decreased motivation and reduced performance. In one study feedback perceived as being strongly negative generally evoked emotional responses, including anger and discouragement [13]. Trainers reportedly often avoid giving feedback, in order to prevent offence or provoking defensiveness [24, 25]. Several studies suggested that maximizing opportunities for training of assessors in giving optimal feedback and administering assessments would improve the quality of feedback. If WBAs are simply used as a box-ticking exercise, without sufficient emphasis on feedback, then any gains will be limited [26].


            This systematic review had some limitations. The studies were uncontrolled thereby limiting the strength of findings but this may be due to the difficulties in assessing the effect of feedback on future performance of doctors. Limitations in our methodology include the grey literature not being reviewed and only including studies in the English language which may have led to bias. Another limitation of the study is the focus on feedback which is only one potentially beneficial aspect of WBA. Others can include on the job training whilst being observed by a senior and documentation of competence in a particular area. [27]


            The relationship between feedback and outcome is not always straightforward and may not always achieve the desired results [28]. Good feedback can lead to increased motivation and confidence in trainees. On the other hand, negative feedback is not aimed to demotivate or demoralise a trainee, but should be taken as constructive criticism for trainees to improve. More studies are required to provide further evidence for the effect of feedback from WBAs on subsequent performance, as the evidence base contains few high quality conclusive studies. Although a difficult area to research, more randomised controlled studies on a change in behaviour following feedback from specific WBAs should be encouraged. There is, however, good evidence that if well implemented, feedback from WBAs, particularly MSF, leads to a perceived positive effect on practice.


            Authors’ Affiliations

            Department of Ophthalmology, Heart of England NHS Foundation Trust
            Department of Otolaryngology, University Hospitals Coventry and Warwickshire
            Department of Vascular Surgery, Freeman Hospital
            Department of Surgery, University Hospitals Coventry and Warwickshire and Warwick Medical School, University of Warwick
            Warwick Medical School, University of Warwick


            1. Van De Ridder JM, Stokking KM, Mcgaghie WC, Ten Cate OT: What is feedback in clinical education? Med Educ 2008, 42:189–197.View Article
            2. Norcini J, Burch V: Workplace-based assessment as an educational tool: AMEE Guide No 31. Med Teach 2007, 29:855–871.View Article
            3. Miller A, Archer J: Impact of workplace based assessment on doctors' education and performance: a systematic review. BMJ 2010, 341:c5064.View Article
            4. Berlin JA, University of Pennsylvania Meta-analysis Blinding Study Group: Does blinding of readers affect the results of meta-analyses? Lancet 1997,350(9072):185–186.View Article
            5. Colthart I, Bagnall G, Evans A, Allbutt H, Haig A, Illing J, McKinstry B: The effectiveness of self-assessment on the identification of learner needs, learner activity, and impact on clinical practice. BEME Guide No 10. Med Teach 2008,30(2):124–145.View Article
            6. Kirkpatrick D: Evaluation of Training. In Training and Development Handbook. Edited by: Craig R, Bittel L. McGraw-Hill, New York; 1967:131–167.
            7. Archer J, McGraw M, Davies H: Republished paper: Assuring validity of multisource feedback in a national programme. Postgrad Med J 2010,86(1019):526–31.View Article
            8. Bullock AD, Hassell A, Markham WA, Wall DW, Whitehouse AB: How ratings vary by staff group in multi-source feedback assessment of junior doctors. Med Educ 2009,43(6):516–520.View Article
            9. Burford B, Illing J, Kergon C, Morrow G, Livingston M: User perceptions of multi-source feedback tools for junior doctors. Med Educ. 2010,44(2):165–76. Epub 2010 Jan 5View Article
            10. Canavan C, Holtman MC, Richmond M, Katsufrakis PJ: The quality of written comments on professional behaviors in a developmental multisource feedback program. Acad Med 2010,85(10 Suppl):S106-S109.View Article
            11. Violato C, Lockyer JM, Fidler H: Changes in performance: a 5-year longitudinal study of participants in a multi-source feedback programme. Med Educ 2008,42(10):1007–1013.View Article
            12. Sargeant J, Mann K, Sinclair D, van der Vleuten C, Metsemakers J: Challenges in multisource feedback: intended and unintended outcomes. Med Educ 2007, 41:583–591.View Article
            13. Sargeant J, Mann K, Ferrier S: Exploring family physicians' reactions to multisource feedback: perceptions of credibility and usefulness. Med Educ 2005,39(5):497–504.View Article
            14. Weller JM, Jones A, Merry AF, Jolly B, Saunders D: Investigation of trainee and specialist reactions to the mini-Clinical Evaluation Exercise in anaesthesia: implications for implementation. Br J Anaesth 2009,103(4):524–30.View Article
            15. Weller JM, Jolly B, Misur MP, Merry AF, Jones A, Crossley JG, Pedersen K, Smith K: Mini-clinical evaluation exercise in anaesthesia training. Br J Anaesth 2009,102(5):633–641.View Article
            16. Holmboe ES, Yepes M, Williams F, Huot SJ: Feedback and the mini clinical evaluation exercise. J Gen Intern Med 2004,19(5 Pt 2):558–561.View Article
            17. James K, Cross K, Lucarotti ME, Fowler AL, Cook TA: Undertaking procedure-based assessment is feasible in clinical practice. Ann R Coll Surg Engl 2009,91(2):110–112.View Article
            18. Marriott J, Purdie H, Crossley J, Beard JD: Evaluation of procedure-based assessment for assessing trainees' skills in the operating theatre. Br J Surg 2010,98(3):450–457.View Article
            19. Murphy DJ, Bruce DA, Mercer SW, Eva KW: The reliability of workplace-based assessment in postgraduate medical education and training: a national evaluation in general practice in the United Kingdom. Adv Health Sci Educ Theory Pract 2009,14(2):219–232.View Article
            20. Cohen SN, Farrant PB, Taibjee SM: Assessing the assessments: U.K. dermatology trainees' views of the workplace assessment tools. Br J Dermatol 2009,16(1):34–39.View Article
            21. Johnson G, Barrett J, Jones M, Parry D, Wade W: Feedback from educational supervisors and trainees on the implementation of curricula and the assessment system for core medical training. Clin Med 2008,8(5):484–489.
            22. Abdulla A: A critical analysis of mini peer assessment tool (mini-PAT). J R Soc Med 2008, 101:22–26.View Article
            23. Hattie J, Timperley H: The power of feedback. Rev Educ Res. 2007, 77:81–112.View Article
            24. Hewson MG, Little ML: Giving feedback in medical education: verification of recommended techniques. J Gen Intern Med 1998, 13:111–116.View Article
            25. Ende J: Feedback in clinical medical education. JAMA 1983, 250:777–781.View Article
            26. Saedon H, Saedon MH, Aggarwal SP: Workplace-based assessment as an educational tool: Guide supplement 31.3--viewpoint. Med Teach 2010,32(9):e369-e372.View Article
            27. Papettas T, Saedon H, Saedon M: Opportunities for learning in the surgical workplace and how they can be exploited: a practical guide. Br J Hosp Med (Lond) 2011,72(12):707–710.
            28. McKinley RK, Williams V, Stephenson C: Improving the content of feedback. Clin Teach 2010, 7:161–166.View Article
            29. Pre-publication history

              1. The pre-publication history for this paper can be accessed here:http://​www.​biomedcentral.​com/​1472-6920/​12/​25/​prepub


            © Saedon et al; licensee BioMed Central Ltd. 2012

            This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://​creativecommons.​org/​licenses/​by/​2.​0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.