- Research article
- Open Access
Peer assessment to improve medical student’s contributions to team-based projects: randomised controlled trial and qualitative follow-up
BMC Medical Education volume 19, Article number: 371 (2019)
Medical schools increasingly incorporate teamwork in their curricula but medical students often have a negative perception of team projects, in particular when there is unequal participation. The purpose of this study is to evaluate whether a novel peer evaluation system improves teamwork contributions and reduces the risk of students “free loading”.
A cluster randomised controlled trial (RCT) with qualitative follow up enrolled 37 teams (n = 223 students). Participating teams were randomised to intervention group (19 teams) or control group (18 teams). The validated Comprehensive Assessment Team Member Effectiveness (CATME) tool was used as the outcome measure, and was completed at baseline (week 2) and at the end of the project (week 10). The team contribution subscale was the primary outcome, with other subscales as secondary outcomes. Six focus group discussions were held with students to capture the team’s experiences and perceptions of peer assessment and its effects on team work.
The results of the RCT showed that there was no difference in team contribution, and other forms of team effectiveness, between intervention and control teams. The focus group discussions highlighted students’ negative attitudes, and lack of implementation of this transparent, points-based peer assessment system, out of fear of future consequences for relationships with peers. The need to assess peers in a transparent way to stimulate open discussion was perceived as threatening by participants. Teams suggested that other peer assessment systems could work such as rewarding additional or floating marks to high performing team members.
Other models of peer assessment need to be developed and tested that are non-threatening and that facilitate early acceptance of this mode of assessment.
Medical practitioners are expected to work effectively in teams and provide peer feedback to ensure quality of care in clinical practice [1,2,3]. Consequently, standards and outcomes for medical education and training worldwide require medical students to learn to work effectively in teams and develop reflective skills of self and peers throughout their careers [4,5,6]. It is therefore important to provide medical students with opportunities to develop and practice teamworking skills. Team learning for medical students is associated with various positive educational outcomes [7, 8]. Consequently, medical school curricula today incorporate active learning modalities to develop such skills. However, many students have a negative perception of team projects which often occur when one or more members of a team do not contribute equal amounts of the work, also known as the free-loader problem .
Peer assessment has been proposed as a possible solution to such problems, when teaching staff cannot directly observe each members’ contributions [10, 11]. Peer assessment, grounded in social constructivist theory , allows learners to consider and specify the level, volume or quality of work completed by other individuals of equal status in learning or professional terms [13,14,15,16,17]. It may be employed in a variety of educational settings and has been demonstrated to be widely used in the assessment of medical students. The use of peer assessment in medical education has been demonstrated to be advantageous from a number of viewpoints . It encourages students to self-appraise performance and that of others, stimulates educational activities and encourages active participation of students [13, 18, 19]. However, there are concerns cited over the approach to peer assessment and the tools used in such assessments [18, 20, 21].
Perhaps of greatest concern is the peer assessment as a social process and the difficulties this may create for participants . Whilst it has been demonstrated to have positive effects for students, interpersonal relationships and social acceptability should be considered as influential dimensions in the process. Bias may occur when friendships or social interactions influences the approach to peer assessment . This relational effect has been reported elsewhere in the literature on small group behaviour and interactionist theory.
The marking system used to assess peers should also be taken into consideration. In a systematic review about student peer assessment in medical education, 22 different tools used mainly in medical education settings were identified. There was great diversity reported between scoring systems and psychometric characteristics of the tools and no golden standard of peer assessment could be identified . None of the assessment tools identified in the trial resemble that employed in this research.
In this study, we developed and evaluated a novel peer assessment system that was designed to stimulate an open and fair distribution of the available marks for team members and to enable group members to directly address the problem of free-loading. We hypothesized that team contribution and functioning would be higher in teams that used this system in comparison to those without it. To our knowledge, this is the first RCT to assess the effect of a peer-assessment system on team effectiveness among undergraduate medical students. The mixed methods design also permitted us to obtain students’ assessments of peer assessment and intervention implementation.
We followed the Consolidated Standards of Reporting Trials (CONSORT) statement for RCT reporting .
Setting of the study
The Royal College of Surgeons in Ireland (RCSI) is characterised by cultural diversity with around 60% of students from the Far East and the Middle East, with less than 20% from Ireland [25, 26]. The study took place during the Population and International Health (PIH) team-based project in semester 3 in the second year of medical undergraduate programme and is compulsory for all students. Over the years several variations of peer assessment have been implemented based on annual student feedback surveys . The current team-based project included a system of peer assessment which requires students to openly discuss and assess team work contributions of peers and achieved by team consensus. Students needed to allocate a fixed overall number of marks which means that if one or more excelling team members were awarded with extra peer marks, this would go at the expense of other (underperforming) team members who would receive reduced marks. While the team-based PIH project was a compulsory part of the module, students were not obliged to take part in the study on this project, and a decision not to participate did not have an effect on their participation in RCSI or future grades. Those who did not want to participate in the study were required to use peer assessment, as this was the usual procedure for the team-based project. All students received project instructions, which require teams to conduct a review of literature on a public health problem and report on teamwork. Students also received training from a librarian on sources of data and searching literature. In addition, teams were asked to keep track of team activities through submitting three team reports during the duration of the project including minutes, team goals, team member roles, team rules and attendance.
Processes, intervention and control
All second-year medical students enrolled in the module in 2014/2015 were invited to participate in September 2014. Students (n = 351) were stratified by gender, nationality and English native speakers in to project teams to create as balanced teams as possible (this is a standard procedure in the team-based project). Randomization was performed with the use of computer-generated random numbers. Consenting students (n = 109) were randomly assigned to teams where students would jointly work on a project without the possibility of assessing each other’s performance. The remaining students (n = 114) were randomly assigned to teams where an additional component of the team-based project was a novel peer-assessment system.
Teams received detailed instructions to provide assessment to peers as to encourage active participation and equal contribution of students in teams. The peer assessment system had two important features: open feedback and balanced peer marking. First, students had to fill in a form in which they gave detailed feedback on their team members’ contributions to team activities in terms of preparation, contribution, respect for others’ ideas and flexibility. It was essential that teams kept good records of meetings called, of team members’ attendances for each meeting, and could demonstrate the contributions (and lack of contributions) of individual team members.
The peer marking section required students to mark each of the other members of the team. Each team had a fixed number of points to allocate (see Table 1). Students could assign a maximum of 6 marks to each team member. This meant that a total of 24 marks could be assigned in a 6-member team. Students were specifically asked to differentiate some of the ratings depending on how team members contributed to the team. This meant that if one member got a score of 5, another team member must be given a score of 3. The final score was to be agreed by the team. If awarding less than 4 marks to a team member it was essential that the team was able to demonstrate, using evidence, sufficient shortfall in the individual’s contribution to justify the reduced peer mark. Where an individual team member wished to challenge a team decision to award him or her with a low mark, s/he needed to be able to demonstrate, using evidence, that s/he merited a higher peer mark than the other team members had proposed.
The primary outcome was change in each team members’ contribution to the team, as assessed by the Comprehensive Assessment Team Member Effectiveness (CATME) Likert Short Version tool  which describes behaviors typical of various levels of performance in five categories (Contributing to the Team’s Work, Interacting with Teammates, Keeping the Team on Track, Expecting Quality, and Having Relevant Knowledge, Skills, and Abilities). Raters rate each teammate on each item using Likert scales (strongly dis- agree–strongly agree). All participating students were asked to complete the CATME at baseline (week 2) and at the end of the project (week 10). This validated questionnaire uses a behaviorally anchored rating scale to measure team-member contributions that are clustered into five broad categories . Secondary outcomes were other CATME subscales: team member interaction, teamwork progress; teamwork quality and having relevant knowledge, skills and abilities. Students allocated to the intervention arm (and those who chose to opt out of the trial) were supplied with detailed instructions on the procedure for peer assessment. Teams were required to submit feedback and allocation of peer marks along with the final project report.
With a median average of 6 students per team, we computed 58 teams conducting the compulsory team-based project. Assuming a response rate of at least 30%, (60 students and 10 groups in each arm), we powered the trial to having 10 teams in each arm. Using the clustersampi command in Stata 13.0, and estimating data from Ohland et al., we were powered for a detectable difference of 0.37, assuming the control group mean was 4.59 (SD 0.62), an average of 6 people per team and an intra-class correlation co-efficient rho = 0.05 [28, 29].
Potential demographic differences between intervention and control groups were assessed using t-tests or chi-square χ2 as appropriate. Linear regression predicted primary and secondary outcomes at follow-up, controlling for the relevant baseline CATME subscale score. Robust variance estimators were used to account for clustering within teams. Data was analysed as per protocol analysis.
Focus group discussions
We conducted focus group discussions with six teams to better understand how the peer assessment system contributed to teamwork and to assess the extent to which the intervention had been conducted according to plans and protocols (i.e. implementation). We stratified the teams according to their ultimate team score (high, intermediate and low) and interviewed three teams that participated in the intervention and three control teams. Focus groups were conducted over a 5-day period, and both the focus group facilitator (M-C K) and the participants were unaware of the CATME results. The focus group discussions were recorded on the free download software Audacity® on a portable recording device, and transcribed verbatim for analysis. The transcripts were exported to NVivo Version 10, read repeatedly to reach data immersion and then thematically analysed . Transcripts were firstly coded with themes emerging from the data following organisation and refinement of the codes.
A total of 37 teams (n = 223 students) participated in the study and 19 teams were randomised to the peer-assessment, with 18 control teams (see Fig. 1).
About one third of the students were from the Middle East and another one third from South-East Asia while less than 15% of the students were Irish. Age, gender and student region of origin was not associated with participation (all p-values>.209), indicating that randomisation was successful. Descriptive statistics are shown in Table 2.
Both in the intervention and control arm, the highest mean scores for the CATME subscales was’ teamwork quality’ while ‘having relevant knowledge, skills and attitudes’ was scored lowest at both baseline and follow up (see Table 3).
Table 4 provides the results of the linear regression, which predicts each outcome by group, controlling for the baseline scale score. There were only small and no statistically significant differences in the means of any of the primary or secondary outcomes between groups.
Three themes were identified in the course of the analysis: anxiety and poor implementation of the intervention, conflicting views whether peer assessment could improve team effectiveness, critical views of the balanced marking system and recommendations for the future (Table 5).
Overall, intervention teams did not implement peer assessment with the main reason being fears and concerns associated with the possibility of having to mark other team members negatively and the tension this may introduce. The need to assess peers in a transparent way to stimulate open discussion was perceived as threatening by participants. There were also disagreements among the teams on the effect of peer assessment on team effectiveness. For example, teams in the intervention group did not think that peer assessment reduced freeloading, alter their approach to the assignment or how they interacted with team members. In contrast, teams in the control arm were more likely to think that peer assessment could have helped them to improve team work effectiveness, especially when some team members disengaged when they had completed their part of the assignment.
Finally, teams also recommended alternative approaches of peer assessment, such as the implementation of additional or floating marks that were not sacrificed by colleagues. It was proposed that such a system would remove the tension and pressure that currently surrounds the peer marking assessment and also award positive contributions to team performance.
This is the first randomised trial to evaluate a peer assessment intervention during a team-based project in undergraduate medical students. Our data does not confirm the hypothesis that team contribution and functioning is higher in teams that used peer assessment in comparison to those without it, mainly because students did not implement peer marking. The focus group discussions suggested that students’ reactions to participate in peer assessment were negative and concerns were related to the open, transparent nature of the peer assessment tool and the manner in which the assessment was conducted.
Students were mainly worried about the impact that open, transparent peer assessment could have on their relationships with each other. The concept of reciprocity, whereby a fear of reprisal from another team member upon allocation of a lower mark to that individual, was described as one of the primary concerns associated with the current peer marking system . Some teams in the intervention arm reported that marks were fixed at the outset, regardless of individuals’ performance, in an effort to minimise conflict within the group. This is in line with other studies that found that a lack of anonymity in peer assessment can lead to disruption of relations between peers, and teams agreeing they would mark each other positively [31, 32]. Others have therefore suggested that student peer assessment is best conducted anonymously and with clearly defined standardized criteria [15, 21]. However, as mentioned before anonymity can lead to other forms of anxiety and gaming and the challenge of ‘friendly assessment’ is a refractory limitation of peer assessment methods and difficult to circumvent even with the amended assessment approach proposed by participants in this study [23, 33].
Secondly, students expressed concerns about the balanced peer assessment system, where peer marks are fixed and must be achieved by team consensus. This requirement increased students’ anxiety particularly when a fellow student was underperforming. Moreover, the negative approach of our peer assessment system did not allow teams to reward some outstanding members without unfairly having to punish other team member. This finding is in agreement with Levine et al.  who found that students in highly functioning teams felt that the points-based peer assessment system unfairly forced them to differentiate. However, in contrast to that study, we did not find that students in more dysfunctional teams felt empowered to score their peers higher or lower based on their performance. In fact, most teams participating in this study elected to assign the same amount of marks to everyone in their team. Finally, novel educational interventions need to bring demonstrable benefits, evaluated using both rigorous randomised trial methodologies and qualitative research, to understand what works, for whom, and when.
There are a number of limitations associated with the design and context of this study. Our undergraduate second-year medical students had no prior experience with peer assessment within the medical school curriculum. This could have resulted in more negative attitudes as has been shown by others . We found that all teams agreed to participate in the research with the hope of avoiding peer assessment. Hostility towards peer assessment is not unique, especially when first experienced . However, others have demonstrated that perceptions do improve once experience is gained in this method of assessment; and a more positive view may appear over time [26, 37].
Second, several teams did not engage in the peer assessment exercise as per the recommended protocol. As previously mentioned, this was partly related to the open, non-anonymous nature of the peer assessment. While students were given some guidance on how to give feedback at the start of the module, the available time-table and staff meant that this was limited in terms of practicing skills in constructive feedback and conflict resolution. Medical schools should introduce peer assessment early in the medical curriculum so that students have the opportunity to develop critical assessment skills and acceptance of being critically evaluated by peers ( . This should preferably be done in a phased approach starting with formative feedback which would be less threatening for students followed by the implementation of (summative) peer marking. As part of this longitudinal approach, students should receive skills training in giving and receiving constructive feedback. Third, given the cultural diversity of RCSI students, intercultural communication challenges could have lead to misinterpretation of students’ activities and reduced their willingness and ability to mark others. Future research should explore whether and how cultural factors play a role in team-based work and peer assessment. Finally, it should be noted that the ultimate benefits of improved skills for evaluating and modifying the performance of one’s peers may not be realised until medical students have graduated and are working in clinical teams, which was the proposition that led to the staff retaining a peer review assessment system over the previous years.
This study highlights the contribution of mixed-methods research to the development of evidence-based medical education. The findings of the RCT show that this model of peer assessment does not improve team effectiveness or reduce free loading. The qualitative follow-up suggests that likely reasons are the risk or fear of negative consequences for students’ future relationships with their peers, with whom they will be studying for a further 4 years. It is recommended for medical schools to implement less threatening forms of peer assessment and provide guidance and training possibilities for developing critical peer assessment skills early in the medical curriculum.
Availability of data and materials
The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.
Comprehensive Assessment Team Member Effectiveness
Consolidated Standards of Reporting Trials
Population and International Health
Royal College of Surgeons in Ireland
randomised controlled trial
Schyve P. The changing nature of professional competence. Jt Comm J Qual Patient Saf. 2005;31(4):185–202.
Weaver SJ, Dy SM, Rosen MA. Team-training in healthcare: a narrative synthesis of the literature. BMJ Qual Saf. 2014;23(5):359–72.
Leonard M, Graham S, Bonacum D. The human factor: the critical importance of effective teamwork and communication in providing safe care. Qual Saf Health Care. 2004;13(Suppl 1):i85–90.
CanMEDS, CanMEDS Framework. 2005, Royal College of Physicians and surgeons in Canada.
General Medical Council. Outcomes for graduates (Tomorrow’s doctors). Manchester: GMC; 2015.
Accreditation Council for Graduate Medical Education. Summative evaluation of competencies for 12-month clinical phase residents 2008 [30 January 2018]; Available from: http://www.acgme.org/acgmeweb/Portals/0/PFAssets/ProgramResources/380_SummativeEvaluation_GPM_AA_04_10_2008.pdf.
Searle NS, et al. Team learning in medical education: initial experiences at ten institutions. Acad Med. 2003;78(10):S55–8.
Thompson BM, et al. Team-based learning at ten medical schools: two years later. Med Educ. 2007;41(3):250–7.
Brooks CM, Ammons JL. Free riding in group projects and the effects of timing, frequency and specificity of criteria in peer assessments. J Educ Bus. 2003;78(5):268–72.
Dannefer EF, et al. Peer assessment of professional competence. Med Educ. 2005;39(7):713–22.
Epstein RM. Assessment in medical education. N Engl J Med. 2007;356:387–96.
Vygotsky LS. Collected Works of L. S. Vygotsky, Problems of General Psychology. New York: Plenum; 1987.
Topping KJ. Peer assessment. Theory Pract. 2009;48(1):20–7.
Loughry ML, Ohland MW, Moore DD. Development of a theory-based assessment of team member effectiveness. Educ Pyschol Measur. 2007;67:505–24.
Norcini JJ. Peer assessment of competence. Med Educ. 2003;37(6):539–43.
Kemper M, et al. Peer Teaching and Peer Assessment are Appropriate Tools in Medical Education in Otorhinolaryngology: Laryngorhinootologie; 2014.
Hulsman RL, Peters JF, Fabriek M. Peer-assessment of medical communication skills: the impact of students’ personality, academic and social reputation on behavioural assessment. Patient Educ Couns. 2013;92(3):346–54.
Speyer R, et al. Reliability and validity of student peer assessment in medical education: a systematic review. Medical Teacher. 2011;33(11):e572–85.
Freeman MD. Peer assessment by groups of group work. Assess Eval High Educ. 1995;20(3):289–300.
Eldredge JD, et al. Student peer assessment in evidence-based medicine (EBM) searching skills training: an experiment. J Med Libr Assoc. 2013;101(4):244–51.
Arnold L, et al. Can there be a single system for peer assessment of professionalism among medical students? A multi-institutional study. Acad Med. 2007;82(6):578–86.
van Gennip NAE, Segers MSR, Tillema HH. Peer assessment as a collaborative learning activity: the role of interpersonal variables and conceptions. Learn Instr. 2010;20(4):280–90.
Magin D. Reciprocity as a source of Bias in multiple peer assessment of group work. Stud High Educ. 2001;26(1):53–63.
Boutron I, et al. Extending the CONSORT statement to randomized trials of nonpharmacologic treatment: explanation and elaboration. Ann Intern Med. 2008;148(4):295–309.
Doyle F, et al. Understanding communication of health information: a lesson in health literacy for junior medical and physiotherapy students. J Health Psychol. 2012:1461–7277.
Byrne E, et al. Peer interviewing in medical education research: experiences and perceptions of student interviewers and interviewees. BMC Res Notes. 2015;8:513.
Doyle F, et al. A survey of student opinion of peer marking in undergraduate medical student team project in RCSI. Dublin: Irish Network of Medical Educators; 2008.
Ohland MW, et al. The comprehensive assessment of team member effectiveness: development of a behaviorally anchored rating scale for self and peer evaluation. Acad Manag Learn Edu. 2012;11:609–30.
Ten-year population education programme reaps good results. Popul Educ Asia Pac Newsl Forum. 1994;(39):12.
Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101.
Arnold L, et al. Use of peer evaluation in the assessment of medical students. J Med Educ. 1981;56(1):35–42.
Arnold L, et al. Medical students’ views on peer assessment of professionalism. J Gen Intern Med. 2005;20(9):819–24.
Williams E. Student attitudes towards approaches to learning and assessment. Assess Eval High Educ. 1992;17(1):45–58.
Levine RE, et al. Peer evaluation in a clinical clerkship: students’ attitudes, experiences, and correlations with traditional assessments. Acad Psychiatry. 2007;31(1):19–24.
Wen ML, Tsai C. University student’s perceptions of and attitudes toward (online) peer assessment. High Educ. 2006;51:27–44.
Sluijsmans D. Establishing learning effects with integrated peer assessment tasks; 2002.
Dochy F, Segers M, Sluijsmans D. The use of self-, peer and co-assessment in higher education: a review. Stud High Educ. 1999;24(3):331–50.
We would like to thank all participating medical students for contributing to this study.
Ethics approval and consent to participate
The study had approval from the RCSI Research Ethics Committee and participants consented to participate by signing a participant consent form.
Consent for publication
Participants gave permission for information collected to be stored or electronically processed for the purpose of scientific research which included conference and peer reviewed publication.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Biesma, R., Kennedy, M., Pawlikowska, T. et al. Peer assessment to improve medical student’s contributions to team-based projects: randomised controlled trial and qualitative follow-up. BMC Med Educ 19, 371 (2019). https://doi.org/10.1186/s12909-019-1783-8
- Randomised controlled trial
- Medical education
- Peer assessment