Skip to main content
  • Research article
  • Open access
  • Published:

Feedback using an ePortfolio for medicine long cases: quality not quantity

Abstract

Background

The evidence for the positive impact of an electronic Portfolio (ePortfolio) on feedback in medicine is mixed. An ePortfolio for medical long cases in a Graduate Medical Program was developed. The purpose of this study was to explore the perceptions of medical students and faculty of the impact of the ePortfolio on the feedback process.

Methods

In total, 130 Year 3 medical students, and six faculty participated in the study. This is a mixed methods study, using a combination of both quantitative and qualitative approaches. Quantitative methods were used to quantify the number of long cases performed. Qualitative methods were used to explore the relationship between quantity and quality of feedback, and provide a rich understanding of both students’ and faculty’s experience and perceptions of the ePortfolio.

Results

Students received a variable quantity of feedback at each of the three studied clinical schools, with an average of between 4 – 5.4 feedback episodes per student. Feedback that was constructive, specific and timely and delivered by a senior academic was important. Quantity was not an essential factor, with two episodes of detailed feedback reported to be adequate. The barriers to the use of the ePortfolio were technical aspects of the platform that interfered with student engagement.

Conclusions

Feedback using the ePortfolio for medical long cases is a valuable tool providing a senior clinician delivers detailed, constructive and personalized feedback in a timely fashion. The ePortfolio system needs to be user-friendly to engage students.

Peer Review reports

Background

Portfolios are one of many tools used in the assessment of students and provide a potentially useful means of providing timely and detailed feedback. They have been increasingly used in medical education and are emerging as a tool for documenting learning and competencies [1]. In Australian medical schools and in the post-graduate training programs, portfolios are not yet compulsory, although a recent Review of Medical Intern Training has recommended their introduction which is likely to impact on medical schools [2]. The literature, however, suggests mixed success in the use of portfolios in health professional education and feedback from mentors appears to be a crucial contributor to success or failure [1, 3, 4].

Development of an ePortfolio

In 2015, limited use of an ePortfolio was introduced into the Sydney Medical Program (SMP), a four-year graduate degree at The University of Sydney. The reason for this was to improve the assessment and feedback provided to students in their medical terms. Another driver was to avoid cumbersome paper files in each of the clinical school. In years three and four of the medical degree, students are based predominantly at one of seven clinical schools for terms in Medicine and Surgery as well as a number of specialty blocks. In Medicine they are attached to a medical team with a specified supervisor and the expectation is that they will be involved in all aspects of clinical work, clerking patients and following them through their admission.

Prior to 2015, students were expected to present one case history (long case) per week to their supervisor during an eight week term as part of a formative assessment. There was no formal requirement of students to submit evidence of presenting long cases during the medical blocks and in an end of year survey of students in 2013, 40 % reported completing 1 – 3 long cases [5] whereas they were expected to have completed eight cases.

Feedback for medical students frequently occurs in the clinical setting, and is thus more informal and less predictable than that occurring in the traditional academic setting [4]. A common complaint from medical students and residents is lack of feedback [6]. In keeping with these observations, evaluation of students in 2014 in their third year of the Sydney Medical Program, found that only 35 % of students agreed that they received helpful feedback about their progress [7].

To address these deficiencies, at the commencement of the 2015 academic year, a required formative assessment was introduced. To sit the summative barrier examination at the end of year 3, students were required to submit evidence of completing eight long cases in their medical block via an ePortfolio system. The assessor completed a criterion-based form, with immediate feedback to the student on their verbal presentation. The student was then required to write up the case, having reflected on the feedback, in a concise way, as they would for a medical admission in a patient’s hospital record. The written case and the feedback form were then uploaded onto an ePortfolio system. A Medical Lead at the students’ clinical school also gave feedback on the written, submitted case electronically. Thus students received feedback on their verbal presentation by their supervisor (or other consultant or resident staff) and by another physician, known as a “medical lead”, on their written case. Students had the option of reflecting on the quality, complexity and presentation of the written case.

The ePortfolio system used was Pebblepad, a proprietary, web-based ePortfolio system supported by e-learning at the University of Sydney [8].

The rationale for this study was that feedback provided using the ePortfolio would improve student engagement and enhance perceived competency at this task, secondly that perceived benefit would be related to quantity of feedback. It was also considered essential to describe challenges and barriers to the development of an ePortfolio. The reason for describing the negative aspects of the system was the concern that the time required using the ePortfolio system would negatively impact on the beneficial aspects of feedback.

The aims of this study were to describe students’ experience of using the ePortfolio, and receiving feedback on written long cases; medical leads’ experience of providing feedback on written long cases, and their perceptions of the value of the learning activity for students; and to explore the relationship between quantity and quality of feedback.

Methods

This is a mixed methods study, using a combination of both quantitative and qualitative approaches. Quantitative methods were used to quantify the number of longcases performed. Qualitative methods were used to explore the relationship between quantity and quality of feedback, and provide a rich understanding of both students’ and medical leads’ experience and perceptions of the ePortfolio.

Description of setting and educational practice

There are seven broad clinical schools to which medical students are assigned, six metropolitan schools and a rural clinical school (four separate sites). Three of the Clinical Schools; coded as A, B and C, were the settings of the study. All are in the Sydney metropolitan region. Participants in the study included 40 students at Clinical School A, 42 students at Clinical School B, and 48 students at Clinical School C.

Data collection and analysis

Mixed methods were used to collect and analyse data. Data collection was broken into four categories:

  1. 1)

    The number of completed long cases uploaded to the e-portfolio and episodes of feedback provided

    The software teaching interface provided data on who had given feedback and how many instances. Descriptive statistics were used to describe the number of episodes of feedback provided and person providing feedback.

  2. 2)

    Students’ perceptions of the process of performing long cases, submitting them and receiving feedback

    An invitation to participate in a focus group was extended via email to all year 3 students attending Clinical Schools A, B and C (n = 130). Convenience sampling was then used to select 8 participants for one focus group at each of the Clinical Schools (A, B, C). The strategy was to take the first 8 students who responded to the invitation. Student focus groups were conducted by an independent researcher who was not involved with the e-portfolio system (see Additional file 1 for student interview guide). Data were transcribed verbatim, with each participant being assigned an anonymous identifier. The identifier of A, B or C was allocated according to the clinical school of the study participant, followed by a unique number. Thematic analysis was used to build an understanding of the students’ experience. A portion of the data was read by the first author and analysed to identify initial themes. Following negotiation of meanings with the second and third authors, a coding framework was developed and applied to the full data set [9].

  3. 3)

    Medical leads’ experience of providing feedback on written long cases and their perception of the value of the learning activity for students

    Following completion of the student focus group data collection, a convenience sample was used to invite two medical leads at each Clinical School (A, B and C) to individual interviews. Medical Lead interviews were carried out by an independent researcher who was not involved with the e-portfolio (see Additional file 2 for medical leads interview guide). Data were transcribed verbatim, with each participant being assigned an anonymous identifier. The identifier of A, B or C was allocated according to the clinical school of the study participant, followed by a unique number. Thematic analysis of the qualitative data was conducted. While this was initially done inductively by three authors, in subsequent analysis of data, we noted that emergent themes from the inductive analysis resonated with the key themes found in student data. At this point the authors discussed the value of using a similar framework, which was applied to a portion of the data to ensure its consistency, and to check for any new and emerging issues that would extend the analysis. Subsequently, the second author coded all of the data to identify recurrent themes and subthemes in the data.

  4. 4)

    Independent review and analysis of the feedback content provided by the medical leads

    Written feedback provided by medical leads on the long cases was randomly sampled from the three clinical schools, until saturation of themes was achieved. Thematic analysis was used to code and categorise data into themes.

Results

All Year 3 medical students from Clinical Schools A, B and C (n = 130) participated in the study. Fifty five percent of the students were male, and 45 % female.

The number of completed long cases uploaded to the e-portfolio and episodes of feedback provided

The details of the numbers of students at the clinical schools and total number of feedback episodes (FE) are provided in Table 1. There were between 40–48 students at each clinical school and the average total number of FE was 191 per clinical school. Each student received between 4 – 5.4 FE for his or her eight cases. Clinical School A had the largest percentage of feedback provided by a clinical academic from the Education Office, 83 % compared with 58 % at clinical school C. The aim was that students would receive at least 1 FE for each of the 4 week terms of medicine in a timely fashion to be of most value to the students. The FE were assessed at the completion of the medical term. In Clinical school A, 100 % of students had achieved this feedback whereas at clinical school B and C, the percentage of students who had received timely feedback reduced to 68 and 46 % respectively.

Table 1 Details of Clinical school student numbers and episodes of feedback

Students’ perceptions of the process of performing long cases, submitting them and receiving feedback

Three focus groups made up of 21 students (8 females, 13 males) were conducted.

Three major themes were identified from analysis of the transcripts of the focus groups:

Authenticity of the activity; Feedback and personal development; and Technology and process. Results from the focus groups are illustrated in Table 2. Importantly, students recognized the benefits of practicing and improving competencies relevant to their future careers; and students appreciated the quality and of the feedback, finding that more than two iterations of feedback was not justified. Additionally, students felt reassured that feedback was provided by faculty members, with knowledge of their summative assessment requirements.

Table 2 Students’ perceptions of their experience of receiving feedback on written long cases, and the use of the eportfolio

Medical leads’ experience of providing feedback on written long cases and their perception of the value of the learning activity for students

In total, six medical leads were interviewed individually. Five of the medical leads interviewed were female, and one was male. Those interviewed included two General Practitioners, two Endocrinologists, one Haematologist and one Advanced Trainee. Four major themes were identified from analysis of the transcripts: Authenticity of the activity; Feedback and personal development; Ability to track students’ progress; and Technology and process. Results from the interviews are illustrated in Table 3. In line with findings from student focus groups, medical leads indicated that student activities should more closely reflect future career situations; and felt given the depth of written feedback, less iterations of feedback were necessary. Importantly, they felt that the written feedback given was qualitatively different to verbal feedback, providing greater emphasis on structure to the long case. Medical leads identified an increase in self-directed learning by students, who became more pro-active in seeking opportunities for patient interaction, and feedback from clinicians. As noted by Medical leads, the e-portfolio system provided a valuable way to systematically record and track students’ progress.

Table 3 Medical leads’ perceptions of their experience of providing feedback on written long cases, and their use of the eportfolio

In summary, our data suggests that students valued the authenticity of the written long case activity. Medical leads felt the task should better reflect a real work situation, with pressure of time in the write-up of the case. Interestingly students perceived the high quality of the feedback contributed to their development as a doctor. Most students felt the incremental benefit from receiving written feedback on more than two of the eight cases was not justified. Of note, there was no difference in the perception of adequacy of feedback between two of clinical schools, A and B. However, students at clinical school C perceived that they received less feedback, and that it was delayed, therefore reducing its value.

While the medical leads felt that feedback provided via the e-portfolio system to be qualitatively different to feedback provided via a face-to-face encounter, it built the capacity of students to understand what is required of them, particularly in terms of the structure of the long case and how to write up a medical admission. Both students and medical leads felt feedback was quite detailed, and that fewer iterations of feedback with greater consistency, at particular time points, would be of greater value to students than more frequent, less detailed feedback.

Major benefits identified by the medical leads were that the long case task increased students’ motivation to engage in clinical activities; and the e-portfolio system provided an efficient method to record and track students’ progress, and identify students in need of remediation. Students and medical leads alike found the ePortfolio system cumbersome. Both parties expressed concern about the detailed steps involved during the process, hindering their active engagement with the feedback. Of concern, some students were not uploading their long cases consistently, and additionally, some medical leads were not providing timely feedback, negating potential benefits.

Review and analysis of the feedback content provided by the medical leads

Analysis of the feedback, provided by the medical leads, demonstrated in Table 4, identified some key findings to inform future use of the ePortfolio. The feedback was student-focused at all times, with encouragement, even in those who had not performed as well as others, emphasising the concept of continuous professional development and lifelong learning. Authenticity again was a prominent theme with feedback reiterating that the written long case should mimic how notes are written in the real context of hospital practice. Medical leads provided commentary on the difference between verbal reporting and written text on the same case, demonstrating how complementary skills were needed and advice was given on how to develop both skills.

Table 4 Feedback provided by medical leads

The main focus of feedback was around getting the structure correct, with examples of how to achieve this rather than “just do it like I want”. Students often narrowly addressed the presenting issue but not other conditions that might be present, and could affect management and long term outcomes. The feedback from medical leads helped to broaden the students’, perspectives, with suggestions on approaches to chronic disease.

In summary the feedback was granular and specific. Medical leads corrected clinical reasoning where needed in case-specific ways, highlighted key features in the history and or examination and incongruences with the written case and final diagnosis/management. At this stage of training, the feedback focused on structure, linking presenting complaints to the differential and problem list to ensure coherence.

Discussion

Feedback is critical for learning in clinical medicine [10]. This study has provided detailed information about feedback using an ePortfolio system. The main finding is that medical students value feedback that is detailed, personalised and from a clinician with some experience. This is consistent with descriptions of “strong feedback” proposed by Van de Ridder, which included: well observable tasks, expert observer, feedback of highly specific information, personal observation and a plan to re-observe [10]. Other factors, which were described as essential for effective feedback, were that it was timely, constructive and actionable [11]. Credible feedback is often that received from a senior physician who is in a supervisory role [12]. Students in our study identified all these factors as important.

Our students particularly valued critical feedback and specific suggestions on how to improve. They were disparaging of the generic feedback usually provided by the university, such as “satisfactory”, and preferred “harsh”, constructive feedback. This is in contrast to the study by Boehler et al. [13], who found that students valued compliments rather than specific feedback on how to improve. In her study she assessed students satisfaction with feedback on surgical knot-tying, with half receiving only compliments and the other half receiving specific, constructive feedback. The main finding was that student satisfaction was higher in the group receiving compliments, whereas the improvement in the task was greater in the group receiving constructive feedback. The conclusion was that student satisfaction was not an accurate measure of quality of feedback, whereas learning was a function of constructive feedback. However, in contrast to Boeler’s study, which focussed on a manual skill, students in this current study were required to complete an authentic and intellectually demanding task.

Triangulation of the data, analysing feedback from the perspective of the teacher and student as well as independent review of the feedback content provided a deeper understanding of the feedback process. Quantity was not seen as important in this study, and most students felt “two was enough”. Analysis of the feedback revealed that high quality and granular initial feedback set the expectations of the student so that subsequent feedback could be very focused and shorter. The perceived benefit of two pieces of personalised, detailed feedback was considered enough, with not much further incremental benefit from more feedback. This certainly is practically important from the point of view of sustainability in providing feedback to approximately 300 students per year.

Details of what is the ideal quantity of feedback have not really been addressed in studies of ePortfolios, assuming that “more was better”. Students using our portfolio system were surprisingly aware and empathetic towards how much time the process of providing feedback was taking clinicians. Additionally, the acceptance and effectiveness of the written feedback appeared to be reliant upon students’ perceived credibility of the person providing the feedback, as has been reported in previous literature [14]. Students felt reassured that their written feedback was provided by a faculty member, familiar with University summative assessment requirements.

The use of an ePortfolio in general facilitated the delivery of feedback but in some cases, because of technical issues, detracted from the feedback process. Some students refused to engage with progressive use of the ePortfolio system, so uploaded all their cases at the end of the block, thus not benefitting from the progressive cycle of reflective practice as described by Kolb [15].

As in other ePortfolios, some students failed to see the purpose of the activity. This is similar to a number of other studies in both medical schools and other health sciences [4, 16]. A few students only appreciated the concept of using the ePortfolio to support the process of learning, to reflect and plan future learning activities. Most saw it as a simple “upload of an assignment”, and did not view the ePortfolio as part of deep learning from feedback.

Although the literature supports the use of an electronic format for providing feedback, previous studies have not looked at the quality and characteristics of the feedback in such detail. In the study of Spickard et al., although the electronic format resulted in more feedback, details of the quality were not explored; the analysis of students’ perception was based on a four-question survey with a five point Likert scale [17]. Belcher et al. [16], describes the use on an ePortfolio in an undergraduate medical program in the UK . Her findings are in contrast with this study in that the students perceived the feedback as either poor-quality or non-existent. This seemed to relate to the delay in provision of feedback and lack of engagement of the supervisors with the feedback process. This differs from two of the clinical schools in our study where the students found the feedback specific, timely and personalised.

Experiential learning and reflective practice are key theories in medical education. In the setting of learning medicine through case-based discussions, such as long cases, students can transform experience into learning [18]. In the medicine ePortfolio, learning was based around real patients, so situated learning complements experiential learning by framing the whole experience in the medical community [19]. The task was authentic which was also important as students could position the task as one required in intern years and beyond. Reflection is also a key aspect of the medicine case history ePortfolio and contributes to the depth of learning; through feedback, the student reflects on this, contextualises it, and improves performance [20].

Another interesting observation was the significant empathy expressed by students towards the medical leads. There was concern about the amount of work required of the medical leads providing feedback to the students. This is in contrast to studies demonstrating a decline in empathy in medical students and residents, perhaps partly due to an inappropriate learning environment, and inadequate role models [21]. A survey of medical students and interns concerning which factors they viewed as affecting empathy during education identified “mentoring and clinical experiences that promote professional growth” to be the most important [22]. It may be that an added benefit of the ePortfolio is providing mentors who can guide development of learning through the ePortfolio.

Barriers

Technical issues with the Pebblepad platform were the main concern with students using the ePortfolio. Time taken to negotiate and upload cases and difficulty finding feedback were some of the issues raised. This is reflected in the literature regarding ePortfolios. One of the main concerns is the substantial time commitment required by students, with perceived detraction from other learning activities. Time was given as the most important factor limiting the use of the ePortfolio by tutors and students in the study by Duque et al. (2006) [23]. In the study by Belcher et al. [16], students failed to see the purpose of the ePortfolio, feeling that it detracted from clinical time. The findings in this study are similar in that students reported spending too much time engaging with the technical aspects of the portfolio, spending up to 5 h on a single case.

Hall et al. similarly reported their experience with an ePortfolio in Ottawa Medical School and the main barrier to its implementation was the complexity of the system’s design which made its use cumbersome and tedious for students to use [24]. One of the main complaints in this study was the technical aspect of using Pebblepad, which frustrated students and impacted on their engagement with the learning process.

Limitations of the study

The limitations of this study include issues relating to design and outcome measures. There are seven major metropolitan clinical schools and three rural clinical schools, and the sampling for this study was from three of the major metropolitan schools. These results may not have reflected the experience of feedback of students at other clinical schools or rural sites. It is planned to organise further focus groups at other sites, including rural clinical schools. The sampling should also be more purposive to ensure equal representation of males to female students as it has been shown that males do not necessarily use feedback as a learning experience as well as females [25].

Conclusions

In conclusion, students valued the feedback provided by the ePortfolio system because of its quality, depth and personalised nature. It was also timely and constructive. However the technical difficulties with the platform had negative impacts on the engagement of some students, and overall the student cohort did not find it “user-friendly”. This study contributes to the literature in outlining the qualities of the feedback which students valued, and that quality was much more important than quantity. The recommendations for further improvements of the system include simplifying the ePortfolio platform, introducing authentic tasks such as discharge summaries and letters to GP and reducing the number of episodes of feedback to two, provided by the same senior clinician, for consistency of feedback. It is also essential that an ePortfolio is integrated both horizontally and vertically across the medical program as a tool for documenting learning, competencies and for reflective practice.

References

  1. Buckley S, Coleman J, Davison I, Khan KS, Zamora J, Malick S, Morley D, Pollard D, Ashcroft T, Popovic C, Sayers J. The educational effects of portfolios on undergraduate student learning: a Best Evidence Medical Education (BEME) systematic review. BEME Guide No. 11. Med Teach. 2009;31(4):282–98.

    Article  Google Scholar 

  2. Wilson A, Feyer AM. Review of Medical Intern Training. 2015. Australian Health Ministers’ Advisory Council, Retrieved from http://www.coaghealthcouncil.gov.au/medicalinternreview. Accessed 27 Apr 2016.

    Google Scholar 

  3. Murray C, Sandars J. e-Learning in medical education: Guide supplement 32.2–Practical application. Med Teach. 2009;31(4):364–5.

    Article  Google Scholar 

  4. Peacock S, Murray S, Scott A, Kelly J. The transformative role of ePortfolios: Feedback in healthcare learning. Int J ePortfolio. 2011;1(1):33–48.

    Google Scholar 

  5. Klein L. Evaluation of the Medicine 3 Core Blocks. Student Evaluation Report, Office of Medical Education, Assessment and Evaluation Unit. 2013 (11 July 2014).

  6. Branch Jr WT, Paranjape A. Feedback and reflection: teaching methods for clinical settings. Acad Med. 2002;77(12, Part 1):1185–8.

    Article  Google Scholar 

  7. Klein L, Comber J. Report on Stage 3, year 3 end of Year survey 2014. Office of Medical Education, Australia: Assessment and Evaluation Unit; 2015 (1 March 2015).

  8. PebblePad. 2016. http://www.pebblepad.com.au/. Accessed 27 Apr 2016.

  9. Creswell J. Educational Research: Planning, Conducting and Evaluating Quantitative and Qualitative Research. Upper Sadler River: Merrill; 2002.

  10. Van de Ridder J, Stokking KM, McGaghie WC, Ten Cate OTJ. What is feedback in clinical education? Med Educ. 2008;42(2):189–97.

    Article  Google Scholar 

  11. Watling CJ. Unfulfilled promise, untapped potential: Feedback at the crossroads. Med Teach. 2014;36(8):692–7.

    Article  Google Scholar 

  12. Watling CJ, Lingard L. Toward meaningful evaluation of medical trainees: the influence of participants’ perceptions of the process. Adv Health Sci Educ. 2012;17(2):183–94.

    Article  Google Scholar 

  13. Boehler ML, Rogers DA, Schwind CJ, Mayforth R, Quin J, Williams RG, Dunnington G. An investigation of medical student reactions to feedback: a randomised controlled trial. Med Educ. 2006;40(8):746–9.

    Article  Google Scholar 

  14. Telio S, Ajjawi R, Regehr G. The “educational alliance” as a framework for reconceptualizing feedback in medical education. Acad Med. 2015;90(5):609–14.

    Article  Google Scholar 

  15. Kolb D. Experiential learning as the science of learning and development. Englewood Cliffs: NPH; 1984.

    Google Scholar 

  16. Belcher R, Jones A, Smith LJ, Vincent T, Naidu SB, Montgomery J, Haq I, Gill D. Qualitative study of the impact of an authentic electronic portfolio in undergraduate medical education. BMC Med Educ. 2014;14:256–65.

    Article  Google Scholar 

  17. Spickard 3rd A, Gigante J, Stein G, Denny JC. Automatic capture of student notes to augment mentor feedback and student performance on patient write-ups. J Gen Intern Med. 2008;23(7):979–84.

    Article  Google Scholar 

  18. Kolb AY, Kolb DA. Learning styles and learning spaces: Enhancing experiential learning in higher education. Acad Manag Learn Edu. 2005;4(2):193–212.

    Article  Google Scholar 

  19. Lave J, Wenger E. Situated learning: Legitimate peripheral participation. UK: Cambridge university press; 1991. p 138.

  20. Mann KV. Theoretical perspectives in medical education: past experience and future possibilities. Medical Educ. 2011;45(1):60–8.

    Article  Google Scholar 

  21. Neumann M, Edelhäuser F, Tauschel D, Fischer MR, Wirtz M, Woopen C, Haramati A, Scheffer C. Empathy decline and its reasons: a systematic review of studies with medical students and residents. Acad Med. 2011;86(8):996–1009.

    Article  Google Scholar 

  22. Winseman J, Malik A, Morison J, Balkoski V. Students’ views on factors affecting empathy in medical education. Acad Psychiatry. 2009;33(6):484–91.

    Article  Google Scholar 

  23. Duque G, Finkelstein A, Roberts A, Tabatabai D, Gold SL, Winer LR. Learning while evaluating: the use of an electronic evaluation portfolio in a geriatric medicine clerkship. BMC Med Educ. 2006;6(S):4.

    Article  Google Scholar 

  24. Hall P, Byszewski A, Sutherland S, Stodel EJ. Developing a sustainable electronic portfolio (ePortfolio) program that fosters reflective practice and incorporates CanMEDS competencies into the undergraduate medical curriculum. Acad Med. 2012;876:744–51.

    Article  Google Scholar 

  25. Sinclair HK, Cleland JA. Undergraduate medical students: who seeks formative feedback? Med Educ. 2007;41(6):580–2.

    Article  Google Scholar 

Download references

Acknowledgements

We wish to acknowledge and thank Tyler Clark, Associate Lecturer, Sydney Medical School, for assistance with collection of qualitative data.

Funding

There was no funding for this study.

Availability of data and materials

Datasets supporting the conclusions of this article are included within the article. Additional data at the level of individual students is not available as per confidentiality agreements approved by the Human Research Ethics Committee, University of Sydney.

Authors’ contributions

JB was responsible for the design of the study, data collection, analysis and interpretation of the data, drafting of manuscript. AB contributed to the collection and analysis of data, and critical revision of the manuscript for important intellectual content. RW contributed to the collection of data and critical revision of the manuscript for important intellectual content. IH contributed to the collection and analysis of data, and critical revision of the manuscript for important intellectual content. All authors read and approved the final manuscript.

Authors’ information

Jane Bleasel is Professor and Co-Director of the Sydney Medical Program. She is a rheumatologist at Royal Prince Alfred Hospital with a PhD on the Genetics of Osteoarthritis.

Annette Burgess is a Senior Lecturer in the Education Office, Sydney Medical School, the University of Sydney.

Ruth Weeks is an Educational Designer in the Education Innovation Team at the University of Sydney.

Inam Haq is Professor and Co-Director of the Sydney Medical Program. His clinical training is in rheumatology and from 2008–2015 was Director of Undergraduate Studies at Brighton and Sussex Medical School in the UK.

Competing interests

Annette Burgess, a co-author, is currently an Associate Editor for BMC Medical Education.

All other authors (Jane Bleasel, Inam Haq, Ruth Weeks) declare that they have no competing interests to declare.

Consent for publication

Not applicable.

Ethics approval and consent to participate

Approval for this study was obtained from the University of Sydney Human Ethics Committee (protocol number 2015/606). Written informed consent was obtained from all study participants.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jane Bleasel.

Additional files

Additional file 1:

Interview guide for students. (DOCX 15 kb)

Additional file 2:

Interview guide for medical leads. (DOCX 13 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bleasel, J., Burgess, A., Weeks, R. et al. Feedback using an ePortfolio for medicine long cases: quality not quantity. BMC Med Educ 16, 278 (2016). https://doi.org/10.1186/s12909-016-0801-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12909-016-0801-3

Keywords