Teaching communication skills in clinical settings: comparing two applications of a comprehensive program with standardized and real patients
BMC Medical Education volume 14, Article number: 92 (2014)
Communication is important for the quality of clinical practice, and programs have been implemented to improve healthcare providers’ communication skills. However, the consistency of programs teaching communication skills has received little attention, and debate exists about the application of acquired skills to real patients. This study inspects whether (1) results from a communication program are replicated with different samples, and (2) results with standardized patients apply to interviews with real patients.
A structured, nine-month communication program was applied in two consecutive years to two different samples of healthcare professionals (25 in the first year, 20 in the second year). Results were assessed at four different points in time, each year, regarding participants’ confidence levels (self-rated), basic communication skills in interviews with standardized patients, and basic communication skills in interviews with real patients. Data were analyzed using GLM Repeated-Measures procedures.
Improvements were statistically significant in both years in all measures except in simulated patients’ assessment of the 2008 group. Differences between the two samples were non-significant. Differences between interviews with standardized and with real patients were also non-significant.
The program’s positive outcomes were replicated in different samples, and acquired skills were successfully applied to real-patient interviews. This reinforces this type of program structure as a valuable training tool, with results translating into real situations. It also adds to the reliability of the assessment instruments employed, though these may need adaptation in the case of real patients.
Studies consistently report improvements in communication skills among participants in programs teaching these competencies in healthcare settings [1–3]. The efficacy of the programs may vary depending on the teaching methods employed. Programs with modules divided into three sections are considered the exemplary practice for teaching these skills. They comprehend cognitive, modeling, and behavioral components, corresponding respectively to didactic, illustrational (e.g., through videos and other examples) and small group role-playing with feedback [4, 5]. Course structure has been offered as an explanation also for the long-term effects observed in communication skills acquired after a program, where the advantages of problem-based over lecture-based curricula are highlighted [6, 7]. In addition, courses with more hours of training that are intensive, practice-oriented and skills-focused tend to result in higher levels of change [1, 3, 5]. Less attention has been paid to the consistency of particular programs in effecting changes with different applications and different samples, that is, program replicability – a point also made in previous research highlighting this limitation in the various reviewed studies .
Improvements in communication skills after programs aimed at these competences are noticeable in participants’ subjective sense of confidence in their clinical practice [8, 9] and reflected in externally rated encounters with patients [10–12]. Programs commonly use standardized or simulated patients (SPs) to train and evaluate communication skills [13–15]. Though these two terms are frequently treated interchangeably, standardized patients are simulated patients trained to deliver identical, standardized performances [16–19], and both have been shown to provide reliable ratings of students’ competence [1, 15, 20]. Most importantly, however, is the application of acquired skills to real-life situations, with real patients (RPs).
Studies have shown that improvements in communication skills after a program are observed in encounters with both simulated and real patients [11, 21, 22], but the more important changes happened in SPs’ interviews. Professionals’ levels of performance with RPs are lower than with SPs . SPs have many recognized advantages over RPs (including their potential for a more objective assessment of interpersonal communication [19, 20] and their knowledge of the situation, which leads to more focused feedback regarding students’ performances ). However, SPs recreate artificial situations that may lack the realism inherent to encounters with RPs [19, 25, 26]. The skills healthcare professionals employ in one and the other situation may thus be different  and, even if adequate, they may escape checklists used to evaluate performance . Though research has reported that students react in similar ways to simulated and real patients , interacting with SPs may thus be different than interacting with RPs in real situations in terms of applying acquired communication skills. Instruments that adapt well to one situation may not fully adapt to the other.
Since effective communication has been shown to be relevant to clinical practice [28–30], and programs are created to improve communication between healthcare professionals and their patients, it is important to assess the consistency of these programs in operating changes and the extent to which these changes transfer to real situations with RPs, for whom they are intended. To these ends, the current study has two goals: (1) to inspect whether the results obtained with a structured, comprehensive program teaching communication skills can be replicated with a different sample of professional practitioners (to examine the effects of aspects such as chance, or sample characteristics, on the results), and (2) to examine whether results obtained with SPs in controlled situations apply to interviews with RPs in real situations.
The program had a duration of nine months and was offered in 2008 and, again, in 2009 as a post-graduate program on communication skills. Each year, the same content was taught by the same faculty members, following the same structure and sequence. The program has been described in detail in a previous work  and is summarized in Table 1. The current study targets basic communication skills, the focus of the initial three months of the program.
Classes met twice a week for five hours each day and typically included theoretical presentations of the materials, role-modeling through video viewing and discussion, and role-playing in small groups of peers (six or seven students), with feedback by both peers and faculty. Participants were divided into the smaller groups according to their professional backgrounds, to maximize group heterogeneity and exposure to professional diversity. Two faculty members per group tutored the same students throughout the program, encouraging participation and discussion. The role-playing practice represented a substantial part of the program (about 67 percent of the course load) and used previously written scripts based on cases of patients from the faculty’s and the students’ clinical experiences.
In this observational study, participants were evaluated at four moments in time: before the program (T0); three months into the program, at the end of the basic communication skills section (T1); six months into the program, at the end of the advanced skills section (T2); and three months later, at the end of the program (T3). In each moment, participants conducted a 25-minute first interview with SPs. All interviews were videotaped with participants’ consent. Participants were also asked to record a 25-minute first interview with one of their RPs, in a real situation, at T1, T2 and T3.
Participants were evaluated on their communication performances in the videotaped interviews with SPs (by two trained faculty members per interview and by the SP after the interview) and with RPs (by one trained faculty member per interview). They were additionally evaluated on their (self-rated) levels of confidence in conducting clinical encounters at T0, T2 and T3. Participants were informed that, with the exception of the interview at T0, all interviews with SPs and with RPs counted towards their grades. Participation was voluntary, and data were used for research purposes with participants’ consent. The study was approved by the Ethics Committee for Health of S. João Hospital/School of Medicine of Porto, and complies with the Declaration of Helsinki.
The interviews in 2008 and 2009 used the same five professional actors (three males and two females) from one theatre company who met with the students during these evaluation times exclusively. The actors shared the same ethnic background as the students (all Caucasian). Three were in their thirties, one male was in his late twenties and one male was in his fifties. Each actor met with an average of three to eight participants per evaluation time each year.
The actors had no previous experience as SPs, though two had majored in psychology and a third had attended several workshops dealing with communication in medical settings. The team of faculty members (all trained and experienced teachers in communication skills) coached the actors on the purposes of the interviews, on the scripts of ‘patients’ created for the SPs and on the specific aspects of the encounters, such as the amount and kind of “talking” and “acting” they should do for the purposes of these interviews. Each script was reviewed with each actor, and possible responses and interactions rehearsed. The initial preparation took about two hours with each actor, plus an average of over two hours actors reported spending with each script at home. At each evaluation point, actors were again briefed on the scripts, the acting and the purposes of the program. Further discussions of the cases and the interviews occurred throughout the programs on demand. Additionally, faculty members systematically monitored and gave SPs feedback about their performances throughout the programs in order to maintain acting quality and consistency, and keep it within the evaluation goals.
Written by faculty members with clinical experience, the scripts featured situations based on real cases of patients. Tailored to the actors (each actor had his or her own scripts) and adapted to the professional areas of the participants in this study, all scripts followed the same format and contained detailed information on patients’ personal circumstances (e.g., age, occupation), sets of symptoms (and associated emotions, beliefs and impact in daily life), health history, family health history, life style (e.g., physical activity, alcohol consumption), and psychosocial aspects (e.g., life satisfaction, preoccupations, support net). Since the main goal of the study was communication, actual clinical symptoms were consistently kept simple across scripts, namely to ensure equivalent situations across the different professional areas involved.
Faculty members assessed participants’ communication skills in interviews with SPs and with RPs using the SEGUE framework . SPs rated students’ performances after each interview using the Interpersonal and Communication Skills Checklist (ICSC) . Participants rated their own levels of confidence in conducting clinical encounters using a modified version of Smith et al.’s self-efficacy questionnaire . All instruments were translated into Portuguese, and item clarity was individually checked with respondents in the initial interviews.
These instruments, as well as the rationale for their choice, are described elsewhere [8, 15, 31–34]. In summary, the SEGUE framework is a 25-item (yes/no) checklist designed by Makoul to facilitate teaching and assessment of critical communication tasks . Easy to use, it has also demonstrated acceptable psychometric characteristics in varied contexts over the years [32, 34].
The ICSC was developed by the eight medical schools in the New York City Consortium for Clinical Competence. It is a 17-item (yes/no) checklist designed to assess interpersonal and communication skills that has shown acceptable psychometric properties .
Developed to assess residents’ attitudes towards psychosocial skills used in medical care, Smith et al.’s questionnaire is a 38-item, 7-point Likert scale that evaluates self-efficacy . The original instrument contains items that assess domains unrelated to our curriculum, and we used the 17 items that do correspond to our learning contents, adapting them to our program. Students evaluated their sense of self-efficacy regarding clinical communication skills (e.g., How confident are you that you can refrain from interrupting the patient? How confident are you that you can identify unexpressed feelings?) in a 7-point Likert scale (from 1 = not at all confident to 7 = totally confident). For the purposes of this study, only the 10 initial items, focusing on basic communication skills, were used (the remaining seven are directed at advanced skills and special situations and were excluded).
Twenty-five healthcare professionals participated in the program in 2008 and another 20 enrolled in 2009. The two groups are comparable in several ways: professional background composition (physicians, nurses, clinical psychologists, physiotherapists, speech therapists, etc., present in both groups in similar proportions, the first two above-mentioned professions making up about 68% of the 2008 group and 65% of the 2009 group), professional experience (mean of approximately three years of practice in both groups, representing the beginning of participants’ careers), and gender (two males in 2008, two males in 2009).
For the SEGUE framework, and for the ICSC, the score was the percentage of items checked ‘yes’ in each interview. The mean of the scores given by the two different SEGUE raters to interviews with SPs was calculated and used. For the measure of confidence, the mean of the ratings (in the Likert scale) was computed for the used items. Total scores for each measurement time were obtained by calculating the mean of all students’ scores in each instrument.
To address the first goal of this study, GLM Repeated-Measures procedures were conducted in PASW (Predictive Analysis Software) to inspect changes over time in the measures of confidence and of performance (SEGUE and ICSC) with SPs, and to examine differences between samples, following a factorial plan of 2 × 4 – (2008/2009 group) × (T0/T1/T2/T3 evaluation times). Bonferroni correction was used for multiple comparisons (alpha = 0.01).
In order to assess whether changes with the program transfer to real situations with RPs (the second goal of this study), another GLM Repeated-Measures procedure was conducted on the measure of performance (SEGUE) according to a factorial plan of 2×3 – (SPs/RPs) × (T1/T2/T3 evaluation times). Participants’ interviews with RPs were added to their SPs’ interviews, leading to a database of 90 subjects. Because the instruction in both situations asked for a first encounter with a patient, students who conducted other types of encounters with RPs were excluded. Thus procedures were carried out for a total n = 70.
Program replicability: the 2008 and 2009 groups
Performances follow similar patterns in the 2008 and 2009 groups (Figure 1). The 2009 group scored lower in all three measures before entering the program (at T0), and reached higher values throughout its course also in all three measures, than the 2008 group. Between-subjects tests show that apparent differences between the two groups throughout the program are statistically non-significant.
Effects of taking the course assessed by SPs are statistically significant for the 2009 group (as shown by within-subjects tests, including corrected Greenhouse-Geisser, Huynh-Feldt and Lower-bound procedures, F(3, 54) = 24.79, p < 0.01), but not for the 2008 group (Table 2). Within-subjects simple contrasts for SPs’ assessments of the 2009 group show significant differences between performances before the course (at T0) and each measured time after: at T1 (F(1, 18) = 19.14, p < 0.01), at T2 (F(1, 18) = 45.75, p < 0.01) and at T3 (F(1, 18) = 30.90, p < 0.01). By contrast, SPs already rated 2008 group performances at high levels before the beginning of the program, approximating them to their subsequent evaluations at T1, T2 and T3.
Effects of taking the course on self-confidence are statistically significant for the 2009 group (as shown by within-subjects tests, including corrected Greenhouse-Geisser, Huynh-Feldt and Lower-bound procedures, F(2, 36) = 17.61, p < 0.01), as well as the 2008 sample (shown by within-subjects tests, including corrected Greenhouse-Geisser and Huynh-Feldt procedures, F(2, 40) = 6.90, p < 0.01). At the end of the program (T3), participants were significantly more confident in their communication skills in clinical interviews than before taking it (2008 group, F(1, 20) = 10.17, p < 0.01; 2009 group, F(1, 18) = 24.76, p < 0.01), and the increase in confidence was statistically significant also at T2 for the 2009 group (F(1, 18) = 11.17, p < 0.01), though it was non-significant for the 2008 sample, reflecting the former group’s greater change with the program, mentioned above. Pairwise comparisons further indicate that, for the 2009 group, the change between T2 and T3 is also statistically significant (mean difference of 0.45, p = 0.009), though it is non-significant for the 2008 group, with its less pronounced increase.
Effects of the program on communication skills assessed by faculty are statistically significant for the 2009 sample (as shown by within-subjects tests, including corrected Greenhouse-Geisser, Huynh-Feldt and Lower-bound procedures, F(3, 57) = 48.69, p < 0.01), as well as the 2008 group (as shown by within-subjects tests, including corrected Greenhouse-Geisser, Huynh-Feldt and Lower-bound procedures, F(3, 72) = 40.31, p < 0.01). Comparing with T0, the increase in communication skills was statistically significant for the two samples at T1 (2008 group, F(1, 24) = 68.52, p < 0.01; 2009 group, F(1, 19) = 46.18, p < 0.01), T2 (2008 group, F(1, 24) = 104.83, p < 0.01; 2009 group, F(1, 19) = 117.85, p < 0.01) and T3 (2008 group, F(1, 24) = 44.02, p < 0.01; 2009 group, F(1, 19) = 93.88, p < 0.01) (Table 2).
Pairwise comparisons indicate that the decline observed by both external observers and SPs in communication skills at T3 is statistically non-significant, when compared with T2, for the 2008 as well as the 2009 sample (as are all other differences between T1, T2 and T3 for the two samples, whether assessed by external observers or SPs). As mentioned before, in both samples, the change in competence at T3 remains statistically significant, comparing with T0, despite the decline (except for SPs’ assessment of the 2008 group, where, as mentioned earlier, no change reached statistical significance).
Comparison between standardized- and real-patient scores
Results show the numerical similarity between ratings for SPs and for RPs throughout the program (Table 3).Participants’ communication skills reach the highest scores at T1 for RPs, with a slight (statistically non-significant) decline afterwards, whereas for SPs, the highest point occurs at T2, with a slight (statistically non-significant) decline at T3 (Figure 2). Differences between SPs’ and RPs’ interviews are statistically non-significant.
Program replicability: the 2008 and 2009 samples
Regarding the first goal of this study, the analyses indicate that the positive outcomes of this structured, comprehensive training program are replicated in different samples in two different years. These positive outcomes are reflected, each year, in statistically significant increases in confidence, self-rated by participants, and in communication skills, assessed by external observers and by SPs, with between-group effects indicating that no differences exist overall between the 2008 and the 2009 groups throughout the program. The slight decay observed for both the 2008 and the 2009 groups at T3 has no statistical significance in either external observers’ or SPs’ assessments. These findings substantiate the robustness of the procedures, validating the program through confirmation of its results with a different sample and in a different year (hence minimizing the possibility that results were due to aspects such as chance, or particular characteristics of the 2008 group, or other circumstances occurring during that particular year, such as faculty’s initial motivation). They also show the stability of the instruments employed, which adds to their potential for comparative use (in 22 evaluations, only one registered a comparatively substantial difference, i.e., the 2008 score by SPs at T0, when compared with their 2009 score at T0).
The fact that SPs’ ratings at T0 are lower in 2009 (close to faculty’s ratings) than in 2008 may partly be an effect of learning. Throughout their first year, SPs were exposed to professionals increasingly more effective at communicating, which may have created new (and higher) standards for them, then applied to the next year’s participants at T0. They may also have become more familiarized with the specificities of the competences taught, moving away from the global ways in which patients tend to evaluate health professionals [34–36] and providing more focused assessments at T0. These possibilities are in line with research indicating that non-trained raters tend to give students higher scores than trained raters [37, 38].
On the other hand, participants in the 2009 class scored lower at T0 than those of the 2008 group in all measures. This could indicate that the 2009 group began with fewer communication skills than did the 2008 sample (to which SPs would have been sensitive, with their ratings). However, the generally smaller standard deviations in 2009 for SPs (as well as for external observers) support the learning hypothesis (Table 2). Still, the standard deviation for SPs at T0 is rather large (and larger than the 2008 one), and further studies on how these patients – as well as RPs – assess their health professionals are needed for a more thorough understanding of these results.
Comparison between standardized and real patients
Regarding the second goal of this study, the analyses indicate that communication competences acquired and used throughout the program, and applied with SPs, were effectively transferred to situations with RPs. These results confirm previous research indicating that after-program improvements in students’ communication skills are observed both in SPs’ and in RPs’ interviews [10, 11, 21, 22]. The similarity of the results (with statistically non-significant differences) between the two groups reinforces the notion that SPs constitute effective proxies for RPs .
One difficulty of using the SEGUE framework in real situations, though, had to do with the duration of the interviews. Within the 25-minute interviews, the parameters of the SEGUE framework applied to encounters with SPs in controlled situations more easily than to encounters with RPs. This observation is in line with the idea that SPs may not provide enough realism in clinical interviews, and experienced clinicians may deviate from checklists while employing adequate interviewing skills . Interviews with some RPs required attention and response to several relevant and emotionally-loaded situations that took most of the interview time, preventing some items of its structure from being discussed within that time frame. Whenever these competencies were correctly used (and indeed necessary in such cases), and given the lack of items in the SEGUE framework to be checked regarding this special attention to the person, “being personally present”, and flexibility , our choice was to mark as “non applicable” parts of the interview that were not covered. This adaptation may partly explain the slightly higher scores participants obtained with RPs than with SPs (contrary to previous research ). For RPs, the number of items marked “yes” was, in several cases, closer to the total number of valid items (those considered “applicable”).
One limitation in this study is that it did not include interviews with RPs at T0, preventing analyses on the magnitude of the changes with the program for those patients. Also, since students voluntarily enroll in this program, they may be particularly motivated to learn and apply these communication skills, limiting the generalization of program effects to other healthcare professionals. Despite these limitations, the study has the advantages of complementing participants’ (subjective) self-assessments with evaluations by external observers and by SPs; including evaluations at different points in time, namely before and after the program; using two different samples of participants; and including SPs in controlled situations, as well as RPs in real situations. Future studies with control groups and participant randomization in a pretest-posttest design can help disentangle the effects of confounding variables, such as participants’ own experience, or the effects of the more advanced classes of the program, on the observed increase in basic communication skills and in self-confidence. Once the outcomes of the overall program are established, its defined components can also be manipulated to assess their specific effects in the results. Qualitative interviews will provide relevant information on participants’ perspectives about what makes the communication training effective and why. Finally, the program’s robustness can further be tested when taught by different instructors in future studies.
The consistent positive results in this study indicate that this type of program structure and procedures can be used to improve communication skills among healthcare professionals. The skills learned and practiced within the program in artificial settings are transferrable to controlled situations with simulated patients, and results are replicated in different samples of participants. They also transfer to real-life situations with real patients, the ultimate goal of these courses. Future follow-up studies with these samples will assess the permanence over time of these communication skills in clinical interviews.
Parts of this work were presented at the International Conference on Communication in Healthcare in Verona, Italy, September 2010.
Aspegren K: BEME guide no.2: Teaching and learning communication skills in medicine – a review with quality grading of articles. Med Teach. 1999, 21: 563-570. 10.1080/01421599978979.
Cegala DJ, Broz SL: Physician communication skills training: A review of theoretical backgrounds, objectives and skills. Med Educ. 2002, 36: 1004-1016. 10.1046/j.1365-2923.2002.01331.x.
Merckaert I, Libert Y, Razavi D: Communication skills training in cancer care: Where are we and where are we going?. Curr Opin Oncol. 2005, 17: 319-330. 10.1097/01.cco.0000167737.72196.f9.
Maguire P, Pitceathly C: Key communication skills and how to acquire them. BMJ. 2002, 325: 697-700. 10.1136/bmj.325.7366.697.
Bylund CL, Brown R, Gueguen JA, Diamond C, Bianculli J, Kissane DW: The implementation and assessment of a comprehensive communication skills training curriculum for oncologists. Psychooncology. 2010, 19: 583-593.
Levenkron JC, Greenland P, Bowley N: Teaching risk-factor counseling skills: A comparison of two instructional methods. Am J Prev Med. 1990, 6 (Suppl 2): 29-34.
Holm U, Aspegren K: Pedagogical methods and affect tolerance in medical students. Med Educ. 1999, 33: 14-18. 10.1046/j.1365-2923.1999.00332.x.
Smith RC, Mettler JA, Stöffelmayr BE, Lyles JS, Marshall AA, van Egeren LF, Osborn GG, Shebroe V: Improving residents’ confidence in using psychosocial skills. J Gen Intern Med. 1995, 10: 315-320. 10.1007/BF02599950.
Doyle D, Copeland HL, Bush D, Stein L, Thompson S: A course for nurses to handle difficult communication situations. A randomized controlled trial of impact on self-efficacy and performance. Patient Educ Couns. 2011, 82: 100-109. 10.1016/j.pec.2010.02.013.
Fallowfield L, Jenkins V, Farewell V, Saul J, Duffy A, Eves R: Efficacy of a Cancer Research UK communication skills training model for oncologists: A randomised controlled trial. Lancet. 2002, 359: 650-656. 10.1016/S0140-6736(02)07810-8.
Delvaux N, Razavi D, Marchal S, Bredart A, Farvacques C, Slachmuylder JL: Effects of a 105 hours psychological training program on attitudes, communication skills and occupational stress in oncology: A randomised study. Br J Cancer. 2004, 90: 106-114. 10.1038/sj.bjc.6601459.
Jenkins V, Fallowfield L, Solis-Trapala I, Langridge C, Farewell V: Discussing randomised clinical trials of cancer therapy: Evaluation of a Cancer Research UK training programme. BMJ. 2005, 330: 400-403. 10.1136/bmj.38366.562685.8F.
Colliver JA, Swartz MH: Assessing clinical performance with standardized patients. JAMA. 1997, 278: 790-791. 10.1001/jama.278.9.790.
Epstein RM, Hundert EM: Defining and assessing professional competence. JAMA. 2002, 287: 226-235. 10.1001/jama.287.2.226.
Duffy FD, Gordon GH, Whelan G, Cole-Kelly K, Frankel R, Buffone N, Lofton S, Wallace M, Goode L, Langdon L: Assessing competence in communication and interpersonal skills: The Kalamazoo II report. Acad Med. 2004, 79: 495-507. 10.1097/00001888-200406000-00002.
Bokken L, Rethans JJ, Scherpbier A, van der Vleuten C: Strengths and weaknesses of simulated and real patients in the teaching of skills to medical students: A review. Sim Healthc. 2008, 3: 161-169. 10.1097/SIH.0b013e318182fc56.
Bokken L, Linssen T, Scherpbier A, van der Vleuten C, Rethans JJ: Feedback by simulated patients in undergraduate medical education: A systematic review of the literature. Med Educ. 2009, 43: 202-210. 10.1111/j.1365-2923.2008.03268.x.
Bosse HM, Schultz J-H, Nickel M, Lutz T, Möltner A, Jünger J, Huwendiek S, Nikendei C: The effect of using standardized patients or peer role play on ratings of undergraduate communication training: A randomized controlled trial. Patient Educ Couns. 2012, 87: 300-306. 10.1016/j.pec.2011.10.007.
Cleland JA, Abe K, Rethans JJ: The use of simulated patients in medical education: AMEE Guide No. 42. Med Teach. 2009, 31: 477-486. 10.1080/01421590903002821.
Fiscella K, Franks P, Srinivasan M, Kravitz RL, Epstein R: Ratings of physician communication by real and standardized patients. Ann Fam Med. 2007, 5: 151-158. 10.1370/afm.643.
Razavi D, Merckaert I, Marchal S, Libert Y, Conradt S, Boniver J, Etienne AM, Fontaine O, Janne P, Klastersky J, Reynaert C, Scalliet P, Slachmuylder JL, Delvaux N: How to optimize physicians’ communication skills in cancer care: Results of a randomized study assessing the usefulness of posttraining consolidation workshops. J Clin Oncol. 2003, 21: 3141-3149. 10.1200/JCO.2003.08.031.
Delvaux N, Merckaert I, Marchal S, Libert Y, Conradt S, Boniver J, Etienne AM, Fontaine O, Janne P, Klastersky J, Mélot C, Reynaert C, Scalliet P, Slachmuylder JL, Razavi D: Physicians’ communication with a cancer patient and a relative: A randomized study assessing the efficacy of consolidation workshops. Cancer. 2005, 103: 2397-2411. 10.1002/cncr.21093.
Reinders ME, Blankenstein AH, van Marwijk HW, Knol DL, Ram P, van der Horst HE, de Vet HC: Reliability of consultation skills assessments using standardised versus real patients. Med Educ. 2011, 45: 578-584. 10.1111/j.1365-2923.2010.03917.x.
Bokken L, Rethans JJ, Jöbsis Q, Duvivier R, Scherpbier A, van der Vleuten C: Instructiveness of real patients and simulated patients in undergraduate medical education: A randomized experiment. Acad Med. 2010, 85: 148-154. 10.1097/ACM.0b013e3181c48130.
Hodges B, Regehr G, McNaughton N, Tiberius R, Hanson M: OSCE checklists do not capture increasing levels of expertise. Acad Med. 1999, 74: 1129-1134. 10.1097/00001888-199910000-00017.
Malhotra A, Gregory I, Darvill E, Goble E, Pryce-Roberts A, Lundberg K, Konradsen S, Hafstad H: Mind the gap: Learners’ perspectives on what they learn in communication compared to how they and others behave in the real world. Patient Educ Couns. 2009, 76: 385-390. 10.1016/j.pec.2009.07.024.
Sanson-Fisher RW, Poole AD: Simulated patients and the assessment of medical students’ interpersonal skills. Med Educ. 1980, 14: 249-253. 10.1111/j.1365-2923.1980.tb02269.x.
Kaplan SH, Greenfield S, Ware JE: Assessing the effects of physician-patient interactions on the outcomes of chronic disease. Med Care. 1989, 275 (Suppl 3): 110-127.
Fallowfield LJ: Psychosocial adjustment after treatment for early breast cancer. Oncology (Williston Park). 1990, 4: 89-97.
Stewart M, Meredith L, Brown JB, Galajda J: The influence of older patient-physician communication on health and health-related outcomes. Clin Geriatr Med. 2000, 16: 25-36. 10.1016/S0749-0690(05)70005-7.
Carvalho IP, Pais VG, Almeida SS, Ribeiro-Silva R, Figueiredo-Braga M, Teles A, Castro-Vale I, Mota-Cardoso R: Learning clinical communication skills: Outcomes of a program for professional practitioners. Patient Educ Couns. 2011, 84: 84-89. 10.1016/j.pec.2010.05.010.
Makoul G: The SEGUE Framework for teaching and assessing communication skills. Patient Educ Couns. 2001, 45: 23-34. 10.1016/S0738-3991(01)00136-7.
Cohen DS, Colliver JA, Marcy MS, Fried ED, Swartz MH: Psychometric properties of a standardized-patient checklist and rating-scale form used to assess interpersonal and communication skills. Acad Med. 1996, 71 (Suppl 1): 87-89.
Schirmer JM, Mauksch L, Lang F, Marvel MK, Zoppi K, Epstein RM, Brock D, Pryzbylski M: Assessing communication competence: A review of current tools. Fam Med. 2005, 37: 184-192.
Franks P, Fiscella K, Shields CG, Meldrum SC, Duberstein P, Jerant AF, Tancredi DJ, Epstein RM: Are patients’ ratings of their physicians related to health outcomes?. Ann Fam Med. 2005, 3: 229-234. 10.1370/afm.267.
Franks P, Jerant AF, Fiscella K, Shields CG, Tancredi DJ, Epstein RM: Studying physician effects on patient outcomes: Physician interactional style and performance on quality of care indicators. Soc Sci Med. 2006, 62: 422-432. 10.1016/j.socscimed.2005.05.027.
Finlay IG, Stott NC, Kinnersley P: The assessment of communication skills in palliative medicine: A comparison of the scores of examiners and simulated patients. Med Educ. 1995, 29: 424-429. 10.1111/j.1365-2923.1995.tb02866.x.
Holmboe ES, Hawkins RE, Huot SJ: Effects of training in direct observation of medical residents’ clinical competence: A randomized trial. Ann Intern Med. 2004, 140: 874-881. 10.7326/0003-4819-140-11-200406010-00008.
The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6920/14/92/prepub
The authors thank Sara Rocha for her invaluable technical support and Ernesto Paulo Fonseca for his assistance with statistical analyses.
No funding was involved in this project.
The authors declare that they have no competing interests.
All authors participated in the conception of the study after RMC’s initial proposal, and in the data collection process. IPC, VGP and FRS conducted the literature review and IPC analyzed the data. IPC wrote a first draft of the manuscript, which was read by the other authors. All authors contributed to the paper’s final version. All authors read and approved the final manuscript.
About this article
Cite this article
Carvalho, I.P., Pais, V.G., Silva, F.R. et al. Teaching communication skills in clinical settings: comparing two applications of a comprehensive program with standardized and real patients. BMC Med Educ 14, 92 (2014). https://doi.org/10.1186/1472-6920-14-92
- Communication program
- Healthcare professionals
- Standardized patients
- Real patients