Skip to main content

Quality of care associated with number of cases seen and self-reports of clinical competence for Japanese physicians-in-training in internal medicine



The extent of clinical exposure needed to ensure quality care has not been well determined during internal medicine training. We aimed to determine the association between clinical exposure (number of cases seen), self- reports of clinical competence, and type of institution (predictor variables) and quality of care (outcome variable) as measured by clinical vignettes.


Cross-sectional study using univariate and multivariate linear analyses in 11 teaching hospitals in Japan. Participants were physicians-in-training in internal medicine departments. Main outcome measure was standardized t-scores (quality of care) derived from responses to five clinical vignettes.


Of the 375 eligible participants, 263 (70.1%) completed the vignettes. Most were in their first (57.8%) and second year (28.5%) of training; on average, the participants were 1.8 years (range = 1–8) after graduation. Two thirds of the participants (68.8%) worked in university-affiliated teaching hospitals. The median number of cases seen was 210 (range = 10–11400). Greater exposure to cases (p = 0.0005), higher self-reports of clinical competence (p = 0.0095), and type of institution (p < 0.0001) were significantly associated with higher quality of care, using a multivariate linear model and adjusting for the remaining factors. Quality of care rapidly increased for the first 100 to 200 cases seen and tapered thereafter.


The amount of clinical exposure and levels of self-reports of clinical competence, not years after graduation, were positively associated with quality of care, adjusting for the remaining factors. The learning curve tapered after about 200 cases.

Peer Review reports


Healthcare systems throughout the world are searching for better ways of delivering high quality care. Attention to quality of patient care has become an important healthcare issue during the last decade, not only for health authorities, policymakers, and managers, but also for physicians and patients. Improving the quality of healthcare involves a broad range of discrete activities such as rigorous evaluation of conventional treatments, incorporating patients' views in healthcare decisions, and audit and feedback of healthcare practices. Physicians are one of the main healthcare providers and are confronted with increasing pressure to provide and improve care. The skills and knowledge of physicians improve through a combination of didactic and experiential learning that can in turn contribute to improving patient care [1]. Learning occurs through repeated experience with many clinical cases. The number of clinical cases seen might be an important factor linked to quality of care.

The number of clinical cases needed to meet optimal levels of proficiency in surgical procedures [2, 3], such as colonoscopy [4], has been evaluated often. However, there is very little literature that evaluates the impact of the amount of clinical exposure on quality of care in internal medicine, that is, the learning curve, especially for residents [5]. Bugelski suggested in the 1970's that the major increment in learning occurs in the early stages of exposure and that less is learned during later stages [6]. This theory has been tested especially for various surgical procedures and invasive diagnostic tests [2, 4, 7], but there was only one study (Medline search) that examined the learning curve for internal medicine training. Day et al. concluded that the rate of increase in self-reports of clinical competence in specific skills was influenced by the number of post-graduation years [5]. Residents in the first post-graduation year (PGY1) reported that their skills improved by an average of 196% during that year compared to less than 50% for residents in the third year. The goal of this study was to examine the relationship between self-estimates of clinical exposure (number of cases seen) and self-reports of clinical competence and quality of care using specially designed clinical vignettes.



An anonymous survey was administered in departments of internal medicine at 11 teaching hospitals in Japan, a convenience sample from 6 university-affiliated and 5 non-affiliated teaching hospitals (with oversampling of university settings). All physicians-in-training within 10 years of graduation from medical school were eligible for the study (n = 375). To avoid contamination, the survey was administered on the same day (03/14/2003) in all institutions.

Measurement method

Quality of care was defined as the delivery of patient care in a manner that leads to better outcomes for individuals and populations [8]. Clinical vignettes have been used to measure variations in quality of care [9]. The scores derived from the vignettes reliably reflected actual levels of physician practice, resulting in higher criterion validity compared to scores derived from chart abstractions. Based on disease prevalence in Japan, we began by selecting six clinical vignettes to measure quality of care: four common outpatient chronic conditions (diabetes mellitus, chronic obstructive pulmonary disease, vascular disease, and depression) and two acute emergency room conditions (subarachnoid hemorrhage and gastrointestinal bleeding).

Two detailed clinical vignettes were developed for each chronic condition, for a total of 8 vignettes. These vignettes were originally developed to measure quality of care in the United States. The vignettes were translated in Japanese and partly revised to match clinical practice in Japan, for example, using equivalent drugs and screening procedures. In addition, we developed two original Japanese vignettes for the two acute conditions. From the 10 vignettes available, each participant received five randomly selected vignettes, one from each condition (4 chronic and 1 acute). The vignettes required open-ended responses to questions that were presented in sections characteristic of a typical patient encounter: presenting complaint, history, physical examination, radiological or laboratory tests, diagnosis, and treatment and management plans. Each section began with the presentation of new information. After answering a given section, participants could not return to previous sections to revise (possibly improve) their answers. Participants were given 85 minutes to complete all 5 vignettes.

Clinical exposure was measured using participant self-estimates of the number of patients seen in in-patient wards, outpatient clinics, and emergency rooms. Data was also collected on the number of years after graduation, type of institution (university-affiliated teaching hospitals or non-affiliated), self-reports of clinical competence (i.e., problem-solving ability, basic procedural skills [e.g., venipuncture, bone marrow aspiration], and basic medical knowledge), and communication ability (i.e., attitude toward patients and their family and cooperativeness with other medical staff). Self-reports were rated using a five-point ordinal rating scale (i.e., unsatisfactory, satisfactory, good, excellent, or outstanding). The overall model consisted of one quality-of-care outcome variable, portrayed by the vignettes, and four predictor variables, that is, self-estimates of total number of patients seen, type of institution, and self-reports of clinical competence and communication ability. The latter two variables were also summed to create a global self-reported competence variable.


The responses to the vignettes were scored by the authors. To ensure consistency in scoring, given conditions were scored by the same author. With regard to chronic conditions, we used the scoring criteria developed by the original American authors who based their criteria on national guidelines [9]. These criteria were then reviewed and ratified by expert panels of academic and community physicians in Japan, in fields relevant to each condition; in the end the original criteria were adopted. Scoring criteria for the acute conditions were developed de novo, using expert panels of Japanese physicians. To verify the equivalence of the Japanese version with the original English version, the 10 vignettes were back-translated into English and verified by the original American authors. Based on their recommendations and consensus among the authors, the vignettes and scoring criteria were finalized. Each vignette contained an average of 37 criteria (range = 26–50). Each criterion was rated according to a three-level quality-of-care scale: adequate, unnecessary, and inappropriate care. A one-point credit was assigned for each criterion when adequate care was proposed. An overall vignette score was assigned by summing the scores from the individual criteria.

First, we used the general linear model to test for a vignette (disease condition) effect; there was no such effect (p = 0.239). Thus the scores from all five vignettes were added for each participant and then converted to a standardized t-score with a mean of 50 and a standard deviation of 10; t-scores were used as the outcome (criterion) variable for quality of care in the analyses. This transformation facilitated the interpretation of the relative importance of the predictor variables by comparing the corresponding t-scores to means of 50.

Predictor variables

The amount of clinical exposure was computed by adding the number of cases that each subject had seen in each setting (inpatient, outpatient and ER) and then scored according to five ordinal categories: 0–100, 201–300, 301–400, 401–600, 601–800, >800 cases. The data distribution was skewed to the left and consequently ordinal categories were used because they fit the model better than log transformations (based on Akaike's information criteria – AIC [10]). We also used a broader range (200 vs. 100) for numbers above 401 because the data were skewed and sparse in those categories. The proportion of cases in the in-patient setting was also calculated and incorporated into the analyses because training occurs mostly in in-patient settings in Japan. The number of years after graduation could influence the amount of clinical exposure and was thus incorporated into the analysis using three ordinal categories (because again they fit the model better than log transformations): one year after graduation (PGY1); two years after graduation (PGY2); and more than 3 years after graduation (≥PGY3).

In addition to examining the relationship between overall clinical exposure and t-scores, we also looked specifically at the exposure to disorders similar to the ones in the vignettes. This was measured using a common disease index (CDI), defined as (numerator:) the number of cases seen that were similar to the diseases in the vignettes (i.e., stroke, gastrointestinal bleeding, COPD, heart failure, ischemic heart disease, depression, and diabetes mellitus), divided by (denominator:) the total number of cases seen. The t-scores were plotted against CDI to interpret graphically the relationship between the CDI and t-scores. (a measure of quality of care). We also examined the relationship between CDI and quality of care, adjusting for overall clinical exposure in order to verify whether simply increasing the proportion of clinical exposure to similar disorders would lead to higher quality for a fixed amount of exposure.

Clinical competence was defined as the sum of the self-assessed ratings for the three elements of competence (i.e., problem-solving ability, basic procedural skills, and basic medical knowledge). CDI and clinical competence scores were each further divided into three-level variables: low (up to 33rd percentile), middle (up to 67th percentile), and high (greater than 67th percentile). The maximum score within each level of competence was 15, that is, the sum of 5 points maximum for each element of competence (e.g., problem solving, procedural skills, and knowledge).

Quality of care can vary depending on the type of institution [9], and thus it was also incorporated into the analyses. Two types of teaching hospitals were included: university-affiliated and non-university-affiliated (community) teaching hospitals. While this variable was included in the analyses, specific nominal results about this factor are not reported because some hospitals did not consent to revealing type of institution.


Descriptive statistics included rates and proportions for categorical data and means and standard deviations (SD) for continuous data. We first performed univariate analyses to evaluate the relationship between predictor variables and quality of care. Analysis of covariance or pooled t-test was used for categorical data. Pearson or Spearman correlation coefficients were used for continuous data.

Multivariate linear regression models were then constructed to examine the association between clinical exposure and quality of care. We incorporated all predictor variables into the model because all of the variables were thought to be important factors that could potentially be associated with levels of quality of care. We tested the interaction between the amount of clinical exposure and type of institution and self-reports of clinical competence as well as for a case (vignette type) main effect. For all analyses, alpha was set at 0.05. Analyses were done using commercially available software (Intercooled STATA 8.0; STATA Corporation, TX, USA). Ethics approval was granted for this study by the Kyoto University Faculty of Medicine Institutional Review Board.


Of the 375 eligible physicians-in-training, 263 (70.1%) consented to participate, the majority of whom were first (57.8%) and second-year residents (28.5%). The mean number of years after graduation was 1.8 years (range: 1–8). Two thirds of the participants (68.8%) worked in university-affiliated hospitals. The median number of cases reported seen was 210 (range: 10–11400). The proportion of cases seen in inpatient settings was 48.6%, but the variance was large (SD = 27.4%). A third of the participants (34.7%) had seen more than 400 cases overall. See Table 1 for details.

Table 1 Baseline characteristics of participants (n = 263), self-reports of clinical competence, and common disease index (CDI)

Univariate analyses revealed that t-scores were significantly associated with the amount of clinical exposure, type of institution, number of years after graduation, and self-reports of clinical competence, but not with CDI, the proportion of inpatient cases, and self-reports of communication. The mean t-scores between types of institution were significantly different (p < .0001): 55.5 (SD = 2.1) compared to 47.5 (SD = 4.5) (the type of institution is concealed because certain institutions did not consent to revealing their identity). The mean t-scores increased with the number of post-graduation years, amount of exposure, and self-reports of clinical competence. Results of univariate and multivariate analyses are shown in Table 2. CDI was not associated with t-scores (Figure 1).

Table 2 Relationship between predictor variables and quality of care (t-scores) according to univariate and multivariable analyses
Figure 1

Association between Common Disease Index (CDI) and t-scores. Each point indicates the relationship between Common Disease Index (CDI) and the t-score for a given institution. The dotted line depicts the regression line (p = 0.5372).

In the multivariate model, the number of post-graduation years was not statistically significant (p = 0.6942) while the proportion of inpatient cases seen was a statistically significant predictor (p = 0.0095). There was no significant interaction between the amount of clinical exposure and type of institution or levels of competence; consequently the interaction term was not included in the final model. CDI was not associated with t-scores, even after adjusting for the other factors.

The slope of t-scores went up sharply for the first 100 to 200 cases and tapered thereafter. The t-scores were higher for one type of institution and for high competence levels, but the slope of the curves was the same among stratified groups, for given levels of amount of clinical exposure. The t-scores and amount of clinical exposure, stratified by type of institution and levels of self-reports of clinical competence, are graphically shown in Figure 2.

Figure 2

Association between the Number of Cases Seen (clinical exposure) and Quality of Care, Stratified by Type of Institution and Levels of Self-reports of Clinical Competence, Adjusted for Other Factors. Each curve indicates the relationship between the amount of clinical exposure and the t-score for each stratified group. Solid curves (A) indicate one type of institution (either university affiliated or non-university-affiliated teaching hospitals – identity kept confidential); dotted curves (B) indicate other type. Triangles indicate high competence levels; rectangles, middle competence levels; circles, low competence levels.


This study shows that quality of care for physicians-in-training in internal medicine in Japan increased as physicians saw more cases, especially during the initial stages, and tapered off thereafter. These results are consistent with those of Day et al. for PGY1 and PGY3 residents [5]. However, the number of years since graduation, which Day et al. suggested was an important predictor, was not significantly associated with quality of care when clinical exposure was included as a variable.

Although we have found that the overall amount of clinical exposure is an important determinant of the quality of care for physicians-in-training, a related problem is still unsolved: which of the amount of clinical exposure and quality of education is more strongly related to quality of care? To answer this question, we determine the association between the proportion having clinical exposure to certain diseases (CDI) and the quality of care (t-score) for those conditions, instead of evaluating the overall conditions. In the present study, the CDI was not associated with quality of care when adjusting for the overall amount of clinical exposure. Figure 1 illustrates this result. The CDI for institution B was among the lowest, indicating that physicians in that institution saw fewer similar diseases to the vignettes than physicians in other institutions. However, the average t-score for institution B was the second highest among all institutions. Although there is no accepted indicator for teaching, this institution is well known for its excellent teaching, and is the first educational hospital to have started residency training system in Japan; all authors agree that this hospital provides an excellent education. This suggests that physicians could have high quality of care in specific fields even if they had limited clinical exposure, if the quality of their education was excellent. The same point is made with Institution J, also known for its good teaching, which has a high t-score and high CDI.

There was a discrepancy between the effect on quality of care of the overall amount of exposure and the case-specific clinical exposure (CDI). The overall amount of clinical exposure was associated with better quality of care, whereas case-specific clinical exposure was not. A possible reason is that the most important skills to be acquired by internists in their training include basic skills and knowledge: history taking, physical examination, interpersonal skills, competence in continuing care, competence in diagnosis, selection of appropriate diagnostic studies, skills in searching evidence, clinical reasoning skills, and problem-solving skills [11, 12]. These skills and knowledge could be acquired by experience in seeing various diseases, and could be applicable to any type of case seen subsequently, so that the amount of case-specific clinical experience is less important given the same amount of clinical exposure.

The present results indicate that self-reports of clinical competence are significantly associated with quality of care. However, some studies suggest that self-reports of competence are not in agreement with objective measures of clinical skills [13, 14]. A possible explanation is misclassification of self-reports due to the anonymous (blinded) nature of the study. The participants might have considered that their self-assessments could influence their future career, so that they rated themselves higher than their actual performance. This hypothesis is supported by the observation of Woolliscroft et al. that the bottom quartile of medical students, according to objective evaluations, rate themselves higher than others [15]. Some other studies [1416] did not state explicitly whether the self-reported ratings were blinded to evaluators. The anonymous self-reports were a strength of the present study, and this might have helped the participants make more accurate self-assessments with no resulting misclassification bias.

A limitation of the present study is that we did not adjust for medical school provenance or academic performance. Some studies suggest that selection of a medical school may influence practice outcomes [17, 18]. This type of adjustment is important because of variations in the quality of education among schools and also because high achievers prefer to be educated in an institution with acknowledged excellent teaching. This could confuse the relation between type of institution and quality of care. Instead we adjusted for self-reported levels of competence, because we believe that the effect of current (self-assessed) competence was greater than that of past educational experiences.


In summary, the overall amount of clinical exposure (number of cases seen) and levels of self-reports of clinical competence, but not the number of years after graduation, were significantly associated with quality of care, after adjusting for the remaining factors. Quality of education (e.g., the number and quality of the faculty) should be taken into account in future studies. It is possible that a selection bias exists whereby better quality students apply to more highly rated institutions.


  1. 1.

    Ogrinc G, Headrick LA, Mutha S, Coleman MT, O'Donnell J, Miles PV: A framework for teaching medical students and residents about practice-based learning and improvement, synthesized from a literature review. Acad Med. 2003, 78 (7): 748-756. 10.1097/00001888-200307000-00019.

    Article  Google Scholar 

  2. 2.

    Grasberger RC, McMillian TN, Yeston NS, Williams LF, Hirsch EF: Residents' experience in the surgery of trauma. J Trauma. 1986, 26 (9): 848-850.

    Article  Google Scholar 

  3. 3.

    Backous DD, Coker NJ, Jenkins HA: Prospective study of resident-performed stapedectomy. Am J Otol. 1993, 14 (5): 451-454.

    Google Scholar 

  4. 4.

    Tassios PS, Ladas SD, Grammenos I, Demertzis K, Raptis SA: Acquisition of competence in colonoscopy: the learning curve of trainees. Endoscopy. 1999, 31 (9): 702-706. 10.1055/s-1999-146.

    Article  Google Scholar 

  5. 5.

    Day SC, Cook EF, Nesson HR, Wolf MA, Goldman L: A learning-curve approach to the self-assessment of internal medicine training. J Med Educ. 1984, 59 (8): 672-675.

    Google Scholar 

  6. 6.

    Bulgelski BR: The Psychology of Learning Applied to Teaching (Second Edition). 1971, New York , Bobbs-Merrill, 13.

    Google Scholar 

  7. 7.

    Hsu LH, Liu CC, Ko JS: Education and experience improve the performance of transbronchial needle aspiration: a learning curve at a cancer center. Chest. 2004, 125 (2): 532-540. 10.1378/chest.125.2.532.

    Article  Google Scholar 

  8. 8.

    Lohr KN: Medicare: A Strategy for Quality Assurance. 1990, Washington, DC , National Academy Press

    Google Scholar 

  9. 9.

    Peabody JW, Luck J, Glassman P, Dresselhaus TR, Lee M: Comparison of vignettes, standardized patients, and chart abstraction: a prospective validation study of 3 methods for measuring quality. Jama. 2000, 283 (13): 1715-1722. 10.1001/jama.283.13.1715.

    Article  Google Scholar 

  10. 10.

    Akaike H: A new look at the statistial model identification. IEEE Transaction on Autonomic Control. 1974, 19: 716-723. 10.1109/TAC.1974.1100705.

    Article  Google Scholar 

  11. 11.

    Kern DC, Parrino TA, Korst DR: The lasting value of clinical skills. Jama. 1985, 254 (1): 70-76. 10.1001/jama.254.1.70.

    Article  Google Scholar 

  12. 12.

    McPhee SJ, Mitchell TF, Schroeder SA, Perez-Stable EJ, Bindman AB: Training in a primary care internal medicine residency program. The first ten years. Jama. 1987, 258 (11): 1491-1495. 10.1001/jama.258.11.1491.

    Article  Google Scholar 

  13. 13.

    Barnsley L, Lyon PM, Ralston SJ, Hibbert EJ, Cunningham I, Gordon FC, Field MJ: Clinical skills in junior medical officers: a comparison of self-reported confidence and observed competence. Med Educ. 2004, 38 (4): 358-367. 10.1046/j.1365-2923.2004.01773.x.

    Article  Google Scholar 

  14. 14.

    Marel GM, Lyon PM, Barnsley L, Hibbert E, Parise A: Clinical skills in early postgraduate medical trainees: patterns of acquisition of confidence and experience among junior doctors in a university teaching hospital. Med Educ. 2000, 34 (12): 1013-1015. 10.1046/j.1365-2923.2000.00730.x.

    Article  Google Scholar 

  15. 15.

    Woolliscroft JO, TenHaken J, Smith J, Calhoun JG: Medical students' clinical self-assessments: comparisons with external measures of performance and the students' self-assessments of overall performance and effort. Acad Med. 1993, 68 (4): 285-294. 10.1097/00001888-199304000-00016.

    Article  Google Scholar 

  16. 16.

    Luck J, Peabody JW: Using standardised patients to measure physicians' practice: validation study using audio recordings. Bmj. 2002, 325 (7366): 679-10.1136/bmj.325.7366.679.

    Article  Google Scholar 

  17. 17.

    Pearson SA, Rolfe I, Ringland C, Kay-Lambkin F: A comparison of practice outcomes of graduates from traditional and non-traditional medical schools in Australia. Med Educ. 2002, 36 (10): 985-991. 10.1046/j.1365-2923.2002.01281.x.

    Article  Google Scholar 

  18. 18.

    Rolfe IE, Andren JM, Pearson S, Hensley MJ, Gordon JJ: Clinical competence of interns. Programme Evaluation Committee (PEC). Med Educ. 1995, 29 (3): 225-230.

    Article  Google Scholar 

Pre-publication history

  1. The pre-publication history for this paper can be accessed here:

Download references



Author information



Corresponding author

Correspondence to Yasuaki Hayashino.

Additional information

Competing interests

Fukuhara Shunichi was partly supported by a grant-in-aid from the scientific fund of the Ministry of Health, Labor and Welfare in Japan.

Authors' contributions

YH developed the Japanese version of clinical vignettes and coordinated the study with SF, KM, YN and TM. YH, SF, KM, YH, TM and JWP conceived the study, and participated in the study design. YH and DB performed statistical analysis and drafted the initial manuscript for journal submission and participated in revisions. JWP developed the original version of clinical vignettes, and contributed to the evaluation of back-translated Japanese version. YM, YH, KK, YF, JH, TI, US, JS, NH, and HI coordinated data collection at each sites. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Hayashino, Y., Fukuhara, S., Matsui, K. et al. Quality of care associated with number of cases seen and self-reports of clinical competence for Japanese physicians-in-training in internal medicine. BMC Med Educ 6, 33 (2006).

Download citation


  • Chronic Obstructive Pulmonary Disease
  • Standard Deviation
  • Clinical Competence
  • Clinical Exposure
  • Clinical Vignette