- Open Access
Translating and validating a Japanese version of the instrument for patient assessment of medical professionalism (J-IPAMP): a cross-sectional survey
BMC Medical Education volume 22, Article number: 641 (2022)
Although there are many tools to assess medical professionalism, they rarely address patients’ perspectives. The instrument for patient assessment of medical professionalism (IPAMP) comprises 11 items and has been established and validated as a valuable tool for assessing trainees’ professionalism from the patient’s perspective. However, there is no instrument to assess professionalism from the patient’s perspective in Japan. The purpose of the present study was to develop a Japanese version of the IPAMP (J-IPAMP) and test its validity and reliability.
We conducted a cross-sectional survey to examine the reliability and validity of the J-IPAMP in two hospitals (one each in an urban and rural area) in Japan. Receptionists or surveyors distributed the anonymous questionnaire to 276 inpatients; all participants were aged above 20 years and assigned to medical trainees. We evaluated its structural and criterion-related validity, as well as internal consistency reliability.
Data of 235 (85.1%) patients were analyzed. Using the split-half validation technique, we performed an exploratory factor analysis (EFA) along with a confirmatory factor analysis (CFA). The EFA showed a one-factor solution. Then, to compare the model fitness between two models (the two-factor model from the original English version vs. unidimensional model suggested by the EFA), the CFA was performed. The CFA showed that almost all of the fit indices met their respective criteria and were approximately the same for the two models. Thus, we adopted a single-factor model. The Pearson correlation coefficients between the total J-IPAMP scores and the global ratings were 0.738, indicating adequate criterion-related validity. The Cronbach’s alpha of the 11 items of the instrument was 0.96 (95% confidence interval: 0.96–0.97) and the omega value was 0.96, demonstrating acceptable internal consistency reliability.
We developed the Japanese version of the IPAMP. Its validity and reliability were verified through analysis. This instrument can be utilized for professionalism education in the postgraduate training setting.
In recent years, the importance of medical professionalism has been increasingly highlighted by scholars [1,2,3]. Nevertheless, it remains one of the most intangible and challenging concepts in medical education . In 1999, it was listed as one of the six general competencies that physicians-in-training must develop, with several foundations in Europe and the United States collaborating to launch the Medical Professionalism Project in the same year. Afterwards, in 2002, the Physician’s Charter on Medical Professionalism was published [4, 5], which is now endorsed by more than 100 professional associations worldwide . Thus, in the field of medical education, medical professionalism has received growing attention over the years.
The assessment of medical professionalism is important because it: (1) promotes learning, (2) facilitates the formation of a professional identity, (3) is vital in the quality improvement cycle, and (4) is crucial to identifying and remediating learners whose performance may endanger patients [7,8,9]. Although there are many tools to assess medical professionalism [10, 11], patients’ perspectives regarding the assessment of such professionalism have rarely been addressed . Meanwhile, in other areas of medicine, such as patient safety and quality improvement, great strides have been made to incorporate patients’ perspectives, such as by assessing patient feedback . Because patients are the key stakeholders for medical professionals , research on medical professionalism assessment using patients’ lived experiences is a matter of high priority [12, 15]. Wilkinson et al., in a recent systematic review aimed at producing a blueprint to assess medical professionalism , described the following specific physician behaviors that could be appropriately assessed by patients: respecting the diversity of patients, showing politeness and patience, having an empathetic and compassionate attitude, and engaging with patients in shared decision making . As these behaviors are difficult to evaluate in official assessment settings, patients’ opinions regarding these behaviors are crucial in that they may complement other sources of information (e.g., paper-based test, simulation, and global view of supervisor).
In 2020, Ratelle et al. developed the Instrument for Patient Assessment of Medical Professionalism (IPAMP), which can be used to rate internal medicine trainees’ professionalism . The assessment method was adapted from Dine et al. and the American Board of Internal Medicine [17,18,19], which were both well-validated for reliability and validity analysis. Ratelle et al. iteratively refined and validated the instrument. In Japan, assessing the medical professionalism of trainees is very important and a matter of great interest among researchers in this field . However, no tools are available for assessing trainees’ professionalism from the patients’ perspectives in Japan. Therefore, we developed the Japanese version of the IPAMP (J-IPAMP) and examined its validity and reliability in a multicentered, cross-sectional study.
Study design, setting, and participants
The IPAMP was subjected to cross-cultural translation and adaptation into Japanese. We conducted a cross-sectional survey to examine its psychometric properties. We contacted two postgraduate clinical training hospitals (Oji Seikyo Hospital and Suwa Central Hospital) in Japan, and both agreed to cooperate. Oji Seikyo Hospital is located in an urban area and has 159 beds, whereas Suwa Central Hospital is in a rural area and has 360 beds. From September 2021 to March 2022, we distributed an anonymous self-administered version of the J-IPAMP to potential participants. The eligibility criteria for patients were: aged 20 years and above, admitted to one of these two hospitals, and assigned to clinical trainees (postgraduate years 1–5) during the survey period. All voluntarily agreed to join this study beforehand. Patients who were expected to find it difficult to answer the paper-based questionnaire on their own because of severe physical (e.g., fracture of the dominant arm) or mental (e.g., severe dementia) disorders, based on hospitalization observations by hospital staff, were excluded from the study. The questionnaires were delivered by the receptionists or surveyors at each institution. The respondents filled out the questionnaire, put it in an envelope, and dropped it in the collection box at each institution.
The original IPAMP has 11 items, each of which is rated on a 5-point Likert scale (1 = poor, 2 = fair, 3 = good, 4 = very good, and 5 = excellent). Factor analysis showed a two-factor model comprising “involvement and respect” (Q5, Q6, and Q8–Q11) and “compassion and rapport” (Q1–Q4 and Q7) .
After obtaining permission from the original author, we translated the IPAMP into Japanese following an international guideline for the cross-cultural adaptation process . First, three forward translations from English to Japanese were conducted independently by three translators (HF, DS, and KK). All three were fluent in English, familiar with the cultures of healthcare in which both languages are used, and had professional experience translating questionnaires in the field of medical education . Second, the translators worked together to synthesize and refine the Japanese translations (Version 1). Third, we requested professional bilingual translators who had no prior knowledge of the questionnaire to back translate Version 1 into English. We compared the back-translated version with the original English version, proofread it, and produced Version 2. Fourth, we asked an expert in medical professionalism (YT) to review Version 2 and modified it based on the feedback (Version 3). Fifth, we asked the original author to review Version 3 to ensure that there were no problems with the translation process. Finally, from July to August 2021, a pilot test was conducted on 11 inpatients at Suwa Central Hospital. The author (HF) interviewed the inpatients on whether Version 3 was intelligible and understood as intended. As the pilot test revealed that there were no problematic items in the translation processes, Version 3 was considered final. The tool’s face and content validity were ensured by all authors.
The structural validity of the J-IPAMP was tested through both exploratory factor analysis (EFA) and confirmatory factor analysis (CFA). We applied the split-half validation approach and randomly split the sample into two independent groups. As this study sought to develop a scale that is optimized for the Japanese healthcare context, we decided to perform the EFA first.
Before the EFA, the Kaiser–Meyer–Olkin (KMO) measure of sampling adequacy and Bartlett’s test of sphericity were performed to assess the suitability of the data for performing EFA. A KMO value greater than 0.60  and a significant Bartlett’s test (p < 0.05) indicates suitability for EFA. EFA was conducted on half of the dataset using the maximum likelihood with promax rotation method. For factor extraction, Kaiser-Guttman Criterion (Eigenvalues greater than 1) and parallel analysis were employed . A cut-off value of 0.40 was adopted for factor loadings.
We conducted CFA using the maximum likelihood estimation approaches on the other half of the data to assess the suitability of the original two-factor model and identify an alternative model. The model fitness of the data was determined by calculating goodness-of-fit indices. We employed the following criteria to assess the model fitness: comparative fit index (CFI) close to 0.90 or higher, Tucker–Lewis index (TLI) close to 0.90 or higher, root mean square error of approximation (RMSEA) close to 0.08 or below, and standardized root mean square residual (SRMR) close to 0.08 or below [23,24,25].
We used the J-IPAMP total scores and the global rating to examine criterion-related validity. The question for the global rating was as follows: “Using any number from 0 to 10, where 0 is the worst doctor possible and 10 is the best doctor possible, what number would you use to rate this doctor as a professional doctor?” Criterion-related validity was assessed using Pearson correlation coefficients between the J-IPAMP total scores and the global rating. Correlation coefficients were considered meaningful if they were above 0.30 . Finally, we utilized Cronbach’s alpha and omega coefficients to examine the internal consistency reliability of the scale. A value of above 0.70 for both coefficients is considered satisfactory reliability [27, 28]. We chose a complete case analysis method because of the small amount of missing data. We analyzed the data using R version 4.2.1 (R Foundation for Statistical Computing, Vienna, Austria; www.R-project.org). We used psych version 2.2.5 and GPArotation version 2022.4–1 to perform EFA and lavaan version 0.6–12 and semPlot version 1.1.6 to conduct CFA [29,30,31,32].
All participants gave their oral consent before the study began. We obtained ethics approval from the Institutional Review Board of the University of Tokyo (2021074NI).
Of the 276 eligible participants, 253 (91.7%) completed the questionnaire. We excluded 18 responses because of missing data. Thus, we included 235 (85.1%) patients in the analysis (Fig. 1). Table 1 summarizes the participants’ characteristics, and Table 2 shows the participants’ responses to each item of the J-IPAMP.
First, we performed EFA on 117 participants to assess the structural validity of the J-IPAMP. The KMO value was 0.945, which exceeded the 0.60 threshold. Bartlett’s test was significant with p < 0.001 (χ2 = 1432.036, df = 55). These results indicated that the data were appropriate for EFA. Table 3 presents the results of the EFA. Both the Kaiser-Guttman Criterion and parallel analysis showed that a one-factor structure for all 11 items fit the data best, and all factor loadings were more than 0.40. The one-factor model explained 73.7% of the variance.
We then conducted CFA on the remaining 118 participants for the following two models: the two-factor model from the original English version and a unidimensional model suggested by EFA. The goodness-of-fit results are presented in Table 4. All the fit indices, except RMSEA, met their respective criteria and were approximately the same for both models. Considering the results of the factor analysis, we decided to adopt a one-factor model for the following two reasons. First, the parallel analysis, a robust method for determining the optimal number of factors, in the EFA demonstrated that a one-factor structure was optimal. Second, the results of the inter-factor correlation in the CFA for two-factor models was very high at 0.926. Figure 2 shows the path diagram for the CFA of the J-IPAMP (the final one-factor model).
To examine criterion-related validity, we calculated the Pearson correlation coefficients between the J-IPAMP total scores and the global rating. The Pearson correlation coefficient was 0.738, which exceeded the 0.30 criterion. The Cronbach’s alpha value of the 11 items was 0.96 (95% confidence interval: 0.96–0.97) and omega value was 0.96, indicating acceptable internal consistency reliability. Thus, we obtained the final version of the J-IPAMP (Table 5).
The English version of the IPAMP was translated into Japanese and examined in various aspects. The J-IPAMP showed good structural and criterion-related validity and internal consistency reliability. To the best of our knowledge, the J-IPAMP is the first validated scale that allows for the measuring of medical professionalism from the patient’s perspective in Japan.
This study has a number of methodological strengths. First, it followed a robust translation methodology. In keeping with international guidelines , we adapted the original English survey into Japanese using the following procedure: translation, synthesis, back translation, review by an expert in medical professionalism, and pilot testing. This robust methodology strengthened the validity of our findings. Second, the response rate was very high, at 85.1%. Although there is no clear definition of a high response rate, it is generally considered excellent when it exceeds 80% . Obtaining a high response rate is essential for high-quality research that produces results with reliability, validity, and generalizability [33,34,35]. However, according to the National Research Council, response rates have continued to decline over time . Research using patient-reported outcome measures has often encountered problems in recent years with low response rates owing to the lack of interest or time . Our study achieved a high response rate, which diminished sampling bias concerns.
The original English version of the IPAMP has two factors (“involvement and respect” and “compassion and rapport”) , whereas an EFA of the J-IPAMP identified only one factor. This may have been the case for three reasons. First, there may have been a problem in the EFA of the original study, which concluded that EFA resulted in a two-factor structure despite the difference in factor loadings between the two factors having been small for all 11 items. It is possible that the original English version, similar to our Japanese version, had a one- rather than a two-factor structure. Second, there could be a translation issue. The original meaning of the instrument items may have changed slightly through the translation processes, which may have affected the respondents’ interpretation of the Japanese version. Third, cultural differences may have affected the results. Japan is generally considered a high-context culture , where behaviors and norms rely heavily on non-explicit rules and non-verbal cues that group members are expected to know implicitly . In Japanese high-context communication, most information is withheld by the speaker, and little is in the explicit part of the message, whereas most information is vested in explicit expressions in low-context communication in the United States . These cultural differences can lead to variations in expectations and assumptions for physician–patient communication. Ruhnke et al. noted that open discussions between doctors and patients are more limited in Japan than in the United States . A comparative quantitative analysis of doctor–patient communication by Ohtaki et al. showed that there were differences between medical encounters in the United States and Japan in terms of interview time, proportion of time spent on various tasks, and use of back-channel responses . Thus, the differences in physician–patient communication styles between both countries may have had an impact on the responses to the survey.
This is the first Japanese version of the IPAMP. To the best of our knowledge, almost all previously developed tools for measuring physician professionalism were responded by healthcare professionals, whereas the IPAMP is responded by patients. In this era of patient-centered medicine, it is crucial to use the IPAMP, which incorporates the patient’s voice, in medical education. The J-IPAMP can be used to educate trainees in medical professionalism based on lived assessments by hospitalized patients in Japan. The development of versions in other languages would be worthwhile and welcome in order to promote international research.
This study had some limitations. First, the fit indices were not fully acceptable. A CFA should be conducted again in future studies when the data are fully accumulated. Further research to validate the fit indices would robust our tool, the J-IPAMP. Second, in this study, although the structural and criterion-related validity and internal consistency reliability of the scale were verified, we did not examine other psychometric properties (e.g., convergent and discriminant validity and test–retest reliability). Further validation of these properties can strengthen the developed tool. Third, the number of hospitals where the validation survey was conducted was relatively small. A larger multi-institute study would strengthen the instrument further. Fourth, as few studies have been found in terms of the assessment of physician professionalism by patients, it was difficult to compare our results with those of existing studies. However, this also underscores the novelty of the present study. Finally, we employed a global rating rather than selecting a specific scale to examine criterion-related validity.
We translated the original English version of the IPAMP into Japanese, and then examined its structural and criterion-related validity and internal consistency reliability. Although further research is required for re-examining its structural validity, this instrument can be utilized for educating trainees on professionalism based on patients’ perspectives.
Availability of data and materials
The data sets generated and analyzed in this study are available from the corresponding author on reasonable request.
Confirmatory factor analysis
Comparative fit index
Exploratory factor analysis
Instrument for patient assessment of medical professionalism
Root mean square error of approximation
Standardized root mean square residual
O’Sullivan H, van Mook W, Fewtrell R, Wass V. Integrating professionalism into the curriculum: AMEE Guide No. 61. Med Teach. 2012;34(2):e64-77.
Birden H, Glass N, Wilson I, Harrison M, Usherwood T, Nass D. Teaching professionalism in medical education: A Best Evidence Medical Education. (BEME) systematic review BEME Guide No. 25. Med Teach. 2013;35(7):e1252-66.
Bajwa NM, Nendaz MR, Posfay-Barbe KM, Yudkowsky R, Park YS. A meaningful and actionable professionalism assessment: validity evidence for the Professionalism Mini-Evaluation Exercise (P-MEX) across 8 years. Acad Med. 2021;96(11S):S151–7.
ABIM (American Board of Internal Medicine Foundation), ACP-ASIM (American College of Physicians-American Society of Internal Medicine) Foundation, European Federation of Internal Medicine. Medical professionalism in the new millennium: a physician charter. Ann Intern Med. 2002;136(3):243–6.
Medical Professionalism Project. Medical professionalism in the new millennium: a physicians’ charter. Lancet. 2002;359(9305):520–2.
ABIM Foundation. Endorsements of The Charter. https://abimfoundation.org/what-we-do/physician-charter/endorsements-of-the-charter. Accessed 25 Apr 2022.
Cruess RL, Cruess SR, Steinert Y. Teaching medical professionalism: supporting the development of a professional Identity. 2nd ed. Cambridge: Cambridge University Press; 2016.
Goldie J. Assessment of professionalism: a consolidation of current thinking. Med Teach. 2013;35(2):e952–6.
Tay KT, Ng S, Hee JM, Chia EWY, Vythilingam D, Ong YT, et al. Assessing professionalism in medicine – a scoping review of assessment tools from 1990 to 2018. J Med Educ Curric Dev. 2020;7(1):2382120520955159.
Veloski JJ, Fields SK, Boex JR, Blank LL. Measuring professionalism: a review of studies with instruments reported in the literature between 1982 and 2002. Acad Med. 2005;80(4):366–70.
Li H, Ding N, Zhang Y, Liu Y, Wen D. Assessing medical professionalism: a systematic review of instruments and their measurement properties. PLoS ONE. 2017;12(5):e0177321.
Hodges B, Paul R, Ginsburg S, The Ottawa Consensus Group Members. Assessment of professionalism: from where have we come - to where are we going? An update from the Ottawa Consensus Group on the assessment of professionalism. Med Teach. 2019;41(3):249–55.
Ward K, Stanyon M, Ryan K, Dave S. Power, recovery and doing something worthwhile: A thematic analysis of expert patient perspectives in psychiatry education. Health Expect. 2022;25(2):549–57.
Wilkinson TJ, Wade WB, Knock LD. A blueprint to assess professionalism: results of a systematic review. Acad Med. 2009;84(5):551–8.
Pirjani R, Rabiei M, Alipour S, Hosseini R, Shahvari Z. Promoting professionalism in practice: using patients’ lived experiences. Med Educ. 2019;53(5):505.
Ratelle JT, Halvorsen AJ, Mandrekar J, Sawatsky AP, Reed DA, Beckman TJ. Internal medicine resident professionalism assessments: exploring the association with patients’ overall satisfaction with their hospital stay. Acad Med. 2020;95(6):902–10.
McLeod PJ, Tamblyn R, Benaroya S, Snell L. Faculty ratings of resident humanism predict patient satisfaction ratings in ambulatory medical clinics. J Gen Intern Med. 1994;9(6):321–6.
Dine CJ, Ruffolo S, Lapin J, Shea JA, Kogan JR. Feasibility and validation of real-time patient evaluations of internal medicine interns’ communication and professionalism skills. J Grad Med Educ. 2014;6(1):71–7.
Tamblyn R, Benaroya S, Snell L, McLeod P, Schnarch B, Abrahamowicz M. The feasibility and value of using patient satisfaction ratings to evaluate internal medicine residents. J Gen Intern Med. 1994;9(3):146–52.
Tsugawa Y, Tokuda Y, Ohbu S, Okubo T, Cruess R, Cruess S, et al. Professionalism Mini-Evaluation Exercise for medical residents in Japan: a pilot study. Med Educ. 2009;43(10):968–78.
Beaton DE, Bombardier C, Guillemin F, Ferraz MB. Guidelines for the process of cross-cultural adaptation of self-report measures. Spine (Phila Pa 1976). 2000;25(24):3186–91.
Fujikawa H, Son D, Kondo K, Djulbegovic M, Takemura Y, Eto M. Translating and validating a Japanese version of the Patient Care Ownership Scale: a multicenter cross-sectional study. BMC Med Educ. 2021;21(1):415.
Tabachnick BG, Fidell LS. Using multivariate statistics. 7th ed. New York: Pearson; 2019.
Vandenberg RJ, Lance CE. A review and synthesis of the measurement invariance literature: suggestions, practices, and recommendations for organizational research. Organ Res Methods. 2000;3(1):4–70.
Brown TA. Confirmatory factor analysis for applied research. 2nd ed. New York: The Guilford Press; 2015.
Cohen J. Statistical Power Analysis. Curr Dir Psychol Sci. 1992;1(3):98–101.
Hair JF, Black WC, Babin BJ, Anderson RE. Multivariate data analysis. 8th ed. Andover: Cengage Learning EMEA; 2019.
DeVellis RF. Scale development: theory and applications. 4th ed. Los Angeles: Sage; 2017.
CRAN. Package ‘psych’. https://cran.r-project.org/web/packages/psych/psych.pdf. Accessed 11 Aug 2022.
CRAN. Package ‘GPArotation’. https://cran.r-project.org/web/packages/GPArotation/GPArotation.pdf. Accessed 11 Aug 2022.
CRAN. Package ‘lavaan’. https://cran.r-project.org/web/packages/lavaan/lavaan.pdf. Accessed 11 Aug 2022.
CRAN. Package ‘semPlot’. https://cran.r-project.org/web/packages/semPlot/semPlot.pdf. Accessed 11 Aug 2022.
Booker QS, Austin JD, Balasubramanian BA. Survey strategies to increase participant response rates in primary care research studies. Fam Pract. 2021;38(5):699–702.
Andres L. Administration of surveys and enhancing response rates. In: Andres L, editor. Designing & Doing Survey Research. Los Angeles: SAGE Publications Ltd; 2012. p. 129–48.
Phillips AW, Reddy S, Durning SJ. Improving response rates and evaluating nonresponse bias in surveys: AMEE Guide No. 102. Med Teach. 2016;38(3):217–28.
National Research Council. Nonresponse in social science surveys: a research agenda. Washington: The National Academies Press; 2013.
Ho A, Purdie C, Tirosh O, Tran P. Improving the response rate of patient-reported outcome measures in an Australian tertiary metropolitan hospital. Patient Relat Outcome Meas. 2019;10(1):217–26.
Hall ET. Beyond culture. New York: Doubleday; 1976.
Hooker J. Working across cultures. Stanford: Stanford Business Books; 2003.
Ohtaki S, Ohtaki T, Fetters MD. Doctor–patient communication: a comparison of the USA and Japan. Fam Pract. 2003;20(3):276–82.
Ruhnke GW, Wilson SR, Akamatsu T, Kinoue T, Takashima Y, Goldstein MK, et al. Ethical decision making and patient autonomy: a comparison of physicians and patients in Japan and the United States. Chest. 2000;118(4):1172–82.
The authors wish to thank all participants. The authors would also like to thank the receptionists at Oji Seikyo Hospital and Suwa Central Hospital for their support with the distribution of the questionnaires. Further, the authors thank Editage for English language editing and proofreading.
This study did not receive any grant from any funding agency in the public, commercial, and not-for-profit sectors.
Ethics approval and consent to participate
This study was performed in keeping with the Declaration of Helsinki. All procedures were performed in keeping with relevant guidelines. All participants received explanations of the questionnaire and were notified that their informed consent would be confirmed if they completed and submitted the questionnaire. This procedure was approved by the Institutional Review Board of the University of Tokyo (IRB approval number: 2021074NI).
Consent for publication
The authors declare no conflict of interest in association with the present study.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Fujikawa, H., Son, D., Aoki, T. et al. Translating and validating a Japanese version of the instrument for patient assessment of medical professionalism (J-IPAMP): a cross-sectional survey. BMC Med Educ 22, 641 (2022). https://doi.org/10.1186/s12909-022-03699-8
- Medical professionalism
- Scale development