Skip to main content

“Seeing inside out”: revealing the effectiveness of otoscopy training in virtual reality enhanced practical exams - a randomized controlled trial

Abstract

Background

The study aimed to assess the impact of different training modalities on otoscopy performance during a practical exam using a high-fidelity simulator and to determine if objective evaluation of otoscopy is feasible using a simulator that records insertion depth and tympanic membrane coverage.

Methods

Participants were assigned to one of four groups: control and three intervention groups with varying training approaches. Participants received otoscopy training and then were assessed through a practical exam on a high-fidelity simulator that uses virtual reality to visualize the ear canal and middle ear. Performance was evaluated using a modified Objective Structured Assessment of Technical Skills checklist and Integrated Procedural Performance Instrument checklist. Insertion depth, tympanic membrane coverage, and correct diagnosis were recorded. Data were tested for normal distribution using the Shapiro-Wilk test. One-way ANOVA and, for non-normally distributed data, Kruskal-Wallis test combined with Dunn’s test for multiple comparisons were used. Interrater reliability was assessed using Cohen’s κ and Intraclass correlation coefficient.

Results

All groups rated their training sessions positively. Performance on the OSATS checklist was similar among groups. IPPI scores indicated comparable patient handling skills. The feedback group examined larger tympanic membrane areas and had higher rates of correct diagnosis. The correct insertion depth was rarely achieved by all participants. Interrater reliability for OSATS was strong. IPPI reliability showed good correlation.

Conclusion

Regardless of training modality, participants perceived learning improvement and skill acquisition. Feedback improved examination performance, indicating simulator-guided training enhances skills. High-fidelity simulator usage in exams provides an objective assessment of performance.

Peer Review reports

Background

Otoscopy is used to diagnose a wide range of ear canal and middle ear diseases. Accurate performance of this procedure is vital for valid diagnoses of otologic diseases, which are common in all age groups and can significantly affect the function of this important sensory organ [1]. Otitis media is one of the most common childhood infections, the leading cause of doctors’ visits by children and the most frequent reason children are prescribed antibiotics or undergo surgery in developed countries [2]. Approximately 25% of primary care complaints are related to otolaryngology, and patients with ear pain often consult a general practitioner or pediatrician. These medical professionals must then examine and evaluate the ear canal and the tympanic membrane [3], making otoscopy a crucial skill not only for otorhinolaryngologists but also for other medical disciplines. However, pediatricians and general practitioners often lack proper training in the clinical assessment of the ear [4,5,6]. Although otoscopy is a crucial procedure in multiple specialties and conditions treated by otolaryngologists account for up to 17% of adult primary care referrals and 50% of pediatric referrals, many medical school curricula in the U.S. do not require a mandatory otolaryngology rotation [3, 7]. Even though an otorhinolaryngology rotation is obligatory during medical school, training is very limited in medical curricula. In addition to didactic lectures and faculty demonstrations, otoscopic examination techniques are often learned and practiced in skills-labs [8, 9]. In peer-to-peer learning, however, there usually are no pathologic findings. The teacher is unable to check whether the student has inserted the instrument correctly, which parts of the tympanic membrane have been examined, and if the structures indicated have actually been recognized. However, to make a correct diagnosis in otoscopy, the otoscope must be inserted to the correct depth and then, due to the otoscope’s limited field of view, the entire ear canal and tympanic membrane must be scanned systematically.

To compensate for these limitations, the use of simulators has increased steadily in the field of medical education in recent years [10], and various simulators for ear examination are available [11,12,13,14,15]. While most current otoscopy simulators are capable of presenting physiological and pathological findings [16], they do not offer accurate depth assessment of the inserted instrument or real-time feedback capability because many simulators still do not allow the instructor to see exactly what their learners are observing. Feedback plays a key role in the acquisition and optimization of clinical skills [17]. Therefore, otoscopy is still a skill difficult to teach [18] and, moreover, remains difficult to assess objectively in the context of a practical examination. The high-fidelity otoscopy simulator investigated in this study [14] displays physiological and pathological findings utilizing virtual reality technology. The instrument’s tip is tracked in three-dimensional space, and a computer determines the corresponding image based on its position relative to the “patient’s” ear. The Image is then displayed on a monitor built into the otoscope. Realistic renderings of the ear canal, the tympanic membrane and middle ear are possible. A training mode enables the user to selectively hide specific structures, like the eardrum, or focus on various areas and structures within the field of view, while being guided during different instructional sessions. In addition, the trainee receives direct feedback on the insertion depth of the instrument. The tutor can follow exactly what the learner sees during the otoscopy on an external monitor and thus provide direct comments and feedback. In addition the simulator records the percentage of the tympanic membrane examined. This feature could be used in a practical examination to objectively evaluate the performance of the actual procedure.

The goal of this study was to determine whether (1) the fidelity of the training tool used to learn otoscopy has an impact on the ability to accurately conduct otoscopy in a practical exam with a high-fidelity simulator and (2) the objective evaluation of otoscopy in a practical exam can be accomplished with the use of a high-fidelity trainer that records the insertion depth of the otoscope and the percentage of the tympanic membrane examined.

Methods

Ethical approval

The study was conducted in accordance with the general terms and conditions and approval of the institutional ethics committee (Ethics committee Freiburg, #21-1646) and was registered in the German Clinical Trials Register on 22/06/2002 under the registration number DRKS00027178. Participation was voluntary and participants were informed of the aims of the study and provided informed consent before participating in the study. Students with previous experience in otoscopy were excluded from the study.

Sample, study design and assignment to study group

173 medical students attending their otolaryngology rotation were recruited to participate in this prospectively designed, cluster-randomized, controlled trial. Participants were trained in otoscopy over the course of three faculty-led instructional sessions followed by a practical exam.

The study design includes three intervention groups (IG) and one control group (CG).

The first skill training session for all groups was structured according to the Peyton’s 4-step approach [19] and included otoscopy training on a peer. Skill training sessions two and three were conducted with a student to teacher ratio of 1:1 and each participant had five minutes to practice otoscopy. During the instructional sessions, the control group practiced otoscopy only on peers. In intervention group one (IG1), a low-fidelity otoscopy trainer (Blue line, Schultes medacta GmbH, Germany) was used to practice otoscopy during training sessions two and three. The low-fidelity trainer is a model of the right half of a head, featuring a silicone molded pinna and auditory canal. Paper strips with distinct tympanic membrane findings can be inserted through an opening on the side. Hair and the natural curvature of the auditory canal in the transition area from the cartilaginous to the bony part are not depicted, and pathologies in the ear canal cannot be visualized. Intervention group two (IG2) practiced with the low-fidelity trainer in session two and were trained on the high-fidelity trainer (Earsi Otoscope, Haag-Streit Group, Germany) in session three. Intervention group three (IG3) practiced with the low-fidelity simulator during the second instructional session and with the high-fidelity trainer during the third session receiving feedback on insertion depth and percentage of tympanic membrane examined. The high-fidelity simulator was used in the practical exam for all groups (Fig. 1). Physiological and pathological tympanic membrane findings were shown in the intervention groups practicing otoscopy on a model.

Fig. 1
figure 1

Study design and study course for control group (CG), intervention group 1 (IG1), intervention group 2 (IG2) and intervention group 3 (IG3). LFT = low fidelity trainer; HFT = high fidelity trainer, Questionnaire (Q) collecting baseline data and skill specific self-assessment; Evaluation of the Skill training sessions (E) on a 6-point Likert scale

Evaluation prior to training (Q)

To account for any confounding variables, each participant gave information on their age, gender, handedness, past clinical experience relating to clerkships, and qualifications as a paramedic or nurse or previous education in another non-healthcare-related profession, to ensure that conditions were the same in all groups. After an introductory lecture on the otolaryngology examination, participants were asked to self-assess their abilities to perform an ear examination correctly.

Evaluation of skills training sessions

After every training session the participants were asked to complete an evaluation form to maximize the comparability of the different training sessions. A 6-point Likert scale with 1 = agree completely to 6 = disagree completely was used to evaluate the following seven statements: (1) The course was clearly structured. (2) The contents of the course were conveyed in an understandable manner. (3) Questions and concerns of the students were addressed. (4) My learning improvement is high. (5) The teachers were engaged. (6) The learning atmosphere was pleasant. (7) I was able to acquire practical skills.

Objective assessment of training

To assess acquired competencies learned during the training sessions, a practical exam was held at the end of the otorhinolaryngology rotation. In the individual exam, participants first had to take an ear-specific medical history with an actor and then examine the ear. Examination was performed on the high-fidelity simulator displaying a pathologic finding matching the anamnesis. Performance of the otoscopy was assessed by two independent raters using a modified Objective Structured Assessment of Technical Skills (OSATS) checklist [14], consisting of the Items (1) the otoscope is held correctly, (2) the otoscope is stabilized, (3) the otoscope is inserted to the correct insertion depth, (4) the auricle is pulled back- and upwards, and (5) the otoscopy is performed atraumatically. In addition, an evaluation of the procedural skills in context was conducted using an Integrated Procedural Performance Instrument (IPPI) checklist [20] by the same two independent raters. The IPPI checklist contained the Items (1) introduced himself/herself to the patient, (2) explained the examination and gave the patient the opportunity to consent, (3) Preparation of the examination, (4) technical performance, (5) being aware of the needs of the patient during examination, (6) professionality (7) overall ability to perform the examination. Time required to examine the tympanic membrane was taken and the percentage of the tympanic membrane examined was recorded. Finally, the participant’s ability to correctly identify the pathological finding was assessed.

Statistical analysis

An a-priori power analysis with the power of 80% was performed with a total sample size of 175 at a 5% significance level assuming a moderate to strong effect.

The data was tested for normal distribution using the Shapiro-Wilk test. In cases where the data was non-normally distributed, the Kruskal-Wallis test was used to compare the data combined with a Dunn’s test in case of multiple comparison. A one-way ANOVA was used to perform a statistical analysis on the normally distributed data. Statistical significance was defined as a p-value less than 0.05. Data are shown as means with standard deviation. Interrater reliability for the OSATS with two raters during the oral-practical exam was assessed using Cohen’s κ, with values between 0.6 and 0.8 considered as strong agreement and > 0.8 as near complete agreement. For the IPPI the intraclass correlation coefficient (ICC) was calculated. Based on the 95% confidence interval of the ICC estimate, values between 0.5 and 0.75, between 0.75 and 0.9, and greater than 0.90 were considered moderate, good, and excellent reliability, respectively [21]. Statistical analyses were performed using SPSS Version 29 (IBM Corp., Armonk, New York; USA).

Results

Participants

The baseline values regarding age, gender, semester, handedness, and education in medical profession are shown in Table 1. None of the participants had received prior teaching in otorhinolaryngology.

Table 1 Group characteristics of the intervention groups 1–3 (IG 1–3) and the control group (CG)

Evaluation of skill training sessions

All training Session were rated on a 6-point Likert scale directly after the session. Overall, the participants rated all training sessions very positively, with a high level of agreement for the different statements describing the courses. The training sessions were rated as clearly structured (≥ 84% agreement). The contents of the course were conveyed in an understandable manner (≥ 91.1% agreement) and questions and concerns of the students were addressed in a pleasant learning atmosphere during all training sessions (≥ 95.3% agreement). With an agreement rate of ≥ 95.8%, all teachers were engaged. Regardless to their group allocation participants rated their learning improvement of each training session as high (≥ 83.3% agreement) and were able to acquire practical skills (≥ 78.3% agreement). Mean values of the results of all three evaluations of the individual groups are shown in Table 2

Table 2 Evaluation of all training sessions. Data presented as mean values (SD) of the three intervention groups (IG1-3) and the control group (CG).

Objective assessment of training

OSATS ratings

Performance of the otoscopy was assessed using a modified OSATS checklist. Despite the insertion depth, the otoscopy procedure on the high-fidelity model was performed very well and comparably by all groups, regardless of their previous otoscopy training. Only the item “otoscope is held correctly“ showed a significant difference between IG1 and the other groups, where there were no significant differences between the other groups and the other Items. The correct insertion depth was rarely achieved in all groups and was often considered to be too shallow. Mean values of the individual items are shown in Fig. 2

Fig. 2
figure 2

Mean values with standard deviation of the individual items of the modified OSATS checklist. Intervention group 1–3 (IG1-3) and control group (CG); (A) Item: the Otoscope is inserted to the correct insertion depth; (B) Item: the otoscope is stabilized; (C) Item: the auricle is pulled back- & upwards; (D) Item: Otoscopy is performed atraumatically; (E) Item: The otoscope is held correctly. **** p < 0.0001; *** p < 0.001

IPPI ratings

The procedural skill in context of the doctor-patient setting during the oral-practical examination was assessed using an IPPI instrument. The items of the IPPI concentrated on the sub-category of communication skills during the procedure, as proposed previously [22], as the technical skills were already covered by the OSATS checklist. All four groups achieved similar scores in the individual Items with no significant differences in “introduced himself/herself to the patient” (IG1: 3.9 (0.5);IG2: 3.6 (0.5); IG3: 3.9 (0.5); CG: 3.8 (0.3), “explained the examination and gave the patient the opportunity to consent” (IG1: 3.8 (0.5); IG2: 3.5 (0.6); IG3: 3.8 (0.5) CG: 3.4 (0.6), “Preparation of the examination”(IG1: 3.9 (0.4); IG2: 3,5(0.6); IG3: 4,0(0.3); CG: 3,6 (0.5), “technical performance” (IG1: 3.8 (0.5); IG2: 3.4 (0.7); IG3: 4.0 (0.2); CG: 3.4 (0.4), “being aware of the needs of the patient during examination” (IG1: 3.6 (0.6); IG2: 3.4 (0.6); IG3: 3.8 (0.4); CG: 3.7 (0.6) and “professionality” (IG1: 3.9 (0.3); IG2: 3.5 (0.1); IG3: 4.0 (0.2); CG: 3.5 (0.3), as shown in Fig. 3

Fig. 3
figure 3

The graphs show the mean and standard deviation of each item on the IPPI checklist. The median and [IQR] of each group are given in parentheses according to the group order shown in the graphs. Intervention group 1–3 (IG1-3) and control group (CG); ns = not significant; (A) Item: introduced himself/herself to the patient (4.0 [0]; 4.0 [1]; 4.0 [0]; 4.0 [0]); (B) Item: explained the examination and gave the patient the opportunity to consent (4.0 [0]; 4.0 [1]; 4.0 [0]; 4.0 [1]); (C) Item: Preparation of the examination (4.0 [0]; 4.0 [1]; 4.0 [0]; 4.0 [0]); (D) Item: technical performance (4.0 [0]; 4.0 [1]; 4.0 [0]; 4.0 [0]); (E) Item: being aware of the needs of the patient during examination (3.5 [0.5]; 3.5 [1]; 4.0 [0.5]; 4.0 [0.5]); (F) Item: professionality (4.0 [0]; 4.0 [1]; 4.0 [0]; 4.0 [0])

Using the Item “overall ability to perform the examination” as a measure for clinical validation, the absolute number of “competent students” (students who received ratings of “1” and “2”), “borderline students” (ratings “3” and “4”), and “incompetent students” (“5” and “6”) was calculated. 148 students were rated as competent, 25 as borderline and 0 as incompetent with no significant differences among groups

Interrater reliability

Performance during the oral-practical exam regarding the procedural skill in context and the practical skill were rated by two independent raters. The interrater reliability for the checklists were very good with a strong or nearly total agreement for the items of the modified OSATS checklist. The individual Cohen’s κ values for the items of the modified OSATS checklist are shown in Table 3. The ICC was calculated for the cumulative IPPI ratings and showed good correlation with an intraclass correlation of 0.867 with a 95% confidence interval between 0.821 and 0.902

Table 3 Interrater reliability for the individual items of the modified OSATS checklist assessed with Cohen’s κ

Examination time and percentage of eardrum assessed

During the practical exam, time required to examine the tympanic membrane and percentage of the tympanic membrane examined were recorded with the examination time representing the time between insertion and removal of the otoscope

The average examination time for all participants was 83 seconds with a minimum examination time of 15 and a maximum of 196 seconds. The mean examination times (in seconds) for the individual groups were IG1 (95 ± 32), IG2 (79 ± 33), IG3 (81 ± 28) and CG (70 ± 25), with a significant longer examination time of IG1 compared to IG2 (p = 0.03) and CG (p = 0.008), but with no significant difference between the CG and the IG2 and IG3 as presented in Fig. 4a

Fig. 4
figure 4

Mean values with standard deviation. Intervention group 1–3 (IG1-3) and control group (CG); ns = not significant; **p < 0.01 * p < 0.05  (A) time required to examine the tympanic membrane; (B) percentage of the tympanic membrane examined; (C) provided the correct diagnosis (0 was assigned for a correct and 1 for an incorrect diagnosis)

The percentage of the examined tympanic membrane was on average 85% with group differences of 88 ± 15% for IG1, 80 ± 18% for IG2, 90 ± 14% for IG3 and 79 ± 21% in the control group. Significant differences were found comparing CG with IG3 (p = 0.007) and IG2 with IG3 (p = 0.003) as presented in Fig. 4b

Finding the right diagnosis

At the end of the examination, participants were asked about their suspected diagnosis based on the collected medical history and examination findings. 85% of the participants were able to provide the correct diagnosis. Participants of IG2 and IG3 were significantly more likely to provide the correct diagnosis as participants of the CG (p = 0.007; p = 0.01), with no difference between IG1 and CG as well as between IG1 and IG2/IG3 as shown in Fig. 4c

Discussion

The present study evaluates the impact of different otoscopy training methods on performance during an exam on a high-fidelity simulator and explores the feasibility of an objective assessment using a high-fidelity trainer that records insertion depth and the percentage of the tympanic membrane examined.

The participating 173 medical students were randomly assigned and shared similar characteristics in terms of their age, gender, semester, handedness, and prior medical education. To ensure a comparable quality of the individual training sessions and thus avoid a potential bias of the performance during the practical examination, the respective training sessions were evaluated by the participants. Participants’ performance on the modified OSATS checklist remained similar across groups. Previous studies have identified otoscopy as an easy-to-learn procedure, with no significant improvement after the third practical teaching session, as measured by instrument handling [18, 23]. In our study, the examination also took place after the third practical lesson. The good results on the OSATS checklist are therefore not surprising and in line with the literature. The significantly poorer performance of IG1 in holding the otoscope could be explained by the fact that this group was trained on the low-fidelity trainer, which only allows examination of the right ear, whereas the high-fidelity trainer simulates examination of the left ear. The correct insertion depth could not be measured in previous studies. As the correct insertion depth of the otoscope is critical for identification of the tympanic membrane, it is interesting to note that it was rarely achieved by the participants and was considered too shallow in 89% of the examinations. Using the IPPI ratings, communication skills during doctor-patient interactions remained consistent across groups, suggesting patient handling was unaffected by training method, which is also reflected in the excellent score in the global item “general ability to perform the examination”, with 85% of the participants being rated competent. Therefore, according to our data, patient handling does not suffer when a skill like otoscopy has been learned on a simulator. Lund et al. were even able to show that teaching of intravenous cannulation skills acquired using simulators in a skills laboratory was superior to bedside teaching when assessed using the IPPI [22]

IG1 took the longest to examine the eardrum, taking an average of 15 seconds longer, without examining a significantly larger area of the tympanic membrane than the other groups. IG2 and IG3 took about the same amount of time, but the feedback group examined the largest part of the eardrum. According to Issenberg, feedback is a key feature of medical simulation that leads to the most effective learning [24]. The higher percentage of tympanic membrane examined could therefore be a result of the feedback IG3 received during their last training session. Although not statistically significant, the highest percentage of correct insertion depths during examination also was achieved by IG3 and possibly more training session with feedback might have yielded a significant difference. The control group took the shortest time for the examination, but also examined the smallest part of the tympanic membrane and was the worst group in determining the correct pathologic finding. Wormald et al. demonstrated in a structured approach to otoscopy training, that the diagnostic ability of study participants tested with photographs of tympanic membranes with chronic otitis media significantly improved after training and concluded, that this teaching approach is likely to be equally beneficial to other otological conditions and to live otoscopy [25]. This statement confirms our observation that IG2 and IG3 were significantly more likely to correctly identify the pathology than the control group, who had never seen a pathological tympanic membrane during their training with peers.

Limitations

This study has limitations, as do all studies. The IG3 group only received feedback during one training session. Previous studies have demonstrated the benefit of repetitive skills training with feedback in the early acquisition of procedural skills [26]. It is therefore reasonable to assume that repeated feedback may improve performance, but this was not investigated in the present study. Repeated feedback may also lead to an overall improvement in the insertion depth of the otoscope.

The study involved three training-sessions with two 5-minute sessions on a simulator to learn otoscopy. While Polk ML [18] identified otoscopy as an easy-to-learn procedure, additional or longer training sessions may have also resulted in better performance. It is also important to mention that the assessment was completed on a high-fidelity simulator that was only accessible to IG2 and IG3 beforehand. This may have given IG2 and IG3 an advantage in recognizing pathologies, despite having comparable results in the OSATS and the IPPI ratings.

The effect on the ability to accurately perform otoscopy in a hands-on examination with a high-fidelity simulator was assessed by the number of correct items in the OSATS. For this purpose, a power analysis was performed a-prior to the study with a total sample size of 175 participants at a 5% significance level. Assuming a moderate to strong effect size of d = 0.75, the power was 80%. This leaves the study with a 20% probability of a second type error.

Conclusion

Our study shows that, regardless of the training modality, participants perceived substantial learning improvement and an acquisition of a practical skill. Otoscopy appears to be effectively learned through various teaching methods. The feedback on the high-fidelity simulator notably enhances examination quality in terms of tympanic membrane assessment and accurate diagnosis of pathological findings. The use of a high-fidelity simulator for otoscopy in a practical exam is possible, even without prior training, without compromising the performance. Incorporating a high-fidelity simulator in exams ensures consistent testing conditions and more realism through pathological findings. It allows objective evaluation of performance by measuring the insertion depth and the percentage of the tympanic membrane examined. With the overall poor performance of all groups with respect to insertion depth, this study suggests that otoscopy seems to be a simple examination, the correct procedure may not be as easy to learn as previously assumed by other authors

Data availability

All relevant raw data from this study will be available from the corresponding author upon reasonable request for non-commercial purposes.

References

  1. Vos T, Flaxman AD, Naghavi M, Lozano R, Michaud C, Ezzati M, Shibuya K, Salomon JA, Abdalla S, Aboyans V, et al. Years lived with disability (YLDs) for 1160 sequelae of 289 diseases and injuries 1990–2010: a systematic analysis for the global burden of Disease Study 2010. Lancet. 2012;380(9859):2163–96.

    Article  Google Scholar 

  2. Rovers MM, Schilder AG, Zielhuis GA, Rosenfeld RM. Otitis media. Lancet. 2004;363(9407):465–73.

    Article  Google Scholar 

  3. Ishman SL, Stewart CM, Senser E, Stewart RW, Stanley J, Stierer KD, Benke JR, Kern DE. Qualitative synthesis and systematic review of otolaryngology in undergraduate medical education. Laryngoscope. 2015;125(12):2695–708.

    Article  Google Scholar 

  4. Fisher EW, Pfleiderer AG. Assessment of the otoscopic skills of general practitioners and medical students: is there room for improvement? Br J Gen Pract. 1992;42(355):65–7.

    Google Scholar 

  5. Pichichero ME, Poole MD. Assessing diagnostic accuracy and tympanocentesis skills in the management of otitis media. Arch Pediatr Adolesc Med. 2001;155(10):1137–42.

    Article  Google Scholar 

  6. Steinbach WJ, Sectish TC, Benjamin DK Jr., Chang KW, Messner AH. Pediatric residents’ clinical diagnostic accuracy of otitis media. Pediatrics. 2002;109(6):993–8.

    Article  Google Scholar 

  7. Boscoe EF, Cabrera-Muffly C. Otolaryngology in the medical school curriculum: current trends in the United States. Laryngoscope. 2017;127(2):346–8.

    Article  Google Scholar 

  8. Bugaj TJ, Nikendei C. Practical clinical training in skills labs: theory and practice. GMS J Med Educ. 2016;33(4):Doc63.

    Google Scholar 

  9. Herrmann-Werner A, Nikendei C, Keifenheim K, Bosse HM, Lund F, Wagner R, Celebi N, Zipfel S, Weyrich P. Best practice skills lab training vs. a see one, do one approach in undergraduate medical education: an RCT on students’ long-term ability to perform procedural clinical skills. PLoS ONE. 2013;8(9):e76354.

    Article  Google Scholar 

  10. Chipman JG, Schmitz CC. Using objective structured assessment of technical skills to evaluate a basic skills simulation curriculum for first-year surgical residents. J Am Coll Surg. 2009;209(3):364–e370362.

    Article  Google Scholar 

  11. Wickens B, Lewis J, Morris DP, Husein M, Ladak HM, Agrawal SK. Face and content validity of a novel, web-based otoscopy simulator for medical education. J Otolaryngol Head Neck Surg. 2015;44:7.

    Article  Google Scholar 

  12. Morris E, Kesser BW, Peirce-Cottler S, Keeley M. Development and validation of a novel ear simulator to teach pneumatic otoscopy. Simul Healthc. 2012;7(1):22–6.

    Article  Google Scholar 

  13. Davies J, Djelic L, Campisi P, Forte V, Chiodo A. Otoscopy simulation training in a classroom setting: a novel approach to teaching otoscopy to medical students. Laryngoscope. 2014;124(11):2594–7.

    Article  Google Scholar 

  14. Albrecht T, Nikendei C, Praetorius M. Face, Content, and Construct Validity of a virtual reality Otoscopy Simulator and Applicability to Medical Training. Otolaryngol Head Neck Surg 2021:1945998211032897.

  15. Xu J, Campisi P, Forte V, Carrillo B, Vescan A, Brydges R. Effectiveness of discovery learning using a mobile otoscopy simulator on knowledge acquisition and retention in medical students: a randomized controlled trial. J Otolaryngol Head Neck Surg. 2018;47(1):70.

    Article  Google Scholar 

  16. Thone N, Winter M, Garcia-Matte RJ, Gonzalez C. Simulation in Otolaryngology: a teaching and training tool. Acta Otorrinolaringol Esp. 2017;68(2):115–20.

    Article  Google Scholar 

  17. Burgess A, van Diggele C, Roberts C, Mellis C. Feedback in the clinical setting. BMC Med Educ. 2020;20(Suppl 2):460.

    Article  Google Scholar 

  18. Polk ML, Lailach S, Kemper M, Bendas A, Zahnert T, Neudert M. [Learning curve of the ENT mirror examination: goal-oriented curriculum planning for psychomotor learning]. HNO. 2020;68(4):238–47.

    Article  Google Scholar 

  19. Krautter M, Dittrich R, Safi A, Krautter J, Maatouk I, Moeltner A, Herzog W, Nikendei C. Peyton’s four-step approach: differential effects of single instructional steps on procedural and memory performance - a clarification study. Adv Med Educ Pract. 2015;6:399–406.

    Article  Google Scholar 

  20. Kneebone R, Nestel D, Yadollahi F, Brown R, Nolan C, Durack J, Brenton H, Moulton C, Archer J, Darzi A. Assessing procedural skills in context: exploring the feasibility of an Integrated Procedural Performance Instrument (IPPI). Med Educ. 2006;40(11):1105–14.

    Article  Google Scholar 

  21. Koo TK, Li MY. A Guideline of selecting and reporting Intraclass correlation coefficients for Reliability Research. J Chiropr Med. 2016;15(2):155–63.

    Article  Google Scholar 

  22. Lund F, Schultz JH, Maatouk I, Krautter M, Moltner A, Werner A, Weyrich P, Junger J, Nikendei C. Effectiveness of IV cannulation skills laboratory training and its transfer into clinical practice: a randomized, controlled trial. PLoS ONE. 2012;7(3):e32831.

    Article  Google Scholar 

  23. Krauss F, Giesler M, Offergeld C. [On the effectiveness of digital teaching of practical skills in curricular ENT education]. HNO. 2022;70(4):287–94.

    Article  Google Scholar 

  24. Issenberg SB, McGaghie WC, Petrusa ER, Lee Gordon D, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005;27(1):10–28.

    Article  Google Scholar 

  25. Wormald PJ, Browning GG, Robinson K. Is otoscopy reliable? A structured teaching method to improve otoscopic accuracy in trainees. Clin Otolaryngol Allied Sci. 1995;20(1):63–7.

    Article  Google Scholar 

  26. Bosse HM, Mohr J, Buss B, Krautter M, Weyrich P, Herzog W, Junger J, Nikendei C. The benefit of repetitive skills training and frequency of expert feedback in the early acquisition of procedural skills. BMC Med Educ. 2015;15:22.

    Article  Google Scholar 

Download references

Funding

There was no financial support or funding for the conducted study. All authors and participants did not receive any financial benefits regarding the study or the simulator.

Open Access funding enabled and organized by Projekt DEAL.

Author information

Authors and Affiliations

Authors

Contributions

T. A.: designed the study and acquired the data; did the data analysis, drafted the manuscript; acknowledged the drafting of the article and approved the version to be published. Agrees to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. N. F.: acquired data and helped drafting the manuscript, acknowledged the drafting of the article and approved the version to be published. Agrees to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. W. R.: helped in interpreting the data and helped drafting the manuscript, acknowledged the drafting of the article and approved the version to be published. Agrees to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. C. N.: helped in interpreting the data and helped drafting the manuscript, acknowledged the drafting of the article and approved the version to be published. Agrees to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. C. O.: designed the study and acquired data, helped drafting the manuscript, acknowledged the drafting of the article and approved the version to be published. Agrees to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Corresponding author

Correspondence to Tobias Albrecht.

Ethics declarations

Ethics approval and consent to participate

The study was conducted in accordance with the general terms and conditions and approval of the institutional ethics committee (Ethics committee Freiburg, #21-1646) and was registered in the German Clinical Trials Register on 22/06/2002 under the registration number DRKS00027178. Participation was voluntary and participants were informed of the aims of the study and provided informed consent before participating in the study. Students with previous experience in otoscopy were excluded from the study.

Informed consent

Not applicable.

Conflict of interests

All authors do not have any conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Albrecht, T., Fehre, N., Ramackers, W. et al. “Seeing inside out”: revealing the effectiveness of otoscopy training in virtual reality enhanced practical exams - a randomized controlled trial. BMC Med Educ 24, 439 (2024). https://doi.org/10.1186/s12909-024-05385-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12909-024-05385-3

Keywords