Skip to main content

The impact of artificial intelligence on clinical education: perceptions of postgraduate trainee doctors in London (UK) and recommendations for trainers

Abstract

Background

Artificial intelligence (AI) technologies are increasingly used in clinical practice. Although there is robust evidence that AI innovations can improve patient care, reduce clinicians’ workload and increase efficiency, their impact on medical training and education remains unclear.

Methods

A survey of trainee doctors’ perceived impact of AI technologies on clinical training and education was conducted at UK NHS postgraduate centers in London between October and December 2020. Impact assessment mirrored domains in training curricula such as ‘clinical judgement’, ‘practical skills’ and ‘research and quality improvement skills’. Significance between Likert-type data was analysed using Fisher’s exact test. Response variations between clinical specialities were analysed using k-modes clustering. Free-text responses were analysed by thematic analysis.

Results

Two hundred ten doctors responded to the survey (response rate 72%). The majority (58%) perceived an overall positive impact of AI technologies on their training and education. Respondents agreed that AI would reduce clinical workload (62%) and improve research and audit training (68%). Trainees were skeptical that it would improve clinical judgement (46% agree, p = 0.12) and practical skills training (32% agree, p < 0.01). The majority reported insufficient AI training in their current curricula (92%), and supported having more formal AI training (81%).

Conclusions

Trainee doctors have an overall positive perception of AI technologies’ impact on clinical training. There is optimism that it will improve ‘research and quality improvement’ skills and facilitate ‘curriculum mapping’. There is skepticism that it may reduce educational opportunities to develop ‘clinical judgement’ and ‘practical skills’. Medical educators should be mindful that these domains are protected as AI develops. We recommend that ‘Applied AI’ topics are formalized in curricula and digital technologies leveraged to deliver clinical education.

Peer Review reports

Background

The healthcare sector is undergoing a digital revolution [1], with a wide range of artificial intelligence (AI) innovations increasingly encountered in routine clinical practice [2]. From designing and training algorithms, to interpreting their output in clinical practice, these technologies require skillful human-machine interaction. The need for formal AI training has been recognized as a priority for government health policy [3, 4]. Although there is an abundance of evidence supporting the safety, accuracy, cost and workflow benefits of AI technologies in healthcare [5,6,7], there is limited research investigating the impact of AI on medical education [8, 9], the new skills required, and in particular the perceptions of trainee doctors who will be frontline users of AI in healthcare.

Increasingly, AI technologies are being developed that are capable of automating many tasks typically performed by doctors as part of their clinical training. So far these advancements are mostly limited to research applications, but it is only a matter of time before commercial uptake results in their routine use in everyday clinical practice. For example, decision support systems can triage patients, suggest diagnoses and alert test results by analysing data faster and more accurately than clinicians [6, 10]. Natural language processing can automate clinical documentation by text summarization [11] and detect diagnoses from scan reports [12]. Computer vision algorithms can detect lesions from radiological scans, make measurements, and pick-up incidental findings, saving time and reducing error [13, 14]. Robots trained by human operators could perform procedures such as venepuncture, and ultrasound [15, 16].

Although these technologies undoubtedly offer considerable benefits to patients, clinicians, and healthcare systems, there are concerns that they could come at a cost such as lack of trust in the algorithmic output due to the black-box effect [17, 18]. AI could also reduce training opportunities for doctors to develop clinical judgement, practical ability and communication skills. Current training programmes may fall short of equipping clinicians with the technical, statistical and analytical skills required to apply AI effectively for themselves and their patients.

The length of postgraduate medical training varies between countries, but the content of medical curricula is largely consistent globally [8]. Traditionally, medical education is centred around knowledge assimilation (by didactic teaching) and vocational training to develop practical, interpersonal and professional skills [19]. With AI algorithms capable of retaining and computing many orders of magnitude more data than humans, there may be less need for clinicians to memorize large amounts of medical information or to hone procedural skills by repetitive practice. Conversely, additional skills are required to interact safely and effectively with AI technologies such as data science, statistics and AI ethics [8]. Tomorrow’s clinicians must be skilled in data input, interpreting algorithmic outputs and communication of AI-derived treatment plans to patients [20]. Although formal education in these skills is recommended in national healthcare policy [2], it does not currently feature in most clinical training curricula.

In this study we evaluate the impact of AI technologies on clinical education, as perceived by doctors in postgraduate training. Based on the results, we make practical recommendations for trainers to maximise the benefits of clinical AI whilst mitigating potential negative impacts to deliver future-proof medical education.

Methods

Survey

A questionnaire was devised to investigate trainee doctors’ perceptions of AI technologies, specifically the impact on their training and education. The survey was sent by email to doctors working in London (UK) at NHS postgraduate training centers at which research team members were based to maximise response rate. These centers, mainly headquartered at tertiary hospitals, administrate both hospital and community-based postgraduate training in London, enabling inclusion of trainees from a wide range of specialities. The survey was conducted between October and December 2020. Respondents provided their informed consent by participating in the anonymous survey. The rationale, protocol and data collection tool were approved by the Guy’s and St. Thomas’ NHS Foundation Trust Institutional Audit and Quality Improvement Review Board (ref. 11,865). All methods were carried out in accordance with relevant guidelines and regulations.

Demographic data such as age, sex, clinical specialism and geographical location were anonymously collected. Five-point Likert-type scales (ranging from ‘strongly disagree’ to ‘strongly agree’) were used to investigate participants’ agreement with statements across a number of domains. The content validity of Likert-type scale statements was assessed by an expert panel of six healthcare professionals with experience of AI in healthcare. Individual items were scored to reach a consensus on which should be included. The survey questions were designed to minimize bias and further refined based on a pilot study of participants. The domains of impact assessment mirrored common themes in UK clinical training curricula: ‘clinical judgement’, ‘practical skills’, ‘research and quality improvement skills’ and ‘curriculum mapping’. Non-mandatory free-text spaces provided respondents the opportunity to comment on the positive and negative impacts of AI on clinical training.

Quantitative analyses

Internal reliability of the survey questions was measured by calculating Cronbach’s alpha. Likert-type data were analysed by (i) calculating the category of the median response for each statement, and (ii) comparing the proportion of agreement between the overall perception statement (‘AI systems being used in healthcare will improve my training and education’) with the other statements in a pairwise manner using Fisher’s exact test. To display the range of responses across each question, waffle plots were generated using the matplotlib library in Python 3.8. K-modes clustering was used to identify key groups of responses to the Likert-type statements [21].

Qualitative analyses

Free-text responses were analysed using thematic analysis [22]. Text data were hand-coded by two independent researchers. Themes and sub-themes were subsequently generated and representative examples identified from the raw data.

Results

Quantitative analyses

Two hundred ten doctors responded to the online survey (response rate 72%; 47% female). Overall there was high internal reliability of the survey questions (Cronbach’s alpha 0.82). 58% perceived an overall positive impact of AI technologies on clinical training (Fig. 1). Trainees agreed that AI would reduce clinical workload (62%) and improve training in research and audit skills (68%). Lower proportions agreed that it would improve training in clinical judgement (46%) and practical skills (32%) (Fig. 1). The majority reported insufficient AI training in their current curricula (92%), and supported more formal AI training (81%) (Fig. 2). Detailed responses to all Likert-type questions in the survey are available in Supplementary Appendix A (Figures 1 and 2). The median responses to the Likert-scale questions are shown in Table 1. There was agreement that AI would have an overall positive impact on training and education. This agreement was sustained for ‘research and audit skills’ and ‘curriculum mapping’. Conversely, respondents tended towards disagreement for the domains of ‘clinical judgement/decision making’ (p = 0.12) and strong disagreement for the domain ‘clinical skills’ (p < 0.01) (Table 2).

Fig. 1
figure 1

Domain-based impact of clinical AI on training and education - waffle plots of responses to Likert-type questions in survey of 210 trainee doctors (each icon represents one respondent to the survey)

Fig. 2
figure 2

Exposure to AI systems and attitudes towards AI training - waffle plots of responses to Likert-type questions in survey of 210 trainee doctors (each icon represents one respondent to the survey)

Table 1 Median response of participants answering the Likert-type statements
Table 2 Proportion of participants agreeing with statements (agree and strongly agree). P-values of pairwise Fisher’s exact test comparisons of % agreement between the overall statement (bold) and other statements

Cluster analyses

Clustering of the Likert-type statement responses revealed four key groups:

  • AI Negative (8 participants) - strongly disagreed with the majority of the statements.

  • AI Indifferent (84 participants) - no preference toward agree or disagree.

  • AI Optimistic (105 participants) - a preference toward agreeing with statements.

  • AI Positive (15 participants) - strongly agreed with the majority of the statements.

The cluster composition of clinical specialisms was calculated (Fig. 3), and shows that community-based and radiology specialisms contained a higher proportion of ‘AI Optimistic’ participants. Acute specialities and child and maternal health specialities contained a higher proportion of ‘AI Indifferent’ participants. The strongly polarized groups of ‘AI Positive’ and ‘AI Negative’ are much smaller in size in all specialisms except for Clinical Radiology in which there was a high proportion of ‘AI positive’ participants. This result should be interpreted in the context of the small proportion of respondents from this specialty.

Fig. 3
figure 3

The Likert-type statement responses cluster composition of different clinical specialities. (Acute specialities include Acute Medicine, Intensive Care Medicine, Anaesthetics and Emergency Medicine; Child and maternal health include Paediatrics, Obstetrics and Gynaecology; Community specialities include General Practice and Psychiatry)

Qualitative analyses

Thematic analysis of free-text responses revealed two main themes (‘positive perceptions’ and ‘negative perceptions’) (Fig. 4). Positive sub-themes included more free time for training and educational activities, directly improving the quality of practical skills training by enabling high-fidelity simulation, enabling more time to engage in interpersonal aspects of clinical practice, and boosting training in research, audit and quality improvement.

Fig. 4
figure 4

Themes and sub-themes identified from thematic analysis of free-text response data, along with representative examples from the raw data

Negative sub-themes included less development of clinical judgement, reduced opportunity for practical skills, workflow intrusion, increased administrative workload and reduced development of clinical accountability and probity.

Discussion

This study is the first of its kind to evaluate the impact of AI technologies in healthcare on medical education and clinical training by analysing perceptions of trainee doctors across a range of hospital and community-based specialities in London, UK. Overall, doctors perceive that clinical AI will have a positive impact on their training (58% agree). Domain-based analysis reveals more mixed perceptions. The overwhelming majority (82%) report insufficient training in AI topics, with strong support for formal training in these topics.

Domains of perceived positive impact

Doctors were most optimistic that clinical AI would improve training in research, audit and quality improvement. These are key education domains and can be challenging to fulfill without significant time and effort commitment outside of work. AI systems can improve the efficiency of research and audit by rapidly and accurately analysing large volumes of data. This may explain trainees’ positive perceptions in this regard [23]. An indirect but desirable effect is that clinical AI could free up doctors’ time to spend doing other educational activities, which was a common positive theme throughout.

Developing skills in evidence-based medicine is a key training requirement, but keeping up with rapidly changing clinical guidelines can be burdensome for doctors. AI-based decision support systems are automatically updated with the latest literature. Trainees perceived this to be a positive impact, enabling them greater exposure to the latest evidence to improve their knowledge and the quality of care they provide.

Domains with skeptical perceptions

Trainees perceived that clinical AI could reduce their training in practical skills, clinical judgement and decision-making. Developing these skills requires iterative practice, formation of heuristics, personal reflection, varied clinical experience and time. Participants reported that decision support systems, robotics and automatic image analysis could reduce training opportunities in these domains leading to deskilling. Another skeptical perception was increased administrative workload leading to information overload. Clinical AI developers should ensure that these technologies do not impede workflow to enable their adoption in clinical practice.

Medical educators should note that there are some areas in which training may be harmed by clinical AI. Involving clinicians in the development of these algorithms (such as ground-truth labelling and procedural training) will help trainees continue to develop these skills, because the AI will depend on their training to mimic behavior. This will also increase trust in the AI technologies and improve explainability to patients.

Impact on interpersonal and ethical development

The impact on interpersonal and ethical skills training featured in both positive and negative perception themes. Doctors envisage that AI will automate tasks, freeing up more time to develop communication skills. Trainees are optimistic that this will enhance their ability to provide patient-centred care. Conversely, doctors express concern that accountability is unclear when AI is part of clinical decisions, which could cause human deskilling in ownership and probity. Both points of view are valid; AI in healthcare has already created a new ethical landscape [24]. Governing bodies and medical educators should work collaboratively to produce ethical and legal frameworks that will protect and enable clinicians to develop these skills effectively in the age of clinical AI.

Perceptual variations by clinical specialism

Medical, surgical and community-based (General Practice, Psychiatry) specialities had a higher proportion of ‘AI optimistic’ trainees Fig. 3). This may be due to a lower clinical acuity in these specialities and higher administrative workload. Trainees’ in these specialities may perceive that AI will improve their workflow to free up time for training and educational activities such as communication skills development.

Acute specialities (Emergency Medicine, Acute Medicine, Anesthesia, Intensive Care) and child and maternal health (Paediatrics, Obstetrics and Gynaecology) had more ‘AI Negative’ and ‘Indifferent’ responses. This might be due to the higher clinical acuity in these specialities, including emergency procedures and diagnostic ambiguity, which may rely on experience or ‘gut-feeling’. These skills are notoriously hard to model for AI development [25]. This mirrors trainees’ perceptions that training in practical procedures and clinical judgement might be reduced by clinical AI.

Clinical Radiology trainees, although a small proportion of respondents, were highly optimistic. This echoes positive attitudes reported previously [26]. Radiology has experienced the most AI advances in clinical practice already [27, 28] so Radiology trainees are most likely to already first-hand experience of AI’s impact on their training. This may explain their positive perceptions compared to other specialities.

Overall, specialities with higher capacity for automation were more optimistic; the implication is that rather than a panacea, delivery of medical education in the AI age will need to be tailored to these subtle variations between specialities.

Implications for clinical curricula and recommendations for trainers

As clinical practice changes so must clinical education. This study brings the need for formal AI training in clinical curricula into sharp relief, and confirms a willing appetite for this from trainee doctors. The impact of AI on medical education can take different routes. Direct routes leverage AI technology to improve the delivery of training itself. Indirect routes benefit education by streamlining workflow to free up more time for education and training. Although the majority (72%) of respondents in our survey were yet to regularly encounter AI systems in their training and education, it is an area of active research ranging from assisted radiology teaching [29] to virtual reality for surgical skills development [30] and automated assessment of procedural performance [31]. Medical curricula should be reviewed to leverage these technologies to directly improve the delivery of clinical education.

We propose that medical curriculum makers consider a new set of AI-specific skills. These include data input and management, mathematics and statistics, communicating AI outcomes to patients and AI-specific ethics. Medical training curricula are already saturated with limited room for new topics so practical training in ‘Applied AI’ would be most feasible. Alongside an overview of common ML architectures, this should include balanced training in clinical AI interpretation including data bias, overfitting and the potential for harm.

Navigating the current ML research landscape is challenging. Common pitfalls include over-optimistic conclusions from ‘human versus clinician’ studies that are usually retrospective and prone to bias [32], lack of standardized benchmarks and no universally accepted AI evaluation metrics [33]. Training in ‘Applied AI’ must additionally equip clinicians with skills in critical literature analysis.

Training in ‘Applied AI’ will need to be supported by e-learning, didactic teaching, assessment in examinations, supervised clinical learning events and personal reflection (Fig. 5). Although this has been considered as a priority for the future health workforce [2, 3], ‘Applied AI’ topics remain widely absent from clinical training curricula [9, 19]. Ultimately this will negatively affect the quality of patient care by missing out on the myriad benefits of clinical AI systems.

Fig. 5
figure 5

Potential new AI-related clinical training domains in future curricula. Novel training and assessment methods to deliver AI-based training. (ML = machine learning). (Created with BioRender.com)

Our survey was not prefaced by any educational material on AI, for two reasons. First, it avoided prolonging the time participants would need to commit, which would have reduced response rates. Second, it would have biased the responses towards taking the point of view implicitly expressed in the educational material. Even if we had striven for neutrality, it would be difficult to achieve it and even more difficult to document that we had achieved it.

The optimistic perceptions of AI reported in this study may lessen as doctors gain more first-hand experience of AI in their clinical practice and training (including its weaknesses such as data bias and the black-box effect) [20]. Conversely, the opposite may occur; indeed in this study the participants working in the speciality most likely to have already experienced AI technologies (Radiology) were actually the most ‘AI optimistic’. Curriculum development must provide a balanced view, recognizing AI limitations, weaknesses and potential for harm.

Limitations

This study provides a snapshot of the opinions of trainee doctors working in the UK NHS. The survey was hosted primarily by postgraduate training centers in London (UK), where research team members were based to maximise response rate (72%). Although these centers administrate postgraduate training for hospital and community based trainees (enabling representation of trainees in General Practice, Psychiatry etc.), the results of our survey could be biased in favor of trainees working in London, who may have specific experiences due to local uptake of AI technologies in urban compared to rural areas. Based on our results, we recommend a survey of all UK NHS postgraduate centers to gain a cross-sectional representation of trainee experience, and to elucidate any geographical variations in experience and opinion. We also recommend the inclusion of medical undergraduate students, since they are the clinical workforce of the future, and most likely to be directly affected by the impact of AI technologies on medical training.

Participants’ level of AI knowledge was not collected in our survey. Participants’ role at the time of responding was collected, with only 2 respondents being currently involved with clinical AI research. The majority of respondents neither encountered AI technologies regularly in their clinical practice (68%) or training (72%). Future work would benefit from participants self-rating their level of AI knowledge to contextualize findings.

The impact on communication and interpersonal skills was not assessed in the Likert-type part of our survey, Thematic analysis of free-test responses revealed important trainee perceptions in these areas. Future evaluation of trainees’ attitudes should further probe the perceived impact on domains such as communication, professionalism, leadership and probity, which are key elements of all clinical training curricula.

Conclusions

Clinical AI will affect medical education and clinical training. Trainee doctors have overall positive perceptions of this impact. Training in practical procedures and clinical judgement may be reduced by clinical AI and educational opportunities in these skills should be protected. There is overwhelming support for formal education in AI-based skills and topics. Medical curricula should be reviewed to include ‘Applied AI’ topics, using digital technologies to deliver training.

Availability of data and materials

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.

Abbreviations

AI:

Artificial intelligence

ML:

Machine learning

References

  1. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. 2019;25(1):44–56. https://doi.org/10.1038/s41591-018-0300-7.

    Article  Google Scholar 

  2. Ross J, Webb C, Rahman F, AoRCM. Artificial Intelligence in Healthcare: Academy of Medical Royal Colleges; 2019. [cited 2020 Jan 12]. Available from: https://www.aomrc.org.uk/wp-content/uploads/2019/01/Artificial_intelligence_in_healthcare_0119.pdf

  3. Topol E. The Topol Review - preparing the healthcare workforce to deliver the digital future: NHS; 2019. [cited 2021 Jan 12]. Available from: https://topol.hee.nhs.uk/wp-content/uploads/HEE-Topol-Review-2019.pdf

  4. Joshi I, Morley J. Artificial Intelligence: How to get it right. Putting policy into practice for safe data-driven innovation in health and care: NHSX; 2019. [cited 2021 Jan 12]. Available from: https://www.nhsx.nhs.uk/media/documents/NHSX_AI_report.pdf

  5. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542(7639):115–8.

    Article  Google Scholar 

  6. Tomašev N, Glorot X, Rae JW, Zielinski M, Askham H, Saraiva A, et al. A clinically applicable approach to continuous prediction of future acute kidney injury. Nature. 2019;572(7767):116–9. https://doi.org/10.1038/s41586-019-1390-1.

    Article  Google Scholar 

  7. Nelson A, Herron D, Rees G, Nachev P. Predicting scheduled hospital attendance with artificial intelligence. NPJ Digital Med. 2019;2(1):1–7.

    Article  Google Scholar 

  8. Paranjape K, Schinkel M, Nannan Panday R, Car J, Nanayakkara P. Introducing Artificial Intelligence Training in Medical Education. JMIR Med Educ. 2019;5(2) [cited 2021 Jan 10] Available from: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6918207/.

  9. Wartman SA, Combs CD. Medical education must move from the information age to the age of artificial intelligence. Acad Med. 2018;93(8):1107–9. https://doi.org/10.1097/ACM.0000000000002044.

    Article  Google Scholar 

  10. Fernandes M, Vieira SM, Leite F, Palos C, Finkelstein S, Sousa JMC. Clinical decision support Systems for Triage in the emergency department using intelligent systems: a review. Artif Intell Med. 2020;102:101762. https://doi.org/10.1016/j.artmed.2019.101762.

    Article  Google Scholar 

  11. Goldstein A, Shahar Y. An automated knowledge-based textual summarization system for longitudinal, multivariate clinical data. J Biomed Inform. 2016;61:159–75. https://doi.org/10.1016/j.jbi.2016.03.022.

    Article  Google Scholar 

  12. Bressem KK, Adams LC, Gaudin RA, Tröltzsch D, Hamm B, Makowski MR, et al. Highly accurate classification of chest radiographic reports using a deep learning natural language model pretrained on 3.8 million text reports. Bioinformatics. 2021;36(21):5255–61.

  13. Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, et al. A guide to deep learning in healthcare. Nat Med. 2019;25(1):24–9. https://doi.org/10.1038/s41591-018-0316-z.

    Article  Google Scholar 

  14. Howard JP, Zaman S, Ragavan A, Hall K, Leonard G, Sutanto S, et al. Automated analysis and detection of abnormalities in transaxial anatomical cardiovascular magnetic resonance images: a proof of concept study with potential to optimize image acquisition. Int J Cardiovasc Imaging. 2020 [cited 2021 Jan 4]; Available from;37(3):1033–42. https://doi.org/10.1007/s10554-020-02050-w.

    Article  Google Scholar 

  15. Groetz S, Wilhelm K, Willinek W, Pieper C, Schild H, Thomas D. A new robotic assistance system for percutaneous CT-guided punctures: initial experience. Minim Invasive Ther Allied Technol. 2016;25(2):79–85. https://doi.org/10.3109/13645706.2015.1110825.

    Article  Google Scholar 

  16. Blaivas M, Blaivas L, Philips G, Merchant R, Levy M, Abbasi A, et al. Development of a deep learning network to classify inferior vena cava collapse to predict fluid responsiveness. J Ultrasound Med. 2021;40(8):1495–504.

  17. Asan O, Bayrak AE, Choudhury A. Artificial Intelligence and Human Trust in Healthcare: Focus on Clinicians. J Med Internet Res. 2020;22(6) [cited 2021 May 26] Available from: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334754/.

  18. DeCamp M, Tilburt JC. Why we cannot trust artificial intelligence in medicine. Lancet Digital Health. 2019;1(8):e390. https://doi.org/10.1016/S2589-7500(19)30197-9.

    Article  Google Scholar 

  19. Wartman SA, Combs CD. Reimagining medical education in the age of AI. AMA J Ethics. 2019;21(2):E146–52. https://doi.org/10.1001/amajethics.2019.146.

    Article  Google Scholar 

  20. Regulating Black-Box Medicine | Michigan Law Review [Internet]. [cited 2021 Jan 10]. Available from: https://michiganlawreview.org/regulating-black-box-medicine/

  21. Huang Z. Extensions to the k-means algorithm for clustering large data sets with categorical values; 1998.

    Google Scholar 

  22. Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101. https://doi.org/10.1191/1478088706qp063oa.

    Article  Google Scholar 

  23. Extance A. How AI technology can tame the scientific literature. Nature. 2018;561(7722):273–4. https://doi.org/10.1038/d41586-018-06617-5.

    Article  Google Scholar 

  24. Gerke S, Minssen T, Cohen G. Ethical and legal challenges of artificial intelligence-driven healthcare. Artificial Intelligence Healthcare. 2020:295–336. https://doi.org/10.1016/B978-0-12-818438-7.00012-5.

  25. Hashimoto DA, Witkowski E, Gao L, Meireles O, Rosman G. Artificial intelligence in anesthesiology: current techniques, clinical applications, and limitations. Anesthesiology. 2020;132(2):379–94. https://doi.org/10.1097/ALN.0000000000002960.

    Article  Google Scholar 

  26. Dumić-Čule I, Orešković T, Brkljačić B, Kujundžić Tiljak M, Orešković S. The importance of introducing artificial intelligence to the medical curriculum - assessing practitioners’ perspectives. Croat Med J. 2020;61(5):457–64. https://doi.org/10.3325/cmj.2020.61.457.

    Article  Google Scholar 

  27. Rajpurkar P, Irvin J, Ball RL, Zhu K, Yang B, Mehta H, et al. Deep learning for chest radiograph diagnosis: a retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med. 2018;15(11):e1002686. https://doi.org/10.1371/journal.pmed.1002686.

    Article  Google Scholar 

  28. Ardila D, Kiraly AP, Bharadwaj S, Choi B, Reicher JJ, Peng L, et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat Med. 2019;25(6):954–61. https://doi.org/10.1038/s41591-019-0447-x.

    Article  Google Scholar 

  29. Cheng C-T, Chen C-C, Fu C-Y, Chaou C-H, Wu Y-T, Hsu C-P, et al. Artificial intelligence-based education assists medical students’ interpretation of hip fracture. Insights Imaging. 2020;11(1):119. https://doi.org/10.1186/s13244-020-00932-0.

    Article  Google Scholar 

  30. Aeckersberg G, Gkremoutis A, Schmitz-Rixen T, Kaiser E. The relevance of low-fidelity virtual reality simulators compared with other learning methods in basic endovascular skills training. J Vasc Surg. 2019;69(1):227–35. https://doi.org/10.1016/j.jvs.2018.10.047.

    Article  Google Scholar 

  31. Winkler-Schwartz A, Bissonnette V, Mirchi N, Ponnudurai N, Yilmaz R, Ledwos N, et al. Artificial intelligence in medical education: best practices using machine learning to assess surgical expertise in virtual reality simulation. J Surg Educ. 2019;76(6):1681–90. https://doi.org/10.1016/j.jsurg.2019.05.015.

    Article  Google Scholar 

  32. Nagendran M, Chen Y, Lovejoy CA, Gordon AC, Komorowski M, Harvey H, et al. Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies. BMJ. 2020;368:m689.

    Article  Google Scholar 

  33. Choudhury A, Asan O. Role of artificial intelligence in patient safety outcomes: systematic literature review. JMIR Med Inform. 2020;8(7):e18599. https://doi.org/10.2196/18599.

    Article  Google Scholar 

Download references

Acknowledgements

None.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

MB and DaC contributed equally to this work. MB, DaC, JS, JR and SZ devised the research question and strategy. KP, IJ, performed data collection. DiC and DF performed data analyses. MB, DaC, NL, GC, DF, JS, JR and SZ contributed to manuscript writing. All authors reviewed the manuscript. The author(s) read and approved the final manuscript.

Corresponding author

Correspondence to Sameer Zaman.

Ethics declarations

Ethics approval and consent to participate

Informed consent was obtained from all participants. Respondents provided their consent by participating in the anonymous survey. The rationale, protocol and data collection tool were approved by the Guy’s and St. Thomas’ NHS Foundation Trust Institutional Audit and Quality Improvement Review Board (ref. 11865) and did not require separate ethics committee approval. All methods were carried out in accordance with relevant guidelines and regulations.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1: Appendix A.

Responses to Likert-type questions in the survey of trainee doctors.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Banerjee, M., Chiew, D., Patel, K.T. et al. The impact of artificial intelligence on clinical education: perceptions of postgraduate trainee doctors in London (UK) and recommendations for trainers. BMC Med Educ 21, 429 (2021). https://doi.org/10.1186/s12909-021-02870-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12909-021-02870-x

Keywords