Dermatology undergraduate skin cancer training: a disconnect between recommendations, clinical exposure and competence
© Aldridge et al.; licensee BioMed Central Ltd 2012
Received: 27 October 2011
Accepted: 8 May 2012
Published: 9 July 2012
Skin cancers are the most common malignancies in Caucasian populations. Non-specialists are responsible for the initial assessment of skin lesions and are required to act as the gatekeepers to dermatological cancer services in many healthcare systems. The majority of such physicians receive very limited formal undergraduate or postgraduate dermatology training. The British Association of Dermatologists (BAD) has produced guidelines that list the lesions that students should be able to diagnose on graduation and the majority of UK medical schools’ operate curricula in keeping with these. There is, however, virtually no evidence as to whether these competencies are being achieved. We set out to determine students’ competence at skin lesion diagnosis and to quantify their clinical exposure to examples of such lesions during their dermatology attachment.
Three linked studies were undertaken. In the first, students’ competence was tested by randomized slideshows of images containing the 16 lesions recommended in the UK guidelines. Students’ accuracy was tested at the beginning (Day 1) and end (Day 10) of their clinical placement, with a random sample of students retested 12 months later. Secondly, students’ exposure to these lesions was recorded during their attachments. Finally a survey of the additional dermatological resources used by the students was undertaken.
Study 1: Students’ diagnostic accuracy increased from 11% on Day 1 to 33% on Day 10 (effect size +2.72). After 12 months half of this effect had disappeared and the students accuracy had dropped to 24%. Study 2: Students’ exposure to the recommended lesions was poor with 82% not even witnessing a single example of each of the 3 major skin cancers. Despite these measurements, only a minority of students reported that they were not confident at diagnosing skin tumours. Study 3: The majority of students use additional resources to supplement their learning.
In the light of what we know about learning in dermatology, our data would suggest, that the current (traditional) undergraduate attachment is inadequate to meet the UK recommendations for graduate competence. As well as critically examining the basis for these recommendations, we need more empirical data on student performance and exposure, in order to improve teaching and learning.
Cutaneous malignancies account for over a quarter of all new ‘cancer’ diagnoses in the UK . Diagnosing skin cancer is largely a perceptual skill, relying little on formal or explicit rules, but rather on prior exposure and feedback either in a training environment or in the clinic. Following Norman’s terminology, the skills involved are largely thought to be those of non-analytical pattern recognition (NAPR), and this core skill can be viewed as being able to attach semantics to images or percepts [2–10].
In the UK general practitioners are responsible for the initial assessment of skin lesions and act as the gatekeepers to dermatological cancer services. For most of these medical practitioners their only formal clinical dermatology teaching is as an undergraduate. Despite the fact that cutaneous disease has been shown to be the most common reason for primary care consultations (at 24% of physicians’ new patient workload) , medical school training in dermatology is relatively limited contributing to only ≈ 1% of undergraduate teaching time .
Whilst historically each UK University has set its own curricula, in the light of recommendations from the General Medical Council (GMC), in 2006 the British Association of Dermatologists (BAD) produced the recommendations “Dermatology in the Undergraduate Medical Curriculum”, outlining what ought to be expected of students upon graduation . These guidelines were based on the results of a modified Delphi study that had been conducted with a multi-disciplinary panel of 66 individuals involved in delivering undergraduate education . Under the “skin cancer” section the guidelines state that it is very important that graduates should be able to recognize the three main skin tumours (basal cell carcinoma (BCC), squamous cell carcinoma (SCC) and melanoma) and that it is fairly important that they can recognize another 13 common skin lesions — diagnosing implicitly means distinguishing the rarer malignant lesions from the more frequent benign ones. A 2009 audit of UK medical schools’ curricula concluded that these “skin cancer” learning outcomes were present in the syllabuses of most universities . Of course because a topic is present in a curriculum does not mean that a particular student competence is actually achieved.
Given the paucity of research work in this area — even of observational studies — we wished to collect evidence of our students’ competencies in this clinical domain and relate them to clinical exposure and student confidence. How much exposure to skin cancers do students actually gain as an undergraduate and what do the prescribed competencies mean in practice? We show that whilst current teaching has a measurable effect on improving students’ diagnostic accuracy, half of that effect is transient, and furthermore students are not exposed to enough examples to realistically meet the suggested UK graduate competencies.
The University of Edinburgh’s undergraduate dermatology teaching program is similar to the majority of other UK Medical Schools ; consisting of an introductory series of lectures, followed by clinical exposure in outpatients. In Edinburgh there are 9 lectures and 10 half-day clinical sessions over a two-week attachment undertaken in the student’s penultimate clinical year. The clinical sessions incorporate 8 small group demonstration clinics where the instructor has no clinical responsibility and where patients are obtained from up to 10 adjacent NHS feeder clinics, a “one-on-one” NHS clinic with a consultant, and a skin surgery session. The whole attachment is undertaken in the regional teaching hospital, with a departmental throughput of in excess of 25,000 new patients per annum.
Study 1: Student competence at skin tumour diagnosis
All 77 students undertaking their dermatology attachment between July and September 2009 were enrolled. Fifty students (65%) were female. Over the ten-week study period, 5 cohorts of between 14–17 students attended. The students’ diagnostic accuracy was examined using a digital slideshow, with the students writing diagnoses on a custom designed answer sheet. We were “generous” in what we accepted as correct answers, allowing spelling mistakes, incomplete terminology, abbreviations and lay terms. Assessment was undertaken prior to seeing any patients on the first morning of the attachment (Day 1 test) and again on the final afternoon (Day 10 test). Note that the students are not formally examined at the end of each two-week dermatology block, but undertake a joint exam with other disciplines at a later date. Five separate test batches of images were constructed so that both the Day 1 and Day 10 tests were different for each group of students. Each test batch consisted of 25 digital images randomly selected in a stratified manner to ensure that each test contained an identical spectrum of lesional diagnoses. The 25 lesions, which were presented for each test batch in a different randomized order, included one or more examples of each of the 16 lesions considered important in the UK guidelines. The test images were randomly selected from 687 suitable digital photographs in the Department of Dermatology’s image database. This image library (now standing at over 4000 lesions) has been prospectively collected for research into diagnostic expertise and automated diagnostics. Although the images are not from consecutive patients the library is designed deliberately not to be a typical departmental library of “classic cases” but to represent cases from routine clinical practice. All the images in the library were captured using the same controlled fixed distance photographic setup; Canon (Canon, Japan) EOS 350D 8.1MP cameras, Sigma (SIGMA, Japan) 70 mm f2.8 macro lens and Sigma EM-140 DG Ring Flash at a distance of 50cms. The digital slideshows were conducted in an identical manner using iPhoto on a 15” Apple Macbook Pro (Apple, California) connected to an Epson H285B high definition projector (Epson, Japan), under similar ambient lighting conditions in the Department’s seminar room. The 25 lesions were all presented at the same scale for 20 seconds each. Students were not given feedback after the Day 1 test, but after completing the Day 10 test underwent an addition tutorial during which feedback was provided. 12 months later, in August 2010, 30 of the original students were randomly selected, contacted by email and asked to return to the department. The students were not informed of the reason that they had been selected and asked to return, but they were advised that they would be helping out with a fellow student’s SSC project and would be remunerated for their time at a level of £10. Nineteen students (63%) were able and willing to attend for retesting (12 Month test). The batches of test images used for the 12 Month test were coordinated so that each student re-attending had not been exposed previously to the same digital photographs. The slideshow was conducted in an identical manner as described above for the Day 1 and Day 10 tests.
Study 2: Clinical exposure to skin lesions during the attachment
Over a six-week period the total clinical exposure for each of the 50 students who attended their undergraduate dermatology attachment from July to August 2010 was assessed. Three cohorts of between 16–17 students were monitored. Every clinical visual encounter that the students were exposed to was recorded on a purpose-designed checklist. This checklist contained the same 16 skin lesions considered important in the UK guidelines and was completed by each student for their “one-on-one” teaching and skin surgery sessions, whereas for the group demonstration clinics it was completed by a dedicated clinical observer. If a diagnosis was not on the checklist additional space was available to record text for “other” diagnoses. If more than one lesion was seen on the same patient each lesion counted as a separate clinical encounter. In contrast, if a patient was being followed-up or being used to demonstrate a “dermatological history”, unless there was also a visual example of a skin lesion, these patients did not count as a clinical encounter under the terms of reference of the present study. Medical staff involved in the teaching clinics were not pre-warned that the monitoring process was being undertaken.
Students were asked to complete an anonymous questionnaire on the final afternoon of their attachment (Day 10). The questionnaire asked them to rate their abilities on 7-point Likert scales across the three key skin cancer learning outcomes; their ability to diagnose melanomas, their ability to diagnose basal cell carcinomas and their ability to diagnose squamous cell carcinomas.
Study 3: Learning exposure outside the clinic
To gain a measure of other non-patient visual dermatology exposure, all students completing their dermatology attachment between July and September 2010 (n = 50) and again between March and May 2011 (n = 61) were asked to complete an anonymous questionnaire about the additional resources they had used to supplement their formal teaching. In total 106 (95%) students completed the questionnaire. The purpose-designed questionnaire had tick boxes covering the most popular textbooks and Internet sites and also space to add supplementary free-text if the resource used was not listed. In addition to assessing the types of resources used the questionnaire also asked the students to estimate how long on average they spent using these resources.
Data was tabulated in Excel (Microsoft, California) then exported into R for graphing and statistical analysis .
The NHS Lothian ethics committee granted permission for the collection and use of the images, and additional permission for studies of student learning were granted from the University of Edinburgh College of Medicine and Veterinary Medicine students’ ethics committee.
Study 1: Student competence at skin tumour diagnosis
Study 2: Clinical exposure to skin lesions during the attachment
Table showing the median number of lesions that the students witnessed, split across the 3 cohorts of students for all 16 lesions contained in the UK guidelines
UK “Recommended lesions”
Student cohort 1 (n=16)median exposure
Student cohort 2 (n=17) median exposure
Student cohort 3 (n=17) median exposure
Overall student median exposure n=50 (Range)
Percentage of students observing 1 or more example of each lesion
Percentage of students observing >3 examples
Percentage of students observing >5 examples
Individual lesions “Very Important”
Individual lesions “Important”
All 16 lesions
Total number of lesional diagnoses seen
426/800 (53%) Overall Observation Rate
Total number of skin lesions seen
Students’ self-assessment of their own abilities at diagnosing the 3 major skin cancers on completion of the dermatology attachment
Questions: How confident are you in your ability to diagnose…
Mean Likert Score (Range)
Median Likert Score
Percentage of students “unconfident” (score <4)
…squamous cell carcinomas?
…basal cell carcinomas?
Study 3: Learning exposure outside the clinic
The supplementary resources used by students (n=106). Answers from an anonymous end of attachment questionnaire
Internet resources (n = 69)
No. of Students (%age)
DermNet NZ 
University of Edinburgh 
Textbook resources (n = 97)
No. of Students (%age)
Davidson’s Principles and Practice of Medicine 
Clinical Dermatology 
Dermatology: an Illustrated Colour Text 
Oxford Handbook of Clinical Specialties 
ABC of Dermatology 
Crash Course 
Physical Signs in Dermatology 
Fitzpatrick’s Dermatology in General Medicine 
Differential Diagnosis in Dermatology 
Rook’s Textbook of Dermatology 
Lecture Notes: Dermatology 
In the first study we showed that students’ diagnostic accuracy for the 16 “important” skin lesions was only 33% after completing their dermatology attachment and that it dropped to 24% a year later. Such absolute measures of diagnostic competence are obviously incomplete and potentially misleading because they are so dependent on the difficulty of the specific tests. However, in the absence of normative data elsewhere in the literature, they provide some sort of benchmark for future work, and other authors are welcome to use our test-sets.
A potential reason for what we take to be the students’ poor accuracy was identified in our second study, where we showed that students’ exposure to the 16 important skin lesions is highly variable but universally limited. Some limitation in clinical exposure is to be expected but, rightly or wrongly, we were shocked by how limited the extent of clinical exposure was. Although there is no systematic data on this issue we suspect that the clinical exposure our students gain is typical of, if not better than that at many other UK centres, and our course duration is actually greater than the UK average . Of note our course scores highly in terms of student feedback compared with other clinical attachments within the University of Edinburgh. In addition, due to the structure of our group demonstration clinics (which allows multiple students to see a single patient) the overall melanoma witnessing rate of 38% was predominately achieved because of a single case -if this one patient had not presented and agreed to be examined at the time of a demonstration clinic the overall melanoma observation rate would have dropped to 18%.
The students’ low exposure demonstrates the major difficulty encountered when teaching students to identify skin cancers and their mimics — a lack of reliable clinical examples. Unlike other specialties where pertinent clinical signs are often longstanding and can be demonstrated repeatedly to different groups of students, suspected skin malignancies are excised in dermatology outpatients as a matter of priority. Therefore if face-to-face patient teaching is to be relied on a constant throughput of new patients is required.
It might be argued that “skin cancer” education is not performed exclusively during students’ dermatology attachments but that additional teaching is happening during their allied clinical attachments. We are sceptical of this and note that our students’ scores, which were lower one year after their dermatology training, followed their 2-month attachment in primary care.
The finding that there appeared a mismatch between objective ability and self-confidence again should not be surprising. Although subjective measures are in widespread use for assessing teaching and learning, there is increasing acknowledgment that students’ (and other health professionals’) self-assessment of their own abilities are often erroneous [36, 37]. An additional reason for the students’ overconfidence could also be that the lesions encountered during their teaching attachment were not suitably representative of the variety of presentations that had been randomly selected in our test slide shows (and could be argued are encountered in real-life). Neither the lesions that the students witness in the group demonstration clinics nor those that they are exposed to in their additional study are unselected. In our experience both textbook images and the demonstration clinic lesions are often chosen precisely because they are “classical” examples. Irrespective of the cause, the mismatch in confidence and competence at the end of students’ only formal dermatology training has implications on the role of non-experts as gatekeepers to cutaneous cancer services.
Given the potential importance of undergraduate dermatology education it is surprising that there has been so little investigation of UK medical school dermatology teaching [12, 15, 38–43]. Whilst additional studies have subjectively shown that both medical students and primary care physicians feel dermatology training could be improved [44–46], there are only a few US studies that have objectively assessed students’ diagnostic acumen after undergraduate dermatology electives [47–49]. These studies are, however, unlikely to be transferable to UK undergraduate students, and are not directly comparable to our study because the difficulty of the questions was not controlled for. Whilst the US studies demonstrated a higher level of improvement in diagnostic competence than we witnessed (71-82% accuracy), there are a number of other potential reasons for this. First, in these studies there could be an element of selection bias; the attachments were not compulsory and the students enrolled had volunteered to undertake dermatology electives, many presumably with the intention of pursuing a dermatology career. Second, the studies used the same images for both the initial and final assessments, which are likely to have artificially raised the final scores. More importantly the test contained not just lesions but also dermatoses and the images used for testing were not randomly selected, instead, the images chosen were often described as “classical” examples. Third, the answers were in multiple choice format rather than free text, which does not correspond to everyday practice. Finally these optional electives were of longer duration than in the UK, being full-time for 2–4 weeks.
We believe the biggest difficulty in interpreting our work is the lack of a coherent and justified framework for the purpose, standards and remit of undergraduate dermatology education. Although the BAD have produced guidelines that have been integrated into UK medical schools’ curricula, there are no specific criteria for any of their recommendations. As a result the current statements, such as “graduates should be able to recognize melanomas”, are as lacking in precision as saying a mathematics graduate should be able to compute. If the guidelines are to be adopted successfully, these criteria need to be objectively defined. This requires that we have to tackle exactly how difficult particular cases are with reference to some common standard.
So how could students’ skin lesion identification be improved? Our own students clearly seek out online resources, and the number of images available online far exceeds that available in the clinic or in textbooks. Online content is, however, very variable in quality and many online images are of poor quality and not infrequently in our experience the diagnosis is questionable, if not wrong. Whilst we know of no experimental work directly comparing learning on real patients versus learning from image databases for skin cancer we strongly suspect that the latter is the way to proceed. Modern imaging techniques already allow for accurate 3D models of skin lesions to be captured simply and efficiently . These models have significant advantages over conventional 2D photography as they allow students to rotate and pan around the images as they would in real-life. A large database of such images could address the main obstacle to effective dermatology teaching by removing our reliance on a constant supply of new example lesions. Furthermore if the images were available online, widespread access could be achieved with relatively low production costs. This would allow standardized exposure across multiple institutions, limiting the intrinsic fluctuations of exposure that occur between students with present teaching arrangements.
This work suggests that the traditional dermatology teaching attachment is inadequate to meet the current UK guidelines for graduate skin cancer competencies. We note that these guidelines have little objective evidence base to ensure their practicality or validity, and realize that some would argue that it is not reasonable to expect students to be able to diagnose skin cancers to a defined standard. We differ, and furthermore would argue that more systematic data might allow us to educate students more efficiently and perhaps even at lower cost in terms of staff and student time.
R Benjamin Aldridge was supported by the Wellcome Trust (Reference 083928/Z/07/Z).
The work was supported by The Wellcome Trust (Reference 083928/Z/07/Z). We are also grateful to the advice and assistance given by Karen Roberston and Yvonne Bisset (Department of Dermatology, University of Edinburgh) regarding the photographic capture and preparation of the digital images.
- Cancer Research UK (CRUK): CancerStats [Internet]. 2011, [Accessed 03/08/2011]. Available from, [http://info.cancerresearchuk.org/cancerstats]Google Scholar
- Smeulders A, Worring M, Santini S, Gupta A, Jain R: Content-based image retrieval at the end of the early years. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2000, 22 (12): 1349-1380. 10.1109/34.895972.View ArticleGoogle Scholar
- Norman G, Eva K, Brooks L: Expertise in Medicine and Surgery. The Cambridge Handbook of Expertise and Expert Performance. Edited by: Ericsson KA, Charness N, Hoffman RR, Feltvoich PJ. 2006, Cambridge University Press, Cambridge, 339-354.View ArticleGoogle Scholar
- Norman GR, Rosenthal D, Brooks LR, Allen SW, Muzzin LJ: The development of expertise in dermatology. Arch Dermatol. 1989, 125 (8): 1063-1068. 10.1001/archderm.1989.01670200039005.View ArticleGoogle Scholar
- Norman GR, Brooks LR: The Non-Analytical Basis of Clinical Reasoning. Adv Health Sci Educ Theory Pract. 1997, 2 (2): 173-184. 10.1023/A:1009784330364.View ArticleGoogle Scholar
- Norman G, Young M, Brooks L: Non-analytical models of clinical reasoning: the role of experience. Medical education. 2007, 41 (12): 1140-1145.Google Scholar
- Norman G: Building on experience–the development of clinical reasoning. N Engl J Med. 2006, 355 (21): 2251-2252. 10.1056/NEJMe068134.View ArticleGoogle Scholar
- Allen SW, Norman GR, Brooks LR: Experimental studies of learning dermatologic diagnosis: the impact of examples. Teaching and Learning in Medicine. 1992, 4 (1): 35-44. 10.1080/10401339209539531.View ArticleGoogle Scholar
- Aldridge RB, Zanotto M, Ballerini L, Fisher RB, Rees JL: Novice identification of melanoma: not quite as straightforward as the ABCDs. Acta Derm Venereol. 2011, 91 (2): 125-130.View ArticleGoogle Scholar
- Aldridge RB, Glodzik D, Ballerini L, Fisher RB, Rees JL: Utility of non-rule-based visual matching as a strategy to allow novices to achieve skin lesion diagnosis. Acta Derm Venereol. 2011, 91 (3): 279-283. 10.2340/00015555-1049.View ArticleGoogle Scholar
- Schofield JK, Fleming D, Grindlay D, Williams H: Skin conditions are the commonest new reason people present to general practitioners in England and Wales. Br J Dermatol. 2011, 165 (5): 1044-1050. 10.1111/j.1365-2133.2011.10464.x.View ArticleGoogle Scholar
- Burge S: British Association of University Teachers of Dermatology. Teaching dermatology to medical students: a survey of current practice in the U.K. Br J Dermatol. 2002, 146 (2): 295-303. 10.1046/j.1365-2133.2002.04522.x.View ArticleGoogle Scholar
- The British Association of Dermatologists: Dermatology in the Undergraduate Medical Curriculum [Internet]. 2006, [Accessed 03/08/2011]. Available from, [http://www.bad.org.uk//site/617/default.aspx]Google Scholar
- Clayton R, Perera R, Burge S: Defining the dermatological content of the undergraduate medical curriculum: a modified Delphi study. Br J Dermatol. 2006, 155 (1): 137-144. 10.1111/j.1365-2133.2006.07190.x.View ArticleGoogle Scholar
- Davies E, Burge S: Audit of dermatological content of U.K. undergraduate curricula. Br J Dermatol. 2009, 160 (5): 999-1005. 10.1111/j.1365-2133.2009.09056.x.View ArticleGoogle Scholar
- R Development Core Team: R: A language and environment for statistical computing. 2009, R Foundation for Statistical Computing, Vienna, Austria, ISBN 3- 900051-07-0, [http://www.R-project.org]Google Scholar
- Norman GR, Streiner DL: Biostatistics: The Bare Essentials. 2008, BC Decker Inc, HamiltonGoogle Scholar
- DermNet NZ: Facts about skin from the New Zealand Dermatological Society. [http://www.dermnet.org.nz/.(accessed 25/4/2012]
- University of Edinburgh Electronic Medical Curricullum (EEMeC): Not available outside University networkGoogle Scholar
- Wikipedia. [http://www.wikipedia.org/]
- The British Association of Dermatologists Website: Undergraduate Education Section. [http://www.bad.org.uk//site/617/default.aspx]
- Emedicine: Medscape Reference. [http://emedicine.medscape.com/]
- Primary Care Dermatology Society (PCDS): Image Atlas. [http://www.pcds.org.uk/pcdshome/image-atlas]
- Patient.co.uk: Comprehensive health information as provided by GPs and nurses to patients during consultations. [http://www.patient.co.uk/display/16777222/]
- Davidson’s Principles and Practice of Medicine: Edited by: Colledge NR, Walker BR, Ralston SH. 2010, Churchill Livingstone, EdinburghGoogle Scholar
- Weller RPJB, Hunter JAA, Savin JA, Dahl MV: Clinical Dermatology. 2008, Blackwell Publishing, OxfordView ArticleGoogle Scholar
- Gawkrodger DJ: Dermatology: an illustrated colour text. 2003, Churchill Livingstone, EdinburghGoogle Scholar
- Collier J, Longmore M, Turmezei T, Mafi AR: Oxford Handbook of Clinical Specialties. 2009, Oxford University Press, OxfordView ArticleGoogle Scholar
- ABC of Dermatology: Edited by: Buxton PK, Morris-Jones R. 2009, Wiley-Blackwell, OxfordGoogle Scholar
- Cheung ST, Islam N, Keegan D, McGuinness J, Stouthidis N: Crash course: Ophthalmology, Dermatology, ENT. Edited by: Horton-Szar D. 2009, Mosby, St. LouisGoogle Scholar
- Lawrence CM, Cox NH: Physical Signs in Dermatology. 2001, Mosby, St. LouisGoogle Scholar
- Fitzpatrick’s Dermatology in General Medicine: Edited by: Wolff K, Goldsmith LA, Katz SI, Gilcrest BA, Paller As, Leffell DJ. 2007, McGraw-Hill, MaidenheadGoogle Scholar
- Ashton R, Leppard B: Differential Diagnosis in Dermatology. 2004, Radcliffe Publishing Ltd, OxfordGoogle Scholar
- Rook’s Textbook of Dermatology: Edited by: Burns T, Breathnach S, Cox N, Griffiths C. 2010, Wiley-Blackwell, OxfordGoogle Scholar
- Graham-Brown R, Burns T: Lecture Notes: Dermatology. 2011, Wiley-Blackwell, OxfordGoogle Scholar
- Eva KW, Cunnington JPW, Reiter HI, Keane DR, Norman GR: How can I know what I don't know? Poor self assessment in a well-defined domain. Adv Health Sci Educ Theory Pract. 2004, 9 (3): 211-224.View ArticleGoogle Scholar
- Davis DA: Accuracy of Physician Self-assessment Compared With Observed Measures of Competence: A Systematic Review. JAMA. 2006, 296 (9): 1094-1102. 10.1001/jama.296.9.1094.View ArticleGoogle Scholar
- Burge SM: Learning dermatology. Clin Exp Dermatol. 2004, 29 (3): 337-340. 10.1111/j.1365-2230.2004.01463.x.View ArticleGoogle Scholar
- Finlay AY, Coles EC, Dawber RPR, Graham-Brown RAC, Hunter JAA, Marks JM: Dermatology examination performance: wide variation between different teaching centres. Medical education. 1994, 28 (4): 301-306. 10.1111/j.1365-2923.1994.tb02716.x.View ArticleGoogle Scholar
- Chiang YZ, Tan KT, Chiang YN, Burge SM, Griffiths CEM, Verbov JL: Evaluation of educational methods in dermatology and confidence levels: a national survey of UK medical students. Int J Dermatol. 2011, 50 (2): 198-202. 10.1111/j.1365-4632.2010.04664.x.View ArticleGoogle Scholar
- Cliff S, Bedlow AJ, Melia J, Moss S, Harland CC: Impact of skin cancer education on medical students' diagnostic skills. Clin Exp Dermatol. 2003, 28 (2): 214-217. 10.1046/j.1365-2230.2003.01237.x.View ArticleGoogle Scholar
- Marks J: Teaching dermatology to undergraduates in the 1980s. Clin Exp Dermatol. 1984, 9 (3): 263-266. 10.1111/j.1365-2230.1984.tb00793.x.View ArticleGoogle Scholar
- Burge SM, Lancaster T: Assessment in undergraduate dermatology. Clin Exp Dermatol. 2004, 29 (4): 441-446. 10.1111/j.1365-2230.2004.01546.x.View ArticleGoogle Scholar
- Hansra NK, O'Sullivan P, Chen CL, Berger TG: Medical school dermatology curriculum: are we adequately preparing primary care physicians?. J Am Acad Dermatol. 2009, 61 (1): 23-29. 10.1016/j.jaad.2008.11.912.View ArticleGoogle Scholar
- Kerr OA, Walker J, Boohan M: General practitioners' opinions regarding the need for training in dermatology at undergraduate and postgraduate levels. Clin Exp Dermatol. 2006, 31 (1): 132-133. 10.1111/j.1365-2230.2005.01914.x.View ArticleGoogle Scholar
- Moore MM, Geller AC, Zhang Z, Hayes BB, Bergstrom K, Graves JE, et al: Skin cancer examination teaching in US medical education. Arch Dermatol. 2006, 142 (4): 439-444. 10.1001/archderm.142.4.439.View ArticleGoogle Scholar
- Enk CD, Gilead L, Smolovich I, Cohen R: Diagnostic performance and retention of acquired skills after dermatology elective. Int J Dermatol. 2003, 42 (10): 812-815. 10.1046/j.1365-4362.2003.02018.x.View ArticleGoogle Scholar
- Sherertz EF: Learning dermatology on a dermatology elective. Int J Dermatol. 1990, 29 (5): 345-348. 10.1111/j.1365-4362.1990.tb04757.x.View ArticleGoogle Scholar
- Whitaker-Worth DL, Susser WS, Grant-Kels JM: Clinical dermatologic education and the diagnostic acumen of medical students and primary care residents. Int J Dermatol. 1998, 37 (11): 855-859. 10.1046/j.1365-4362.1998.00537.x.View ArticleGoogle Scholar
- Aldridge RB, Xi L, Ballerini L, Fisher RB, Rees JL: Teaching dermatology using 3-dimensional virtual reality. Arch Dermatol. 2010, 146 (10): 1184-1185.View ArticleGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6920/12/27/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.