- Technical advance
- Open Access
- Open Peer Review
Good exams made easy: The item management system for multiple examination formats
© Hochlehnert et al.; licensee BioMed Central Ltd. 2012
- Received: 24 February 2012
- Accepted: 26 July 2012
- Published: 2 August 2012
The development, implementation and evaluation of assessments require considerable resources and often cannot be carried out by a single faculty/institution. Therefore some medical faculties have founded cooperation projects which mainly focus on the exchange of multiple choice questions (MCQs).
Since these cooperation projects do not entirely support all relevant processes in terms of preparation, implementation and evaluation of assessment, in 2006 the Medical Assessment Alliance (MAA) was founded for mutual support. In addition to MCQs the MAA started to develop innovative assessment formats and facilitate content through a coordinated exchange of experiences. To support cooperation within this network, the web-based Item Management System (IMS) was developed which supports all processes of the assessment workflow as an all-in-one working platform.
At present, the Alliance has 28 partner faculties in Europe. More than 2.800 users in 750 working groups are collaborating. Currently 90.000 questions have been stored in the IMS. Since 2007, nearly 4.600 examinations have been successfully conducted.
This article describes in detail the unique features of the IMS and contrasts it with the item management systems of other associations.
- Assessment alliance
- Quality control
Although medical faculties have considerable freedom in designing their assessments , the development, implementation, and evaluation of such assessments represent a huge challenge. The design of “good” exams not only includes the creation of tasks  but also conducting examinations, reviewing examination papers for content and accuracy, defining the grading system, statistically analyzing the exams, and reporting results. Many faculties use multiple-choice questions because moving forward to more innovative question types is supposed to be a time-consuming and expensive procedure. In the same way, the implementation of review processes in many faculties is proving very difficult due to limited resources . Automated exam correction, grading, statistical analysis, the use of document reading systems, evaluation programs, or the implementation of computer-based assessments have resulted in resource savings. However, for question design, quality assurance of the tasks and exam blueprints there has been only limited progress.
In English-speaking countries there are consortia which have developed item pools to share assessment content (e.g., IDEAL, an international coalition of faculties  with an item database which currently contains approximately 32.000 items (personal communication with Prof. R. Hays, May 2012) or UMAP, a coalition of medical schools in Great Britain ). While they include, for example, the development of common standards, they do not cover all areas of assessment design and implementation in comparable depth (e.g., detailed exam blueprints and implementation of a pre-review ).
The Medical Assessment Alliance (MAA) represents, in contrast to the IDEAL and UMAP consortia, a significant development in using a shared web-based platform, called the Item Management System (IMS). This supports staff in the preparation, implementation, evaluation, and reporting of assessments according to existing guidelines.
This article describes the MAA/IMS and critically evaluates experience with it, future challenges, and possible solutions.
There are annual two-day consensus meetings with the leaders of staff responsible for assessments from all partner faculties. These, for example, jointly develop innovative assessment formats, exchange OSCE checklists, and construct guidelines for the review process. Joint decisions are made on awarding external contracts, adjusting processes, and improving the user-friendliness of the system. In 2010 and 2011 the following topics were addressed in Alliance meetings: Review activity, development of a composite internal progress test and selection tests, exchange of OSCE checklists, and the development of encounter cards.
Since regular formative feedback promotes competency-based teaching , the Alliance is also involved in collaborative research projects which support the faculties (e.g., formative assessments and progress test ).
To strengthen inter-faculty collaboration, the web-based IMS was implemented. It is an all-in-one platform which supports the work of staff responsible for the preparation, administration, evaluation and reporting of assessments . The cost of operating and developing this platform is shared by the entire Alliance as a non-profit organization.
Current figures on the IMS assessment platform
Use of the assessment platform
Currently 2.800 users are registered with IMS who are working together in 750 groups. In these groups, people are no longer just limited to their own faculty but have found inter-faculty colleagues with whom they design items and examinations and cooperate on other aspects of assessment processes.
To date 4.600 examinations composed of a total of 123.000 questions have been successfully administered to more than 800.000 medical students using IMS. 584.000 statistical test parameters (e.g. item difficulty, selectivity, etc.) have been calculated and these help in the compilation of future exams. These have been uploaded to the IMS database.
Development of the exam database
Description of the individual IMS features
Each IMS user receives their own protected user account and can access the IMS based on clearly defined roles and responsibilities with a personalized view. To support collaborative work, colleagues can set up working groups and design exams as a team or conduct quality control processes.
In IMS, the complete assessment sequence is laid out as described by the Assessment Committee of the German Medical Association (Gesellschaft für Medizinische Ausbildung GMA) and the guidelines of the Centre of Excellence for Medical Assessments . In the following, the eight key components of this process are explained:
Assessment content (items) can be created and saved in IMS in various formats, including long menu, free text, MCQs, key feature cases, OSCE checklists, and structured oral examinations . Entering the test items is supported by format-specific input forms. Graphics and figures can be easily integrated into the questions; the integration of sound and video material will soon be available.
Task classification in IMS
In addition to classification by subject and sub-subject, every faculty can create their own classification system in the IMS to classify tasks according to their specific learning targets. This allows, for example in the case of a Swiss Faculty, for classification into a Bachelor’s and Master’s system which is differentiated into several course units and modules. The implementation of a classification system linked to the German National Competency-based Learning Targets Catalogue in Medicine (NKLM)  is in progress.
Task evaluation prior to assessment (pre-review)
According to existing guidelines, all examination questions should be evaluated in a pre-review prior to their use in a live assessment . This can be done in the IMS using a standardised review checklist which was jointly designed by the Alliance. The review can be carried out by colleagues either individually or as groups , which means that more than two reviewers come together in a meeting and review items directly in the IMS.
Collating exam tasks
Examinations can easily be assembled by dragging and dropping questions into exams. Apart from tasks created personally by the authors, they also have access to a public pool available to all Alliance members.
To further promote efficiency in the organisation of assessments, it is possible to choose various blueprint sub-functions. For example, 1) a table of contents blueprint can be used for targeted exam compilation (e.g., defining the proportion of cardiology tasks in an exam on internal medicine), 2) an evaluation blueprint can be used to define the distribution of points, and 3) a date blueprint can be used to monitor all examinations of a faculty over a fixed period of time (e.g. for compiling an exam in a timely manner before the exam date).
Compiling an assessment
IMS allows the user to create printable exam sheets (in PDF format). The items can be permuted using flexible algorithms to create several versions of an examination which could be sorted by subject, item-type, etc.. In addition, standard interfaces for transferring data to computer-based assessment programs or document readers are integrated.
Conducting an assessment
An examination can be conducted on paper, as traditionally the case, computer-supported using machine-readable answer sheets, computer-based in a computer lab, as an oral exam, or as an OSCE. By using IMS interfaces, each partner faculty can use its own established systems or standard Alliance tools, such as the document reading system Klaus®  or the computer-based exam system Campus® , which are available free of charge to Alliance members.
Following the examination, the raw data can be evaluated using IMS. The system has interfaces for importing this data from a document reading system or a computer-based examination system. The system calculates scores for subtests and total test scores based on parameters which can be individually set, such as the maximum number of points of a question or the grading algorithm (e.g. application of automatic adjustment rules, etc.). IMS also calculates detailed test statistics, including question and exam indicators such as difficulty, accuracy or the index of discrimination .
Potentially wrong tasks can be identified through conspicuous statistical test parameters. In such a case, the points awarded in such tasks can be corrected as part of a post-review or tasks can be removed from the overall exam score. Such information is stored in the IMS in order to document the deficiencies and to facilitate revision by the author. Furthermore, comments by teaching staff or test participants can be recorded.
Apart from direct support for the aspects listed above, IMS offers further infrastructure assistance to study deaneries such as monitoring upcoming exams and keeping a record of actors used as standardized patients.
Actor database additional module
The IMS also provides support for the development and implementation of standardised objective clinical or oral-practical examinations (e.g. OSCE). The organisation and management of actor-patients is often an additional challenge due to the large number of people involved. The module integrated into IMS for this purpose can be used to manage events, assign roles, and reimburse the actor-patients. In addition to its use in OSCE examinations, this module can also be used to organise the use of actor patients in the teaching of the regular curriculum. This not only frees up considerable resource in the planning of teaching activities but also improve quality assurance.
Exchanging assessment experiences and exam tasks created at other faculties form an important basis for improving medical education. The use of a web-based platform to support assessment significantly eases such cooperation. Both the MAA and the IMS assessment platform are in line with the Consensus Statement and Recommendations from the 2010 Ottawa conference which states that “the medical education community still needs to develop a deeper understanding of how technology can underpin and extend assessment practices” .
Within 5 years it has been possible to establish a successful and stable inter-faculty cooperation among 28 faculties. Despite differences in the human and technical resources of the partner faculties, MAA has a high acceptance which, amongst other things, is clearly demonstrated by the speed with which it has grown to include 28 participating faculties and the intensive use of the system. The accession of the medical faculty of Geneva as a French-speaking university and of the faculties in the Netherlands as Dutch-speaking universities is a further signal that the IMS not only offers extra value due to its extensive collection of questions (which for the most part available in German only) but is a complete system with a high degree of functionality and the workflow assistance. This differentiates the IMS from other known assessment associations such as the IDEAL Consortium, which focuses mainly on building a database of exam questions  and UMAP, which involves a commitment to the quality assurance of tasks. A faculty conducts quality assurance reviews as a contribution and in return receives access to a certain number of reviewed tasks. However, a department can only view the descriptors of each question prior to using it, not the question itself . Other international networks are either not focused on assessment (FAIMER [19, 20] or do not support the entire assessment process using IT .
One of the advantages of the IMS is that authors can decide whether they want to share items with others. Some people choose not to publish their items because they only want them to be used locally. The full benefit of the system will be realized as authors begin to see the quality and efficiency inherent in cross faculty collaboration.
The analysis of user experience shows that the further development of the IMS promotes quality development processes at participating faculties. For example, the review activity of IMS participants increased significantly during the past year following changes to the process which included e.g. more flexible ways to perform a review of an item or a better presentation of the review results. To further expand review activity, cooperation in specialised assessment associations (e.g. paediatrics, occupational health) is consistently promoted and supported by the MAA (e.g., advice, presentations at professional conferences, and help with networking).
A technical challenge is the development of powerful search algorithms. At present, more than 17.000 questions are managed in the shared pool and the number is rising steadily. Using additional filters which take statistical test parameters, review results, usage frequencies of the tasks, and MeSH terminology into account , there are plans to increase the efficiency of the search functions in future to further improve access.
Currently, in spite of the variety of question formats (long menu, free text, MCQs, key feature cases, OSCE, etc.) available in the IMS, the main type used are MCQs (96%). This mirrors results which showed that MCQs are still the predominant format at German medical faculties . In this regard, the IMS offers the possibility of supporting further implementation of urgently needed competency-based assessment formats and content due to its support features for practical and oral assessments.
The development of competency-based assessment formats and the creation of the relevant content are increasingly seen as a major challenge internationally. Qualification frameworks like the competencies developed by the Accreditation Council for Graduate Medical Education (ACGME)  or the National Competency-based Learning Targets Catalogue in Medicine (NKLM) for Germany  which are in line with the CanMEDS roles  will contribute to giving further impetus to competency-based assessment. As these competency-based Catalogues are applied nationwide, the individual learning targets can be deposited in the IMS and linked directly with the exam papers. Additionally, the Alliance regularly submits joint applications for joint research projects, for example focussing on the question of how to integrate other test formats which are now on the horizon. To catalyse these changes, the German Federal Ministry of Education and Research has funded a five-year project starting 2012 to develop competency-based examination formats and generate content.
In the past, assessments at individual faculties have focused on summative examinations. It is crucial for steering the students’ learning process to receive feedback on their current knowledge through formative tests. For example, the progress test is a format, growing in popularity, demonstrates growth in knowledge over time . As the creation and evaluation of progress tests is very resource intensive for a single faculty, the MAA has already done several steps to realize this project as a joint endeavour together with the faculty of Maastricht (Netherlands).
The concept of having many partners who contribute to improving quality by providing their special knowledge is one of the main reasons for the success of MAA. Through the consensus meetings and the direct exchange between doctors, psychologists, and IT developers, the system enhances the quality development processes at the faculties (see Table 1). Additionally the participation in the MAA provides economic benefits for each participating faculty because this way of cooperating directly results in savings for every partner.
Over a three year period, nearly 450.000€ were needed to develop the first version of the IMS. Following this, the development and maintenance of the IMS is financed by the fees of the participating faculties. The costs for one faculty in the pilot phase is 10 € per student. Thereafter, the cost per year is 30 € per student to a maximum of 25.000 € per year per faculty. This includes organizational and technical support as well as further development of the IMS. Participating faculties benefit economically because they do not have to develop a similar system but only pay for the running costs of the IMS.
Through the cooperation of various university partners, the task of creating fair, high quality, competency-based assessments can be shared. Overall, the MAA has fulfilled the expectations in terms of the creation and design of exams. The increase in the number of users and the continuous increase in the number of items suggest that the IMS has been accepted and is used extensively. The IMS has proven itself as a practical and efficient solution for the entire assessment workflow, so that new challenges can be explored and addressed.
We would like to thank the teaching staff and the assessment coordinators at our participating faculties for their support and valuable suggestions for further development of the IMS, in particular: Waltraud Georg, Markus Stieg (Berlin), Sybille Bergmann, Anke Graupner, Matthias Hinz (Dresden), Urte Fahron, Thomas Rotthof (Düsseldorf), Meral Franz, Michael Scheib (Erlangen), Peter Brüstle, Ferdinand Berger, Irmgard Streitlein-Böhme (Freiburg), Bernard Cerutti, Mauro Serafin, Nu Viet Vu (Geneva), Henrik Kropp, Joachim Kreuder (Giessen), Christian Münscher, Andreas Fuhrig, Kathrin Lagodny (Göttingen), Richard Melamed (Frankfurt), Wolfgang Hampe, Christian Tolasch (Hamburg), Irmgard Braik, Cordula Harter, Sören Huwendiek, Martina Kadmon, Markus Krautter, Petra Schling, Barbara Schneider, Stefan Titz (Heidelberg), Harald Affeldt, Anna Klein (Mainz), Havanna Gencer, Harald Fritz (Mannheim), Klaus-Jochen Klose, Christine Schäfer (Marburg), Daniel Bauer, Martin Fischer, Matthias Holzer (LMU Munich), Friederike Holderried, Kim-Alexandra Schweizer, Eduard Vlad (Tübingen), Claudia Grab, Anna Vander Beken (Ulm), Ingrid Koscholleck-Szappanos, Sebastian Schubert (Witten/Herdecke).
The development of the Item Management System was supported financially by the Ministry of Science and Arts, Baden-Württemberg, Germany. The German Federal Ministry of Education and Research has funded a five-year project starting 2012 to develop competency-based examination formats and generate content.
- Frenk J, Chen L, Bhutta ZA, Cohen J, Crisp N, Evans T, Fineberg H, Garcia P, Ke Y, Kelley P, et al: Health professionals for a new century: transforming education to strengthen health systems in an interdependent world. Lancet. 2010, 376 (9756): 1923-1958. 10.1016/S0140-6736(10)61854-5.View ArticleGoogle Scholar
- Möltner A, Duelli R, Resch F, Schultz JH, Jünger J: Fakultätsinterne Prüfungen an den deutschen medizinischen Fakultäten. GMS Z Med Ausbild. 2010, 27 (3): Doc44.Google Scholar
- Jünger J, Möltner A, Lammerding-Köppel M, Rau T, Obertacke U, Biller S, Narciß E: Durchführung der universitären Prüfungen im klinischen Abschnitt des Medizinstudiums nach den Leitlinien des GMA-Ausschusses Prüfungen: Eine Bestandsaufnahme der medizinischen Fakultäten in Baden-Württemberg. GMS Z Med Ausbild. 2010, 27 (4): Doc57.Google Scholar
- Prideaux D, Gordon J: Can global co-operation enhance quality in medical education? Some lessons from an international assessment consortium. Med Educ. 2002, 36 (5): 404-405. 10.1046/j.1365-2923.2002.01240.x.View ArticleGoogle Scholar
- Freeman A, Nicholls A, Ricketts C, Coombes L: Can we share questions? Performance of questions from different question banks in a single medical school. Med Teach. 2010, 32 (6): 464-466. 10.3109/0142159X.2010.486056.View ArticleGoogle Scholar
- Fischer MR: Leitlinie für Fakultäts-interne Leistungsnachweise während des Medizinstudiums: Ein Positionspapier des GMA-Ausschusses Prüfungen und des Kompetenzzentrums Prüfungen Baden-Württemberg. GMS Z Med Ausbild. 2008, 25 (1): Doc74.Google Scholar
- Brass K, Hochlehnert A, Jünger J, Fischer MR, Holzer M: Studiumbegleitende Prüfungen mit einem System: ItemManagementSystem für die Medizin. GMS Z Med Ausbild. 2008, Jahrestagung der Gesellschaft für Medizinische Ausbildung (GMA), Greifswald, Doc37.Google Scholar
- Jünger J, Lauber H, Schultz JH, Herzog W: 5 Jahre Kommunikations- und Interaktionstraining von Medizinstudenten (Medi-KIT). GMS Z Med Ausbild. 2007, 24 (1): Doc68.Google Scholar
- Iobst WF, Sherbino J, Cate OT, Richardson DL, Dath D, Swing SR, Harris P, Mungroo R, Holmboe ES, Frank JR: Competency-based medical education in postgraduate medical education. Med Teach. 2010, 32 (8): 651-656. 10.3109/0142159X.2010.500709.View ArticleGoogle Scholar
- Coombes L, Ricketts C, Freeman A, Stratford J: Beyond assessment: feedback for individuals and institutions based on the progress test. Med Teach. 2010, 32 (6): 486-490. 10.3109/0142159X.2010.485652.View ArticleGoogle Scholar
- Hochlehnert A, Brass K, Schultz JH, Jünger J, Fischer M, Georg W: IT-unterstützte Prüfungen mit einem qualitätssichernden System. GMS Z Med Ausbild. 2008, Jahrestagung der Gesellschaft für Medizinische Ausbildung (GMA); German Medical Science GMS Publishing House, Greifswald, Doc08gma18.Google Scholar
- Tekian A, Yudkowsky R: Assessment in health professions education. Oral examinations. Edited by: Downing SM, Yudkowsky R. 2009, Routledge, New York, 269-286.Google Scholar
- Hahn EG, Fischer MR: Nationaler Kompetenzbasierter Lernzielkatalog Medizin (NKLM) für Deutschland: Zusammenarbeit der Gesellschaft für Medizinische Ausbildung (GMA) und des Medizinischen Fakultätentages (MFT). GMS Z Med Ausbild. 2009, 26 (3): Doc35.Google Scholar
- Hochlehnert A, Brass K, Feistner L, Kurtz W, Möltner A, Schultz JH, Fischer MR, Jünger J: Fortschritte durch Vernetzung: Präsentation der aktuellen Entwicklungen im Prüfungsverbund Medizin und Vorstellung der Weiterentwicklungen an der Prüfungsplattform Item-Management-System (IMS). 2010, Jahrestagung der Gesellschaft für Medizinische Ausbildung (GMA). German Medical Science GMS Publishing House, Ruhr-Universität Bochum, Doc10gma40.Google Scholar
- [http://www.blubbsoft.de]Google Scholar
- Ruderich F, Bauch M, Haag M, Heid J, Leven FJ, Singer R, Geiss HK, Jünger J, Tonshoff B: CAMPUS–a flexible, interactive system for web-based, problem-based learning in health care. Medinfo. 2004, 11 (Pt 2): 921-925.Google Scholar
- Möltner A, Schellberg D, Jünger J: Grundlegende quantitative Analysen medizinischer Prüfungen. GMS Z Med Ausbild. 2006, 23 (3): Doc53.Google Scholar
- Amin Z, Boulet JR, Cook DA, Ellaway R, Fahal A, Kneebone R, Maley M, Ostergaard D, Ponnamperuma G, Wearn A, et al: Technology-enabled assessment of health professions education: Consensus statement and recommendations from the Ottawa 2010 conference. Med Teach. 2011, 33 (5): 364-369. 10.3109/0142159X.2011.565832.View ArticleGoogle Scholar
- Burdick W, Amaral E, Campos H, Norcini J: A model for linkage between health professions education and health: FAIMER international faculty development initiatives. Med Teach. 2011, 33 (8): 632-637. 10.3109/0142159X.2011.590250.View ArticleGoogle Scholar
- Norcini J, Burdick W, Morahan P: The FAIMER Institute: creating international networks of medical educators. Med Teach. 2005, 27 (3): 214-218. 10.1080/01421590500126379.View ArticleGoogle Scholar
- Prideaux D, Teubner J, Sefton A, Field M, Gordon J, Price D: The Consortium of Graduate Medical Schools in Australia: formal and informal collaboration in medical education. Med Educ. 2000, 34 (6): 449-454. 10.1046/j.1365-2923.2000.00568.x.View ArticleGoogle Scholar
- Coletti MH, Bleich HL: Medical subject headings used to search the biomedical literature. J Am Med Inform Assoc. 2001, 8 (4): 317-323. 10.1136/jamia.2001.0080317.View ArticleGoogle Scholar
- Lurie SJ, Mooney CJ, Lyness JM: Measurement of the general competencies of the accreditation council for graduate medical education: a systematic review. Acad Med. 2009, 84 (3): 301-309. 10.1097/ACM.0b013e3181971f08.View ArticleGoogle Scholar
- Frank JR, Danoff D: The CanMEDS initiative: implementing an outcomes-based framework of physician competencies. Med Teach. 2007, 29 (7): 642-647. 10.1080/01421590701746983.View ArticleGoogle Scholar
- Schuwirth L, Bosman G, Henning RH, Rinkel R, Wenink AC: Collaboration on progress testing in medical schools in the Netherlands. Med Teach. 2010, 32 (6): 476-479. 10.3109/0142159X.2010.485658.View ArticleGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6920/12/63/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.