Question-writing as a learning tool for students – outcomes from curricular exams
© Jobs et al.; licensee BioMed Central Ltd. 2013
Received: 12 August 2012
Accepted: 14 June 2013
Published: 21 June 2013
Writing exam questions can be a valuable learning tool. We asked students to construct multiple choice questions for curricular exams in Internal Medicine. The questions for the particular exams were chosen from a pool of at least 300 student-written questions. The uncorrected pool was accessible to all students. We studied the influence of this approach on the students’ learning habits and their test results. We hypothesized that creating a pool of their own questions for the exams could encourage students to discuss the learning material.
All students had to pass 4 exams in 7 fields of Internal Medicine. Three exams were comprised of 20 questions, and we applied the new method in one of these exams. The fourth exam was comprised of 30 questions, 15 of which were chosen from a students’ pool. After all exams had been completed we asked the students to fill in a web-based questionnaire on their learning habits and their views on the new approach. The test-results were compared to the results of the lecturers’ questions that defined high and low performing students.
A total of 102 students completed all four exams in a row, 68 of whom filled in the questionnaire. Low performing students achieved significantly better results in the students’ questions. There was no difference in the number of constructed questions between both groups of students. The new method did not promote group work significantly. However, high performing students stated a stronger wish to be rewarded by good performance.
Creating a curricular exam by choosing questions from a pool constructed by students did not influence the learning habits significantly and favored low performing students. Since the high performing students sought to be rewarded for their efforts, we do not consider the approach applied in our study to be appropriate.
KeywordsQuestion-writing Multiple choice questions Curricular exams Learning behavior Internal medicine
Most lecturers have probably experienced the difficulties of constructing valid questions for written exams. Apart from the different possible question formats  the content of the questions ought to be correct and able to withstand scientific challenge. In some cases a thorough review of literature is necessary in order to exclude misconceptions. This process can be used as a learning tool for students. It has been reported that constructing questions enhances the recall for a studied text . It may also foster a more active and self-determined way of learning  which is more likely to promote a deeper understanding of the subject . Moreover, applying self-written questions in exams may create a constructive learning climate and reduce anxiety .
The construction of questions has occasionally been used as a tool in medical education: In a contest during a physiology class, 37 out of 100 students submitted a total of 912 questions . These questions were graded and discussed in the class. The authors state that question writing motivated students to study and that lively discussions occurred. In another study the construction of multiple choice questions (MCQs) was used as a stimulus for learning in clinical surgery . The questions submitted were of high quality; however, the studied group of students did not achieve better results in the exams than a control group. The method was unpopular at first but was rated more favourable after the exercise.
More recently, Yu and Liu studied the effects of question-posing as compared to question-answering during weekly post-lecture sessions in a cohort of 69 civil engineering sophomores . While there was no difference in academic achievement the question-posing group had significantly higher abilities in cognitive and metacognitive learning strategies after the intervention. Denny and co-workers developed a web-based system in which students create MCQs and answer those created by their peers (http://peerwise.cs.auckland.ac.nz). Students using this system have been shown to produce good quality MCQs and use higher order thinking skills while taking an active role in their learning .
Applying self-written questions in curricular medical exams could have substantial impact on the students’ learning behaviour. In a recent study, 25% of the MCQs in an end-of-year formal examination were derived from a bank of student-generated questions . Many students, however, chose to memorise the question bank as a “high‒yield” strategy for mark inflation, favouring surface rather than deep learning. To date, the literature on the impact of student-generated questions in curricular exams is scarce. In 2010, the student council of our university asked the lecturers of Internal Medicine whether students could submit questions for the exams in this field. The student council would collect the questions and provide a pool of at least 300 appropriate questions per exam to the lecturers. The lecturers would then select the exam questions from the pool. The eventual motive was to achieve a more problem-based approach to learning thereby promoting team work. After a discussion, the advisory board agreed to apply this method in two of four exams. Hence, we had a unique opportunity to study the effects of this procedure in comparison to questions constructed by the lecturers. The main goal of the study was to evaluate the effect on the students’ learning habits in comparison to the conventional approach. We were also interested in the impact of the method on the grade distribution.
In our faculty, Internal Medicine is taught in the fourth year of medical studies. The subjects in this field are allocated into four sections and there is a written exam for each section: Endocrinology and Nephrology (EN), Gastroenterology, Hematology, and Oncology (GHO), Respiratory Medicine and Cardiology (RMC), and Rheumatology (Rh). The students can choose to be tested in the particular subjects in the fourth year or later. During the studied period (2011), the exams consisted of 20 (EN, RMC, and Rh) and 30 (GHO) MCQs. All the questions in EN and RMC, and 15 of the questions in GHO were constructed by the lecturers themselves and underwent a review process by the teaching advisory board of Internal Medicine. All 20 questions in Rh and 15 of the questions in GHO were derived from a pool of at least 300 questions which had been submitted by students three weeks before the exam.
All students had been instructed in writing MCQs using an approved manual, which was made accessible to the students by the students’ council. The advice given by the manual referred to the question type (type-A questions with five choices and one correct answer), the structure (how to weigh stem and options, how to avoid cues), the content (reference to learning objective, relevance), and the hierarchy of knowledge (preference of higher-order skills). Questions were only eligible for the pool if they were of acceptable quality. The selection from this pool was made by the lecturers and the selected questions underwent the same review process through the advisory board as the questions constructed by the lecturers themselves. All the questions were assessed by their correctness (with regard to form, language and content) and their relevance (according to the stated learning objectives). The questions for the exam were chosen from the best proposals in order to cover a broad section of the field and to exclude cues by question interference. The reviewers felt that the quality of the selected questions was not significantly different between the students' and the lecturers' proposals. However, they did not rate the questions numerically. To avoid the possibility that students could guide the selection process by intentionally submitting incorrect questions (which then would be excluded), both lecturers and the advisory board were allowed to make minor corrections. The pool of the submitted (possibly incorrect) questions was accessible to all students.
After all exams had been completed, the tested students were asked via e-mail to fill in a web-based questionnaire. In the questionnaire the students were asked to state the number of questions they had constructed and the hours they had spent studying for the particular exam. They also reported the amount of time they spent using variable learning methods: studying lecture slides or textbooks, working in a group, creating questions, studying the pool of the students’ questions and others (studying clinical guidelines, other pools of questions, web-based resources or miscellaneous). Additionally, they could rate their attitude towards the exam in a 5-point-true-false-scale. In order to combine the students’ results of the exams with the answers given in the questionnaire, their matriculation number was used. All students who filled in the questionnaire gave informed consent for processing their data. The pooled data were further processed without identification of the students. The study was approved by means of the ethics committee of the University of Lübeck. The work was carried out in accordance with the Declaration of Helsinki, and the anonymity of all participants was guaranteed.
Performance of the three groups of students divided into tertiles by the results of the lecturers’ questions
Relative performance [%] of students who took all four exams (n = 102)
Relative performance [%] of students who took all four exams and filled in the questionnaire (n = 68)
(mean ± SD)
(mean ± SD)
(mean ± SD)
(mean ± SD)
87.8 ± 3.4
93.9 ± 3.3
86.8 ± 3.7
92.9 ± 4.2
78.5 ± 2.1
90.9 ± 4.5
77.9 ± 1.5
91.0 ± 3.1
69.4 ± 5.3
88.5 ± 5.1
68.9 ± 6.1
87.9 ± 6.5
Results of the questionnaire
High performer (n=21)
Low performer (n=26)
The grade in internal medicine is important to me.
I consider the grade in internal medicine to be important for future job applications.
I find it important that the exams in internal medicine differentiate between high and low performing students.
I consider a good grade be a reward for my learning efforts.
I have learned more in the system with the students’ questions than in the system with the lecturers’ questions.
Number of constructed questions
Data management, statistical analysis, and graphical presentation were performed using the R software environment . The results of the exams are given as percentage of the achievable correct answers (relative performance; mean ± SD). The Welch test was applied for the comparison of high and low performing students. The results of the ordinal scaled questionnaire and the numbers of constructed questions per student are presented as median [interquartile range] and compared by the Mann–Whitney U test. The total time spent in preparation for the exams was compared using the analysis of variance (ANOVA) followed by pairwise testing with adjustment for multiple testing according to Bonferroni. The particular amount of time in hours spent with different learning methods was compared between high and low performing students using the Welch test. A P-value < .05 was considered significant.
Distribution of students who constructed up to three questions versus students who constructed six or more questions to the high and low performing group, respectively
≥6 constructed questions
≤3 constructed questions
Hours spent preparing for the four exams according to different learning methods as recalled in the questionnaire
Time spent preparing for an exam in hours
Construction of questions
0 ±0 *
pool of questions
The calculation of the difficulty and discrimination index of students' (n = 35) and lecturers' questions (n = 55) revealed that the lecturers' questions were more difficult (0.78 versus 0.90, respectively; P < .001) and discriminated better (discrimination index 0.18 versus 0.12, respectively; P < .05) than the students' questions.
In this study, we report on the effects of students writing questions for their own exams. Writing questions is challenging and might encourage a more intense approach to learning because the writer has to scrutinize and question his own views. All students’ questions were accessible to the learners, however, only in the original, uncorrected students’ version. It was therefore an additional idea to create a discussion among the students about the correctness of the questions, hence promoting team work.
The main result of the study is that our approach did not measurably influence the students’ learning habits in the way we intended. Little time was spent constructing the questions and studying in a group was a consistently unpopular learning method. The time studying the pool of questions by far exceeded the time constructing questions. In preparation for the exam with both types of questions (GHO) the students did not schedule their time differently compared to the exams with lecturers' questions. This suggests that when it comes to answering unknown questions the students would rather rely on their conventional approach to learning than on getting involved in writing and discussing their own questions. The overall time studying for the exam with students' questions only (Rh) was significantly lower than for the other exams (P < .001) and the students spent a great deal of their time memorising the question pool. This may indicate that medical students indeed are strategic learners trying to be time efficient in their preparation for the exams  rather than seeking deeper understanding. However, there may be alternative explanations for the reduced time the students spent learning for the latter exam: Firstly, only one subject in the field of Internal Medicine was tested and, secondly, the exam took place at the end of the semester competing with other exams in the same group of students. Nevertheless, the studied method did not promote learning with a peer group as it intended.
With regards to the test results, low performing students achieved better results while the high performing students did not have any advantage. This may be interpreted as a desirable result because the weaker students were supported. However, due to the fact that the same results occurred in the mixed exam with both students’ and lecturers’ questions (GHO) it is more likely that the weaker students achieved better by learning the pool of questions by heart rather than by understanding the material on a deeper level. The high performing students also spent a lot of time studying the pool, however, they did not achieve a significantly better score. This may be due to a ceiling effect implying that they had already obtained their best possible score.
Exams can serve several purposes: they may be used to assign a grade, and/or to provide feedback to the students and information to the teacher about what the students did not understand . From this point of view the review of the students’ questions allowed the lecturers a good overview of the students’ deficits, and, moreover, of what they perceived to be important. This knowledge can be used to realign the lectures. On the other hand, the self-written exam was neither able to assign a correct grade nor give informative feedback to the students because the greater part of the corrective was in the students’ hands in advance of the exam.
Superordinate goals of an assessment in medical education are to optimize the capabilities of all learners by providing motivation and direction for future learning, to protect the public by identifying incompetent candidates, and to provide a basis for choosing applicants for advanced training . Hence, the most appropriate concept of fairness of an exam is probably whether it is testing what makes the better professional . The grades of the self-written exams did not reflect the appreciation of the lecturers. In our setting all lecturers were experienced physicians. As we have previously shown, physicians have a different view of what is important in medical education than students . From a professional point of view, physicians may be in a better position to oversee the requirements for a medical graduate than the medical students themselves. Given that their questions reflected these requirements, their questions would better fulfil the criteria of fairness than the students’ questions. The disadvantage for the high performing students was also more obvious due to the fact that there was a significantly greater wish for discrimination between high and low achievement as well as a greater sense of the importance of the grade among these higher performing students.
There are, however, several limitations with this study. The number of students studied was limited. We only included the students who chose to take all four exams in a row. Less than half of the cohort fulfilled this criterion, two thirds of whom filled in the questionnaire. However, we believe that as there was no detectable positive effect on learning habits in a group of this size, the overall benefit of the method is likely to be too small to justify the effort of curricular implementation. Secondly, while question writing seems to be a valuable learning tool, we can only speak for the studied method. The procedure was decided on before the study; hence, the study had an observational character. A different approach might have led to a different outcome. A web-based platform to create and discuss the questions as previously described  could have promoted a broader interest in the process of scrutinising the correctness of the questions. Also, a modification in handling the pool (e.g. number of questions, introducing a rating, limiting to reviewed questions) could have changed the results. Letting the students write the questions for their own exam might therefore indeed be a valuable learning tool . However, we were not able to find a positive influence of our method on the learning behaviour and abandoned it in the curriculum.
Letting students write their own exam in the described manner does not appear to have a beneficial influence on learning habits. Low performing students achieve better by learning the questions by heart. Since the high performing students in our study wished their achievements to be acknowledged we considered the method unfair.
Multiple choice question
Endocrinology and nephrology
Gastroenterology hematology, and oncology
Respiratory medicine and cardiology
- Bloom BS, Madaus GF, Hastings JT: Evaluation to improve learning. 1981, New York: Mcgraw-HillGoogle Scholar
- Frase LT, Schwartz BJ: Effect of question production and answering on prose recall. J Educ Psychol. 1975, 67 (5): 628-635.View ArticleGoogle Scholar
- Brown IW: To learn is to teach is to create the final exam. Coll Teach. 1991, 39: 150-153. 10.1080/87567555.1991.9933419.View ArticleGoogle Scholar
- Rash AM: An alternative method of assessment: Using student-created problems. Primus. 1997, 7 (1): 89-95. 10.1080/10511979708965848.View ArticleGoogle Scholar
- Smith K: Let your students write their own tests. 1990, Ireland: Paper presented at the Annual Meeting of the International Association of Teachers of English as a Foreign Language Conference in DublinGoogle Scholar
- Sircar SS, Tandon OP: Involving students in question writing: a unique feedback with fringe benefits. Am J Physiol. 1999, 277 (6 Pt 2): S84-S91.Google Scholar
- Palmer E, Devitt P: Constructing multiple choice questions as a method for learning. Ann Acad Med Singapore. 2006, 35 (9): 604-608.Google Scholar
- Yu FY, Liu YH: The comparative effects of student question-posing and question-answering strategies on promoting college students' academic achievement, cognitive and metacognitive strategies use. J Educ Psychol. 2008, 31 (3): 25-52.Google Scholar
- Bottomley S, Denny P: A participatory learning approach to biochemistry using student authored and evaluated multiple-choice questions. Biochem Mol Biol Educ. 2011, 39 (5): 352-361. 10.1002/bmb.20526.View ArticleGoogle Scholar
- Papinczak T, Peterson R, Babri AS, Ward K, Kippers V, Wilkinson D: Using student-generated questions for student-centred assessment. Assess Eval High Educ. 2011, 37 (4): 439-452.View ArticleGoogle Scholar
- Development Core Team R: A language and environment for statistical computing. 2011, Vienna, Austria: R Foundation for Statistical Computing, http://www.R-project.org/, 3-900051-07-0.Google Scholar
- Mattick K, Dennis I, Bligh J: Approaches to learning and studying in medical students: Validation of a revised inventory and its relation to student characteristics and performance. Med Educ. 2004, 38 (5): 535-543. 10.1111/j.1365-2929.2004.01836.x.View ArticleGoogle Scholar
- Bouville M: Exam fairness. 2008, General Physics, http://arxiv.org/pdf/0803.4235.pdf.Google Scholar
- Epstein RM: Assessment in medical education. N Engl J Med. 2007, 356 (4): 387-396. 10.1056/NEJMra054784.View ArticleGoogle Scholar
- Weitz G, Twesten C, Hoppmann J, Lau M, Bonnemeier H, Lehnert H: Differences between students and physicians in their entitlement towards procedural skills education - a needs assessment of skills training in internal medicine. GMS Z Med Ausbild. 2012, 29 (1): 10.3205/zma000777.
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6920/13/89/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.