Skip to main content

Standard setting Very Short Answer Questions (VSAQs) relative to Single Best Answer Questions (SBAQs): does having access to the answers make a difference?

Abstract

Background

We investigated whether question format and access to the correct answers affect the pass mark set by standard-setters on written examinations.

Methods

Trained educators used the Angoff method to standard set two 50-item tests with identical vignettes, one in a single best answer question (SBAQ) format (with five answer options) and the other in a very short answer question (VSAQ) format (requiring free text responses). Half the participants had access to the correct answers and half did not. The data for each group were analysed to determine if the question format or having access to the answers affected the pass mark set.

Results

A lower pass mark was set for the VSAQ test than the SBAQ test by the standard setters who had access to the answers (median difference of 13.85 percentage points, Z = -2.82, p = 0.002). Comparable pass marks were set for the SBAQ test by standard setters with and without access to the correct answers (60.65% and 60.90% respectively). A lower pass mark was set for the VSAQ test when participants had access to the correct answers (difference in medians -13.75 percentage points, Z = 2.46, p = 0.014).

Conclusions

When given access to the potential correct answers, standard setters appear to appreciate the increased difficulty of VSAQs compared to SBAQs.

Peer Review reports

Background

Single Best Answer Questions (SBAQs) are widely used in medical assessment including high stakes licensing exams such as the US Medical Licensing Examination, membership examinations of many of the UK Royal Colleges, and final examinations of UK medical schools. However there has been criticism of this question format, for being subject to cueing and not reflecting real life clinical practice [1, 2].

Compared to SBAQs, Very Short Answer Questions (VSAQs) are a relatively novel assessment method that has been proposed as a solution to this problem. Like SBAQs, VSAQs have a clinical vignette followed by a lead-in question. However instead of having a list of answer options to choose from, the candidate provides their own answer of between one and five words in length. The candidate’s answers are marked against a set of preapproved answers [2,3,4,5,6] and, due to recent advances in technology, they can be delivered and marked electronically [3,4,5,6]. Any answers that do not match the preapproved options can then be reviewed to consider if they should be marked correct and added to the future lists of approved answers [2,3,4,5,6]. The cueing associated with SBAQs is mitigated with VSAQs as the answer options are removed [3]. Student performance in exams has been shown to be affected by using this question format, indicating that students find this question type more challenging [3,4,5,6]. VSAQs have been shown to be a better representation of candidates’ unprompted level of knowledge, with a recent study showing that the average student scored 21 percentage points lower on the VSAQ compared to the SBAQ of the same stem [3].

There is considerable variation in the means of assessment and methods of standard setting across medical schools, with the Angoff method being reported as the most widely used in high stakes written examinations [7,8,9]. Standard setting can be categorised into relative (or norm-referenced) and absolute (or criterion-referenced). Relative approaches are established based on a comparison of those who take the assessment to each other, for example a pass mark created based on the number of examinees that will pass [10]. They are useful for when the assessment is used for selection and the number of places is limited [10]. Absolute methods are set by determining the amount of exam content that must be answered correctly in order to pass, for example candidates must answer 60% of items to successfully pass [10]. These are more commonly used in high stake examinations as they are useful for determining whether examinees meet requirements for a standard [10]. It is possible for all examinees to pass or fail using this type of standard [10]. The Angoff is a test centred, absolute method of standard setting [11]. The basic Angoff method involves a panel of expert judges making estimates on the proportion of borderline candidates that would answer each item correctly. These estimates are then averaged across all items and judges to create a standard cut-off score. This method depends on the panel’s familiarity with the hypothetical borderline group, their characteristics and response to exam items [11]. They must also be familiar with the standards that students are expected to meet at the level they are taking the test in order to pass [11].

Whilst several standard setting methods have been examined empirically for SBAQs [11], standard setting methodology for VSAQs has not been studied to date. As VSAQs have been successfully introduced into undergraduate assessment [2,3,4,5,6], the question of how to set the pass mark for this assessment method needs to considered.

It has previously been shown that standard setting estimates for SBAQs are significantly affected by a judge knowing, or not knowing the answer to the item [12]. Verheggen et al. found that a judge’s knowledge of the subject and their stringency as a judge impacted on the standard set for an item and therefore the standard was not purely a reflection of the difficulty of the item [12]. As VSAQs are designed to have a range of accepted answers, provision of these for the judges when standard setting may be even more significant. Bourque et al. (2020) looked at standard setting SBAQs for a national postgraduate exam (using the Ebel method) and found no difference in scoring regardless of whether the answers were provided to the judges or not [13].

Using a set of common stem items, we set out to study whether the question format (VSAQs versus SBAQs) affects the pass mark. We also investigated whether having access to the answers had an effect on the pass marks set for both SBAQ and VSAQ formats.

Methods

Two 50-question assessment papers were created using the same question vignettes, one paper using the VSAQ and the other paper in an SBAQ format with five answer options. These items had previously been used in a formative assessment of 1,417 volunteer final year medical students [3]. The papers were standard set using the Angoff method based on the guidelines used by the Medical Schools Council Assessment Alliance (MSCAA) [9, 10].

Twenty three teaching faculty from Imperial College School of Medicine were trained on standard setting in undergraduate examinations using the Angoff method, through a face-to-face workshop. This allowed participants to arrive at a common understanding of what constitutes a borderline candidate. Participants were randomised into four groups to standard set the papers in different formats, as per Table 1.

Table 1 Session design by group

Eleven partcipants judged the paper without access to the answers, as the student would see it, and twelve received the correct answers with a justification for that answer, as typically happens in standard setting practice (Fig. 1).

Fig. 1
figure 1

Question formats: A—SBAQ without answer; B—SBAQ with answer; C—VSAQ without answer; D—VSAQ with answers

In order to account for the impact of the order in which standard setters saw VSAQs and SBAQs half the groups saw VSAQs first before SBAQs and the other half saw SBAQs before VSAQs. This meant that the 50 items were judged twice by each group of standard setters, once in VSAQ and once in SBAQ format. A washout period of six weeks between session 1 and session 2 was created to prevent standard-setters being subject to cueing from previous standard setting sessions.

Data Collection

Each participant was asked to standard set the paper using the Angoff method, by judging each item in the paper using the question “What proportion of consistently just safe newly qualified doctors would get the question correct?”. Scores were submitted using an electronic survey tool. For the VSAQs they were able to submit a response between 0 and 100%. For the SBAQs this was adjusted to a response between 20 and 100% to allow for the 20% chance that the candidate can guess the correct answer from the five options provided with no prior knowledge.

Data Analysis

The anonymised standards data for each participant were downloaded from the electronic survey tool into an Excel file. The data were transferred into Stata V.16 for analysis. The mean of each participant’s standard from the Angoff method for the 50 items was calculated to produce their overall pass mark. For each group the median of all members’ pass marks was calculated for the VSAQ exam and the SBAQ exam.

To determine if the question format influenced the standard set, a paired-data Wilcoxon Sign Rank test on participants’ overall pass marks by question format was carried out separately for the participants that had the question answers (groups A & B) and the participants that did not (groups C and D). To determine if having access to the answers influenced the standard set, an independent samples Mann Whitney U test was carried out separately on the overall pass marks set for the VSAQ and for the SBAQ papers (Groups A + B vs. Groups C + D). For all statistical significance testing a critical p-value of 0.0125 was used given the use of multiple comparisons. Ethical approval for the study was granted by the Imperial College London Medical Education Ethics Committee (MEEC) (MEEC1920-178).

Results

Question format

Table 2 presents summary statistics comparing standards set for the SBAQ and VSAQ formats of the assessment, with and without the answers. There was a statistically significant difference between VSAQ and SBAQ pass marks set by the groups that had access to the answers. The SBAQ pass mark was set higher than the VSAQ pass mark, with a median difference of 13.85 percentage points (Z = -2.82, p = 0.002). There was not a statistically significant difference between the VSAQ and SBAQ pass marks set by the groups that did not have access to the answers (median difference -1.90 percentage points, Z = 0.45, p = 0.700).

Table 2 Summary statistics comparing standards set for SBAQs and VSAQs

Access to Answers

For VSAQs, having the answers resulted in a statistically significant reduction in the pass marks set, as shown in Table 2 (difference in medians -13.75 percentage points, Z = 2.46, p = 0.014). For SBAQs, having the answers did not make a significant difference to the pass mark set (difference in medians -0.25 percentage points, Z = 0.06, p = 0.952).

Discussion

In this study, the question format affected the pass mark set by standard setting judges when they were given access to the answers (as is usual in standard setting practice). When standard setters were shown the answers for VSAQs, they produced a lower pass mark for the VSAQ paper. It has been shown that students score an average 21 percentage points lower on VSAQs [3], and this study suggests that this is taken into account to some extent by standard setters with access to the answers, who set an median pass mark of almost 14 percentage points lower for the VSAQs.

In addition, we investigated if there are different pass marks set when standard setting judges do or do not have access to the answers in VSAQs vs SBAQs. Standard setters who could see the answer and justification for that answer set a lower median pass mark for the VSAQs. We hypothesise that this is related to having access to the range of accepted VSAQ answers, which gives a indication of the degree of difficulty of the question. This is in contrast to studies with SBAQs, where it has been suggested that access to the answers is likely to cause judges to underestimate the difficulty of the question [12], or access to the answers made no difference to the standard set using the Ebel method [13]. Our study found that for SBAQs having the answers did not make a significant difference to the pass mark set, which supports previous findings by Bourque et al [13].

As judges were provided with both the correct answers and the explanatory justification for that answer, it is not clear what the relative contribution of each of these is to the variation in standard setting judgements. Being given the correct answer could make the a judge perceive the question is easier, but the justification might highlight the complexity, and this may have a differing effect on VSAQs and SBAQs. Further research is needed to understand this relationship.

A limitation of our study was that we were not able to hold a group discussion, as is considered best practice when standard setting for high stakes assessment. Group standard setting meetings result in sharing of information and discussion of questions that often results in constructive revision of scores [8]. It has also been shown to improve method reliability and reduces the number of judges required [14]. This is a valuable part of the process, especially when members of the standard setting panel may be less familiar with VSAQs, and so are likely to benefit from sharing of experience. Providing standard setters with question facility for VSAQs, or typical performance differences for VSAQs versus SBAQs, may also help when setting standards in this unfamiliar question type. This was not done in our study as the questions were used formatively so performance data is not likely to be a true reflection of summative assessment.

A further limitation of our study is the small sample size, which was due to finding suitable members of faculty within one institution, and must be kept in mind when interpreting the findings. We limited eligiable participants to those who had experience in final medical school examinations and were involved in delivery the undergraduate currciculum, to ensure the highest quality judges in our standard setting panel. A future study of standard setters across a wider cohort of medical schools would allow a larger sample size and could also look at judges’ characteristics (average age, years standard setting, years spent in undergraduate teaching for example) that may affect the standards they set.

As far as we are aware, our study is the first to look at standard setting for VSAQs. It is also the first study to demonstrate the importance of the standard setting panel having access to the answers when scoring VSAQs. As VSAQs are increasingly introduced into undergraduate medical assessments, it opens the discussion for what must be considered when identifying the ideal standard setting method for this novel question format. Our study demonstrates the feasibility of using the Angoff method to standard set this novel question type in undergraduate medical education. In addition, it provides a platform for further research, including comparing other recognised methods of standard setting – the Cohen method which would consider setting a standard in relation to the performance of the cohort, and the Ebel method which asks judges to consider the importance of the knowledge tested as well as the difficulty of the question.

Conclusions

The potential benefit of integrating VSAQs into undergraduate medical assessments has already been demonstrated, so it follows that a validated standard setting method is needed if they are to be used in high stakes examinations. To our knowledge this is the first study comparing standard setting with and without the answers in VSAQs. Further research on a larger scale is warranted to determine if the effect we have seen persists in a larger and more varied population of standard setters. Based on the present study, we recommend that answers should be provided to the standard setters to help them arrive at a valid standard.

Availability of data and materials

The datasets used and analysed during the current study are available from the corresponding author on reasonable request.

Abbreviations

VSAQs:

Very short answer questions

SBAQs:

Single best answer questions

MSCAA:

Medical schools council assessment alliance

References

  1. Veloski JJ, Rabinowitz HK, Robeson MR, Young PR. Patients don’t present with five choices: An alternative to multiple- choice tests in assessing physicians’ competence. Acad Med. 1999;74(5):539–46.

    Article  Google Scholar 

  2. Sam AH, Hameed S, Harris J, Meeran K. Validity of very short answer versus single best answer questions for undergraduate assessment. BMC Med Educ. 2016;16(1):266.

    Article  Google Scholar 

  3. Sam AH, Westacott R, Gurnell M, Wilson R, Meeran K, Brown C. Comparing single-best-answer and very-short-answer questions for the assessment of applied medical knowledge in 20 UK medical schools: Cross-sectional study. BMJ Open. 2019;9(9):e032550.

    Article  Google Scholar 

  4. Sam AH, Peleva E, Fung CY, Cohen N, Benbow EW, Meeran K. Very short answer questions: A novel approach to summative assessments in pathology. Adv Med Educ Pract. 2019;10:943–8.

    Article  Google Scholar 

  5. Sam AH, Fung CY, Wilson RK, Peleva E, Kluth DC, Lupton M, et al. Using prescribing very short answer questions to identify sources of medication errors: A prospective study in two UK medical schools. BMJ Open. 2019;9(7):e028863.

    Article  Google Scholar 

  6. Sam AH, Field SM, Collares CF, van der Vleuten CPM, Wass VJ, Melville C, et al. Very-short-answer questions: reliability, discrimination and acceptability. Med Educ. 2018;52(4):447–55.

    Article  Google Scholar 

  7. MacDougall M. Variation in assessment and standard setting practices across UK undergraduate medicine and the need for a benchmark. Int J Med Educ. 2015;31(6):125–35.

    Article  Google Scholar 

  8. Yeates P, Cope N, Luksaite E, Hassell A, Dikomitis L. Exploring differences in individual and group judgements in standard setting. Med Educ. 2019;53(9):941–52.

    Article  Google Scholar 

  9. Taylor CA, Gurnell M, Melville CR, Kluth DC, Johnson N, Wass V. Variation in passing standards for graduation-level knowledge items at UK medical schools. Med Educ. 2017;51(6):612–20.

    Article  Google Scholar 

  10. McKinley DW, Norcini JJ. How to set standards on performance-based examinations: AMEE Guide No. 85. Med Teach. 2014;36(2):97–110.

    Article  Google Scholar 

  11. Bandaranayake RC. Setting and maintaining standards in multiple choice examinations: AMEE Guide No. 37. Med Teach. 2008;30(9–10):836–45.

  12. Verheggen MM, Muijtjens AMM, Van Os J, Schuwirth LWT. Is an Angoff Standard an Indication of Minimal Competence of Examinees or of Judges? Adv Heal Sci Educ 2006 132. 2006;13(2):203–11.

    Google Scholar 

  13. Bourque J, Skinner H, Dupré J, Bacchus M, Ainslie M, Ma IWY, et al. Performance of the Ebel standard-setting method in spring 2019 royal college of physicians and surgeons of canada internal medicine certification examination consisted of multiple-choice questions. J Educ Eval Health Prof. 2020;20:17.

    Google Scholar 

  14. Fowell SL, Fewtrell R, McLaughlin PJ. estimating the minimum number of judges required for test-centred standard setting on written assessments. Do discussion and iteration have an influence? Adv Heal Sci Educ. 2008;13(1):11–24.

    Article  Google Scholar 

Download references

Acknowledgements

Not applicable

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

AHS, KRM and CAB designed and delivered the standard setting sessions. KRM and CAB carried out the statistical analysis, with support from AHS. All authors were involved in the overall design of the study, and all provided major contributations to writing the manuscript. All authors reviewed and approved the final manuscript.

Corresponding author

Correspondence to Amir H. Sam.

Ethics declarations

Ethics approval and consent to participate

Ethical approval for the study was sought from and granted by the Imperial College London Medical Education Ethics Committee (MEEC) on 22/11/2019 (ref number: MEEC1920-178). All methods were performed in accordance with the Declarations of Helsinki and informed consent was obtained from all participants for the study.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Sam, A.H., Millar, K.R., Westacott, R. et al. Standard setting Very Short Answer Questions (VSAQs) relative to Single Best Answer Questions (SBAQs): does having access to the answers make a difference?. BMC Med Educ 22, 640 (2022). https://doi.org/10.1186/s12909-022-03693-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12909-022-03693-0

Keywords

  • Assessment
  • Undergraduate
  • Standard setting