Skip to main content

Item pre-knowledge true prevalence in clinical anatomy - application of gated item response theory model



Computer and paper examinations in our days are constructed from an item pool which is regularly updated. Given the way that exams are created, one of the major concerns is the security of the items that are being used in order to ensure a good estimation of abilities. The aim of this study is to measure the prevalence of item pre-knowledge in our medical school.


The Deterministic, Gated Item Response Theory Model (DGM) was applied to estimate the prevalence of students who have had item pre-knowledge from six multiple choice examinations of the Clinical Anatomy course at the Faculty of Medicine of University of Porto. Each examination consisted of 100 items with an average of 200 students and 20% repeated items per examination. The estimation of the sensitivity and specificity was based on a simulation study. The sensitivity and specificity estimates, and apparent prevalence were used to estimate true prevalence of cheating students in the examinations under study.


The specificity in the DGM for different simulation scenarios was between 68 and 98%, while the sensitivity ranged from 60 to 91%. The apparent prevalence was between 0.0 and 3.4%, while the true prevalence ranged from 1.2 to 3.7%.


The true prevalence was much lower compared to the students self-reported copying of responses from other students; however, it is important to keep monitoring the pre-knowledge prevalence in order to enforce measures in case an increase occurs.

Peer Review reports


Over a medical school course, it is very common to assess developed skills through multiple choice examinations [1]. An exam is constructed from an item pool which is regularly updated. Given the way that exams are created, one of the major concerns is the security of the items that are being used to ensure a good estimation of abilities. In some situations, students have item pre-knowledge either by over exposure or by item illicit access, and their item responses deviate from the underlying item response theory (IRT) by inflating their test scores [2].

Illicit access to items would be considered academic cheating. Academic cheating is defined as unethical or unauthorized academic activity, and is usually related to examinations [3]. A coordinated and purposeful exposure of items is very worrisome and would magnify examination scores for students who have gained examination pre-knowledge, while punishing honest students and consequently threaten the examination’s validity [4]. Additionally, item responses may move away from the subjacent IRT model [5].

Modeling potential behavior for students possessing prior item knowledge is further compounded by the issue of whether this knowledge is actually used to gain some advantages on the examination [4]. That is, modeling the impact of prior item knowledge is difficult because we need to identify disclosed items and we cannot disregard students who may have access to this information [6].

Several studies have shown that innocuous repeating of a small set of items within a larger examination had little impact on performance [2, 7, 8]. For example, in a national USA certification test in radiography, the same test or a different test form were assigned for the individuals that repeated the examination and indistinguishable score gains between the two groups were found [9]; a similar result was observed for the Medical Council of Canada Evaluating Examination [10]. Normally, the test-maker can control the proportion of reused items when assembling the test, however due to lack of time or economical pressure this is not always done.

The self-reported prevalence of item pre-knowledge was about 25%, while the self-reported prevalence of copying answers during an examination at least once during medical school has ranged from 52% [11] to 67% [12].

The most common way to detect copying answers or item pre-knowledge is using Classical Testing Theory (CTT) or Rasch IRT modeling to identify miss fitting response patterns. These miss fitting response patterns, especially on lower ability candidates on the examination overall, although not conclusive evidence of “cheating” per se, suggest that irregular behavior might have been engaged in order to achieve the correct responses on difficult questions (something we would not expect from low ability candidates). In this context, several classical statistics [3, 13,14,15] and software [16] have been developed to detect cheating on multiple-choice examinations.

Furthermore, several item pre-knowledge detection statistics have been recently developed [17, 18] and those that showed better efficiency were the posterior shift and the Shu Deterministic, Gated Item Response Theory Model (DGM) [19].

In 2013, Shu et al. proposed the DGM that classifies students as cheaters or non-cheaters according to score gain in the exposed items (e.g.: repeated items on previous examinations) compared to the non-exposed items (e.g.: new items) [18].

The proposed DGM consists of a two One-parameter Logistic (1-PL) model mixture [20,21,22] which classifies students into two groups, cheater and non-cheater by conditioning them to two types of items; the first type includes the items that are probably exposed, and the second type, the non-exposed items. The DGM allows item pre-knowledge detection through the analysis of the variation between students’ item pre-knowledge ability and their true ability.

Although, previous studies have measured the apparent prevalence (AP) (percentage of students classified as having item pre-knowledge), no studies have measured the true prevalence (TP) (percentage of students who truly have item pre-knowledge) as they did not take into account the sensitivity (SEN) and specificity (SPE) of the detection method.

In the case of high pre-knowledge item prevalence, the design of the examinations of Clinical Anatomy will need to be restructured.

The aim of this study was to estimate the item pre-knowledge true prevalence among medical students in the course of Clinical Anatomy at the Faculty of Medicine of University of Porto (FMUP) through the application of the DGM.


All multiple choice examinations from the Clinical Anatomy course between 2008 and 2011 were analyzed to estimate the prevalence of students who had item pre-knowledge.

In each year, there were two final examinations which comprised a total of eight examinations. Each examination consisted of 100 standard multiple choice questions (MCQ) (five response options where only one was the correct answer), for a total of 800 items.

Each of the 100 items in each examination was compared with all other examination items in order to verify whether the item had been reused. The year 2008 was considered as the starting year and was excluded from the analysis because it did not contain any reused items. The items classified as reused were treated as exposed items, since students may have memorized items from a previously provided examination. The items used for the first time in the examination were treated as non-exposed items.

Initially, the data description was carried out using CTT in order to better comprehend the items’ characteristics; 1-PL and Two-parameter Logistic (2-PL) IRT models [20,21,22] were applied in order to validate the 1-PL model choice used in the DGM.

The 1-PL and 2-PL models were estimated using the marginal maximum likelihood estimation and the Expectation-Maximization (EM) algorithm [23, 24]. The chosen 1-PL model was the logistic model in which the discrimination parameter was estimated to be identical in all items.

In this study, the selected model was defined according to the Akaike Information Criterion (AIC) [25, 26], the Bayesian Information Criterion (BIC) [27, 28] and Convex Hull (CHull) method [29]. The model that better fits the data has the lowest AIC and BIC values, and the highest CHull value.

The difficulty (percentage of students who correctly answered the item) and discrimination index (biserial correlation between the item and the number of correct answers to the other items) of item examination were described using the mean and standard deviation (SD).

To assess whether there were significant differences between the examinations or number of repetitions, in the difficulty and discrimination indexes, mixed effect models were used with a fixed factor (examination or number of repetitions) and the item-level random intercept to account for the residual correlation within the same (reused) items.

Data were aggregated by item in order to eliminate the residual correlation within students that repeated the examinations; therefore, the previous model needed to include a student- level random intercept. The main reasons for aggregating data were data sparsity due to the reduced number of items reused and the small number of students that repeated the examinations; furthermore, item-level characteristics (e.g. the number of repetitions of the items) were the features of interest in this study.

The recommendations for the interpretation of the difficulty index suggest that values between 0 and 30% indicate a difficult item; values ranging from 31 to 80% imply an item with medium difficulty; values between 81 and 100% can be labeled as an easy question [30]. The recommendations for the interpretation of the discrimination index suggest five categories: values between − 1.00 and − 0.19 indicate negative discrimination; values ranging from − 0.20 to 0.19, weak discrimination; values between 0.20 e 0.29, sufficient discrimination; values from 0.30 to 0.39, good discrimination; and between 0.40 and 1.00, very good discrimination [31].

Cronbach’s alpha was used to assess the examination reliability. Recommendations suggest that examinations with 50 or more items have a good reliability if Cronbach’s alpha value is equal to or greater than 0.8 [32]. The alternative coefficient ωh and ωt of McDonald [33] was used as well to evaluate the reliability (general factor saturation and the inter-consistency, respectively) of the examinations.


As referred previously, DGM is composed by a mixture of two 1-PL models which allows students to be classified into two groups. This classification takes into account the students results obtained in the secure and exposed items. Thus, DGM uses, on the one hand the true ability, θtm, to characterize the real skill of the mth student ,m = 1, …, M, and on the other hand, his/hers cheating ability, θcm to estimate cheating efficiency.

Therefore, DGM classifies each student with item pre-knowledge (cheater) or without item pre-knowledge (non-cheater) according to a specific threshold value.

Each item of the test is classified either as compromised or secure according to the fact that it is a reused item or not. Thus, for each item, i, the item exposure status, Gi, is dichotomously defined as

$$ {G}_i=\left\{\begin{array}{c}1,\mathrm{compromised}\ \mathrm{item}\ i\ \\ {}0,\mathrm{secure}\ \mathrm{item}\ i\end{array}\right.\left(i=1,\dots, I\right) $$

Assuming that true and cheating abilities are known, student can be classified as a cheater if his/her true ability is lower than his/her cheating ability. Therefore, for each student is considered the dichotomous indicator variable Tm

$$ {T}_m=\left\{\begin{array}{c}1,{\theta}_{tm}<{\theta}_{cm}\ \\ {}0,\mathrm{otherwise}\end{array}\right.\left(m=1,\dots, M\right) $$

where Tm = 1 represents that the m th examinee is a cheater.

The goal of conditioning the two item types is to use the information provided from the secured items to infer the level of item-compromise contained in the exposed items. The probability that the m th examinee answers correctly to the i th item is

$$ {P}_i\left({\theta}_{tm},{\theta}_{cm}\right)=P\left({y}_{mi}=1|{\theta}_{tm},{\theta}_{cm},{b}_i\right),\left(m=1,\dots, M,i=1,\dots, I\right), $$

where bi represents the item difficulty index.

Both Gi and Tm are dichotomously defined, therefore, the DGM can be further broken down to four conditional models:

$$ P\left({y}_{mi}=1|{\theta}_{tm},{\theta}_{cm},{T}_m,{G}_i,{b}_i\right)=\left\{\begin{array}{c}\begin{array}{c}P\left({y}_{mi}=1|{\theta}_{tm},{b}_i\right)\ \mathrm{for}\ {T}_m=0,{G}_i=0\\ {}P\left({y}_{mi}=1|{\theta}_{tm},{b}_i\right)\ \mathrm{for}\ {T}_m=1,{G}_i=0\\ {}P\left({y}_{mi}=1|{\theta}_{cm},{b}_i\right)\ \mathrm{for}\ {T}_m=1,{G}_i=1\end{array}\\ {}P\left({y}_{mi}=1|{\theta}_{tm},{b}_i\right)\ \mathrm{for}\ {T}_m=0,{G}_i=1\end{array}\right. $$

When the student is classified as a non-cheater, Tm = 0, the responses to all items are based only on his/her true ability, θtm, and therefore do not depend on θcm. However, when Tm = 1, that is, for students that are cheaters, it is necessary to take into account whether the items are exposed or not. Student answers to the unexposed items (G = 0) are based on their true ability (θtm), while responses to the exposed items (G = 1) are based on their cheating ability (θcm). Accordingly, cheating ability only influences the response probability of cheating students in the exposed items.

Taking into consideration the Gi and Tm values, the probability of the mth student correctly answering item i can be written as a unique expression

$$ P\left({y}_{mi}=1|{\theta}_{tm},{\theta}_{cm},{T}_m,{G}_i,{b}_i\right)=P{\left({y}_{mi}=1|{\theta}_{tm},{b}_i\right)}^{1-{T}_m}\times {\left[\left(1-{G}_i\ \right)P\left({y}_{mi}=1|{\theta}_{tm},{b}_i\right)+{G}_iP\left({y}_{mi}=1|{\theta}_{cm},{b}_i\right)\right]}^{T_m} $$

emphasizing the mixture structure of the model used.

In order to discriminate if the student is classified as cheater or non-cheater, it is necessary to fix a value representing the cut-off point. This threshold was defined according to the probability of a student being a cheater (T = 1), Pc (0 < Pc < 1), by the DGM. Shu et al [11] used the fixed value of 90% as the cut-off point Pc, while in the present study, we also used a classification tree to identify the best cut-off point value of Pc to classify students with or without item pre-knowledge. Classification trees are a statistical method used to construct binary trees, by successive divisions of data according to a rule that divides the data into groups as uniform as possible [34]. Homogeneity between the two constituted subgroups is defined by impurity – a measure that takes the zero value in completely homogeneous subgroups. In classification trees (the response variable is qualitative); impurity can be measured by the amount of entropy, which must be minimized since it measures heterogeneity within groups. Thus, the criterion used to choose the best cut-off point from all possible cut-off point values was the one that minimized entropy.

Simulation study

This subsection aims to describe the conditions of the simulation study that supported the analysis of sensitivity and specificity of the DGM as well as the best choice of the cut-off point that distinguishes cheaters from non-cheaters.

The simulation study was carried out considering the closest conditions to the ones verified in the Clinical Anatomy course examinations. In real data, there were an average of 20 reused items and 200 students per examination, and those values were used in the simulation study. The simulation study must take into account the item pre-knowledge characteristics, including the proportion of item pre-knowledge and the effectiveness of item pre-knowledge. The proportion of item pre-knowledge refers to the percentage of students who have pre-knowledge of the exposed items. The effectiveness of item pre-knowledge is the effective score gain as a result of prior knowledge of the exposed items. According to the score gain level, the most effective students (high-effective) obtain the most effective gain and low effective (low-effective) obtain a lower effective gain. We considered four scenarios with four levels of proportion of item pre-knowledge, 5, 10, 35 and 70%, and two levels of cheating efficacy of item pre-knowledge, high-effective and low-effective. For each of the scenarios, we simulated 100 replicates.

The items’ difficulty (b) was simulated according to a standard normal distribution. The student’s true ability (θt) was simulated according to the standardized normal distribution, N(0, 1) and student’s cheating ability (θc) was obtained by the sum of the effective score gain, (), to true ability. In a non-cheating student, the effective gain is zero, while for a cheating student; it is simulated from a beta distribution. When the cheating category is high-effective, the score gain is characterized by Beta(9, 4)  3 and when it is low-effective, it is simulated according to Beta(5, 5)  3.

Thus, we can summarize the distributions used in the simulation of the parameters related to items and students of the DGM as

$$ {\displaystyle \begin{array}{c}{\theta}_t,b\sim N\left(0,1\right),\\ {}{\theta}_c={\theta}_t+\Delta , \end{array}} $$

with  = 0 for the non-cheater, ~Beta(9, 4)  3 for the cheater high-effective and ~Beta(5, 5)  3 for the cheater low-effective.

Let Ymi, m = 1, …, 200, i = 1, …, 100, be the response of student m to item i. Ymi were generated using the equations

$$ P\left({Y}_{mi}=1\right)={P}_i\left({\theta}_{cm}\right)=\frac{1}{1+{e}^{\theta_{cm}-{b}_i}} $$

for the exposed items and cheaters, and

$$ P\left({Y}_{mi}=1\right)={P}_i\left({\theta}_{tm}\right)=\frac{1}{1+{e}^{\theta_{tm}-{b}_i}} $$

for all other cases.

Estimation of the DGM

The parameters of the DGM were estimated using Markov chain Monte Carlo (MCMC) [35, 36] methods through Gibbs algorithm [37]. The following prior distributions were considered:

$$ {\displaystyle \begin{array}{c}{Y}_{mi}\sim Bernoulli\left({P}_i\left({\theta}_{cm}\right)\right),\\ {}{\theta}_{tm},b\sim N\left(0,1\right),\\ {}{\theta}_{cm}\sim N\left(1,2\right),\\ {}{T}_m=1\kern0.75em \mathrm{when}\kern0.5em {\theta}_{tm}<{\theta}_{cm.}\end{array}} $$

These variables are i.i.d for m = 1, …, 200, i = 1, …, 100

Since the distributions of θtm and θcm do not depend on the considered student, for simplification, considerer θtm = θt and θcm = θc. WinBUGS’ DGM commands are available in Additional file 1.

For each DGM, sample parameters were generated, with dimension 110,000 from the posterior distribution, which include a burn-in period of 10,000 observations to ensure the convergence of Markov chains in the sampling process. Only observation parameters with a 100 iterations jump in order to obtain a sample, with dimension 1,000, of approximately uncorrelated observations were stored.

Estimation of the true prevalence

In real data, we do not know if a student is a cheater or not. When we apply a DGM, it tells us which students were classified by the model as cheaters (positive test). The percentage of those students is referred to as the apparent prevalence (AP) and is obtained by

$$ AP\ \left(\%\right)=\frac{\# positives}{\# total\ students}\times 100 $$

We want to know the percentage of students who are truly cheaters; the true prevalence (TP) [38] is

$$ TP\ \left(\%\right)=\frac{\# cheaters}{\# total\ students}\times 100 $$

A Bayesian approach can be used to estimate the TP [39] using the following relationship with the AP and taking into account the sensitivity (SEN) and specificity (SPE) of the DGM.

$$ AP= TP\times SEN+\left(1- TP\right)\times \left(1- SPE\right). $$

The SEN is the percentage of students who were correctly classified as cheaters and the SPE is the percentage of students who were correctly classified as non-cheaters [40].

To obtain the TP, we used the SEN and the SPE means and SD computed in the simulation study. The minimum SEN and SPE for the uniform distribution were fixed for the DGM classification as the minimum and the maximum mean for all scenarios in the simulation study.

The R software [41] was used for statistical analysis and for programming.

Furthermore, the estimation of parameters was performed by Gibbs algorithm, implemented in WinBUGS through the R2WinBUGS package [42], the rpart package [43] for the classification trees, the ltm package [44] to see which model best fit the data, and for the algorithm distributions display and convergence study, we used the coda packages [45] and mcmcplots [46].


Simulation study

The SEN and SPE for the cut-off point of 90% were obtained by computing the 100 replicates of the simulations for the different scenarios showed in Table 1. The SPE was higher than 90%, while the SEN ranged from 60.3 to 90.7%.

Table 1 Specificity, sensitivity, positive and negative predictive value in each scenario of the simulation study using the cut-off value of 90%

The AP in all scenarios was different compared to the TP (Table 1).

The simulation study showed that for high prevalence, the cut-off value should be decreased, and for low prevalence, the cut-off value should be increased.

Figure 1 presents the estimated gain for each one of the scenarios. We can observe that a cheating student obtains a much higher effective score gain than a non-cheating student. For the non-cheating student, the score gain is very close to 0. If we analyze Fig. 1a and b we can observe that for the same proportion of item pre-knowledge (35%), students obtain a higher effective score gain when it is high-effective; the same happens for the proportion of item pre-knowledge (70%) (Fig. 1c and d).

Fig. 1

Example of the effective gain of the cheating versus non-cheating students. a Proportion 35% & medium-effective. b Proportion 35% & high-effective. c Proportion 70% & medium-effective. d Proportion 70% & high-effective

Application to real data

Data description

A total of 1008 students completed the examination between 2008 and 2011, from those 774 (76.8%), 218 (21.6%), 14 (1.4%) and 2 (0.2%) completed the examination 1, 2, 3 and 4 times, respectively.

Table 2 shows the number of reused items, the number of items reuses, the students’ mean score, and the items’ difficulty and discrimination mean levels and respective Cronbach’s alpha and McDonald’s ωh and ωt for each examination.

Table 2 Data description according to CTT

From a total of 800 items, 84 (10.5%) were reused once, 13 (1.62%) twice, and the percentage of repetitions ranged from 4 to 26%. The mean items’ difficulty index was between 0.57 and 0.66, there were statistically significant differences in the difficulty index by examination (p = 0.0471), and all examinations showed a medium difficulty level. The mean items’ discrimination index ranged from 0.30 and 0.37, and there were statistically significant differences in discrimination index by examination (p = 0.008); however, all examinations presented good discrimination. Cronbach’s alpha was above 0.8 in all examinations, which showed that all examinations have a good reliability. The ωt showed high internal consistency and the ωh moderated the general factor saturation for all examinations.

The index of difficulty increased 3.5% (p = 0.013) in the first repetition and 6.9% (p = 0.036) in the second repetition compared to the first time, meaning that with repetitions, the items were easier for the students (Table 3).

Table 3 Mixed effect models to measure repetition effects on difficulty and discrimination indices using CTT

Goodness-of-fit of 1-PL model

In order to assess if we could use the 1-PL model to fit the data we compared the 1-PL and 2-PL models to verify which one gives the best fit to the real data. Table 4 presents a summary of the goodness-of-fit index for year and period. Using BIC and CHull, the 1-PL model better fits the data in the eight examinations. Using AIC, in five of the eight examinations, the model that fits better is the 2-PL model (Table 4).

Table 4 The goodness-of-fit of the 1-PL and 2-PL models to real data by year and period

Item pre-knowledge prevalence

The DGM estimated that the AP ranged from 0.00 to 3.30%, and the TP after using the information SEN and SPE from the simulation study was between 1.20 and 3.70% for all examinations (Table 5).

Table 5 Mean and SD values obtained for the apparent prevalence and the true prevalence for the examinations between 2009 and 2011

This situation happens in all studied examinations and can be seen in Fig. 2, where for students considered not cheaters regardless of item exposure or not, the percentage of the students correct answers practically does not change; the same cannot be said for students considered cheaters. In this case, the percentage of correct answers in the exposed items increases very significantly when compared to the percentage of correct answers in the unexposed items. This was expected considering that the DGM model more easily detects the students with item pre-knowledge with low ability. Those students will have a high gain in the number of correct answers compared to students with high ability where the gain would be smaller, and consequently more difficult to detect. Additionally, these students (with low ability) will be more effective compared to the high ability students in the exposed items, since the main focus will usually be items memorization from past examinations compared to high ability students that use all types of information and so will not be so effective in memorizing items.

Fig. 2

Percentage of correct answers by item and student type 2011 examination period 1

Focusing only on the non-exposed items, there is a considerable difference between the two groups of students revealing differences in their true skills that can also be explained by the arguments referred to above.


In this work, the DGM was applied to six multiple choice examinations of FMUP’s Clinical Anatomy course. The proportion of pre-knowledge items in the analyzed examinations ranged from 1.2 to 3.7%, that is, in this course, the proportion of item pre-knowledge is low compared to the self-reported prevalence of copying answers during an examination at least once during medical school, which has ranged from 52% [8] to 67% [9] and the self-reported prevalence of item pre-knowledge was about 25%. When compared to the prevalence using detection statistics for copying answers, the prevalence was high, for example, in 11 examinations held by the Royal College of Pediatrics and Child Health, there was a prevalence of 0.1% [3]. In a low-stake test for measuring student proficiency in Grade 4 English, the prevalence of item pre-knowledge was about 9% [11].

The low prevalence in this study may firstly be due to the fact that students do little study by previously provided examinations or to the fact that students study by the previously provided examinations but also simultaneously through other sources and therefore there is no big difference between the students’ true ability and their cheating ability because these students have a high true ability. The second hypothesis is supported by the fact that no differences in the difficulty index were detected between exposed and unexposed items within the examination itself, however, over the years, significant differences in the difficulty index were detected and exposed items became increasingly easy.

This is the first study that tries to measure the TP of item pre-knowledge; other studies have used the AP determined through a diagnostic test, which will differ from the TP. In our case, we showed that the “apparent” prevalence would underestimate/overestimate the TP depending on the examination.

The simulation study was required to assess the effectiveness of the DGM under the same conditions of the real data and the DGM was applied to the real data in order to estimate the TP of cheating students per examination. The simulation study showed the effectiveness of the DGM when the number of items per test is high (100), the proportion of the exposed items is low (20%) and the number of students is small (200). The absolute agreement of the DGM with these conditions was more than 76%. In the previous study by Shu, the effectiveness of the DGM was studied for an examination with 40 items, a proportion of committed items higher or equal to 30% and 15,000 students, and the cut-off point was set at 0.9. Our study showed that in the case of a high pre-knowledge prevalence, the SEN was lower compared to the SPE, thus increasing the bias between the AP and the TP. Changing the cut-off value from 90% to lower values would decrease the difference between the SEN and the SPE, thus decreasing the bias between the AP and the TP (data not shown). If the test-maker has a priori information that the pre-knowledge prevalence is high, they should lower the threshold in order to use the AP as an estimate of the TP.

One possible constraint of this study was the fact that the 1-PL model used by the DGM could not fit the real data and diminished the diagnostic capacity; however, the BIC showed that the 1-PL model had a better fit compared to the 2-PL model.

The analyzed examinations had a medium difficulty, good discrimination and good reliability scores using both the CTT and the IRT, showing that the low prevalence of item pre-knowledge did not have a large impact on the quality of the examinations.

Moreover, it is worthwhile to mention that one restriction of the present work is the small scale of the study. Surely, it would be of interest to apply DGM in a larger scale with the increase of response sample size and the inclusion of clinical courses in which item re-usage is more common. This remains a topic for future research.


The DGM has proved effective in item pre-knowledge detection and the prevalence of item pre-knowledge is low. The simulation study showed that the DGM underestimates/overestimates the TP.

Thus, the threshold established should be lower in the case of high pre-knowledge prevalence in order to diminish the bias between the AP and TP.

We will keep monitoring the pre-knowledge prevalence in order to take measures in the case of an increase. These could be by the exclusion of exposed items for the next examinations or to provide seminars to increase the awareness of this problem.

Availability of data and materials

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.



One-parameter Logistic


Two-parameter Logistic


Akaike Information Criterion


Apparent Prevalence


Bayesian Information Criterion


Convex Hull


Confidence Interval


Classical Test Theory


Deterministic, Gated Item Response Theory Model




Faculty of Medicine of the University of Porto


Item response theory




Markov chain Monte Carlo


Multiple choice questions


Standard deviation






True Prevalence


  1. 1.

    Epstein RM. Assessment in medical education. N Engl J Med. 2007;2007(356):387–96.

    Article  Google Scholar 

  2. 2.

    O'Neill TR, et al. The impact of repeated exposure to items. Teach Learn Med. 2015;27(4):404–9.

    Article  Google Scholar 

  3. 3.

    McManus I, Lissauer T, Williams S. Detecting cheating in written medical examinations by statistical analysis of similarity of answers: pilot study. BMJ. 2005;330(7499):1064–6.

    Article  Google Scholar 

  4. 4.

    Yi Q, Zhang J, Chang H-H. Severity of organized item theft in computerized adaptive testing: a simulation study. Appl Psych Meas. 2008;32(7):543–58.

    Article  Google Scholar 

  5. 5.

    McLeod L, Lewis C, Thissen D. A Bayesian method for the detection of item preknowledge in computerized adaptive testing. Appl Psych Meas. 2003;27(2):121–37.

    Article  Google Scholar 

  6. 6.

    Yi Q, Zhang J, Chang HH. Severity of organized item theft in computerized adaptive testing: an empirical study. ETS Res Rep Ser. 2006;2006(2):i–25.

    Article  Google Scholar 

  7. 7.

    Feinberg RA, Raymond MR, Haist SA. Repeat testing effects on credentialing exams: are repeaters misinformed or uninformed? Educ Meas Issues Pract. 2015;34(1):34–9.

    Article  Google Scholar 

  8. 8.

    Lake E, Chambers DW. Effects of posting examinations and answer keys on students’ study methods and test performance. J Dent Educ. 2009;73(5):601–13.

    Google Scholar 

  9. 9.

    Raymond MR, Neustel S, Anderson D. Same-form retest effects on credentialing examinations. Educ Meas Issues Pract. 2009;28(2):19–27.

    Article  Google Scholar 

  10. 10.

    Wood TJ. The effect of reused questions on repeat examinees. Adv Health Sci Educ. 2009;14(4):465–73.

    Article  Google Scholar 

  11. 11.

    Hrabak M, et al. Academic misconduct among medical students in a post-communist country. Med Educ. 2004;38(3):276–85.

    Article  Google Scholar 

  12. 12.

    Mortaz Hejri S, et al. Academic disintegrity among medical students: a randomised response technique study. Med Educ. 2013;47(2):144–53.

    Article  Google Scholar 

  13. 13.

    Angoff WH. The development of statistical indices for detecting cheaters. J Am Stat Assoc. 1974;69(345):44–9.

    Article  Google Scholar 

  14. 14.

    Wollack JA. Comparison of answer copying indices with real data. J Educ Meas. 2003;40(3):189–205.

    Article  Google Scholar 

  15. 15.

    Zopluoglu C, Davenport EC. The empirical power and type I error rates of the GBT and ω indices in detecting answer copying on multiple-choice tests. Educ Psychol Meas. 2012;72(6):975–1000.

    Article  Google Scholar 

  16. 16.

    Wentzel C. A review of INTEGRITY software: an online application to analyze multiple-choice tests and detect test-taking deception. J Sci Educ Technol. 2006;15(3–4):314–9.

    Article  Google Scholar 

  17. 17.

    Belov DI. Detecting item preknowledge in computerized adaptive testing using information theory and combinatorial optimization. J Comput Adapt Test. 2014;2(3):37–58.

    Google Scholar 

  18. 18.

    Shu Z, Henson R, Luecht R. Using deterministic, gated item response theory model to detect test cheating due to item compromise. Psychometrika. 2013;78(3):481–97.

    Article  Google Scholar 

  19. 19.

    Belov DI. Comparing the performance of eight item preknowledge detection statistics. Appl Psych Meas. 2016;40(2):83–97.

    Article  Google Scholar 

  20. 20.

    s dichotomously definedBaker FB, Kim S-H. Item response theory: parameter estimation techniques. In: Statistics: A Series of Textbooks and Monographs. Boca Raton: CRC Press; 2004.

  21. 21.

    Downing SM. Item response theory: applications of modern test theory in medical education. Med Educ. 2003;37(8):739–45.

    Article  Google Scholar 

  22. 22.

    Hambleton RK, Swaminathan H, Rogers HJ. Fundamentals of itemresponse theory, vol 2. Newbury Park: Sage; 1991.

  23. 23.

    Bock RD, Aitkin M. Marginal maximum likelihood estimation of item parameters: application of an EM algorithm. Psychometrika. 1981;46(4):443–59.

    Article  Google Scholar 

  24. 24.

    Bock RD, Lieberman M. Fitting a response model forn dichotomously scored items. Psychometrika. 1970;35(2):179–97.

    Article  Google Scholar 

  25. 25.

    Akaike H. Information theory and an extension of the maximum likelihood principle, in breakthroughs. In: Kotz S, Johnson NL, editors. Statistics: foundations and basic theory. New York: Springer New York; 1992. p. 610–24.

    Google Scholar 

  26. 26.

    Akaike H. Akaike’s Information Criterion, in International Encyclopedia of Statistical Science. New York: Springer; 2011; p. 25.

    Google Scholar 

  27. 27.

    Schwarz G. Estimating the dimension of a model. Ann Stat. 1978;6(2):461–4.

    Article  Google Scholar 

  28. 28.

    Weakliem DL. A critique of the Bayesian information criterion for model selection. Sociol Methods Res. 1999;27(3):359–97.

    Article  Google Scholar 

  29. 29.

    Bulteel K, et al. CHull as an alternative to AIC and BIC in the context of mixtures of factor analyzers. Behav Res Methods. 2013;45(3):782–91.

    Article  Google Scholar 

  30. 30.

    Tavakol M, Dennick R. Post-examination analysis of objective tests. Med Teach. 2011;33(6):447–58.

    Article  Google Scholar 

  31. 31.

    Ebel RL. Essentials of educational measurement. 3rd ed. Englewood Cliffs: Prentice-Hall; 1979. p. 622.

    Google Scholar 

  32. 32.

    Kehoe J. Basic item analysis for multiple-choice tests. Pract Assess Res Eval. 1995;4(10):20–4.

    Google Scholar 

  33. 33.

    McDonald RP. Test theory: a unified treatment. Hillsdale: Erlbaum; 1999.

    Google Scholar 

  34. 34.

    Breiman L, et al. Classification and regression trees. Boca Raton: CRC Press; 1984.

  35. 35.

    Chib S, Greenberg E. Understanding the metropolis-hastings algorithm. Am Stat. 1995;49(4):327–35.

    Google Scholar 

  36. 36.

    Junker BW, Patz RJ, VanHoudnos NM. Markov chain Monte Carlo for item response models. In: Handbook of item response theory, volume two: statistical tools, vol. 21; 2016. p. 271–325.

    Google Scholar 

  37. 37.

    Smith AF, Roberts GO. Bayesian computation via the Gibbs sampler and related Markov chain Monte Carlo methods. J R Stat Soc Ser B Methodol. 1993;55(1):3–23.

    Google Scholar 

  38. 38.

    Rogan WJ, Gladen B. Estimating prevalence from the results of a screening test. Am J Epidemiol. 1978;107(1):71–6.

    Article  Google Scholar 

  39. 39.

    Speybroeck N, et al. Misclassification errors in prevalence estimation: Bayesian handling with care. Int J Public Health. 2013;58(5):791.

    Article  Google Scholar 

  40. 40.

    Porta MS, et al. A dictionary of epidemiology. Oxford: Oxford University Press; 2014. p. 320.

  41. 41.

    R Core Team. R: A language and environment for statistical computing. Vienna: R Foundation for Statistical Computing; 2017.

  42. 42.

    Sturtz S, Ligges U, Gelman A. R2WinBUGS: a package for running WinBUGS from R. J Stat Softw. 2005;12(3):1–16.

    Article  Google Scholar 

  43. 43.

    Therneau T, Atkinson B, Ripley B. rpart: recursive partitioning and regression trees. In: R package version; 2015.

    Google Scholar 

  44. 44.

    Rizopoulos D. ltm: an R package for latent variable modelling and item response theory analyses. J Stat Softw. 2006;17(5):1–25.

    Article  Google Scholar 

  45. 45.

    Plummer M, et al. CODA: convergence diagnosis and output analysis for MCMC. R News. 2006;6(1):7–11.

    Google Scholar 

  46. 46.

    Curtis SM. mcmcplots: create plots from MCMC output; 2015.

    Google Scholar 

Download references




This work was supported in part by the Portuguese Foundation for Science and Technology (FCT-Fundação para a Ciência e a Tecnologia), through CIDMA - Center for Research and Development in Mathematics and Applications, within the project UID/MAT/04106/2013

Author information




MS had a major contribution in conception and design of the study and revised the manuscript critically and supervised the data analysis. FS-P contributed with the data acquisition and analysis and revised the manuscript critically. MAF revised the manuscript critically. MM and IP had contributed in conception and design of the study and revised the manuscript critically. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Milton Severo.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Additional file

Additional file 1:

WinBUGS DGM Model Commands. (DOCX 17 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Severo, M., Silva-Pereira, F., Ferreira, M.A. et al. Item pre-knowledge true prevalence in clinical anatomy - application of gated item response theory model. BMC Med Educ 19, 284 (2019).

Download citation


  • Deterministic Gated item response theory model
  • Item pre-knowledge
  • Item sharing
  • Item exposure
  • Repeat items