Skip to main content

Assessing professional competence in optometry – a review of the development and validity of the written component of the competency in optometry examination (COE)



Credentialing assessment for overseas-educated optometrists seeking registration in Australia and New Zealand is administered by the Optometry Council of Australia and New Zealand. The aim was to review the validation and outcomes of the written components of this exam to demonstrate credentialing meets entry-level competency standards.


The Competency in Optometry Examination consists of two written and two clinical parts. Part 1 of the written exam comprises multiple choice questions (MCQ) covering basic and clinical science, while Part 2 has 18 short answer questions (SAQ) examining diagnosis and management. Candidates must pass both written components to progress to the clinical exam. Validity was evaluated using Kane’s framework for scoring (marking criteria, item analysis), generalization (blueprint), extrapolation (standard setting), and implications (outcome, including pass rates). A competency-based blueprint, the Optometry Australia Entry-level Competency Standards for Optometry 2014, guided question selection with the number of items weighted towards key competencies. A standard setting exercise, last conducted in 2017, was used to determine the minimum standard for both written exams. Item response theory (Rasch) was used to analyse exams, produce reliability metrics, apply consistent standards to the results, calibrate difficulty across exams, and score candidates.


Data is reported on 12 administrations of the written examination since 2014. Of the 193 candidates who sat the exam over the study period, 133 (68.9%) passed and moved on to the practical component. Ninety-one (47.2%) passed both the MCQ and SAQ exams on their first attempt. The MCQ exam has displayed consistently high reliability (reliability index range 0.71 to 0.93, average 0.88) across all 12 administrations. Prior to September 2017 the SAQ had a set cutscore of 50%, and the difficulty of the exam was variable. Since the introduction of Rasch analysis to calibrate difficulty across exams, the reliability and power of the SAQ exam has been consistently high (separation index range 0.82 to 0.93, average 0.86).


The findings from collective evidence support the validity of the written components (MCQ and SAQ) of the credentialing of the competency of overseas-educated optometrists in Australia and New Zealand.

Peer Review reports


Credentialing or professional licensing is a key regulatory activity of many professional bodies. It serves many functions, including protection of the public, protection of a profession’s scope of practice, and is an important means for individuals to enter a profession and gain employment [1]. One of the key applications of the credentialing processes is to ascertain the competence and suitability of professionals trained outside a particular jurisdiction, when those competencies are not automatically recognised by virtue of the granting institution [2,3,4]. Credentialing assessments aim to ensure applicants are able to provide healthcare that meets the competency standards in the country they are seeking registration, while meeting workforce requirements, facilitating global mobility of practitioners, and safeguarding both the public and the professions [5, 6].

The Competency in Optometry Examination (COE) is one such credentialing examination that is administered by the Optometry Council of Australia and New Zealand (OCANZ). OCANZ is a not-for-profit company tasked with protecting the eye health of the Australian and New Zealand public by applying quality standards to local optometry education and training in these two countries, as well as assessing overseas trained optometrists against competency standards. OCANZ has been conducting the COE twice-yearly since 1997. The COE tests the ability of overseas trained optometrists to meet the Optometry Australia Entry-level Competency Standards for Optometry [7], excluding the specific skills to be endorsed for ocular therapeutics, which are assessed separately in the Assessment of Competence in Ocular Therapeutics (ACOT) examination or an accredited program in ocular therapeutics following successful completion of the COE. The COE examination comprises two written examinations, a competency-based assessment of clinical skills and patient consultation examinations (Fig. 1). Candidates who pass the COE are eligible to apply for limited registration with the Optometry Board of Australia (OBA) and provisional registration with the Optometrists and Dispensing Opticians Board in New Zealand (ODOB).

Fig. 1

Flow diagram showing the entry pathways for overseas-trained optometrists to undertake the Competency in Optometry Exam (COE) and the possible outcomes. 1A full list of eligible qualifications can be found at 2To be eligible to sit the COE an applicant must provide evidence of successful completion of a course of 4 years’ full-time study (including at least one equivalent full time academic year, spent primarily in direct contact with patients to experience and learn about clinical practice, including diagnosis and management of patients), or a course of 3 years’ full-time study followed by one year supervised clinical practice after which a professional examination is passed. 3The multiple choice question (MCQ) and short answer question (SAQ) examinations are undertaken in the same sitting over two consecutive days. 4Limited registration with the Optometry Board of Australia (OBA) and provisional registration with the Optometrists and Dispensing Opticians Board in New Zealand (ODOB). The Assessment of Competence in Ocular Therapeutics (ACOT) examination or an accredited program in ocular therapeutics must then be completed within specified timeframes to gain full registration

Eligible candidates are required to pass the written examination before being able to proceed to the clinical section of the examination (Fig. 1). The written examination comprises two examination papers. The first paper, a multiple-choice question examination (MCQ) focuses on clinical science. The second paper, a short-answer question examination (SAQ) focuses on diagnosis and management. The written examination is held face to face simultaneously, over two consecutive days, at up to eight invigilated venues in Australia and overseas depending on candidate demand. At their first attempt candidates must sit both the clinical science examination MCQ and the diagnosis and management examination SAQ at the one sitting. If one of the two papers is failed, the candidate will have one further opportunity to repeat only the failed paper at a second sitting. If at the second sitting they fail that paper again, they will need to re-sit both papers at their next and at any subsequent attempts. There is no limit to how many times an eligible candidate can sit the written examination.

Entry-level competency standards for optometry in Australia were first developed in 1993 [8], and have been revised and updated in 1997 [9], 2000 [10], 2008 [11] and 2014 [7]. These competency standards are examined in, and inform the development of, all components of the COE. In this paper we describe the development, continual improvement and validation processes that have been undertaken by OCANZ to ensure the COE appropriately addresses the above competencies and credentials overseas-trained optometrists to practice in Australia and New Zealand.

Validity is a core requirement in any assessment, but perhaps especially in credentialing testing where the stakes are high and the consequences significant. Credentialing assessments must be conducted in a transparent, defensible and equitable manner in order to produce results that are reliable, valid and fair. While exam reliability can be measured quantitatively [12] and fairness can be evaluated against widely accepted criteria [13, 14], validity is best conceptualized as a qualitative judgement, presented as a reasoned argument based on evidence from multiple sources [15]. For many years validity was regarded as an inherent property of a test; the test characteristics were usually assessed through reference to psychometric indicators such as reliability coefficients, test means and pass rates [16]. This approach ignores the context and potential consequences of the test. A test might be valid for some purposes and contexts, but not for others. Accordingly, modern test validity theory now focuses on the use and interpretation of the test [17, 18]. Rather than a test being inherently valid or non-valid, it is the use and application of the test that needs to be evaluated. While psychometric indicators can still provide valuable information about a test, these are now seen as just part of the validation process, which in contemporary approaches is depicted as a form of argument or claim about the decisions or inferences which are based on the results of a particular test.

Several frameworks have been developed to operationalise such an approach to validation. Two of the most influential in the field of health practice are those developed respectively by Messick [16] and Kane [17]. Messick’s framework organises the potential sources of validity evidence into five categories, namely: content, internal structure, relationships with other variables, response process and consequences [16]. Kane’s framework goes beyond sources of evidence to represent the types of inferences which reflect the overall argument and chain of reasoning, namely: scoring, generalisation, extrapolation and implications [17]. While either approach can offer a useful framework for contemporary validation processes, Kane’s approach provides a particularly useful way of prioritising and situating the different forms of evidence within the overall argument and inferential structure [19]. In this framework, attention is given to a competency-based argument that involves linking test scores to statements about competence, and then to conclusions about expected performance in practice [20]. Furthermore, decisions about licensure or certification within this framework are based on expected performance in practice [21].

Accordingly, in this paper we organise and present the relevant evidence pertaining to the OCANZ assessment process based on Kane’s approach. The primary aim of this paper is to review and describe the validation process of one example of a credentialing examination in the health professions field. This information should be useful for other health professionals developing a credentialing system utilising written examinations as part of their process.


According to each level of Kane’s framework [19], the relevant arguments for a written examination such as the COE include reference to the following components of the test development:

  1. i.

    Scoring: e.g. choice of item format and scoring method; process of item development and peer review for determination of correct response (MCQ) and marking guide (SAQ); marker training and calibration; item analysis and descriptive statistics of candidate performance; resolution of any marker discrepancies.

  2. ii.

    Generalisation: e.g. representativeness of the test in relation to the OCANZ competencies (i.e. the test blueprint); relative weightings of items across the relevant domains; the internal consistency of the results, including reliability coefficient and item correlations.

  3. iii.

    Extrapolation: e.g. authenticity of tasks and the relationship between a test result and real-world proficiency, and/or performance on other related measures; the acceptability and appropriateness of the designated minimum standard (as reflected in the pass mark).

  4. iv.

    Implications: e.g. broader or ‘downstream’ effects or impact of the test, such as pass rates, subsequent performance of successful candidates on the next stage of assessment, and ultimate impact on patient safety and clinical performance in optometric practice; appropriate consideration of impact on candidates including cost, transparency, preparation requirements and support.

Such considerations are presented and discussed in greater detail below, as part of the validation process for the use of the COE exam for determining minimum acceptable competence of overseas-trained optometrists for practice in Australia and New Zealand.

Examination content and constructs

The written component of the COE consists of two exams: an MCQ exam focused on clinical science and an SAQ exam focused on diagnosis and management (Fig. 1). The MCQ exam contains 144 items to be answered in 180 min, consisting of 120 scored questions and 24 non-scored ‘pilot’ questions undergoing validation for future examinations. (Prior to 2018, the MCQ examination consisted of 132 questions, 12 of which were non-scored questions.) The candidates are not made aware of which questions are scored and which are non-scored. All MCQ items are written in the single best answer format, consisting of a ‘stem’, a single correct answer (‘key’), and three incorrect options (‘distractors’). Candidates are instructed to determine the single response that best answers the question.

The MCQ exam assesses candidates’ foundational knowledge of basic biomedical, vision, optical, and clinical sciences, along with their ability to apply this knowledge in a clinical scenario. The questions vary between knowledge recall questions, to more contextual questions that rely on clinical reasoning and integration. The questions for each exam are drawn from an item bank containing approximately 600 questions. The questions in the bank have been written over the course of 12 years by subject matter experts commissioned by OCANZ. Pre-existing questions in the item bank were blueprinted against the 2008 competencies [11]. The questions in the bank were re-blueprinted in 2017/2018 to cover the appropriate clinical science competencies from the revised Optometry Australia Entry-level Competency Standards for Optometry 2014 [7] that were adopted in 2017. A list of the 2014 competencies that are assessed, and the approximate number of questions from each competency found in a typical MCQ exam, can be found in Additional file 1: Appendix 1.

The SAQ exam consists of 18 short answer questions to be completed in 180 min. Each question carries an equal weighting, is scored out of 10 marks (with half-marks), and is scored to a marking rubric. The SAQ exam has been designed to test the candidates’ higher level diagnostic decision making processes [22, 23], in conjunction with the foundational knowledge assessed in the clinical science MCQ exam. Candidates are variously required to: describe abnormal or normal features; discuss observations in anatomical, biochemical, microbiological and/or pathological terms; offer a diagnosis or diagnoses to account for observations and provide justifications for the diagnoses; suggest appropriate treatment or management, including criteria for referral or monitoring; and list systemic, ocular and visual signs and symptoms associated with the condition. The SAQ questions consist of multiple parts based upon short clinical vignettes, often accompanied by a photograph of the clinical condition. Unlike the MCQs that are blueprinted against a single competency per item, the SAQs assess candidates simultaneously across multiple competencies. A list of the 2014 competencies that are assessed, and the approximate number of questions from each competency found in a typical SAQ exam, can be found in Additional file 1: Appendix 2.

All examination scripts are anonymised for marking, with only a candidate identification number provided to the optometry school contracted to undertake the marking. Individual questions are marked by the same examiner across all candidates, and papers from individual candidates that are judged as ‘borderline’ are reviewed by an independent marker. The raw score for each question is used for psychometric analysis as outlined in detail below. Prior to 2018 the marking of the SAQ scripts was rotated amongst the optometry education providers in Australia and New Zealand, but since then, following an open call for expressions of interest, it has been assigned to a single optometry school to promote marker consistency and assessment expertise.

Quality assurance and standard setting

New items are regularly produced and specifically commissioned when areas are identified as having insufficient cover in the item bank as a result of changing professional competencies. All item writers are trained by an educational expert prior to producing items. A formal Question Writing Guide was commissioned to establish approved guidelines for the writing of new items. New items are reviewed, and edited where necessary, by the Written Sub-committee of the OCANZ Examination Committee, then piloted as non-scored items to validate them before they are included in the question bank as scored items. Any questions that perform poorly as a pilot item are reviewed by the Written Sub-committee and either discarded or edited and re-piloted.

The entire COE process was externally audited by an experienced educational consultant in 2015/2016. This review concluded that the COE was consistent with international best practice for this type of credentialing examination, with minor improvements suggested around the examination processes.

The examination cutscores are determined through a formal process of standard setting. The minimum passing standard was re-established at a workshop led by an experienced psychometrician and attended by content experts from Australia and New Zealand in February 2017. Two standard setting approaches were used for the MCQ items based on the two most recently administered exams: 30 items from the September 2016 exam were assessed using a modified Angoff procedure with Beuk and Hofstee adjustment; 26 items from the April 2016 exam were assessed using a bookmark method [24]. The cutscores derived from the standard setting exercise were highly correlated with the cutscores that were implemented in the April and September 2016 exams, thereby validating the cut-score locations. The SAQ items were standard set using nine questions from the September 2016 exam using an Angoff borderline rating method. Follow-up standard setting workshops will be held every 3–4 years to ensure appropriate maintenance of the OCANZ scale and cut-scores for the COE.

Psychometric analysis of results

Excel Psychological and Educational Consultancy Pty Ltd. (EPEC; are contracted by OCANZ to perform psychometric analysis of the COE written exam results. The Rasch model, which estimates and takes into account individual question difficulty and person ability, is used to determine cutscores based on variations in these from one exam to another (reviewed in [25]).

The MCQ results are analysed using both classical statistics and Rasch analysis. In order to maintain comparable cutscores over different exams, a linking and equating process is performed through the use of 20 common items (out of the 120 scored items) between successive examination papers. The calculated difficulty estimates and errors of the common items in the exam being analysed are standardised and adjusted to equate them to the common items in the previous exam paper. Chi-square significance is calculated for each of these linking items, and if any item falls outside of the 95% confidence bands it is removed from the exam calibration. The difficulties of the remaining items are anchored to the results of the previous exam, and the cutscore is set such that the candidate ability required to pass is maintained across examinations.

Rasch analysis is also undertaken on the SAQ scores following a log transformation to achieve equality of intervals on a logit scale given the polytomous nature of the raw data for each question. As for the MCQ examination, the SAQ contains two common, linking questions from the previous exam, and along with content expert standard setting scores this enables equating of the cutscores to maintain the same required candidate performance levels for a pass across exams. Any questions from the MCQ or SAQ exam that are identified as performing poorly through the Rasch analysis are referred back to the content experts in the Written Examination Sub-committee for checking prior to final result release.

Ethics and data statement

This was a retrospective analysis of the outcomes of the written examinations based on data collected by OCANZ over the preceding 6 years. The research was conducted in accordance with the tenets of the Declaration of Helsinki and was approved by The University of Melbourne, the School of Health Sciences Human Research Ethics Committee. A waiver was granted as written informed consent was not able to be retrospectively obtained from the candidates that had sat the examinations. The data are not publicly available due to it containing information that could compromise the privacy of the examinees or the integrity of the credentialing examination. The data was deidentified such that individual examinees privacy could not be compromised. The authors declare that the aggregated data supporting the findings of this study are available within the article.


Candidate results and demographic data

Data from 6 years of administration of the COE, from 2014 to 2019, were analysed for this report, representing 12 administration rounds in total. There were 193 candidates who sat at least one component of the exam during this period, with 272 total administrations over the 6 years as a result of some candidates sitting multiple times (see Fig. 1 for progression pathways). There were 133 candidates (68.9%) who passed both the MCQ and SAQ components of the COE and moved on to the practical component across the six-year period (12 administrations) analysed. A detailed analysis of the candidate results and demographic data can be found in Additional file 1: Appendix 3.

MCQ exam performance

As a consequence of the linking and equating undertaken as part of the Rasch analysis of the results the overall difficulty of the exam has remained relatively stable across administrations, with mean candidate scores ranging from 51.5 to 73.3% (Table 1 and Fig. 2). The MCQ exam has displayed highly consistent reliability, with an average reliability index of 0.88 across 12 administrations (range 0.71 to 0.93; Table 1).

Table 1 Descriptive statistics for all multiple choice question (MCQ) exam administrations
Fig. 2

Mean candidate scores and corresponding cutscores and pass rates for each administration of the multiple choice question (MCQ) component of the Competency in Optometry Exam (COE)

The relationship between cutscores and candidate performance is fairly consistent, with increases in the cutscore mirroring increases in the mean candidate score (Figs. 2 and 3). A strong correlation between the MCQ mean candidate score and the cutscore was observed (r2 = 0.74, p < 0.001). No correlation was observed, however, between the pass rate and the cutscore (r2 = 0.02, p = 0.627). This is possibly the result of low pass rates in April 2016 (38.5%), when there were only 13 candidates, and August 2018 (47.4%), when 8 out of 19 candidates were repeating, suggesting pass rate is influenced by a cohort effect.

Fig. 3

Correlation between candidate performance and cutscore for the multiple choice question (MCQ) and short answer question (SAQ) exams. A strong relationship was observed for the MCQ exams (r2 = 0.74, p < 0.001) and the SAQ exams from September 2017 to August 2019 (r2 = 0.95, p = 0.005). The SAQ exams from April 2014 to April 2017 used a standard cutscore of 50%. The high correlation for the SAQ exams from September 2017 to August 2019 highlights the benefit of using Rasch analysis to set the cutscore compared to the 50% cutscore used previously

SAQ exam performance

Prior to the September 2017 offering of the SAQ exam Rasch analysis was not conducted, linking questions not included and the cutscore was set at 50% for all administrations. As such the overall difficulty of the SAQ exam prior to September 2017 was more variable compared to the exams after the introduction of Rasch analysis (Table 2 and Fig. 4). Across all 12 administrations the mean candidate scores ranged from 38.1 to 60.2%. Prior to September 2017 the mean scores ranged from 38.1 to 60.2%, while the mean scores from September 2017 onwards ranged from 45.5 to 59.5% showing less variability (Fig. 4). Since the introduction of Rasch analysis, the SAQ exams have displayed consistent reliability and good power, with an average separation index of 0.86 across five administrations (range 0.82 to 0.93; Table 2). The questions in the SAQ since September 2017 have been well targeted to the abilities of the candidates, with mean person abilities averaging around zero logits. The SAQ exams also closely fit the Rasch model with Chi-square item-trait interaction probabilities that are highly insignificant. These data all indicate that the reliability of the SAQ exam has improved following the introduction of Rasch analysis in September 2017 compared to the greater variability seen with a 50% cutscore in earlier iterations.

Table 2 Descriptive statistics for all short answer question (SAQ) exam administrations
Fig. 4

Mean candidate scores and corresponding cutscores and pass rates for each administration of the short answer question (SAQ) component of the Competency in Optometry Exam (COE). Rasch analysis was introduced from September 2017 to set the cutscore and analyse the exam. Prior to this a cutscore of 50% was used

The relationship between cutscores and candidate performance in the SAQ following the introduction of Rasch analysis is very consistent, with increases in the cutscore mirroring increases in the mean candidate score (Figs. 3 and 4). A strong correlation between the SAQ mean candidate score and the cutscore was observed (r2 = 0.95, p = 0.005). As with the MCQ exams, no correlation was observed between the pass rate and the cutscore (r2 = 0.60, p = 0.126).

MCQ item analysis

The selection of MCQ items for each exam is based on desired weightings for each of the competencies examined in the MCQ exam (see Additional file 1: Appendix 1). Since April 2015, 1198 MCQ questions have been administered across ten 120 item exams (two items were removed from scoring following expert review, see Table 1), with most questions having been administered in more than one exam. Across all 1198 questions the average item facility was 64.8 ± 21.1%, with an average discrimination index of 0.24 ± 0.25 across 1056 of these questions (items with either 0% or 100% correct responses having no discrimination index). When the items were grouped by competency, the mean facilities (range 56.3 to 94.8%) and discrimination indices (range 0.18 to 0.32) indicated that all competencies were being assessed at the appropriate level by the MCQ items, and performing as expected in separating the stronger candidates from those with less knowledge. There were no competencies that candidates performed extremely poorly on, and only competencies that had very few cumulative question administrations displayed mean facilities above 80%. Overall this suggests that the MCQ is well balanced and is achieving the desired outcome of assessing the clinical science focus appropriately.

Each SAQ exam question addresses more than one competency (see Additional file 1: Appendix 2 for the approximate weighting of each competency across all SAQ questions in each exam). It is therefore not possible to separate out and analyse the individual performances of each competency with the SAQ items as has been undertaken for the MCQ items.


Credentialing of health professionals trained outside of a particular jurisdiction is an important regulatory function for many professional bodies. A number of different approaches to credentialing have been used by a wide range of health professions [2,3,4]. A core requirement of credentialing assessment is the validity of the assessment. This paper described the approach that OCANZ took with the COE to credential overseas trained optometrists within Kane’s validity framework [17]. The use of Kane’s framework to understand each of the key components of the test development (Scoring, Generalisation, Extrapolation and Implications [19]) has resulted in a valid, repeatable and feasible written exam. The use of Rasch analysis alongside Kane’s framework led to an improved understanding of the exam metrics, the quality of individual questions and comparison of exam difficulty from one administration to the next [17, 25], ensuring the assessment was both valid and reliable [12].

Individual candidate performance was not dependent on sex, age, or years since graduation, but did vary with the region where the primary optometry degree was obtained. The outcomes of the linking questions were used to determine exam difficulty for each administration and vary the cut-scores used to determine the pass mark if required. Cut scores for the MCQ varied from 50.8 to 70% and for the SAQ from 46.1 to 57.6%. This represents relatively small variations in exam difficulty that were accounted for in determination of the score required by candidates to show competency with the content. The reliability of the MCQ exam ranged from 0.71 to 0.93 (mean 0.84) but was 0.78 or greater for all but one exam. This represents acceptable to good reliability [26]. Separation index for the SAQ ranged from 0.82 to 0.93 (mean 0.87) confirming that the SAQ was able to distinguish candidates in terms of their ability on this exam. Separation index values greater than 0.8 indicate that there is good fit of the Rasch model to the scores, which confirms the SAQ exams have good power and reliability [26]. The analysis shows that the exams were able to discriminate candidates on their knowledge and that variations in exam difficulty were accounted for. As candidates may pass one exam and fail the other this suggests the MCQ and SAQ exams are testing different competencies and that both exams types are important measures. Candidates should have confidence in this robust and fair assessment process that means that they have the knowledge required to progress to the next stage of the COE.

Content experts were able to use the information about individual questions to replace poorly performing questions and to identify competency areas for which new questions needed to be developed. New questions were piloted before inclusion in the question bank and this ensured only high quality questions were added. Training on how to write MCQs has been shown to lead to better quality questions [27], as does peer review [28], and both of these are undertaken when producing new MCQs and SAQs for the COE.

As the skills and competencies of the profession change [7,8,9,10,11] the exam content will continually need to be updated to ensure it remains aligned to these, as seen in other professional exams [29, 30]. The method described here is a manageable way of ensuring an up-to-date, appropriate, content-specific exam. The analysis of content alignment to professional competencies does require a range of expertise and is relatively time intensive [31] and costly, but quality control processes are important. The quality control processes implemented by OCANZ in setting the COE gives candidates sitting the exams confidence in their outcomes. Indeed, the processes undertaken in the validation of the COE closely align with those detailed in both Kane’s framework [19], and the steps undertaken to ensure validity by other credentialing bodies [2,3,4]. The optometry profession and general public can be assured through the validation of the credentialing exam that overseas-trained optometrists registered in Australia and New Zealand have met the appropriate competency standards to practice.

The scopes and competencies of the practice of Optometry in Australia and New Zealand continue to expand. The challenge for the validation process of the written component of the COE is to ensure it regularly reviews itself so that its outcomes continue to reflect the changing needs of the profession. The end result of the process should be to continue to assess the ability for overseas applicants to meet contemporary Australian and New Zealand competencies, especially when these competencies may differ from those of the home country. We believe the process of change we describe reflects the evolution of the OCANZ validation process over time and the robustness of these changes.

The small number of candidates undertaking each offering of the exam is a limitation of our analysis. Despite the low participant numbers, the consistently high reliability scores, with low inter-examination variability, should give confidence that the validation of the exams is appropriate. The small candidate numbers do mean, however, that the content experts need to be cautious when assessing the reliability of individual multiple-choice questions from a single examination instance. A further limitation of the present study is that it focused solely on the written component of the credentialing examination process. Additional analysis of the clinical examinations will need to be undertaken to ensure validity of the entire COE process.


The findings from collective evidence from the past 6 years of administration of the COE provide support for the validity of the written components (MCQ and SAQ) of the credentialing of the competency of overseas-educated optometrists undertaken in Australia and New Zealand. Candidates sitting the exam can have confidence in their outcomes. While the presented case is specific to the credentialing of optometrists, the processes described provide a blueprint that can be adopted by a wide range of health professions undertaking credentialing assessments.

Availability of data and materials

The data are not publicly available due to it containing information that could compromise the privacy of the examinees or the integrity of the credentialing examination. The data was deidentified such that individual examinees privacy could not be compromised. The authors declare that the aggregated data supporting the findings of this study are available within the article.



Competency in Optometry Examination


Multiple choice question


Optometry Council of Australia and New Zealand


Short answer question


  1. 1.

    Davis-Becker S, Buckendahl CW. Testing in the professions : credentialing policies and practice: Routledge; 2017.

    Google Scholar 

  2. 2.

    Chiavaroli NG, Beck EJ, Itsiopoulos C, Wilkinson P, Gibbons K, Palermo C. Development and validation of a written credentialing examination for overseas-educated dietitians. Nutr Diet. 2018;75(2):235–43.

    Article  Google Scholar 

  3. 3.

    Ozkan UR. Foreign qualification recognition regimes for internationally trained professionals: the case of pharmacists. J Int Migr Integr. 2018;19(2):367–89.

    Article  Google Scholar 

  4. 4.

    Price T, Lynn N, Coombes L, Roberts M, Gale T, de Bere SR, et al. The international landscape of medical licensing examinations: a typology derived from a systematic review. Int J Health Policy Manag. 2018;7(9):782–90.

    Article  Google Scholar 

  5. 5.

    Norcini JJ, Boulet JR, Opalek A, Dauphinee WD. The relationship between licensing examination performance and the outcomes of care by international medical school graduates. Acad Med. 2014;89(8):1157–62.

    Article  Google Scholar 

  6. 6.

    Rizwan M, Rosson NJ, Tackett S, Hassoun HT. Opportunities and challenges in the current era of global medical education. Int J Med Educ. 2018;9:111–2.

    Article  Google Scholar 

  7. 7.

    Kiely PM, Slater J. Optometry Australia entry-level competency standards for optometry 2014. Clin Exp Optom. 2015;98(1):65–89.

    Article  Google Scholar 

  8. 8.

    Kiely PM, Chakman J. Competency standards for entry-level to the profession of optometry 1993. Clin Exp Optom. 1993;76(5):150–81.

    Article  Google Scholar 

  9. 9.

    Kiely PM, Horton P, Chakman J. Competency standards for entry-level to the profession of optometry 1997. Clin Exp Optom. 1998;81(5):210–21.

    Article  Google Scholar 

  10. 10.

    Kiely PM, Chakman J, Horton P. Optometric therapeutic competency standards 2000. Clin Exp Optom. 2000;83(6):300–14.

    Article  Google Scholar 

  11. 11.

    Kiely PM. Optometrists association Australia universal (entry-level) and therapeutic competency standards for optometry 2008. Clin Exp Optom. 2009;92(4):362–5 366-386.

    Article  Google Scholar 

  12. 12.

    Downing SM. Reliability: on the reproducibility of assessment data. Med Educ. 2004;38(9):1006–12.

    Article  Google Scholar 

  13. 13.

    Cizek GJ, Germuth AA, Schmid LA. A checklist for evaluating credentialing testing programs. Kalamazoo: The Evaluation Center, Western Michigan University; 2011.

    Google Scholar 

  14. 14.

    Joint Committee on Testing Practices. Code of fair testing practices in education (revised). Educ Meas Issues Pract. 2005;24(1):23–6.

    Article  Google Scholar 

  15. 15.

    Kane MT, Clauser BE, Kane J. A validation framework for credentialing tests. In: Davis-Becker S, Buckendahl, editors. Testing in the professions: credentialing policies and practice. New York: Routledge; 2017. p. 21–40.

  16. 16.

    Messick S. Validity. In: Linn RL, editor. Educational measurement. 3rd ed. New York: American Council on Education; 1989. p. 13–103.

    Google Scholar 

  17. 17.

    Kane M. Validating the interpretations and uses of test scores. In: Lissitz RW, editor. The concept of validity: revisions, new directions and applications. Charlotte: Information Age Publishing Inc.; 2009. p. 33–64.

    Google Scholar 

  18. 18.

    von Bergmann H, Childs RA. When I say … validity argument. Med Educ. 2018;52(10):1003–4.

    Article  Google Scholar 

  19. 19.

    Cook DA, Brydges R, Ginsburg S, Hatala R. A contemporary approach to validity arguments: a practical guide to Kane's framework. Med Educ. 2015;49(6):560–75.

    Article  Google Scholar 

  20. 20.

    Kane MT. The assessment of professional competence. Eval Health Prof. 1992;15(2):163–82.

    Article  Google Scholar 

  21. 21.

    Kane MT. Validating interpretive arguments for licensure and certification examinations. Eval Health Prof. 1994;17(2):133–59.

    Article  Google Scholar 

  22. 22.

    Hrynchak P, Glover Takahashi S, Nayer M. Key-feature questions for assessment of clinical reasoning: a literature review. Med Educ. 2014;48(9):870–83.

    Article  Google Scholar 

  23. 23.

    Cambron-Goulet É, Dumas J-P, Bergeron É, Bergeron L, St-Onge C. Guidelines for creating written clinical reasoning exams: insight from a Delphi study. Health Professions Educ. 2019;5(3):237–47.

    Article  Google Scholar 

  24. 24.

    Cizek GJ, Bunch MB. Standard setting: a guide to establishing and evaluating performance standards on tests. Thousand Oaks: Sage publications; 2007.

    Google Scholar 

  25. 25.

    Boone WJ. Rasch analysis for instrument development: why, when, and how? CBE Life Sci Educ. 2016;15(4):rm4.

    Article  Google Scholar 

  26. 26.

    Smith EV Jr. Evidence for the reliability of measures and validity of measure interpretation: a Rasch measurement perspective. J Appl Meas. 2001;2(3):281–311.

    Google Scholar 

  27. 27.

    Jozefowicz RF, Koeppen BM, Case S, Galbraith R, Swanson D, Glew RH. The quality of in-house medical school examinations. Acad Med. 2002;77(2):156–61.

    Article  Google Scholar 

  28. 28.

    Malau-Aduli BS, Zimitat C. Peer review improves the quality of MCQ examinations. Assess Eval High Educ. 2012;37(8):919–31.

    Article  Google Scholar 

  29. 29.

    Caulk SS, Plaus K. Development of the certification examination by the American Association of Nurse Anesthetists (1933–2012). In: Eger Ii EI, Saidman LJ, Westhorpe RN, editors. The wondrous story of anesthesia. New York: Springer New York; 2014. p. 471–82.

    Google Scholar 

  30. 30.

    Newton DW, Boyle M, Catizone CA. The NAPLEX: evolution, purpose, scope, and educational implications. Am J Pharm Educ. 2008;72(2):33.

    Article  Google Scholar 

  31. 31.

    Marson SM, DeAngelis D, Mittal N. The Association of Social Work Boards’ licensure examinations:a review of reliability and validity processes. Res Soc Work Pract. 2010;20(1):87–99.

    Article  Google Scholar 

Download references


Excel Psychological and Educational Consultancy Pty Ltd. (EPEC) was contracted to conduct psychometric analysis of these data.


The Optometry Council of Australia and New Zealand provided access to the deidentified data and through the individual author contribution of employee TM provided input to the study conception, design, data analysis, and interpretation of the data in writing the manuscript.

The Optometry Council of Australia and New Zealand provided funding support for the presentation of this research at the 2019 American Academy of Optometry annual meeting in Orlando, Florida and for open access publication fees.

Author information




Conception (SB, NGC, KS, TM, AC, GP, IJ), Design (SB, NGC, KS, TM, AC, GP, IJ), Acquisition of data (SB, NGC, TM, IJ), Analysis of data (SB, NGC, TM, IJ), Interpretation of data (SB, NBC, KS, TM, IJ), Drafting of the article (SB, NGC, KS, TM, AC, GP, IJ), Revision of the article (SB, NGC, KS, TM, AC, GP, IJ), Final Approval (SB, NGC, KS, TM, AC, GP, IJ). All authors were equally involved in the conception and design of the work. All authors contributed to the drafting and revision of the work.

Corresponding author

Correspondence to I. Jalbert.

Ethics declarations

Ethics approval and consent to participate

The research was conducted in accordance with the tenets of the Declaration of Helsinki and was approved by The University of Melbourne, Human Research Ethics Committee. A waiver of informed consent was granted.

Consent for publication

All authors have read and approved the manuscript for submission. All authors have agreed to its content and are accountable to all aspects of the accuracy and integrity of the manuscript in accordance with ICMJE criteria.

Competing interests

There are no competing interests. TM is employed by the Optometry Council of Australia and New Zealand (OCANZ). All other authors are members of the OCANZ Examination Committee and receive sitting fees for committee meetings.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1:

Appendix 1

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Backhouse, S., Chiavaroli, N.G., Schmid, K.L. et al. Assessing professional competence in optometry – a review of the development and validity of the written component of the competency in optometry examination (COE). BMC Med Educ 21, 11 (2021).

Download citation


  • Competence
  • Examination
  • Optometry
  • Rasch analysis