ensuring fairness in difficulty and content among parallel...

56
Copyright © 2020 by James R. Parry ENSURING FAIRNESS IN DIFFICULTY AND CONTENT AMONG PARALLEL ASSESSMENTS GENERATED FROM A TEST-ITEM DATABASE James R. Parry, M.Ed., CPT ABSTRACT This paper presents research and provides a method to en- sure that parallel assessments, that are generated from a large test-item database, maintain equitable difficulty and content coverage each time the assessment is presented

Upload: others

Post on 06-Aug-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

ENSURING FAIRNESS IN DIFFICULTY AND

CONTENT AMONG PARALLEL ASSESSMENTS

GENERATED FROM A TEST-ITEM DATABASE

James R. Parry, M.Ed., CPT

ABSTRACT This paper presents research and provides a method to en-

sure that parallel assessments, that are generated from a large test-item database, maintain equitable difficulty and

content coverage each time the assessment is presented

Page 2: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

1

Ensuring Fairness in Difficulty and Content Among Parallel Assessments

Generated from a Test-item Database

James R. Parry, M.Ed., CPT

Owner/Chief Executive Manager

Compass Consultants, LLC, Virginia, United States of America

May 2020

ABSTRACT This paper presents research and provides a method to ensure that parallel assessments, that are

generated from a large test-item database, maintain equitable difficulty and content coverage each

time the assessment is presented. To maintain fairness and validity it is important that all instances

of an assessment, that is intended to test the same subject content, are presented to each test-taker

without bias. The method described demonstrates how each test-item in an item bank1 (i) is as-

signed a difficulty rating using a recognized test-centered2 legally defensible cut-score3 determi-

nation method, (ii) assigned to either hard, moderate, or easy categories, and (iii) then selected in

a stratified random selection4 fashion, by content and difficulty, to ensure equal content coverage

and difficulty level within an acceptable range of the calculated cut score.

1 An item bank is an organized collection of items. 2 Test-centered methods rely on judgments about test-items, whereas the examinee-centered methods rely on judgments about examinees. 3 To be legally defensible and meet the Standards for Educational and Psychological Testing, a cut score cannot be arbitrarily determined, it must be empirically justified. 4 Stratified sampling divides the population or data into groups or blocks. Random samples or selections are taken from each group or block.

Page 3: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

2

EXECUTIVE SUMMARY With the proliferation of computer based, online and paper-based assessment/testing platforms the

storage of test-items (questions) in a database has become commonplace. Without proper prepara-

tion of the database, which includes the initial assignment of key fields within the structure of the

database, a danger of unfair assessment/testing exists. This unfairness is created by using random

selection of test-items from a database without regard to item difficulty and coverage of content

that supports the objectives or competencies being assessed. This unfairness is amplified when

randomly selecting items from a database to produce parallel forms5 of assessments/tests to be

given simultaneously to a group or as make-up or retests.

The purpose of this research study is:

• to present evidence that when test-items are selected at random from a database, the re-

sulting parallel forms of the assessments/tests will NOT:

o cover content equally with each iteration of a parallel assessment/test

o be equivalent in regard to difficulty of test-items

• to present evidence that when using stratified random selection of test-items from a data-

base, the resulting parallel forms of the assessments/tests will:

o cover all content equally and at the same difficulty with each iteration of a parallel

assessment/test

o maintain an overall assessment/test difficulty within an acceptable range of the

calculated cut score

The paper provides a brief background on test development and procedures for establishing defen-

sible cut or passing scores. In order to conduct the study, three experiments were conducted using

both hypothetical and real client test-item difficulty data. The test-item data was entered into a

spreadsheet tool developed by James R. Parry that:

• calculates item difficulty based upon the results of a cut-score rating session and assigns a

rating of hard, moderate, or easy,

• calculates a cut or passing score for the entire database, and

• provides a final assessment/test design criterion that will maintain equality in both content

coverage and difficulty for parallel test forms.

The overall findings conclude that pure random selection of test items will not produce fair parallel

forms of assessments/tests but stratified random selection will.

5 Parallel forms are different versions of a test that measure the same objectives and yield similar results. (Shrock & Coscarelli, 2007)

Page 4: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

3

Recommendations of the study are:

In order to maintain fairness and ensure parallel forms of assessments/tests are valid, reliable,

without bias and defensible when generated from a test-item database:

• Test-items must be constructed using universally recognized standards

• Cut scores should be established using a recognized test-centered method or, if appropriate,

a test-taker centered method, because arbitrary methods are not defensible

• Each item in a test-item database should be evaluated by a panel of expert judges and a

difficulty score or rating established based upon the agreed upon MAC level of the target

test-taker

• Test items should be selected using stratified randomization based upon both topic cover-

age as well as item difficulty to ensure equitable parallel assessments are generated

• Tests should not be generated in a pure random fashion from a test-item database without

regard to content because content coverage will be erratic

• Tests should not be generated in a pure random fashion from test-item database without

regard to difficulty of test-items because difficulty among tests will be erratic

Page 5: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

4

TABLE OF CONTENTS ABSTRACT ...................................................................................................................................................... 1

EXECUTIVE SUMMARY .................................................................................................................................. 2

INTRODUCTION ............................................................................................................................................. 5

RESEARCH QUESTIONS .................................................................................................................................. 7

THE TESTING PROCESS .................................................................................................................................. 7

TEST DESIGN .................................................................................................................................................. 8

ESTABLISHING A DEFENSIBLE CUT SCORE OR DIFFICULTY RATING ............................................................ 11

Angoff/Modified Angoff Method ............................................................................................................ 12

ITEM DATABASES ........................................................................................................................................ 15

EXPERIMENTAL PROCEDURES ..................................................................................................................... 16

Basic description of the Questionmark OnDemand assessment platform:............................................ 16

Design philosophy of the Compass Consultants, LLC spreadsheet tool: ................................................ 16

Experiment #1A – Random Selection – Hypothetical Data..................................................................... 17

Experiment #1B – Stratified Randomization – Hypothetical Data .......................................................... 21

Experiment #2A – Random Selection – Real Client #1 Data ................................................................... 25

Experiment #2B – Stratified Randomization – Real Client #1 Data ........................................................ 30

Experiment #3A – Random Selection – Real Client #2 Data ................................................................... 33

Experiment #3B – Stratified Randomization – Real Client #2 Data ........................................................ 37

CONCLUSIONS ............................................................................................................................................. 40

RECOMMENDATIONS ................................................................................................................................. 43

LIST OF FIGURES .......................................................................................................................................... 44

LIST OF TABLES ............................................................................................................................................ 45

APPENDIX A ............................................................................................................................................... A-1

Description of the spreadsheet tool: .................................................................................................... A-1

Design philosophy of the Compass Consultants, LLC spreadsheet tool: .............................................. A-2

Function of the Spreadsheet Tool ......................................................................................................... A-2

LIST OF FIGURES IN APPENDIX A ............................................................................................................... A-8

LIST OF TABLES IN APPENDIX A ................................................................................................................. A-8

REFERENCES ............................................................................................................................................... R-1

ACKNOWLEDGEMENTS .......................................................................................................................... Ack-1

ABOUT THE AUTHOR .............................................................................................................................. Ack-1

Page 6: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

5

INTRODUCTION Assessments are important evaluation tools used in educational, industrial, government, medical,

military and other organizations throughout the world. With the proliferation of electronic/com-

puter databases used to store test-items and generate randomized assessments, comes the inordi-

nate possibility of unfairness. When assessments are intended to be parallel, this unfairness can be

either in (i) difficulty, i.e. some assessments are more difficult than others; or (ii) content, i.e.

selecting more or less items from one or more topics for every instance of the assessment. If par-

allel assessments are not equal in both difficulty and content, both the test-taker and the testing

organization are at risk. An unfair assessment that consists of mostly “easy” test-items may serve

to indicate that a minimally competent candidate is qualified when in fact, they just “got lucky”,

whereas an assessment consisting of mostly “hard” test-items, on the same topic or topics as the

easy assessment, may deny a candidate, who is minimally qualified or competent, a position or

promotion that they truly deserve. Alternatively, what if, because of random item selection, all

instances of an assessment do not test the same content? This would provide incorrect assumptions

that all candidates “knew” all of the required objectives or competencies when, in fact, they were

not tested on all requirements.

Example:

A hypothetical 20-item end of unit assessment on electrical safety is intended to test four topics of

equal importance:

a) Grounding

b) Lock-out-tag-out

c) Personal safety equipment

d) Insulation

Because all topics are considered to be of equal importance, we assume that five test-items should

be presented for each topic. The test-item data base contains 100 test-items pertaining to electrical

safety, not sorted by topic or by difficulty. The assessment is administered via computer with each

student receiving a randomly generated version with a selection criterion of “select 20 questions

at random from topic ‘electrical safety’”.

Think about all of the possible outcomes of this item selection criterion in terms of both test diffi-

culty and content and the problems associated with this method. With the possibility of generating

an unfair assessment, why would test administrators even want to generate a random assessment

instead of relying on a single fixed form6?

6 A fixed-form assessment asks all test-takers to respond to the same questions or tasks in the same order.

Page 7: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

6

There are several advantages for using randomization to select test-items from a database instead

of a fixed form which include:

• Anti-cheating – all test-takers do not receive the same test-items so looking at the computer

screen or paper of a nearby test-taker will not present an advantage

• Ability to add or retire7 test-items – new test-items can be added to or retired from the item

database as references or competencies change

• Provides the ability to administer the assessment to the same test-taker multiple times with

different test-items

• Allows test administrators to deliver the assessment at different times without fear of pre-

test communications among test-takers (i.e. test-taker on East coast of the U.S. calling test-

taker on West coast of U.S. about the test contents)

An alternative to pure random selection of test items to generate parallel forms is to used a fixed

form but randomize both the test-items and alternatives within the assessment/test. This will pro-

duce parallel forms each time and maintain both content and difficulty fairness. Disadvantages to

this method are possible overexposure of items to test takers and the inability to produce make-up

or retests that would present alternative test items. Because the stem of the test-items remains

unchanged there is still a danger of pre-test compromise due to communication among test-takers.

So, what are the dangers of using pure random selection to generate assessments/tests?

• Randomization may produce measurement error8 in that some test takers may be presented

with a difficult assessment/test and others may receive an easy one. When an assess-

ment/test is used to classify test-takers into groups, two kinds of wrong decisions can occur

(Livingston & Zieky, 1982):

o A test test-taker who actually belongs in the lower group can get a score above the

passing score

o A test-taker who actually belongs in the higher group can get a score below the

passing score

Classifying someone into the wrong group could lead to less than qualified individuals being pro-

moted or certified or, those who are qualified being denied advancement or certification. These

situations could lead to legal challenges that may be difficult to defend.

The item selection method that I propose to alleviate possible unfairness when using randomization

to select test-items is described in detail in this paper.

7 Once a test-item has been used on an actual assessment it should never be deleted entirely from an item data-base due the possibility of future legal proceedings. Retiring the item (if available in the software) retains the item and all statistics in their original form. 8 Measurement error in education generally refers to either (1) the difference between what a test score indicates and a student’s actual knowledge and abilities or (2) errors that are introduced when collecting and calculating data-based reports, figures, and statistics related to schools and students. (Great Schools Partnership, 2013)

Page 8: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

7

RESEARCH QUESTIONS 1. When test-items which have been assigned a difficulty rating using a recognized test-cen-

tered, empirically justified procedure and placed in appropriate topic classification areas at

various difficulty levels (easy, moderate, or hard) are randomly selected from a computer

based test-item bank, will every iteration of the test generated be presented at a similar

difficulty level and cover all topics within objectives adequately?

2. When test-items which have been assigned a difficulty rating using a recognized test-cen-

tered, empirically justified procedure and placed in appropriate topic classification areas at

various difficulty levels (easy, moderate, or hard) are selected in a stratified-random man-

ner from a computer based test-item bank, will every iteration of the test generated be pre-

sented at a similar difficulty level and cover all topics within objectives adequately?

Before any item selection process is attempted it is important to design both effective test-items

and test instruments. I will begin with a brief overview of the testing process and design that is the

basis for the stratified random selection method that I propose.

THE TESTING PROCESS I begin with a quote from Steven M. Downing, University of Illinois at Chicago:

“Effective test development requires a systematic, well-organized approach to ensure suf-

ficient validity evidence to support the proposed inferences from test scores.”

(Downing, 2006)

The testing process has many stakeholders which may include members of management, admin-

istrators, facilitators, instructors, teachers, test administrators, etc. but the most important stake

holder is the test-taker. If the test is unfair or biased, the test-taker is placed at a disadvantage.

This leads to another quote:

“All tests should be well developed and testing practices, beneficial. There is extensive

evidence documenting the effectiveness of well-constructed tests in relation to supporting

the validity of the test. The proper use of tests can result in making wiser decisions about

individuals and programs than those made without using tests. The improper use of tests,

however, can cause considerable harm to test-takers and others affected by test-based de-

cisions.”

(AERA; APA; NCME, 2014)

Page 9: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Designing Criterion-Referenced Tests Create

Cognitive

Items

h Develop Course

Material

Establish

Content Validity of Objectives

Jr Analyze Job

Content

Establish Content

Validity of Items

Conduct Initial Test

Pilot Create Rating

Instruments

Establish Rellabil I ly of

Cognitive Hems

Document Entire Process as you Proceed

Perform II Ertl

Analysis

Create Parallel Forms

Set Cut Score

Report Scores

Establish

Inter-Baler

Adapted from Shred. & Coscareill.W (2557), page 45. Figure 3.1 Reliability

Copyright © 2020 by James R. Parry

8

Whether an assessment/test is norm-referenced9 or criterion-referenced10, it must be fair to all test-

takers. If a test is being administered to individual or multiple participants it must be developed

and constructed to ensure it is both valid – measures what it is supposed to measure, and reliable -

effectively measures anything at all.

TEST DESIGN The first, and most important step in creating fair, defensible assessments or tests is to ensure

validity. All test-items must match the required job skills or certification requirements. Referring

to the Designing Criterion Referenced Tests flow chart (Figure 1) as a guidance tool, it should be

noted that test-item design comes right after the job analysis phase, before course material is cre-

ated.

Figure 1 – Designing Criterion-Referenced Tests

The next big step is to determine how many items to develop. To solve this dilemma, ask the

following questions:

• How critical are decisions based upon the results of the test?

• What resources (time, money, and personnel) are available for testing?

• How big is the overall objective that is being tested?

• How closely related are the objectives that are being tested?

Table 1, which was adapted from Criterion-Referenced Test Development, Technical and Legal

Guidelines for Corporate Training (Shrock & Coscarelli, 2007), provides a starting point to help

determine the number of test-items required per objective being tested. Once the number of items

9 A norm-referenced test compares people in relation to the test performance of one another. It is generally com-posed of items that will separate the scores of test-takers from one another and is typically used to rank-order se-lect top performers. 10 A criterion-referenced test compares people to a standard. It is composed of items based on specific objectives or competencies. It defines the performance of each test-taker without regard to the performance of others. It is a test of mastery.

Page 10: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Number of Test Items per Objective

If And And Then

The performance objective is:

Critical to safety. life, limb. legal requirements. etc.

From a large objective domain

Unrelated 10-20

Related I0

From a small objective domain

Unrelated 5-10

Related 5

NOT critical to safety. life. limb. legal requirements. etc.

From a la rge objective domain

Unrelated 6

Related 4

From a small objective domain

Unrelated 2

Related 1

Adapted from 9....2r.sAc & & Coscareif, W.2007

Copyright © 2020 by James R. Parry

9

that are required to adequately test an objective has been determined, based on the table, I suggest

that number be increased three to five times to allow for an adequate number of items to be avail-

able for stratified randomized selection from a test-item database. This ensures that there are an

adequate number of items available in each topic to generate several iterations of a parallel assess-

ment without repeating the same items.

Table 1 - Number of Test-Items per Objective

Typically, there are several topics within an objective, some that are considered more important or

critical than other topics. Table 1 can be used as a guide to assist in this decision process. An

example may be derived from the hypothetical 20-question end of unit assessment on electrical

safety presented in the introduction to this paper. This assessment is designed to test four topics of

equal importance:

a) Grounding

b) Lock-out-tag-out

c) Personal safety equipment

d) Insulation

Because the topics are considered to be of equal importance, 25% of the test-item database is

dedicated to each of the topics. To allow for stratified randomization to generate a 20-item assess-

ment, the number of items available in the database for each topic would be as follow:

a) Grounding 15 - 25

b) Lock-out-tag-out 15 - 25

c) Personal safety equipment 15 - 25

d) Insulation 15 - 25

This requires a test-item database size of between 60 to 100 items to select from to provide five

items for each topic on each assessment and ensure different but equal items on each randomly

generated assessment.

Page 11: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

10

If all topics are not considered to be equally important or critical then the number of items available

per topic, within each objective, would be adjusted, for example if Grounding is considered the

most important topic it may be ranked as 50% of the assessment with Lock-out-tag-out ranked

second at 30% and both Personal safety equipment and Insulation ranked equally important at

10% each. This would require each assessment to be planned and generated as illustrated in table

2.

Topic % of Test Number on 20

Question Test

Number Available

in Database

Grounding 50 10 30 - 50

Lock-out-tag-out 30 6 18 - 30

Personal safety equipment 10 2 6 - 10

Insulation 10 2 6 - 10

Table 2 - Topic Weights for Electrical Safety Assessment

As can be derived from table 2, it is easy to see how test-item databases can become quite large to

ensure randomization. Seems pretty straight forward at this point but what about difficulty? It is

important to try to maintain a good balance of easy, moderate and hard items to select from, and

include on each test or assessment. When the test-item authors, in consultation with subject matter

experts (SMEs), are building the items, they generally will have an idea as to the difficulty of each

item with more complex items being more difficult than simple items in most cases. The actual

difficulty will not be confirmed until a cut-score rating session is convened.

Another consideration is test length vs. time available. What if, using table 1 as a guide, the test

length is determined to be 100 questions. How much time is available for testing? Research and

best practice indicated that typical response time for each test-item, on a written test, without ref-

erences provided, ranges from 42 seconds for an easy item to 90 seconds for a very difficult item.

Table 3, based on information in a paper by Phil Higgins (Higgins, 2009), provides a tool to assist

in determining time to be allotted for administration of an assessment, not including administrative

time (attendance, paperwork, reading test procedures/rules, etc.). Referring to table 3, a typical

100 item 4-alternative multiple-choice assessment that is considered to be of moderate difficulty

would require approximately 5,500 seconds or about 1 hour 32 minutes hours to complete, plus

administrative time for a total estimated time of about 2 hours. The table is useful as an initial

planning tool. Using stratified randomization that defines the actual number of easy, moderate,

and hard items selected for each iteration of the assessment, could further refine initial estimates

of test time allotted (e.g. 50 item assessment consisting of 20 easy, 20 moderate, and 10 hard test

items would yield an approximate test time of 43 minutes). The time it takes for individual test-

taker to respond to each item is affected by several factors including, but not limited to language

comprehension, education level, reading ability, previous exposure to test item, test-taker motiva-

tion, test-taker fatigue, unfamiliarity with test administration system, test anxiety, disabilities, etc.

A possible problem with rule of thumb time estimates is that the legality may be challenged unless

there is reliable data to back up the estimate. How can the test designer prove that the time allotted

is sufficient? A review of statistical information related to the time of response by item, which is

Page 12: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

11

typically available with online test administration software suites, will provide a more accurate

estimate once the assessments have been in use.

Time Allotted per Test Item

Overall Test or Item Difficulty

Easy Moderate Hard

Time per test item

43 Seconds 55 Seconds 61 Seconds

50 Item Test Time

2150 seconds (approx. 36

minutes)

2750 seconds (approx. 46

minutes)

3050 seconds (approx. 51

minutes)

100 Item Test Time

4300 seconds (approx. 72

minutes)

5500 seconds (approx. 92

minutes)

6100 seconds (approx. 102

minutes)

Table 3 - Test Time Tool

It is interesting to note that the U.S. Coast Guard found that if references are allowed to be used

during the test, the average time spent per item increases by about 36% (United States Coast Guard,

2015). If the time required is excessive, consideration must be given to omit some topics or rethink

the purpose of the test outcome –

• Are the objectives being tested covering too large of a content domain?

• Are all of the topics being tested actually required and supported by the objectives or cer-

tification requirements?

• Are the “topics” too far ‘down in the weeds’? In other words, is it a step of a topic at a

higher level? (e.g. Unscrewing a light bulb is a step in a larger topic of changing a light

bulb).

ESTABLISHING A DEFENSIBLE CUT SCORE OR DIFFICULTY RATING Establishing a cut score on an assessment or test is also known as standard setting. The ‘standard’

may be in the form of pass/fail, issue/non-issue of a license or certification, award/withhold a

credential, etc. The cut score, in order to be legally defensible, cannot be established arbitrarily,

it must be empirically justified11 (AERA; APA; NCME, 2014). There are several recognized meth-

ods, both test centered and test-taker centered12, to establish defensible cut scores. Gregory J. Cizek

(Cizek, 2006) has done a great job describing and summarizing many of these in his paper Stand-

ard Setting. One thing to keep in mind, according to the Standards for Educational and Psycho-

logical Testing, “There can be no single method for determining cut scores for all tests or for all

purposes, nor can there be any single set of procedures for establishing their defensibility” (AERA;

APA; NCME, 2014) pg. 100.

Most, if not all, of the recognized methods for establishing a cut score are designed around fixed

form tests or assessments where the test-items are selected manually and a cut score established

on the contents of a single test. If a second, third or subsequent test is generated the cut score

11 Something empirically justified can be provable or verifiable by experience or experiment (Dictionary.com Unabridged, 2020) 12 Test-taker centered methods require judges to make decisions based on their knowledge of the examinees and their performance

Page 13: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

12

would typically vary because each iteration of the test would be judged individually. The method

I propose is a variation of the Modified Angoff Method that evaluates the entire database as a

whole, establishes a cut score and difficulty rating for all items within the database and produces

a recommended stratified randomized test design that maintains both the cut score as well as covers

all topics equally and at similar difficulty with each iteration of a test. The Angoff/Modified An-

goff Method with my variations is described as follows:

Angoff/Modified Angoff Method The Angoff method of determining and setting a cut score for an assessment uses subject matter

experts (SMEs) as judges to review each item and assign a score or weight to the item based on

the judge’s conjecture that a minimally competent performer or a test-taker who is at the minimum

acceptable competence (MAC) level required for the job or certification would answer the item

correctly. It should be noted that the Angoff Method is sensitive to both difficulty and importance,

thus, for a welder, putting on safety glasses is easy, but all would be expected to be able to perform

this task – while landing a plane in wind shear is difficult, but all pilots would also be expected to

pass this task (Coscarelli, Barrett, Kleeman, & Shrock, 2005). These scores, based on 100 test-

takers, are then summed and averaged to assign a difficulty rating to each item. i.e. Six judges

score an item as .65, .70, .65, .75, .60, .70. The average of the scores is .675 or 67.5% which

translates to an estimation that 67.5% of test-takers at the MAC level would respond correctly to

the item. The Angoff Method establishes both a floor and ceiling score for each item. Typically,

the floor score (the lowest a judge is permitted score an item) is based on the number of alternatives

for the test-taker to choose from and the ceiling is based upon the results of the judges, who are

considered to be experts, average test scores. The ceiling cannot be higher than the judges average

scores because a test-taker at the MAC level could not be expected to respond correctly if an expert

cannot. The Angoff Method, as with most test-centered methods, is designed to rate the test-items

on a single form of a test, so, in theory, each form of the same assessment, all testing the same

content, could have a different cut score because each item has a unique Angoff weight (Coscarelli,

Barrett, Kleeman, & Shrock, 2005) (e.g. test form A has a cut score of 82%, test form B has a cut

score of 80%, and test form C has a cut score of 85%). Each of the cut scores would probably be

legally defensible because a recognized standard setting method was used to set the scores but the

testing organization may have difficulty explaining the score difference both internally and exter-

nally.

In the paper The Problem of the Saltatory Cut-Score: Some Issues and Recommendations for Ap-

plying the Angoff to Test-item Banks (Coscarelli, Barrett, Kleeman, & Shrock, 2005) the item da-

tabase was divided into two groups – the first containing all items with an Angoff score less than

or equal to the median score and the second group containing all items with scores greater than the

median score. The paper explains how item selections were made using simple randomization as

well as stratified randomization to generate a fair test. The stratified randomization described, al-

ternates item selection from those items equal to the median Angoff score, those less than or equal

to the Angoff score and those greater than the median score. This method guarantees an assessment

that is neither extremely difficult or extremely easy. Their conclusions were:

• For low stakes tests randomly sample within the item bank

Page 14: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

13

• For medium stakes tests, one can probably sample within the bank if the distribution is

statistically normal, but stratification is safer

• For high stakes tests, one should consider stratification of the sample for increased preci-

sion

The methods and results described in the study indicate that random sampling/selection generally

will produce tests with the approximate same difficulty when drawn from a small sample but as

the size of the sample (database) increases, some sort of stratified selection should be used. The

authors warn that as the size of the database increases, so does the chance for errors in selection.

Additionally, they recommend stratified random selection as the criticality of the assessment in-

creases. The stratified random selection method that I propose provides greater accuracy in both

maintaining the established cut score, presenting test-items at the same difficulty level and selects

the same number of items from multiple topics within the item database for each iteration of the

parallel assessment. The authors warn, and I concur, the algorithm (or any algorithm) used to select

items is only as good as the quality of the Angoff weights (scores) that have been assigned to each

item.

My method, which I will call the Parry Method for simplicity, is based on the Angoff/Modified

Angoff Methods with the following differences (Note: There are several variances or adaptations

of the Modified Angoff Method in use throughout the testing community):

• The Angoff Method allows any score between 0 and 1 (0% and 100%)

o Parry Method sets the floor score at the chance guess probability for the number of

plausible alternatives available in a multiple-choice (MC) style item (i.e. 4 alterna-

tive MC item floor score would be .25 (25%), 3 alternative MC item floor score

would be .33 (33%), etc.

o Parry Method sets the ceiling score at .95 (95%) for all items based on the assump-

tion that if all test-takers (100%) who are considered to be at the MAC level of

competence would answer the item correctly the item is a ‘wasted’ item and does

not provide much value in discrimination

• Angoff Method requires each judge to “take” the assessment as a typical test-taker would

to assign a ceiling score

o Parry Method eliminates the requirement for the judges to “take” the assessment

and sets the ceiling score for each item at .95 (95%). This is probably the most

difficult task for the judges because they must put themselves in the mindset and

knowledge level of the test-taker at whatever level they have agree upon as the

MAC

o The reason for not having each judge “take” the test is because of the time required

to answer all items in large databases and the fact that, if stratified random selection

is used to generate each test, there is not a single “test” to take.

Page 15: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

14

• Angoff Method assigns weight or score to each item in a fixed form and sets a cut score

based on the average item scores of the panel of judges

o Parry Method assigns a weight or score to each item in the test-item database. (i.e.

if the database has 300 items, each item is evaluated individually without regard to

what the final test form will look like) and establishes a cut score for the entire

database as a whole

• Angoff Method generally allows the judges to rate the items at any score between 0 and 1

(0% - 100%)

o Parry Method only allows the judges to rate the items at fixed intervals beginning

with the floor score and ending at the .95 ceiling at intervals of .05 (.25, .30, .35,

.40, .45, .50, .55, .60, .65, .70, .75, .80, .85, .90, .95)

• Angoff Method requires judges to disregard any alternative/distractor that even someone

at the MAC level would not choose before deciding their score

o Parry method assumes that all stems13 and alternatives were designed using recog-

nized item development and review techniques so there should be no “wasted” dis-

tractors but requires the judges to comment on the plausibility14 of existing distrac-

tors and recommend changes if necessary

• Angoff Method allows the judges to come together after their initial individual scoring

sessions to discuss differences in their ratings and come to a consensus

o Parry method calculates the standard deviation (SD)15 among judges scores for each

test-item and requires the judges to discuss their ratings as a group if the SD is 10

or greater to attempt to come to a consensus to bring the SD below 10 if possible.

If the judges cannot come to a consensus the item is either retired or the judge(s)

with the outlier score(s) is/are eliminated from the calculation.

13 The stem of a test item presents a single definite and explicit question or problem statement 14 A plausible alternative is one that has the appearance of being credible or believable, not frivolous. 15 The standard deviation is a measure of how spread out numbers are from each other.

Page 16: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

15

ITEM DATABASES With the proliferation of automated (computer) test-item databases and test generation software,

many tests are most likely generated randomly, sometimes without regard to difficulty or coverage

of required objectives. The most important step in designing test-item databases is the initial topic

structure which must allow test-items to be stored with separation of topics, by objective or com-

petency, and further down as necessary. This provides a mechanism for both stratified randomized

selection of test-items as well as future psychometric analysis. A typical topic structure is illus-

trated below:

REPOSITORY NAME

OBJECTIVE 1.0

TOPIC 1.1

SUB-TOPIC 1.1.1

Test-item 1.1.1/1

Test-item 1.1.1/2

TOPIC 1.2

SUB-TOPIC 1.2.1

SUB-TOPIC 1.2.2

OBJECTIVE 2.0

TOPIC 2.1

SUB-TOPIC 2.1.1

Test item 2.1.1/1

Test-item 2.1.1/2

SUB-TOPIC 2.1.2

Test-item 2.1.2/1

The sub-topics can be further divided into three difficulty “buckets” to make selection of test-items

at various levels of difficulty clear:

SUB-TOPIC 1.1.1

1.1.1 HARD

Test-item 1.1.1/1

1.1.1 MODERATE

Test-item 1.1.1/5

1.1.1 EASY

Test-item 1.1.1/9

An alternative way to identify difficulty levels of individual test-items may be through the use of

metatags16 of Hard, Moderate, Easy if the test-item database software supports this feature.

16 Assigning a metatag is a way to index items by specific job tasks, knowledge, skills and abilities, difficulty, etc. to allow for more flexible management and selection of test-items within a large database.

Page 17: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

16

EXPERIMENTAL PROCEDURES All experiments described in this paper were conducted using either hypothetical or real client test-

item data entered into a Questionmark® OnDemand® assessment platform (www.question-

mark.com) with test design generated by a proprietary spreadsheet tool designed by James R.

Parry, Owner/Chief Executive Manager at Compass Consultants, LLC (www.gocompassconsult-

ants.com). The methods described should work on any test development and delivery platform

capable of stratified randomization using either metatags or selection by sub-topic.

Basic description of the Questionmark OnDemand assessment platform: Can assess an un-

limited number of test-sitters, from anywhere in the world. The platform provides a range of as-

sessment formats including ‘drag and drop’, ‘multiple choice’ and many more. Organizations can

conduct a range of assessments across different courses and ability ranges. Tests are automatically

marked. Results are instantly compiled. Trends and patterns are easy and quick to spot.

Design philosophy of the Compass Consultants, LLC spreadsheet tool: The tool, described

fully in Appendix A, is designed to assist in setting a cut score for an assessment based on the

results of a test-centered cut-score rating session that uses a panel of judges or experts to evaluate

the difficulty of each test item. It can be used to assist in the design of assessments using the correct

response P-value17 scores returned using classical test theory (CTT)18 or item response theory

(IRT)19. statistics Additionally, it will determine the number of items from each section at each

level of difficulty (hard, moderate or easy) as set by the cut-score rating of each item. This as-

sumption is made for 4-choice, multiple choice items with a floor of 25% and a ceiling of 95%.

The difficulty is then assigned based on dividing the difference between 25% and 95% by 3 to

arrive at the three difficulty levels. The workbook is designed to accommodate up to ten (10) re-

viewers on the rating panel. The totals required from each section are based upon the numbers of

each level of difficulty available in each section as well as the total number of items available. An

assumption is made that if there are fewer items available in any particular section(s) than in other

section(s), then that section is of less importance or has significantly fewer objectives. As data for

each item is entered in each section, the final test design worksheet is updated automatically.

The tool is designed to establish an initial cut score for a new test-item database. I recommend that

after an appropriate number of statistical correct response P-value results are reported, after the

database is in use, the recommended cut score be revisited and possibly adjusted using another

test-centered method based on actual scores, such as the Bookmark Method20.

17 P-value is the percentage of test-takers who selected each response. Typically, the correct response p-value is referred to as the difficulty index. (Shrock & Coscarelli, 2007) 18 Classical test theory (CTT), also known as the true score theory, refers to the analysis of test results based on test scores. 19 Item response theory (IRT) is a statistical way to analyze responses to tests or questionnaires with the goal of improving measurement of validity and reliability 20 Bookmark Method places all test-items in a booklet, ranging from high to low (easiest to hardest) correct re-sponse P-value. Judges review each and place a bookmark at the point they feel the MAC will not answer the next harder item correctly. This point, after discussion and averaging all judges selected break point, becomes the cut score.

Page 18: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

. • • • • CUT SCORE CALCULATION TOOL

co.,-,./c.rcifi,aior, tlarneil COURSE/CERTIFICATION NAME I Test flarne:1 TEST NAME

I Compass

Facilitator Name/Date: F aril atOr Name/Date Revision 1 Facilitator Name/Date Oate: mmidd/rAry Consultants, lIC Enter TopicreO/Su bier[ ID: 1.0 Grounding Revision 2 Facilitator Name/Date

Th.; ,,,,ad.,,,,ct tool , tAc int lc,t,,, property of and Copyright 0 2010-2013 by Comp.: Con:ultan, LLC. U limited to the term, of the End Wu- Licen:c Ag cement (EUL...).Thi: copy icen:cd to: 30 CAN DEMO

Test Item CUD

Enter* If

New, R i Retiredf

Difficulty M etatag

Average

Pla,..Fan'a

Arm Expert 1

Name Expert 2 Name

Expert 3 Name

Expert d Name

Expert5 Name

Expert 6 Name

Expert? Name

ScoreRating)

n

Expert g Name

Expert 9 Name

Expert 10 Name

Standard Deviation

Topic Moderate

CUt 58.00 Dillieultii

10 ES 17.50 75 80 80 85 75 70 5.24 Approximate Dittiruity Rating

10 E2 90.00 95 95 95 85 80 90 632 25-48.3 Hard

1.0 E3 79.17 80 70 85 70 85 85 7.36 48.4 -71.: Moderate

1.0E4 I

80.83 90 90 75 75 75 80 7.36 71.8 - 95 Easy

1.0 E5 , 75.83 90 65 70 70 80 80 9.17

10 H1 I•

48.33 45 45 50 50 55 45 4.08 Standard Deviation

10 42 29.17 35 25 25 30 35 25 492 A sta n d and deviation ot more

1.0 H3 40.83 25 40 50 50 40 40 9.17 than 10 will trigger an alert.

1.0 H1 4833 50 45 45 50 50 50 258 Discuss the outliers with the

1.0 H5 35.83 30 35 30 35 40 45 5.85 judges who set them to

1.0 M1 68.33 75 70 70 65 65 65 4.08 determine why. Change as

10 M2 5501 60 50 60 55 60 45 832 1e1e650r5.

1.0 MB ante 60.83 55 55 60 60 55 70 5.85

1.0 M4 .e rate 70.83 80 70 70 71) 65 70 4.92 20 In this section 33%

10MS 650) 65 65 65 65 55 65 000 20 Moderate In this section 33%

2.0 El 82.50 90 85 80 85 80 75 524 20 I n this section 33%

.20 ES 5500 95 95 95 95 95 95 C...7.0 60 TOTAL 102%

Topic

Topic Cut

Score&

Difficulty

Items in

Topic

% of Total

Items

Avaiable

Hard

%Frani

Topic

Available

Mod

%From

Topic

Available

Easy

% From

Topic

Total 8

Needed

From

Tapir

Use Hard

(Calculated}

Use Hard

[Actual./

Use Mod

(Calculated/

Use Mod

(Actual)

Use Easy

(Calculated)

Use Easy

(Actua l) Topic

1.0 58 50 100.00, 20 11% 20 33% 20 33% 20.00 6.67 7 6.67 6 6.67 7 1.0

Copyright © 2020 by James R. Parry

17

Experiment #1A – Random Selection – Hypothetical Data Referring to the example in the introduction section of a test on electrical safety that covers four

topics (Grounding, Lock-Out-Tag-Out, Personal Safety Equipment, Insulation), assume all test-

items are placed in a single data file called Fairness Research. All four topics are assumed to be of

equal importance so the final test design should test all topics equally.

Using the Modified Parry Method of cut score setting, each item was reviewed21 and assigned a

difficulty rating with data entered into the spreadsheet tool (Table 4 – partial view)

Table 4 -Experiment # 1- Cut Score Calculation Tool (partial view) showing hypothetical data

The cut score for the entire database of 60 items was determined to be 58.02%

The spreadsheet tool recommended that 6.67 items at each difficulty level (hard, moderate, and

easy) be used to ensure a cut score of approximately 58.02% was maintained. (Table 5)

Table 5 – Experiment #1 - Recommended Test Difficulty Distribution for Hypothetical Data – Random Selection

21 For this phase of the experiment all Angoff/Parry Method difficulty data was hypothetical and assigned to force equal numbers of hard, moderate, and difficult items.

Page 19: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Qu estio-n selections

20 random qu estio-n (s) fro-m to-pie 'FAIRNESS RESEARCH' including subto-pics (Avo-id previo-usty d ehrered)

Attempt

CND

1

SCORE

Attempt

DID

Experiment

2

SCORE

#1- Random

Attempt

OID

3

SCORE

Selection of

Attempt

DID

20 items

4

SCORE

from all

OLD

Attempts

4 topics.

SCORE

Desired

Attempt

DID

target

6

SCORE

d iffi cu lty

Attempt

DID

is 58.02

7

SCORE

with 5 items

Attempt

DID

from

8

SCORE

each topic.

Attempt

OID

9

SCORE

Attempt

OID

10

SCORE

1.0 E3 79.17 1.0 11 77.50 19 El 77.50 1.0 El 77.50 1.0 E3 79.17 1.0 E4 80.83 1.0 H2 29.17 1.0 E2 90.00 1.0 E2 90.00 1.0 E2 90.00

1.0 ES 75.83 1.0 E5 75.83 1.0 E2 90.00 1.0 ES 75.83 1.0 5 75.83 19 115 35.83 19 M1 6833 1.0 E3 79.17 1.0 E3 79.17 1.0 E4 80.83

1.0111 48.33 1.0 Ml 68.33 1.0 111 48.33 1.0114 48.33 1.0 115 35.83 1.0 M3 60.83 1.0 M3 60.83 1.0 E4 80.83 1.0 E4 80.83 19113 40.83

1.0115 35.83 1.0 M2 55.00 19113 41183 1.0 M1 68.33 20 ES 72.50 1.0 M4 71185 1.0 M4 7033 1.0111 48.33 1.0 H2 29.17 1.0 M4 70.83

1.0 M1 68.33 1.0 M3 60.83 1.0 114 4833 2.0 El 82.50 20 E4 82.50 2.0 ES 72.50 1.0 M5 6590 1.0112 29.17 1.0 H3 40.83 2.0 E2 95.00

2.0 E2 95.00 2.0 El 8250 1.0 Ml 6833 2.0 ES 72.50 2.0 Ml 67.50 29 E4 82.50 29 El 95.00 1.0113 40.83 1.0 H4 48.33 29 113 46.67

29 113 46.67 2.0 H1 25.00 1.0 M5 6590 2.0 113 46.67 2.0 M4 49.17 2.0 ES 77.50 29 E3 72.50 19 M3 60.83 1.0 H5 35.83 2.0 Ml 67.50

29114 44.17 2.0 Ml 6750 29 El 8250 2.0 H4 44.17 30 E3 72.50 2.0 Hl 2590 29 E4 82.50 1.0 M4 70.83 1.0 Ml 68.33 2.0 MS 61.67

2.0 M3 61.67 2.0 M3 61.67 29 E4 8250 2.0 M2 68.33 3.0 115 25.83 2.0 H5 3090 2.0 114 44.17 19 MS 65.00 1.0 M3 60.83 3.0 El 80.83

29 M2 58.33 29 M4 49.17 29 H1 2510 2 0 M5 65110 39 Ml 65.00 29 M5 65110 29 115 3090 29 E5 7250 19 M5 6500 39 E2 84.17

2.0 M5 65.00 3.0 E2 84.17 2.0 H3 4667 3.0 E4 82.50 3.0 M2 49.17 3.0 El 80.83 2.0 M4 49.17 2.0115 30.00 2.0 E5 77.50 3.0 E4 82.50

39 El 80.83 39 E3 7250 2 0 H4 44.17 3.015 7250 39 M3 53.33 3.014 82.50 29 M5 6590 20 M2 58.33 29 H1 2590 39 114 33.33

3.0 E2 84.17 39 H5 2583 2.0 H5 30.00 39114 33.33 3.0 M4 49.17 3.0 E5 72.50 39 El 80.83 29 M3 51.67 2.0 H4 44.17 39 115 25.83

39 H3 28.33 3.0 Ml 6590 3.0 E4 8250 3.0 M3 53.33 3.0 M5 61.67 SO H3 2833 39 ES 7250 29 MS 65.00 2.0 M3 61.67 3.0 M5 61.67

3.0 H5 25.83 4.0 11 7250 5.0 E5 7230 5.0 M4 49.17 4.0 El 72.50 3.0 H5 2185 30114 3333 3.0 El 84.17 3 0 El 84.17 4.0 E3 72.50

3.0 M3 53.33 4.0 E3 7250 3.0 H4 3333

ffl 72.50 4.0 E5 94.17 SB Ml 65.00 3.0 M2 49.17 3.0111 26.67 30 H2 4833 4.0 E4 82.50

3.0 M4 49.17 4.0 H2 30.83 4.0 El 72.50 94.17 4.0 H1 28.33 39 M3 5333 49 El 7230 39 M1 65.00 49 E4 82.50 49 112 30.83

4.0 E4 82.50 4.0 H3 4590 4.0 E4 82.50 28.33 4.0 H3 45.00 SB M4 49.17 4.0 115 45.00 4.0 E2 85.00 4.0 ES 94.17 4.0 Ml 50.00

40 H2 30.8.3 49 H4 31.67 40 H.3 45110 49.17 40 M1 50.00 4013 7250 40115 25 83 49112 30 83 40 H5 25.85 40 M2 50110

4.0 M2 50.00 4.0 M5 54.17 4.0 M4 49.17 54.17 4.0 M2 50.00 4.0 Ml 50.110 4.0 M2 5090 49 MS 54.17 4.0 M4 49.17 4.0 M4 49.17

Difficulty 58.67 Dilficulty 58.88 Difficulty 59.33 Difficulty 61.92 Dilficulty 58.96 Difficulty 59.04 Difficulty 58.08 Difficulty 60.42 Difficulty 59.54 Difficulty 62.83

Easy 6 Easy 7 Easy 8 Easy 8 Easy 7 Easy 8 Easy 6 Easy 6 Easy 7 Easy 8

Moderate 7 oderat 8 oderat 3 oderat 7 oderat 9 oderat 7 oderat 8 oderat 8 oderat 5 oderat 7

Hard 7 Hard 5 Hard 9 Hard 5 Hard 4 Hard 5 Hard 6 Hard 6 Hard Hard 8 5

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 7 Topic 1 4 Topic 1 3 Topic 1 4 Topic 1 5 Topic 1 9 Topic 1 10 Topic 1 4

Topic 2 6 Topic 2 5 Topic 2 5 Topic 2 6 Topic 2 4 Topic 2 5 Topic 2 7 Topic 2 5 Topic 2 4 Topic 2 4

Topic 3 6 Topic 3 4 Topic 3 3 Topic 3 5 Topic 3 7 Topic 3 8 Topic 3 4 Topic 3 3 Topic 3 2 Topic 3 6

Topic 4 3 Topic 4 6 Topic 4 4 Topic 4 5 Topic 4 6 Topic 4 2 Topic 4 4 Topic 4 3 Topic 4 4 Topic 4 6

Copyright © 2020 by James R. Parry

18

Ignoring the recommendation of even distribution to maintain the cut score, the assessment design

function of Questionmark OnDemand was instructed to select 20 items at random from the topic

Fairness Research which contained all 60 available test-items (Figure 2).

Figure 2 - Questionmark OnDemand Item Selection Criteria for Hypothetical Data

Thirty (n=30) iterations of a 20-question test were generated as illustrated in tables 6, 7 and 8

Table 6 - Experiment #1A - Item Selection - Attempts 1 - 10

Page 20: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Table 7 - Experiment #1A - Item Selection - Attempts 11 - 20

Table 8 - Experiment #1A - Item Selection - Attempts 21 - 30

Copyright © 2020 by James R. Parry

19

Copyright © 2020 by James R. Parry

19

Table 7 - Experiment #1A - Item Selection - Attempts 11 - 20

Table 8 - Experiment #1A - Item Selection - Attempts 21 - 30

Page 21: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Standard Distribution of Experiment #1 Samples

Random Data

Sample Difficulty Statistics

Target Difficulty/Cut Score 58.02

Mean Difficulty 58.54

Median 58.65

Minimum 52.67

Maximum 65.63

Variance Target vs. Mean 0.13

Standard Deviation all Averages 3.10

95% Confidence Score 1.10849085

Kurtosis 0.308023894

Skewness 0540272895

Copyright © 2020 by James R. Parry

20

The item selection difficulty varied widely with each iteration (see ‘Difficulty’ columns in tables

6, 7, and 8). The target cut score/difficulty was 58.02. The randomization produced a difficulty

range between 52.67 and 65.63 with average of 58.54. The standard deviation of the scores was

3.10 with a 95% confidence interval22 of 1.10849 which means that the true population mean is

between 56.9 and 59.13 of the 30 samples. The kurtosis23 of the average difficulty is 0.308 and the

skewness24 is 0.543. The number of items at each difficulty level from each topic varied with each

iteration. Table 9 provides a summary of the statistics for the sample. Figure 3 illustrates the stand-

ard distribution curve of the sample.

Table 9 - Difficulty Statistics for Experiment #1A

Figure 3 - Standard Distribution of Test Difficulty - Experiment #1A - Random Selection

The content (topic) coverage was erratic as illustrated by the four-color (delineated by sections)

display and ‘Total From Topic’ in tables 6, 7, and 8.

Conclusion: All topics were not covered equally in either difficulty or content.

22 A confidence interval is a range of values that you can be fairly sure contains the true mean of the population. 23 Most often, kurtosis is measured against the normal distribution. If the kurtosis is close to 0, then a normal dis-tribution is often assumed. A low kurtosis indicates a lack of significant outliers. A high kurtosis indicates significant outliers. 24 Skewness is usually described as a measure of a dataset’s symmetry – or lack of symmetry. A perfectly symmet-rical data set will have a skewness of 0 which is referred to as “normal” distribution. Negative skew indicates data is skewed left and positive indicates data is skewed right when referring to the “tail”.

Page 22: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

FAIRNESS RESEARCH

1.0 TOPIC I

ILI 1.4 EASY

LI 1.4 HARD

Li 1.0 MODERATE

A 2.0 TOPIC 2

LI 2.4 EASY

LI 2.4 HARD

LI 2.0 MODERATE

A 3.4 TOPIC 3

LI 3.G EASY

LI 3.4 HARD

LI 3.4 MODERATE

A f 4.4TOPIC 4

Li4.4EASY

LI 4.4 HARD

LI 4.0 MODERATE

Topic

Topic Cut

Score g.

Difficulty

Items in % of Total Avaiable

Hard

%From

Topic

Available

Mod

%From

Topic

Available

Easy

%From

Topic

Total It

Needed

From

Topic

TopicTopicItems

Use Hard

lCalculated)

Use Hard

'Actual)

the Mod

(Calculated)

Use Mod

lActual)

Use Easy

(Calculated)

Use Easy

Actuall

10 52 15 25.00% 5 33% 5 33% 5 33% 5.00 1.67 1 1.67 2 1.67 2 1.0

2 50 15 25.00% 5 33% 5 33% 5 33% 5.00 1.67 2 1.67 1 1.67 2 2

3 55 15 25.00% 5 33% 5 33% 5 33% 5.00 1.57 2 1.67 2 1.67 1 3

4 55 15 25.00% 5 33% 5 33% 5 33% 5.00 1.67 1 1.67 2 1.67 2 4

Copyright © 2020 by James R. Parry

21

Experiment #1B – Stratified Randomization – Hypothetical Data Using the same item data from experiment #1A, the items were divided into four topics within the

main topic and further subdivided into difficulty sub-sub topics based upon the results of the cut

score rating tool (Figure 4)

Figure 4 - Topic, Sub-topic Structure Used in Experiments

The final test design recommendations to maintain fairness were as shown in table 10

Table 10 - Recommended Test Design - Stratified Randomization - Experiment #1B

Note: Each topic has a different cut score/difficulty rating but the overall database difficulty re-

mains at 58.02.

Page 23: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Question selections

2 random question(s) from topic 'FAIRNESS RESEARCH/1.4 TOPIC 1 /1 . G EASY' excluding subtopics

2 random question(s) from topic 'FAIRNESS RESEARCH/1.4 TOPIC 111.4 MODERATE' excluding subtopics

1 random question(s) from topic 'FAIRNESS RESEARCH/1.4 TOPIC 111.4 HARD' excluding subtopics

2 random question(s) from topic 'FAIRNESS RE SEARCHf2. TOPIC 212.4 EASY' excluding subtopics

2 random question(s) from topic 'FAIRNESS RE SEARCHf2. TOPIC 212.4 HARD' excluding subtopics

1 random question(s) from topic 'FAIRNESS RESEARCH/2.4 TOPIC 2/2.4 MODERATE' excluding subtopics

1 random qu estion Cs) from topic 'FAIRNESS RESEARCH/3.4 TOPIC 3/3.4 EASY' excluding subtopics

2 random question(s) from topic 'FAIRNESS RESEARCH/3.4 TOPIC 313.4 HARD' excluding subtopics

2 random question(s) from topic 'FAIRNESS RE SEA.RCH.G. TOPIC 313.0 MODERATE' excluding subtopics

2 random question(s) from topic 'FAIRNESS RESEARCH/4.4 TOPIC 4/4.4 EASY' excluding subtopics

1 random ciu estion (s) from topic 'FAIRNESS RESEARCH/4.4 TOPIC 4/4.4 HARD' excluding subtopics

2 random question(s) from topic 'FAIRNESS RESEARCH/4.4 TOPIC 4/4.0 MODERATE' excluding subtopics

Attempt

RID

1

SCORE

Experiment

Attempt

COD

2

SCORE

til- Directed

Attempt

COD

Random

3

SCORE

Selection

Attempt

RID

of 20 items

4

SCORE

from

Attempt

RID

all 4 topics.

5

SCORE

Attempt

RID

Desired to

6

SCORE

get difficulty

Attempt

RID

7

SCORE

is 58.02 with

Attempt

RID

5 items

8

SCORE

from each

Attempt

RID

topic.

9

SCORE

Attempt

OLD

10

SCORE

1.0 E3 79.17 1.0 E 79.17 1.0 El 77.50 1.0 El 77.50 1.0 E 79.17 1.0 E2 90110 1.0 E2 90110 1.0 El 77.50 1.0 E2 90.00 lb El 7750

1.0 E4 80.83 1.0 E4 80.83 1.0 E2 9010 lb E2 9010 lb 5 75.83 LO ES 79.17 LO E4 80.83 LO E2 90110 lb 5 75.83 LO 5 7583

LO H5 35.83 LO H3 4083 LO H2 29.17 L011.2 29.17 L0115 35.83 L0113 40.83 L0112 29.17 L0113 40.83 L0112 29.17 LO H1 4833

1.0 Ml 68.33 1.0 M3 60.83 1.0 M3 60.83 1.0 M3 60.83 LO M2 55110 LO M1 613.33 LO M3 60.83 LO M1 613.33 LO M4 70.83 LO M1 61333

1.0 M2 551* 1.0 M4 70.83 1.0 M4 70.83 1.0 M5 55110 lb M5 5510 lb M2 55110 lb M5 55110 lb M2 55110 lb M5 65.00 lb M3 6083

20 E2 9500 29 E2 95110 29 ES 7250 29 El 8230 29 El 8250 29 ES 7250 29 ES 7250 29 El 8230 29 El 8250 29 E3 7250

29 E4 82.50 29 E4 82.50 29 ES 77.50 10 E4 82.50 /0 E4 82.50 /0 E4 82.50 /0 E4 82.50 /0 E4 8250 /0 ES 77.50 /0 5 7750

20 114 44.17 20 114 44.17 20 114 44.17 20 Hl 25110 29 Ill 25110 29 114 44.17 29 Ill 2510 29 Ill 2510 29 113 46.67 29 H1 25013

/0 115 30.00 /0 H5 30.00 /0 115 30.00 2.0 H2 33.33 20 114 94.17 20 115 30.00 20 113 46.67 20 113 46.67 20 114 94.17 20 H4 44.17

20 M3 61.67 20 M5 65.00 2.0 M5 65.00 2.0 M3 61.67 29 M3 61.67 29 M5 65.03 29 M1 6750 29 M3 61.67 29 M2 6833 29 M2 61333

3.0 E4 82.50 3.0 El 80.83 39 E2 84.17 3.0 El 80.83 5.0 ES 72.50 5.0 ES 72.50 5.0 E2 84.17 5.0 E4 82.50 5.0 ES 72.50 5.0 5 7250

3.0 Hl 26.67 3.0 Hl 26.67 39 114 33.33 39 Ill 26.67 3.0 Ill 26.67 3.0 112 48.33 3.0 Ill 26.67 3.0 Ill 26.67 3.0 111 26.67 3.0 H3 28.33

30 H4 3333 30 H2 4833 30 H5 25.83 3.0 113 2833 39 115 25.83 39 114 33.33 39 115 25.83 39 113 28.33 39 115 25.83 39 H4 33.33

30 M4 49.17 30 M1 6510 30 M2 49.17 3.0 M4 49.17 39 M1 6510 39 M2 49.17 39 M1 6510 39 M1 6510 39 Ml 6500 39 Ml 65.00

30 M5 61.67 30 M5 61.67 30 M3 5333 3.0 M5 61.67 39 M4 49.17 39 M3 5333 39 M5 61.67 39 M4 49.17 39 M3 5333 39 M3 53.33

4.0 El 7250 4.0 E2 85.00 4.0 E4 82.50 49 El 72.50 49 E2 8510 49 El 72.50 49 El 72.50 49 El 72.50 49 E2 85.00 49 El 7250

4.0 E4 8250 4.0 E4 8250 4.0 5 94.17 49 ES 72.50 49 E5 94.17 49 ES 72.50 49 E4 82.50 49 E2 8510 49 5 94.17 49 5 94.17

4.0 Hl 2833 4.0 H3 45.00 4.0 H3 4510 49 114 31.67 49 Hl 28.33 4.0 Hl 28.33 4.0 115 25.83 4.0 112 30.83 4.0 H1 28.33 4.0 H2 3083

40 M1 50.00 40 M2 50.00 40 M1 5010 40 M2 50110 4.0 M3 49.17 4.0 M1 50110 411M3 49.17 4.0 M4 49.17 4.0 M1 50.00 4.0 M3 49.17

40 M3 49.17 40 M5 54.17 40 M3 49.17 40 M4 49.17 4.0 M1 5010 4.0 M4 4917 4.0 M5 54.17 4.0 M5 54.17 4.0 M4 49.17 4.0 M2 5000

Difficulty 5842 Difficulty 62.42 Difficulty 59.2L Difficulty 56.50 Difficulty 57.63 Difficulty 57.83 Difficulty 5838 Difficulty 58.67 Difficulty 60.013 Difficulty 5837

Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7

'Moderate 7 Modest 7 Moderate 7 Moderate 7 IA oderat 7 IA oderat 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7

Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5

Copyright © 2020 by James R. Parry

22

Following the recommended test design, the assessment design function of Questionmark

OnDemand was instructed to select 20 items in a stratified random fashion from the topic Fairness

Research, with five items from each topic, with equal numbers from each level of difficulty (Figure

5).

Figure 5 - Item Selection Criteria - Experiment #1B

Thirty (n=30) iterations of a 20-question test were generated as illustrated in tables 11, 12, and 13

Table 11 - Experiment #1B Item Selections - Attempts 1 - 10

Page 24: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt 11 Attempt 12 Attempt 13 Attempt 14 Attempt 15 Attempt 16 Attempt 17 Attempt 18 Attempt 19 Attempt 20

QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE (ID SCORE QID SCORE QID SCORE QID SCORE

1.0 E2 9E100 1.0 ES 75.83

1-1-•

6

00

66

rn

3

3 2

Nu

l," 75.83 1.0 E4 80.83 1.0 El 77.50 1.0 E4 80.83 1.0 ES 75.83 1.0 ES 75.83 1.0 E4 8083 1.0 E3 79.17

1.0 El 77.50 10 El 77.50 77.50 1.0 El 77.50 1.0 E3 79.17 1.0 E2 90.00 10 E2 90.00 1.0 ES 79.17 1.0 E2 9000 1.0 El 77.50

1.0 M3 60.83 1.0 M4 70.83 68.33 10 M3 60.83 1.0 M1 6833 1.0 M3 60.83 1.0 M1 68.33 1.0 M5 65.00 1.0 MS 6500 1.0 M4 70.83

1.0 MS 6510 1.0 MS 65.00 6500 1.0 MS 6500 10 M5 65.00 1.0 M2 55.00 1.0 M4 70.83 1.0 M2 55.00 10 M3 6083 1.0 M2 55.0D

1.0 H3 40.83 19115 35.83 35.83 1.0 H3 40.83 1.0115 35.83 1.0114 48.33 1.0 Hl 43.33 1.0 115 35.83 1.0 H2 29.17 1.0 Hl 48.33

20 E4 82.50 2.0 El 82.50 2.0 ES 72.50 2.0 El 82.50 2.0 E3 72.50 2.0 E5 77.50 2.0 ES 72.50 10 E2 95.00 2.0 E3 7250 2.0 E2 95.0D

2.0 El 82.50 2.0 ES 77.50 2.0 ES 77.50 2.0 E2 95.00 2.0 E4 82.50 2.0 El 82.50 2.0 E4 82.50 29 ES 77.50 29 E2 9500 29 ES 77.50

2.0 H1 25.00 2.0114 44.17 20 111 25.00 2.01-4 44.17 2.0 H1 25.00 2.0111 25.00 20 114 44.17 20 115 30.00 10 H1 25.00 10 H3 46.67

20 H4 44.17 20 113 46.67 20 115 30.00 2.0 H1 2500 2.0114 44.17 2.0112 33.33 2.0 112 alas 29 114 44.17 2.0 115 3000 2.0 H2 33.33

20 M3 61.67 2.0 M2 68.33 20 M3 61.67 2.0 M4 49.17 29 Ml 57.50 2.0 MS 61.67 2.0 M2 68.33 2.0 Ml 57.50 29 M5 6500 2.0 M2 68.33

3.0 E3 72.50 39 El 80.83 3.0 El 80.83 3.0 E4 82.50 3.0 ES 72.50 39 El 80.83 3.0 ES 72.50 3.0 E2 84.17 3.0 E4 8250 3.0 E2 84.17

3.0 H3 28.33 39 H4 3333 39 H2 4833 3.0 H1 26.67 3.0 H4 33.33 3.0 H3 28.33 39 113 28.33 3.0 111 26.67 3.0 H2 48.33 3.0 H5 25.83

3.0 Hl 26.67 39 H3 2833 30 H4 33.33 3.0 H2 48.33 3.0 H2 48.33 3.0 H5 25.83 39 H5 25.83 30 H5 25.83 3.0 H1 26.67 3.0 Hl 26.67

3.0 M3 53.33 3.0 M3 5333 39 M2 49.17 3.0 MS 61.67 10 Ml 6510 3.0 M4 49.17 30 M4 49.17 30 M3 53.33 30 M3 53.33 3.0 M2 49.17

3.0 M4 49.17 3.0 M2 49.17 3.0 Ml 6500 3.0 M2 49.17 10 M5 61.67 3.0 M5 61.67 30 Ml 65110 30 M2 49.17 30 M4 49.17 3.0 Ml 65.00

4.0 El 72.50 49 El 72.50 4.0 ES 94.17 4.0 El 72.50 4.0 E4 82.50 49 E4 82.50 4.0 E2 85.00 4.0 El 72.50 4.0 E2 8500 4.0 E4 8250

4.0 E4 82.50 49 ES 94.17 4.0 E2 8500 40 E2 85.00 4.0 El 72.50 49 E3 72.50 49 ES 94.17 4.0 ES 94.17 4.0 ES 94.17 4.0 E2 85.0D

4.0 Hl 28.53 49112 30.85 49 H2 30.83 4.0115 2185 4.0111 28.33 4.0115 25.83 49 112 30.83 49 112 30.83 4.0 H1 2855 4.0 H4 31 67

4.0 M5 54.17 4.0 M4 49.17 4.0 M3 49.17 4.0 MS 54.17 40 M2 50.00 4.0 M2 50.00 4.0 M2 50110 40 MS 54.17 40 M3 49.17 4.0 M4 49.17

4.0 Ml 50.00 49 M5 54.17 4.0 M4 49.17 4.0 MS 49.17 49 MS 54.17 4.0 M3 49.17 4.0 M3 49.17 49 M3 49.17 49 M2 5000 4.0 M5 54.17

Difficulty 5738 Difficulty 59.50 Difficulty 58.71 Difficulty 58.79 Difficulty 59.29 Difficulty 57114 Difficulty 60.21 Difficulty 58.25 Difficulty 59.01 Difficulty 60.25

Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7

1 c de rat. 7 1,1oderat. 7 1,1oderat 7 Moderat 7 Moderate 7 Mod erat 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7

Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5

Attempt 21 Attempt 22 Attempt 23 Attempt 24 Attempt 25 Attempt 26 Attempt 27 Attempt 28 Attempt 29 Attempt 30

QID SCORE QID SCORE Q1D SCORE QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE

6▪ 0

06

6

=3

3r9

r2

N3 A

75.83 1.0 E4 80.09 1.0 E3 79.17 1.0 El 77.50 1.0 E4 80.83 1.0 El 77.50 1.0 E3 79.17 1.0 ES 79.17 1.0 ES 79.17

2 2 2 2

0 0

0

75.83

900E1 10 ES 79.17 10 E5 7583 10 E2 90.011 10 El 77.50 10 E4 80 83 10 E5 7583 10 E5 75 83 10 El 7750 79.17

70.83 1.0 M5 65.00 10 M5 65.00 10 M5 65.00 10 M3 60.83 1.0 M9 70.83 10 MS 60.83 1.0 Ml 68.33 1.0 M5 65.00 70.83

55.00 1.0 M3 60.83 1.0 M3 6083 10 M2 55.00 10 M2 55.00 1.0 M2 55.00 10 Ml 66.33 1.0 M2 55.00 1.0 M2 55.00 60.83

48.33 10 H2 29.17 10112 29.17 1.0 HS 40.83 1.0 H5 35.83 1.0 112 29.17 1.0 1-15 35.83 1.0111 48.33 1.0111 48.33 29.17

20 E4 82.50 20 ES 7250 20 E2 9500 20 E5 7730 20 E2 95 00 20 ES 7750 20 El 8250 20 E2 9500 20 E6 7250 20 E5 77.50

2.0 ES 7750 20 E4 8250 10 E3 7250 10 E3 7250 10 El 82.50 20 E4 82.50 2.0 ES 77.50 2.0 El 82.50 20 ES 7750 10 E3 72.50

20 112 3133 2.0111 25.013 20 114 44.17 10 H4 44.17 10 H3 46.67 10 111 25.00 20 111 2500 2.0113 46.67 2.0114 44.17 20 111 25.00

2.0115 30.00 2.0112 3333 20 113 46.67 10 H2 3333 10 Hl 25110 10 113 46.67 10 1-14 44.17 2.0114 44.17 2.0112 3333 20 112 33.53

20 M4 49.17 2O M5 65.00 2_0 M3 61 67 20 M1 57 50 20 M5 65.00 20 M4 49.17 20 M4 49.17 20 MS 61.67 2_0 M5 65 00 2_0 M3 61.67

3.0 E2 84.17 30 ES 72.50 3.0 El 80.83 3.0 ES 72.50 3.0 E4 82.50 3.0 E2 84.17 3.0 E4 82.50 3.0 E3 72.50 30 E2 84.17 3.0 ES 72.50

30114 33.33 3.0 Hl 26.67 30 111 26.67 30 H3 2833 ao H4 33.33 30112 48.33 30 H4 33.33 3.0 H4 33.33 3.0113 28.33 30 112 48.33

30 112 48.33 3.0 H3 28.33 30 112 43.33 30 Hl 26.67 3.0 H2 48.35 3.0 111 26.67 3.0 1-13 2833 3.0 H3 28.33 3.0112 48.33 30 113 28.55

30 M5 61.67 30 M5 61 67 30 M1 6500 30 M5 51.57 30 M3 53.33 30 M5 61.67 30 M1 65.0(1 30 M2 49.17 3_0 M4 4917 3_0 M3 53.33

3.0 M4 49.17 3.0 M3 53.33 3.0 M2 49.17 30 M4 49.17 10 MS 61.67 3.0 M3 53.33 3.0 M5 61.67 10 M3 53.33 3.0 M1 65.013 3.0 Ml 65.00

4.0 E2 8510 40 E2 85.00 40 E4 82.50 40 ES 94.17 40 E2 85110 40 El 72.50 4.0 E2 85.00 4.0 ES 94.17 40 El 72.50 40 ES 94.17

4.0 ES 94.17 40 ES 94.17 40 E2 8500 40 E2 85.00 40 ES 94.17 40 E2 85.00 4.0 E3 72.50 4.0 E4 82.50 40 ES 94.17 40 El 72.50

40 H5 25.83 40115 25.83 40114 31.67 90 H2 30.83 40 H1 28.33 40112 30 83 401-12 30.83 40115 25.83 40115 25.83 40 H3 45 00

4.0 M4 49.17 4.0 M4 49.17 4.0 M1 50.00 4.0 M1 50.00 4.0 M1 50.00 4.0 M2 50.00 4.0 M1 50.00 4.0 M2 50.00 4.0 M2 50.00 4.0 M3 49.17

4.0 M3 49.17 4.0 M5 54.17 4.0 M4 49.17 40 M5 54.17 40 M2 50.00 4.0 M9 49.17 4.0 M4 49.17 40 M5 54.17 4.0 M5 54.17 4.0 Ml 50.00

Difficulty 59.63 Difficulty 57.17 Difficulty 59.92 Difficulty 58.79 Difficulty 60.54 Difficulty 57.79 Difficulty 5783 Difficulty 6010 Difficulty 59.46 Difficulty 58.21

Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7 Easy 7

Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7 Moderate 7

Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6 Hard 6

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5 Topic 2 5

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5 Topic 4 5

Copyright © 2020 by James R. Parry

23

Table 12 - Experiment #1B Item Selections - Attempts 11 - 20

Table 13 - Experiment #1B Item Selections - Attempts 21 - 30

The item selection difficulty remained relatively constant as desired with each iteration (see ‘Dif-

ficulty’ columns in tables 11, 12, and 13). The target cut score/difficulty was 58.02. The stratified

randomization produced a difficulty range between 56.57 and 62.42 with average of 58.84. The

standard deviation of the scores was 1.24 with a 95% confidence interval of 0.4456 which means

that the true population mean is between 58.39 and 59.29 of the 30 samples. The kurtosis of the

Page 25: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Sample Difficu lty Statistics

Target Cut Score 58.02

Mean difficulty 58.84

Median 58,75

Minimum 55.50

Maximum 62.42

Variance Target vs. Mean 0.34

Standard Deviation all Averages 1.24

95% Confidence Score 0.445089638

Kurtosis 0.986867811

Skewness 0.537266918

Standard Distribution of Experiment #1 Samples

Stratified-Random Data

A 40 45 50 55 60 ES 70 7. SC

Average Test Difficulty

3556

3056

25%

,S 2096

is%

1056

9%

096

Copyright © 2020 by James R. Parry

24

average difficulty is 0.987 and the skewness is 0.537. The number of items at each difficulty level

from each topic varied with each iteration. Table 14 provides a summary of the statistics for the

sample. Figure 6 illustrates the standard distribution curve of the sample.

Table 14 - Difficulty Statistics for Experiment #1B

Figure 6 - Standard Distribution of Test Difficulty - Experiment #1B - Stratified Random Selection

The content (topic) coverage was equal as stratified for each iteration as illustrated by the four-

color (delineated by sections) display and ‘Total From Topic’ columns in tables 11, 12, and 13.

Conclusion: All topics were covered equally as desired in both difficulty and content. Comparing

the distribution of test difficulty scores in figures 3 and 6 shows that the stratified randomization

consistently produced tests well within an acceptable range to meet the desired cut score of 58.02.

Page 26: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

TEST NAME Course/Certification Name: I COURSE/CERTIFICATION NAME I Teri Named

Test Name:1 TEST NAME Course/Certification Name: I COURSE/CERTIFICATION NAME

Revision 1 Facilitator Name/Date

Revision 2 Facilitator Name/Date me:Irnrniddirm

CUT SCORE CALCULATION TOOL

F aci I ator Name/Date

Topic 1

11, car, ix te 30 DAV C. MO ONLY

Test Item COD

Enter .1f

New, Rif

Retired

Difficulty

Metatag

Average

Percentage

Correct {Angoff

Rating)

Expert 1

Name

Expert 2

Name

Expert 3

Name

Expert4

Name

Expert5

Name

Expert 6

Name

Expert 7

Name

Expert

Name

Expert 9

Name

Expert 19

Name

Standard

Deviation

1.0 M1 53.00 70 65 60 50 70 8.37

1.0 El 3000 75 75 so 75 95 8E0

1 0 E2 75.00 75 85 so 50 85 14_58

1.0 M2 57.00 70 50 70 70 75 9.75

70 50 70 70 ss

1.0 E3 77.00 75 65 75 95 75 10.95

1.0 MO 59.00 70 60 75 70 70

1.0 E4 76.00 70 70 85 90 55 10.84

1.0 E5 9400 90 95 95 95 95

1.0 E6 89.00

80 50 75 50 50

85 95 90 90 85 4.18

10 E7 78.00 80 so 70 so so 4.47

1.0 M4 71.00 75 60 so 60 80

1 0 E3 91.00 85 95 95 95 85 5.48

10 1/ 83.00 90 95 80 70 80 975

TO ESO 77.00 75 75 65 90 80 908

10E11 moo 65 85 75 90 75 975

1.0 82 AEI 7500 70 80 70 85 70 7.07

Facilitator Name/Date:

Enter Topic/IRO/Su bject ID:

Approximate Ditfii

25 - 48 3 Hard

48.4 -71: Moderate

71.8 -95 Easy

CCO:Pul:nis, LLC

Topic

Cut 78.05

Score

Standard Deviation

A standard deviation of more

than 10 will trigger an alert.

Discuss the outliers with the

judges who set them to

determine why. Change as

necessary.

la In this section 78%

4 Moderate In this section 27%

0 In this section CI%

18 TOTAL 100%

CUT SCORE CALCULATION TOOL

Facilitator Name/Date:

Enter Topic/I-PO/Subject ID:

Facilator Name/Date Revision 1 Facilitator Name/Date

Revision 2 Facilitator Name/Date IrrkMidd

Topic 2

Test Item DID

Enter. If

New, R if

Retired

Difficult/

Metatag

Average

p...r.,eflug... Correct fAngoff

Esti ngl

expert 1

Name

rt Expe 2

Name

rt Expe 3

Marne

apart 4

Name

Experts

Name

Expert 6

Name

Expert 7

Name

Expert 8

Name

Expert 9

Name

Expert10

Name

Standard

Deviation

2.0 El a2,5 0 80 80 95 90 70 9.75

2.0 E2 92.00 85 95 95 95 90 4.47

2.0 E3 7600 70 65 75 90 80 9.62

2.0 E4 75.00 70 85 70 80 70 507

2 O M1 6300 65 75 60 60 55 7.58

2.0 M2 48.75 65 50 40 40 11.81

2.0 E5 7400 70 70 75 85 70 6.52

2.0 ES 3000 80 75 70 95 80 9.35

2.0 E7 7500 75 85 65 80 70 7.91

2.0 E.2 81.00 85 90 65 80 85 9.62

20 El 8900 85 95 85 95 85 548

20 E1O 8300 80 85 80 90 80 4.47

2.0 M3 67.00 75 70 60 70 60 6.71

20 Ell 8230 80 70 95 85 1E141

2.0 E12 90.00 85 95 90 95 85 5.00

2.0 E13 7900 75 90 75 75 80 6.52

2.0 E14 90.00 75 95 95 95 90 8.66

2.0 EIS 8200 75 95 75 85 80 8.37

Era*, ac

22 In this section

10 Moderate In th is section

1 In this section

33 TOTAL

Topic Moderate

COT 74.00 Difficulty

Score

Approximate Difficulty Rating

25-48.3 Hard

48.4-71: Moderate

71.8-95 Easy

Standard Deviation

A standard deviation of more

than 10 will trigger an alert.

Discuss the outliers with the

Judges who set them to

determine why. Change as

necessary.

Copyright © 2020 by James R. Parry

25

Experiment #2A – Random Selection – Real Client #1 Data Note: Client Test-item QIDs replaced to protect confidentiality

Using cut score (Parry Method) results from real client data (Tables 15, 16, & 17 – partial views),

the items were divided into three topics within the main topic and further subdivided into difficulty

sub-sub topics based upon the results of the cut score rating tool. The cut score/difficulty for the

entire database (71 items) was determined to be 76.13% by averaging all three topic cut scores.

Table 15 - Experiment #2A - Difficulty Calculations - Real Data - Topic 1

Table 16 - Experiment #2A - Difficulty Calculations - Real Data - Topic 2

Page 27: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

C

pass CEOintsuttanis,11C

Topic

Cut 77.01

Score

17 In this section 85'i

Moderate In this section 15'i

In this section

20 TOTAL 1CV

Approximate Difficulty Rat r

25-18.3 Hard

48.4 - Moderate

71 8 - 95 Easy

Standard Deviation

A sta nda rd deviation of more

than 10 will trigger an alert.

Discuss the outliers with the

judges who set them to

determine why. Change as

necessary.

CUT SCORE CALCULATION TOOL

Course/Certification Name:I COURSE/CERTIFICATION NAME Test Name:) TEST NAME

Facilitator Name/Date:

Enter TopiciTPOISu bier, ID:

Facilator Name/Date Revision 1 Facilitator Name/Date

Revision 2 Facilitator Name/Date Topic 3

Test Item COD Enter. If New.

etired R If

R

Difficult/ Metatag

Average Percentage

IA Corgoff rea

n

Rating)

Expert]. Name

Expert 2 Name

Expert 3 Name

Expert 4 Name

Expert 5 Name

Expert 6 Name

Expert 7 Name

Expert 8 Name

Expert 9 Name

Expert 10 Name

Standard Deviation

1.0 El 90.00 80 95 90 95 90 6.12

1.0 E2 87.00 75 95 90 90 85 7.51

1.0 E3 84.00 75 95 75 95 80 10.25

1.0E4 74.00 75 85 60 SO 70 0.62

3.0 ES 7 '900 80 ss 65 90 75 0.62

3.0 E6 73.0C 80 75 75 65 70 5.72

2.0 E7 83.00 80 90 70 90 85 8.37

R /5 70 60 95 80

3.D ES 72.00 80 75 65 80 60 9.08

3.01E9 89.00 80 95 90 95 85 6.52

3.0 Ell) 85.00 so 85 85 95 80 6.12

3.0 N11 5750 70 50 60 50 9.57

30 Ell MOO RO 95 90 70 85 9.62

3.0 312 L 85.00 85 95 70 90 85 P.35

3.0 642 61.00 65 70 50 50 70 10.25

3.0 E13 72.00 75 85 70 60 70

3.0 314 83.00 80 90 70 95 80

3.0 EIS 71.00 70 60 80 75 8.0

Topic

Topic Cut

Score &

Difficulty

Items in

Topic

% of Total

Items

Avaiable

Hard

% From

Topic

available

Mod

'A From

Topic

Available

Easy

% From

Topic

Topic 1 78 18 25.35% 0 0% 4 22% 14 78%

Topic 2 7L 33 45.48% 1 3% 10 30% 22 57%

Topic 3 77 20 28.17% 0 0% 3 15`(:- 17 8596

Total 41

Needed

From

Topic

Use Hard

(Calculated)

Use Hard

(Actual)

Use Mod

(Calculated)

Use Mod

(Actual)

Use Easy

(Calculated)

Use Easy

(Actual) Topic

5.07 0.00 0 1.13 1 3.94 4 Topic 1

9.30 0.28 1 2.82 3 6.20 6 Topic 2

5.03 0.00 0 0.85 1 4.79 4 Topic 3

Copyright © 2020 by James R. Parry

26

Table 17 - Experiment #2A - Difficulty Calculations - Real Data - Topic 3

Topic 1 consisted of 18 test-items with a difficulty rating of 78 (Easy). Topic 2 consisted of 33

test-items with a difficulty rating of 74 (Moderate). Topic 3 consisted of 20 test-items with a dif-

ficulty rating of 77 (Easy). Each topic had a mix of hard, moderate, and easy items. (Table 18)

Table 18 - Experiment #2 - Item Difficulty Distribution by Topic

Referring to the design philosophy of the Compass Consultants spreadsheet tool as to the number

of items drawn from each section, it appears that topic 2 was considered to be the ‘most’ important

with 46.48% of the available items, topic 3 was the next ‘most’ important with 28.17% and topic

1 was the least important with 25.35%.

The final test design to maintain fairness in both content and difficulty is shown in table 19.

Table 19 - Experiment #2 - Recommended Test Design

Page 28: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt

QID

1

SCORE

Attempt

CID

Experiment

2

SCORE

112

Attempt

QM

- Random

3

SCORE

Selection

Attempt

CID

of 20

4

SCORE

items from

Attempt

CND

all 3

5

SCORE

topics. Real

Attempt

QM

Client

6

SCORE

Data. Desired

Attempt

0/ID

7

SCORE

target

Attempt

COD

difficulty is

8

SCORE

76.13.

Attempt

0/ID

9

SCORE

Attempt

QID

10

SCORE

1.0110 77.00 1.0 El 8000 1.0110 77.00 1.0 El 8000 1.0 E12 75.00 1.0 E10 77103 1.0112 75.00 1.0 E12 75110 1.0 12 75.00 1.0 El 80.00

1.0113 76.00 1.0 El0 77.00 1.0 Ell 78.00 1.0 Ell 78.00 1.0 E2 75.00 1.0 El4 79.60 1.0114 79.00 1.0 El3 7510 1.0 E4 76.00 1.0 El0 77110

10 E.3 7700 10 Ell 7800 10 El 7500 10 E7 78 CC 10 E7 7800 LO E5 9400 10 E9 8.360 10 El 7910 10 E7 7800 10 El4 7900

1.0 14 76.00 1.0 E5 94.00 1.0 E6 89.00 1.0 E8 91.CC 1.0 E9 83.00 1.0 E5 8910 1.0 M1 63.00 1.0 E3 77.00 1.0 M3 59.00 1.0 E4 75.60

1018 91.00 LO E9 83.00 2.0 El 83.00 1.0 Ml 63.CC 2.0 El0 83.00 LO E9 83.00 2.0 El0 83.00 LO E7 7810 2.0115 83.00 1.0 Ml 63.00

29 El 8300 io mi 63.90 20E13 7900 10 M4 71 CC 20116 8.300 10 M3 6900 29 Ell 8250 29 El 8390 29 E2 9200 10 M3 6900

2.0 Ell 82.50 2.0 E14 90.00 2.0 E2 9200 2.0 E14 9 C.CC 2.0 E20 80.00 2.0 El 83.00 2.0 113 79.00 2.0 (10 83.00 2.0 E3 76.00 2.0 El 83.00

2.0112 90.00 2.0 E15 82.0(1 20120 80.00 2.0 E16 83.01 2014 75.00 2.0 E10 83.03 2.0 E14 90.00 2.0 El3 7910 2.0 14 75.00 2.0 El7 79.00

2.0115 82.00 2.0 El6 83.00 2.0 E3 76.00 2.0 (19 86.00 2.0 ES 74.00 2.0 E12 90.00 2.0 El7 79.00 2.0 E3 76.00 2.0 E8 81.00 2.0 (18 81.00

29 12 9200 20 El7 7900 20 E4 7500 20 E2 9200 2018 8.160 20 El7 7910 29 E21 7860 2065 7410 20 E9 8900 20 E9 7400

2014 75.00 2.0 E21 78.0(1 2.0 E7 75.00 2.0 E3 75.00 2.0 M10 56.25 2.0 E20 8010 2.0 E5 74.00 2.0 08 81.00 2.0 Ml 53.00 2.0 E5 80.00

2.0 E5 74.00 2.0 E4 75.011 2.0 E9 80.00 2.0 E4 75110 2.0 M1 63.00 2.0 E5 7410 2.0 lb 80.00 2.0 H1 45.25 2.0 M3 57.00 2.0 E8 81.00

2.0 E7 75.00 2.0 E5 80.00 2.0 H1 46.25 2.0 E5 74.03 2.0 M3 57.00 2.0(8 8100 2.0111 46.25 2.0 M8 52.50 2.0 M5 58.00 2.0 M1 63.00

2.0 E9 89.00 2.0111 45.25 2.0 M3 57.00 2.0 E7 75.011 2.0 M6 53.75 2.0 M3 67.011 2.0 M3 57.00 3.0 El 90.0(1 3.0 El0 85.00 2.0 M3 67.011

2.0 M9 70.00 2.0 M4 53.00 2.0 ME. 52.50 2.01.14 53.00 2.0 M7 66.25 2.0 M6 53.75 2.0 M4 53.00 3.0 (10 8500 3.0 E2 8100 2.0 M4 53.00

3.0117 72.00 2.0 M8 52.50 3.0110 85.00 2.0M9 70.00 3.0112 85.00 3.0 EIS 73.00 2.0 M6 53.75 3.0 E14 83.00 3.013 84.00 2.0 MB 52.50

3.0 12 87.00 3.0 E12 85.00 3.0 11 90.00 3.0 E13 72.00 30114 8310 3.0 EIS 90.03 2.0 M8 52.50 3.0 116 75.00 3.0 E4 74.00 3.0 E33 72.00

3.0 E9 86.00 3.0 (14 83.00 3.0112 8500 3.0 (14 83.00 3.0 E4 74.00 3.066 73.00 3.0112 85.00 3.0 E37 72.00 3.0 E7 83.00 3.0 E1.5 73.00

30 M1 5750 30 El5 7300 39 17 8300 30 El6 7500 30 E5 7900 30 E8 7200 39 115 7300 3965 7900 3018 7200 3062 8700

3.0M3 57.50 3.0M3 57.50 39 M2 61.00 3.016 73.00 39 M3 57.50 3.0 M2 6110 3.0 E7 83.00 3.0 MI 57.50 3.0 M3 57.50 3.0 E3 84.00

Difficulty 78.63 Difficulty 74.61. Difficulty 7604 Diffkulty 76.90 Difficulty 73.59 Difficulty 77.54 Difficulty 72.95 Difficulty 74.86 Difficulty 76.73 Difficulty 73.68

Easy 17 Easy 15 Easy 16 Easy 15 Easy 14 Easy 15 Easy 14 Easy 17 Easy 15 Easy 14

Moderate 3 Moderate 4 Moderate 3 Moderate 4 Moderate 6 Moderate 4 Moderate 5 Moderate 2 Moderate 5 Moderate 6

Hard 0 Hard 1 Hard 1 Hard 0 Hard 0 Hard 0 Hard 1 Hard 1 Hard 0 Hard 0

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 6 Topic 1 9 Topic 1 6 Topic 1 4 Topic 1 6 Topic 1 4 Topic 1 5 Topic 1 4 Topic 1 6

Topic 2 10 Topic 2 10 Topic 2 11 Topic 2 10 Topic 2 11 Topic 2 9 Topic 2 13 Topic 2 8 Topic 2 9 Topic 2 10

Topic 3 5 Topic 3 4 Topic 3 5 Topic 3 4 Topic 3 5 Topic 3 5 Topic 3 3 Topic 3 7 Topic 3 7 Topic 3 4

Attempt

4310

11

SCORE

Attempt

101D

12

SCORE

Attempt

(310

13

SCORE

Attempt

0/10

14

SCORE

Attempt

1010

15

SCORE

Attempt

1010

16

SCORE

Attempt

4)10

17

SCORE

Attempt

QM

18

SCORE

Attempt

QID

19

SCORE

Attempt

0/10

20

SCORE

1.0 (14 7910 1.0 E13 75.00 1.0110 77.00 1.0 E10 77.00 1.0 E12 75.00 1.0 El 80.00 1.0 E4 75.03 1.0 E12 75.00 1.0 E10 77.00 1.0110 77.00

1.0 E3 7710 1.0 E8 91.00 1.0 E14 79.00 LO E12 75.00 1.0 E3 77.00 1.0 E2 75.00 1.0 E9 83.00 1.0 E7 7800 1.0 E12 75.00 1.0 E4 76.00

1.0 E9 83.00 LO E9 83.00 LO E5 89.00 1.0 E4 76.00 1.0 E5 94.00 1.0 E3 77.00 2.0 E13 79.00 LO (8 91.00 LO E12 8000 1.0 E5 94.00

LO M1 6300 10101 6300 10 ED 63 00 10 ES 910 10 E7 7800 10 M2 6700 20 E14 9000 20 E10 83 00 10E3 77 OD 10 E6 0900

1.0 M4 71.0 1.0 M3 69.00 2.0 Ell 81.50 1.0 M1 53.00 1.018 91.00 1.0 M4 71.00 2.0 E1? 79.03 2.0 Ell 82.50 19E5 94.00 1.0 M1 53.00

2.0 El 83.00 1.01414 71.00 2.0 E12 90.60 1.0 M2 57.00 1.019 83.00 2.0 EIS 82.0 2.0 E2 92.0 2.0 E12 4060 1.0 EG 89.0 1.0 M2 57.00

20 Ell 9090 29 El 8300 29 E15 6200 20 Ell] 6300 20 El0 8390 20 E22 7500 20 E20 80 .0 29 E14 90 00 10 E7 780 20 Ell 9000

2.0 EIS 82.0 2.0 Ell 82.50 2.0 E17 79.00 2.0 E12 40.00 2.0 Ell 82.50 2.0 E5 7410 2.0 E21 7800 2.0 (15 82.00 1.0 Ml 53.00 2.0 E13 74.00

20 E16 8310 20 Ell 90 00 20 E19 6600 20 E2 92 00 20 E14 9000 20 E6 8010 20 ES 8100 20 El6 83 00 10 M2 6790 20 E2 92 00

2.0 E22 7510 2.0 E14 90.00 2.0 E5 80.00 2.016 80.00 2.0E15 83.00 2.0 E9 89.03 2.0111 45.25 2.0 E19 86.00 2.0 E18 81.0 2.016 80.00

2 0 E6 8000 2 0 El7 790 20E7 75 00 20 M3 6700 2 0 E20 8000 20111 4625 20 M5 6600 20E2 92 00 20E19 660 20 ES 8100

2.0111 45.25 2.0 E3 75.00 2.0111 46.25 2.0 M4 53.00 2.0 E8 81.00 2.0 M1 63011 2.0 M7 65.25 2.0 E5 7400 2.0 E22 76.00 2.0111 46.25

2.0 M10 55.25 2.0 E5 7410 2.0 M4 53.00 2.0 M6 53.75 2.071 46.25 2.0 M6 53.75 3.0 E12 85.00 2.0 E7 75.00 2.0 E3 76.0 2.0 M10 56.25

20 M2 46.75 20 E7 7500 3_0 El0 8500 20 M7 6625 20 M5 6600 20 M7 6925 3 0 E3 8.400 2069 8900 20E5 740 20 M2 98.75

2.0 M7 65.25 2.0 M7 65.25 3.0 EIS 73.00 3.0 E12 85.00 2.0 M9 70.00 2.0 M9 70.00 3.0 E4 74.00 2.0 M2 48.75 2.0 M4 53.00 2.0 M7 56.25

3.0E12 85.00 3.0 El 90.00 3.0 E7 83.00 3.0 EIS 73.00 3.0 E13 72.00 3.0E14 83.0 3.0 E5 79.00 2.0 M5 68.00 2.0 M8 52.50 3.0 El 90.00

3.0 E14 83.00 3.0 E12 85.00 3.0 (8 72.00 3.0 E16 75.00 3.0 E13 90.00 3.0 E5 79.00 3.0 E5 73.00 2.0 M5 53.75 3.0 EIO 85.00 3.0 E4 74.00

3.0(14 90.0 3.0 E14 83.00 3.0 M1 57.50 3.0 13 64.00 3.0E2 87.00 3.0 E6 73.90 3.0 E8 72.00 3.0 EIO 85.00 3.0 E12 65.00 3.0(8 72.00

3 0 E6 7300 3065 7900 30 M2 6100 30 E6 73 00 30 E3 8400 3018 7210 3 0 E9 890 30912 8500 30 EIS 73 OD 3O M1 5750

30 E9 8900 30 E7 8300 30 M3 5750 30 E7 8300 30 M2 6100 3019 8900 30 M3 5750 30614 83 00 30E5 790 30 M2 6100

Difficulty 75.23 Ciffivulty 7904 Difficulty 74.54 Difficulty 75.35 Difficulty 78.79 Difficulty 73.31 Difficulty 76.60 Difficulty 79.70 Difficulty 76.03 Difficuky 73.00

Easy 14 Easy 16 Easy 15 Easy 14 Easy 16 Easy 13 Easy 16 Easy 17 Easy 16 Easy 12 Moderate 5 Moderate 4 Moderate 4 Moderate 6 Moderate 3 Moderate 6 Moderate 3 Moderate 3 Moderate 4 Moderate 7

Hard 1 Hard 0 Hard 1 Hard 0 Hard I Hard I Hard 1 Hard 0 Hard 0 Hard 1

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 6 Topic 1 4 Topic 1 6 Topic 1 6 Topic 1 5 Topic 1 2 Topic 1 3 Topic 1 9 Topic 1 6

Topic 2 10 Topic 2 9 Topic 2 9 Topic 2 8 Topic 2 9 Topic 2 10 Topic 2 10 Topic 2 14 Topic 2 7 Topic 2 9

Topic3 5 Topic3 5 Topic3 7 Topic3 6 Topic3 5 Topic3 5 Topic3 8 Topic3 3 Topic3 4 Topic3 5

Copyright © 2020 by James R. Parry

27

Ignoring the recommendation of item distribution to maintain the cut score, the assessment design

function of Questionmark OnDemand was instructed to select 20 items at random from the single

topic containing all of the test-items. Tables 20, 21, and 22 present the results.

Table 20 - Experiment #2A - Item Selection - Attempts 1 - 10

Table 21 - Experiment #2A - Item Selection - Attempts 11 - 20

Page 29: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt 21 Attempt 22 Attempt 23 Attempt 24 Attempt 25 Attempt 26 Attempt 27 Attempt 28 Attempt 29 Attempt 30

COD SCORE C)10 SCORE C)ID SCORE OID SCORE CID SCORE 919 SCORE OID SCORE COD SCORE CND SCORE C)ID SCORE

1.0 E14 79.00 1.0 E12 7500 1.0113 76.00 LO El3 76.00 1.0 E12 75.00 1.0 El 80.00 1.0 Ell 78.00 LO Ell 7810 1.0 E10 7700 1.0113 76.00

10 El 77 00 11 Eli 7600 10 El 80 OD 10 E2 7500 10 E4 7600 10 E9 83 00 10 E4 76 00 LO E4 7600 10 Ell 7800 10 E2 75 OD

1.0 14 75.00 1.0 E4 7500 1.0 E3 77.0D 1.0 E7 78.00 2.0 116 8300 2.0 E12 90.00 1.0 E5 94.00 1.0 E5 94.00 1.0 E2 7500 1.0 E4 76.00

10E5 9400 10 E7 7800 10E5 9401) 1 0 M1 6300 20E17 79I8) 20E15 8200 10E8 9100 10E6 8900 10E4 7600 10E5 94(8)

1.0 E3 89.00 1.0 E9 8300 1.0 M4 71.00 2.0 E10 8310 2.0 E21 78330 2.0 (18 81.00 1.0 E9 83.00 LO E8 SLOO 1.0 E9 8300 1.0 M3 69.00

113 E7 78.00 2.0 E13 79130 2.0 E10 83.00 2.0 EIS 81.00 2.014 7500 2.0 E21 78.00 1.0 M1 53110 21E13 79.00 1.0 M1 63130 2.0 E13 79.0D

1.018 72.00 2.0 E17 79.00 2.0 E20 80.00 2.0 E2 92.00 2.0 M1 63330 2.013 76.011 2.0113 79.00 2.0114 90.00 1.0 M2 67.011 2.0 E17 79.00

LC) 19 89.00 2.0 E22 7300 2.0121 78.0D 2.0 E20 80.00 2.0 M10 5325 2.0 E6 80.00 2.0 E17 79110 2.0 117 79.00 1.0 M4 71.00 2.0 E2 92.00

2.0 E1.3 79.00 2.0 E3 75.00 2.0 E5 74.00 2.0 E22 76.013 2.0 M2 48.75 2.0 E7 75.00 2.0118 81.00 2.015 74.00 2.0 E12 90.011 2.0 E21 78.00

20115 8200 2 0 E5 7400 20E7 7500 20E4 7500 20 M3 6700 20E9 8900 20E19 8600 20 Ea 8100 20E19 8600 20E9 8900

2.0 E16 8300 2.0 E8 81.00 2.0 M10 56.25 2.0 E5 74.00 2.0 M5 6800 2.0111 46.25 2.0 E2 92.00 2.0111 46.25 2.0 E21 78.00 2.081 46.25

20 E22 7600 20 M1 6300 20 M2 4875 20 ES 8100 20 M7 6625 20 M5 6800 20 E3 7600 20 M6 5375 20 E3 7600 20 M3 67 00

2.0 E5 74.00 3.0 Ell 84.00 2.0 M8 52.50 20111 46.25 2.0 M9 7000 2.0 M7 66.25 20111 46.25 2.0 M9 70.00 2.0 E4 75.011 3.0 El 90.00

21 E7 75.00 3.0 E17 72.00 3.0 Ell 84.0D 2.0142 48.75 3.0 (1 90.00 3.0 E10 85.017 2.0 M10 56.25 3.0 (1 90.00 2.0 E9 89.00 3.0 E13 72.00

30111 8400 30E4 7400 30E13 72 00 20 M7 6625 30 Ell 8400 30E13 72 00 20 M2 48 75 30E13 7200 2_0 M8 5250 30 EIS 73 00

3.0 E14 83.00 3.0 E5 79130 3.0 E3 84.00 2.0 M9 70.00 311 112 8500 3.0 E14 83.00 2.0 M8 52.50 3115 79.00 3.0 E12 85.011 3.0 E16 75.00

30117 7200 3 0 E6 7300 30E7 8300 30E10 8500 30E13 7200 30E15 7300 30E14 8300 3016 7300 30E15 7500 30E4 7400

3.0 12 87.00 3.0 E9 8900 3.0 E9 89.0D 3.0 E12 85.00 3.0 El6 75.00 3.0 EIS 75.00 3.0 EIS 73110 3.0 E7 8310 3.0 E2 87.00 311E6 73.00

3.0 19 89.00 3.0 M2 6100 3.0 M1 57.50 3.0 E6 73.00 3.0 E17 72.00 3.0 E2 87.00 3.016 73.00 3118 72.00 3.0 E8 72.01) 3.018 72.0)

30M2 6100 30 M3 5750 30 M2 619) 30E9 8900 30 E3 8400 30E4 7400 30M2 6100 30M1 57 50 30 M1 5750 30 M1 57 50

Difficulty 79.95 Difficulty 75.28 Difficulty 73.80 Difficulty 74.86 Difficulty 73.36 Difficulty 77.18 Difficulty 7359 Difficulty 76.58 Difficulty 75.65 Difficulty 75.34

Easy 19 Easy 17 Easy 14 Easy 15 Easy 13 Easy 17 Easy 14 Easy 16 Easy 15 Easy 16

Moderate 1 Moderate 3 Moderate 6 Moderate 4 Moderate 7 Moderate 2 Moderate 5 Moderate 3 Moderate 5 Moderate 3

Hard 0 Hard 0 Hard 0 Hard 1 Hard 0 Hard 1 Hard 1 Hard 1 Hard 0 Hard 1

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 3. Topic 1 5 Topic 1 5 Topic 1 4 Topic 1 2 Topic 1 2 Topic 1 6 Topic 1 5 Topic 1 8 Topic 1 5

Topic 2 6 Topic 2 7 Topic 2 8 Topic 2 12 Topic 2 11 Topic 2 31 Topic 2 10 Topic 2 8 Topic 2 7 Topic 2 7

Topic 3 6. Topic 3 E. Topic 3 7 Topic 3 4 Topic 3 7 Topic 3 7 Topic 3 4 Topic 3 7 Topic 3 5 Topic 3 8

Sample ❑ifficulty Statistics

Target Cut Score 76.13

Mean difficulty 75.87

Median 7534

Minimum 73.(X)

Maximum 79.95

Variance Target vs. Mean 0.03

Standard Deviation all Averages 2.17

95% Confidence Score 11777910235

Kurtosis -0.52653425

Skewness 0.613319589

Copyright © 2020 by James R. Parry

28

Table 22 - Experiment #2A - Item Selection - Attempts 21 - 30

The item selection difficulty varied widely with each iteration (see ‘Difficulty’ columns in tables

20, 21, and 22). The target cut score/difficulty was 76.13. The randomization produced a difficulty

range between 73.00 and 79.95 with average of 75.87. The standard deviation of the scores was

2.17 with a 95% confidence interval of 0.778 which means that the true population mean is be-

tween 75.1 and 76.64 of the 30 samples. The kurtosis of the average difficulty is -0.527 and the

skewness is 0.613. The number of items at each difficulty level from each topic varied with each

iteration. Table 23 provides a summary of the statistics for the sample. Figure 7 illustrates the

standard distribution curve of the sample.

Table 23 - Difficulty Statistics for Experiment #2A

Page 30: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

67 69 71 73 75 77 79

Average Test Difficulty

at tn SS

Standard Distribution of Experiment #2 Samples Random Data

Copyright © 2020 by James R. Parry

29

Figure 7 - Standard Distribution of Test Difficulty - Experiment #2A - Random Selection

The content (topic) coverage was erratic as illustrated by the three-color (delineated by sections)

display and ‘Total From Topic’ columns in tables 20, 21, and 22.

Conclusion: All topics were not covered equally in either difficulty or content.

Page 31: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

au estion selections

4 random question(s) from topic 'FAIRNESS RESEARCH 211.4 TOPIC 111.4 EASY' excluding subtopics (Avoid previously delivered}

1 random question(s) from topic 'FAIRNESS RESEARCH 211.4 TOPIC 111.4 MODERATE' excluding subtopics (Avoid previously delivered}

6 random question(s) from topic 'FAIRNESS RESEARCH 212.4 TOPIC 212.4 EASY' excluding subtopics (Avoid previously delivered}

3 random question(s) from topic 'FAIRNESS RESEARCH 212.4 TOPIC 212.4 MODERATE' excluding subtopics (Avoid previously delivered}

1 random question(s) from topic 'FAIRNESS RESEARCH 212. 4 TOPIC 212.4 HARD' excluding subtopics (Avoid previously delivered}

4 random question(s) from topic 'FAIRNESS RESEARCH 213.4 TOPIC 313.4 EASY' excluding subtopics (Avoid previously deltvered)

1 random question(s) from topic 'FAIRNESS RESEARCH 213.4 TOPIC 35.0 MODERATE' excluding subtopics (Avoid previously deltvered)

Attempt

flit)

1

SCORE

Attempt

COD

Experiment

2

SCORE

#2 - Directed

Attempt

COD

3

SCORE

Random

Attempt

CID

Selection

4

SCORE

of 20 items

CID

Attempts

from

SCORE

all 3 topics.

Attempt

CID

Real

6

SCORE

Client Data.

Attempt

CID

Desired

7

SCORE

target

Attempt

CID

difficulty

8

SCORE

is 76.13.

Attempt

COD

9

SCORE

Attempt

COD

10

SCORE

10 E13 73E0 10 E13 7600 10 E2 7500 10 E10 77 00 10 E8 9100 10 El 6000 10 E9 63 00 10 E12 75.00 10 E7 7800 10 E13 7300

10 El EOM 10 E9 6300 10 E12 7500 10 El 8000 10 ES 9400 10 Ell 7800 10 E12 75.00 10 El 8000 10 E4 7600 10 El 8000

10 ES SLED 10E14 7900 10 ES 9100 LO Ell 7800 10E14 7900 10E7 7800 10E7 78.00 1 0 E7 7800 10E5 9400 10E8 9100

1.0 E4 75.00 1.0 El SODO 1.0 E3 77.00 LO E3 77.00 LO E4 75.00 1.0 E9 3300 1.0 Ell 7803 1.0 E8 91.00 LO E13 76.00 1.0 E3 77.00

10M1 6300 10 M1 6300 10 M2 67.00 10M2 6700 10M2 6700 10M3 6900 10 M4 7100 10 M3 6900 10M4 7100 10M4 7100

20 E15 8200 20 E22 7600 20 E3 7600 20 E9 8900 20 E11 6230 20 E15 3200 20 E20 6000 20 E7 7500 20 E15 82 00 20 E9 3900

20E14 900M) 20E17 7900 20E7 7500 20E13 7900 20 EIS 6300 20E7 7500 20E6 6000 20E13 8100 20E13 7900 20E4 7500

20 ES 6100 2 0 El3 7900 2 0 E21 78.00 20E18 81 OD 20E10 83.00 20E9 3900 2 0 E2 9200 2 0 El4 90.00 20E17 7900 20 El 8300

2.0 E2 9200 2.0 E9 6900 2.0 E17 79.00 2.0 E12 90.00 2.0 E17 79.00 2.0 E14 93.00 2.0 E9 6900 2.0 E20 80.00 20 E5 74.00 2.0 E5 7400

2.0 El8 61.00 2.0 E7 75.03 2.0 E10 8300 2.0 ES 81.00 2.0 El3 79.00 2.0 E19 35.00 2.0 E15 32.00 2.0 E15 82.03 211E7 75.00 20 E15 82.00

2.0 E17 79.00 2.0 E15 62.03 2.0 E14 913.00 2.0 ES 74.00 2.0 E22 7500 2.0 E4 7300 2.0 E13 79.03 2.0 E4 75.00 2.0 E4 75.00 2.0 E2 9200

2.0 M5 66.00 2.0 M9 70.00 20 M4 53.00 2.0 M7 66.25 2.0 M7 65.25 2.0 M4 53.00 20 M9 70.03 20 M9 70.00 2.0 M9 70.00 2.0 M10 55.25

2.0 M10 55.25 2.0 M8 52.50 2.0 M2 4835 2.0 M2 4835 20 M10 55.25 2.0 M9 7000 2.0 M7 56.25 2.0 M10 56.25 2.0 M10 56.25 2.0 MS 7000

2.0 M1 6300 2.0M6 5335 2.0 MG 5335 2.0 M3 530D 2.0 M9 70.00 2.0 M10 55.25 2.0 M2 4835 2.0 MG 5335 2.0 M1 63.00 2.0 ME. 5335

20111 45.25 2.0111 46.25 2.0111 46.25 2.0111 46.25 2.0111 45.25 20111 45.25 2.0111 46.25 2.0111 46.25 2.0111 46.25 2.0 H1 45.25

3.0 E9 3900 3.0 El 90.03 3.0 Ell) 85.03 3.0 E6 73.00 3.0 El0 6500 3.0 E13 7200 3.0 El0 65.03 3.0 Ell MOO 3.0 E17 72.00 3.0 El5 7300

3.0 E17 7200 3.0 El0 65.03 3.0 Ell 85.00 3.0 E4 74.00 3.0 E4 7400 3.0 E14 6300 3.0 E7 63.03 3.0 E5 79.03 3.0 E3 84.00 3.0 E7 63.00

3.0 Ell 6.4.130 3.0 E14 63.00 3.0 Ell 34.00 3.0 E1.7 72.00 3.0 Ell 8400 3.0 E9 89.00 3.0 El7 72.00 3.0 E13 72.00 3.0 E12 65.00 3.0 ES 74.00

3.0 E8 72.00 3.0 E2 3300 3.0 El 90.00 3.0 E5 79.00 3.0 E8 7200 3.0 E3 84110 3.0 E3 84.00 3.0 EG 73.00 3.0 E10 85.00 30 E 8400

3.0 M3 5750 3.0 M1 57.50 3.0 M2 6100 3.0 M1 57.50 3.0 M2 61.00 3.0 M3 5750 3.0 M1 57.50 3.13 M2 6100 3.0 M3 57.50 3.0 M1 57.50

Difficulty 74.95 Difficulty 74.30 Difficulty 73.64 Difficulty 72.84 Difficulty 75.21 Difficulty 74.80 Difficulty 74.99 Difficulty 73.56 Difficulty 73.90 Difficulty 74.74

Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14

Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5

Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1

Total From Topic Total From Topk Total From Topk Total From Topic Total From Topic Total From Topic Total From Topk Total From Topk Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Copyright © 2020 by James R. Parry

30

Experiment #2B – Stratified Randomization – Real Client #1 Data Using the same data from experiment #2A, following the recommended test design, the assessment

design function of Questionmark OnDemand was instructed to select 20 items in a stratified ran-

dom fashion from the sub-topics, following the recommended design shown in table 19 (figure 8).

Figure 8 - Item Selection Criteria - Experiment #2B

Thirty (n=30) iterations of a 20-question test were generated as illustrated in tables 24, 25, and 26

Table 24 - Experiment #2B Item Selections - Attempts 1 - 10

Page 32: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt

OID

11

SCORE

Attempt

COD

12

SCORE

Attempt

COD

13

SCORE

Attempt

OID

14

SCORE

Attempt

ORD

15

SCORE

Attempt

OID

16

SCORE

Attempt

OM

17

SCORE

Attempt

111D

18

SCORE

Attempt

111D

19

SCORE

Attempt

1111)

20

SCORE

1.0 E7 7600 1.0 El 8000 1.0 E4 75.00 1.0 E7 78.00 1.0 E12 75.60 1.0 E14 79.00 1.0 15 94.00 1.0 15 94.00 1.0 E8 91.00 1.0 E9 8300

10 E2 75.00 10 ES 91.00 10 E9 83.00 1.0 E9 8300 1.0 E4 7500 1.0 11 80.00 1.0 E7 78.00 10113 75.00 1.0 E7 78.00 1.0 E8 91 00

1.0 Ell 78.00 1.0 E7 78.00 1.0 E3 77.00 1.0 Ell 78.09 1.0 E9 83.03 1.0 ES 91.00 1.0 E13 76.03 3.0 E4 76.00 1.0 E2 75.00 1.0 E5 4400

LO E10 77 00 LO ES 8900 10 ES 8900 10 E3 7700 10 E13 7500 10 E7 7800 10 E4 7600 10 E2 7500 10 Ell 7800 10 E2 7500

1.0 M3 59.00 1.0 M4 71.00 1.0 M2 57.00 1.0 M4 71.00 1.0 M3 69.00 1.0 M4 71.00 1.0 M1 63.00 1.0 M1 53.00 1.0 M3 59.00 1.0 M3 59 00

2.0 E12 90.00 2.0 (13 79.00 2.0 E5 74.00 2.0 E7 75.00 2.0 E8 8100 2.0 E3 75.013 2.0 E12 90.00 2.0 E22 76.011 2.0 ES 81.00 2.0 E10 83.00

20 E7 75 00 20 Ell 9000 20 El6 8301) 20 E15 8200 20 El2 9010 20 E14 9000 20 Eli 7900 20 EIS 82061 20 Eli 7900 20 Ell 82.50

20 Ell 7600 20 Ell 7800 20 Eli 7901) 20 E13 7900 20 Ell 6250 20(5 7400 20(5 7400 2007 7500 20 E21 7800 20 E20 8.000

2.0 (10 83.00 2.0 EIS 82.00 2.0 El 8300 2.0 E9 8900 2.0 E19 65.00 2.016 80.60 20120 60.00 2.0 E19 86.00 2.0 E19 85.00 2.0 (15 8200

20 (22 76.00 20 E2 92.00 20 (17 79.00 2.0 E5 74.00 2.0 E20 80.00 2.0 E14 8310 2.0 El0 63.00 2.0 E9 89.00 2.0 E4 7300 2.0 El 63.00

2.0 E15 82.00 2.0 (19 86.00 2.0 (14 9000 2.0 En 8300 2.0 (18 81.03 2.0 E22 7500 2.0 E19 86.00 2.0 E8 81.00 2.0 E10 83.00 2.0 E3 76.00

20 M5 68.01) 20 M10 5625 2O M5 6801] 20 M10 5325 2 0 M5 6300 2 0 M9 7000 20 M10 5625 20 M10 5625 20 M7 5625 20 M4 5300

2.0 M9 70.00 2.0 M8 52.50 2.0 M3 57.00 2.0 M4 53.00 2.0 M10 55.25 2.0 M4 53.00 2.0 M4 53.00 2.0 M6 53.75 2.0 M1 53.00 2.0 M2 48.75

2.0 M2 48.75 2.0 M5 53.75 2.0 M9 70.01] 2.01419 7000 2.0 M9 70I10 2.0 M2 4835 2.0 M9 70.00 2.0 M8 52.50 2.0 M1O 56.25 2.0 M1O 56.25

2 0111 46 25 20(11 4625 20111 4325 20(11 4325 20(11 4325 2081 45 25 2.081 4625 20 H1 4625 20111 4625 20111 46.25

3.0 E15 73.00 3.0 E2 87.00 3.0 [17 72.00 3.0 E4 74.00 3.0 E7 83.00 3.0 E3 84.00 3.0 E17 72.00 3.0 115 75.00 3.0 E6 7300 3.0 E13 7200

3.0 E14 83.00 3.0 E14 83.00 3.0 E13 72.00 3.0 E13 72.00 3.0 E17 7210 3.0 E14 63.60 3015 79.00 30116 90.00 30112 85.00 3.0 E4 74.00

3.0 E3 84.00 3.0 Eli 72.00 3.0 E14 83.00 3.0 E9 54.00 3.0036 7500 3.0 E2 87.00 3.0 E6 73.03 3.0013 72.00 3.0 Ell 84.00 3.0 E11 84 CC

30 E9 8300 30010 8500 3008 7200 30 EIS 7300 30 E9 6900 30 Eli 7200 30 E2 6700 30 E12 8.500 3002 8.700 30 E10 85.00

3.0 M3 57.50 3.0 M3 5750 3.0 M3 5730 3.0623 5730 3.0 M1 57.50 3.0 M3 57.50 3.0 M1 57.50 3.0 M2 51.00 3.0 M2 51.00 3.0 M1 57.50

Difficulty 74.03 Difficulty 75.46 Difficulty 74.39 Difficulty 73.00 Difficulty 74.83 Difficulty 73.98 Difficulty 73.65 Difficulty 73.24 Difficulty 74.74 Difficulty 73.76

Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14

Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5

Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Attempt

CND

21

SCORE

Attempt

CND

22

SCORE

Attempt

OID

24

SCORE

Attempt

OID

i 25

SCORE

Attempt

CND

26

SCORE

Attempt

OID

27

SCORE

Attempt

CND

28

SCORE

Attempt

01D

29

SCORE

Attempt

OID

30

SCORE

Attempt

OID

23

SCORE

1.0 (13 75.00 1.0 (10 77.00 1.0 (5 94.110 1.0 E4 76.00 1.0 E6 89.00 1.0 E5 94.00 1.0 E9 83.00 1.0 Ell 78.00 1.0 (14 79.00 1.0 E4 76.00

1.0 E3 77.00 1.0 E3 77.00 10113 76.00 1.0 E7 78.00 1.0 Ell 76.00 1.0 E12 75.00 1.0 (5 94.00 1.0 E8 911.10 1.0 (10 77110 1018 91.00

1.0 E2 75.00 1.0 E13 76.00 1.0 E3 77.00 1015 94.00 10 E9 83 DC 10 E3 7700 1.0 El 80.00 1.0 E9 83.00 1.0 Ell 78110 10114 79.00

1.004 75.00 1.0 E9 83110 1012 75.00 1.0 Ell 78.00 10013 76 CC 10 ES 9100 1.008 91.00 1.004 75.00 1.002 75110 1.0 E9 83.00

1.0M4 71.00 1.0M3 59.00 1.0 M1 53.03 1.0 M2 67.00 1.0M4 71 CC 1.0 M1 63.00 1.0M3 69.00 1.0 M1 63.00 10 M2 57.00 1.0 M2 57.00

2.0 E10 83.00 2.004 75.00 2014 7510 2.0 E22 76.00 2.0 El0 83 CC 2.08 76 CC 2.0 E21 78.00 2.0 E3 7500 2.0 E22 75.00 20119 86.00

2.0 E4 7500 2.0 E2 92.00 20113 79.00 20119 85.00 2.0 ES 81 CC 2.0015 83 CC 2.0 E22 75.00 2.0 E22 75.00 2.0 E14 9000 2014 75.00

2.002 92.00 2.005 74.00 20118 81.00 20113 79.00 2.0 [20 80 CC 2.0 E4 75 CC 2.0 ES 7400 2.0 E17 79.00 2.0 E19 85.00 2011 83.00

2.0 E10 83.00 2.007 75.00 2.0E15 82.00 2.0110 83.00 2.0 E6 80 CC 2.0 E14 90 CC 2.004 75.00 2.0 E16 63.00 2.018 81.00 2012 92.00

2D E5 74131) 2D E12 90.01) 2006 8.0.00 2D E7 7500 20 E12 90 CC 20 E10 83 CC 2D E13 79.00 2D E7 75.00 2.003 7600 20015 8200

20 E20 8000 20 E10 83.01) 20 ES 8.100 20012 9000 20 El 8300 20 El 8300 20 E15 6200 20 E9 8900 20 05 7400 20E14 9000

20M1 6300 20 M1 5300 20 M10 5625 20 M10 5625 20 M10 5625 20 M10 5625 20 M7 6525 20 M7 6525 20 M3 57010 20 M3 5700

2.0 M5 6800 2.0 MS 5800 20 M9 7000 20 M7 6625 20 M4 5300 20 M6 5375 2.0 M1 6301) 2.0102 4875 20102 4835 20 M5 5800

2.0M5 53.75 2.0M4 53.00 2.0 M3 57.00 2.0 M6 53.75 2.0 M6 53.75 2.0M4 53.00 2.0 M6 53.75 2.0 M1 63.130 2.0 M5 58.00 2.06)8 52.50

20(11 4625 20 H1 4625 20111 4625 20111 4625 20111 4625 20111 4625 2061 4525 20(11 4625 20(11 4625 20 H1 4625

3O E14 8300 3D E3 8.4.00 30 E9 8.9.00 30 E2 8200 30E13 7200 30E10 8500 30E17 7200 3O E13 7200 3D Ell 8400 30 E16 7500

30 EIS 7300 30 El 9000 30 E3 8400 30(16 7500 30E7 8300 30E13 7200 3D E6 7300 30E10 8500 30E13 7200 30E6 7300

3.0 E7 83.00 3.0 E5 73.00 3.0 (5 7910 3.0 E10 85.013 3.0 E17 72.00 3.0 El 90.00 3.0 E15 75.130 3.0 E7 83.00 3.0 (1 90.00 3.0 E9 89.00

3.0 E5 73.00 3.0 Ell 84.00 3.0 E13 72.00 3.0 E7 83.03 3.0 E5 79.00 3.0 E6 73.00 3.0 E2 87.130 3.0 (14 83.00 3.0 E2 87.00 3.0 E14 83.03

3_0 M3 5750 3_0 M2 5100 3_0 M3 5750 30 M1 5750 30 M3 5750 3_0 M1 5750 30 M3 5750 3_0102 6100 30 M3 5750 3_0 M1 5750

Difficulty 73.13 Difficulty 74.66 Difficulty 74.20 Difficulty 74.60 Difficulty 73.34 Difficulty 73.84 Difficulty 73.74 Difficulty 73.86 Difficulty 73.98 Difficulty 75.76

Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14 Easy 14

Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5 Moderate 5

Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1 Hard 1

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 5

Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10 Topic 2 10

Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5 Topic 3 5

Copyright © 2020 by James R. Parry

31

Table 25 - Experiment #2B Item Selections - Attempts 11 - 20

Table 26 - Experiment #2B Item Selections - Attempts 21 - 30

The item selection difficulty remained constant as stratified with each iteration (see ‘Difficulty’

columns in tables 24, 25, and 26). The target cut score/difficulty was 76.13. The stratified random-

ization produced a difficulty range between 73.00 and 75.76 with average (mean) of 74.11. The

standard deviation of the scores was 0.74 with a 95% confidence interval of 0.2635 which means

that the true population mean is between 73.85 and 74.37 of the 30 samples. The kurtosis of the

average difficulty is 0.117 and the skewness is 0.579. The number of items at each difficulty level

Page 33: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Sample Cifriculty Statistics

Target Cut Score 76.13

Mean difficulty 74.11

Median 73.98

Minimum 73.00

Maximum 75.76

Variance Target vs_ Mean 2.04

Standard Deviation all Averages 0.74

95% Confidence Score 0.263545877

Kurtosis 0.117166773

Skewness 0.579229905

Standard Distribution of Experiment #2 Samples

Stratified-Random Data

67 69 71 73 75 77 79 rict 33 SS

Average Test Difficulty

Copyright © 2020 by James R. Parry

32

from each topic varied with each iteration. Table 27 provides a summary of the statistics for the

sample. Figure 9 illustrates the standard distribution curve of the sample.

Table 27 - Difficulty Statistics for Experiment #2B

Figure 9 - Standard Distribution of Test Difficulty - Experiment #2B - Stratified Random Selection

The content (topic) coverage was equal as stratified for each iteration as illustrated by the three-

color (delineated by sections) display and ‘Total From Topic’ columns in tables 18A, 18B, and

18C.

Conclusion: All topics were covered equally as desired in both difficulty and content. Comparing

the distribution of test difficulty scores in figures 7 and 9 shows that the stratified randomization

consistently produced tests well within an acceptable range to meet the desired cut score of 76.13.

Although the simple randomization in experiment 2A produced tests with an average (mean) dif-

ficulty (75.87) closer to the desired difficulty (76.13) than the stratified randomization in experi-

ment 2B (74.11), the standard deviation of the results in experiment 2B (0.74) indicated signifi-

cantly less variance from attempt to attempt than the standard deviation produced by experiment

2A (2.17).

Page 34: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

CUT SCORE CALCULATION TOOL

Course/Certiikation Name:I FAIRNESS RESEARCH 3 I Test Name:I TEST NAME

Compass

Facilitator Name/Date: For it ator Name/Date Revision 1 Facilitator Name/Date Date: mmiddiyyyy Consnliants, LLC Ente Topic/TM/Subject ID: Topic 1 Revision 2 Facilitator t or Name/Date

Th, cprcad:heet tool is the intellc tual property of a d Copyright .2010 2019 by Compacc Con:ultant., LLC, U:e i: limited to the term:, the End er Licence Agr ement (EL,A). TM: copy i: ken, to: 30 DAY DEMO ONLY

Test Item CND En'r I. it New, R if Retired

Difficulty Metatag

Average PercentageExpert

ct Corre

lAcgoff Rating)

1 Nance

Expert 2 Name

Expert 3 Name

Expert 4 Name

Experts Name

Expert 6 Name

Expert? Name

Expert 8 Name

Expert 9 Name

Expert tO Name

Standard Deviation

Topic

Clit ga.00

Score

Moderate

Difficulty

1.0 M1 uederate 48.57 60 60 40 45 40 55 40 9.45 Approximate Difficulty Rating

1.0 M2 :Aerate 62.14 55 70 60 75 65 60 50 8.59 25-483 Hard

1.0.M.5 51.04 50 60 70 60 60 60 40 9.51 48.4- 71.; Moderate

R 50 65 60 65 60 60 40 71.8-45 Easy

1.0 M4 delete 62.86 70 70 70 60 50 60 50 7.56

10 M5 . 4 d erate 60.00 60 70 70 50 50 70 50 10013 Standard Deviation

1.0 El 77.14 70 85 80 75 70 93 70 8.09

1.0 E2 75.00 70 80 90 75 70 so 60 9.57 A standard deviation of more than

1.0 E3 77.14 70 80 90 75 75 so 70 6.99 10 will trigger an alert. Discuss the

1.0 E4 84.29 75 90 95 90 75 SO 75 8.86 outliers with the judges who set

1.0 M6 Moderate 62.14 55 70 50 50 70 70 70 9.94 them to determine why. Change as

1.0 H1 111111, 4143 30 50 35 35 50 50 40 8.52 necessary.

1.0 M7 Moderate 63.57 60 65 70 50 65 65 70 6.90

1.0 M8 Moderate 5500 55 60 65 50 50 65 4o 9.13 7 In this section 17%

R 70 80 75 go 70 90 40 28 Moderate In this section 67%

10 M9 Moderate 57.19 50 50 70 50 50 70 60 7 11111. In this section 17%

1.0 M10 Moderate 48.57 50 45 50 35 55 55 50 42 TOTAL 100%

1.0 M11 Moderate 57.86 50 70 50 53 65 60 60

CUT SCORE CALCULATION TOOL

Course/Certification Name:l FAIRNESS RESEARCH 3 I Test Name:I TEST NAME

Compass CCOnmsPultants, Facilitator Name/Date: Facil.ator Name/Date Revision 1 Facilitator Name/Date Date:Immiddhyyy LLC

Ente Topic/TPD/Subject ID: Topic 2 Revision 2 Facilitator Name/Date

TA c cpreadcheet tool is the intellectual property of a d Copyright 02010 2019 by Compacc Concultan, L., Uce , limited ° the term: of the End U.er Licence Agreement (EULP.).Thic copy is ken:, to: 30 DRY DEMO ONLY

Test Item DID Enter • If New, Rif Retired

. . DI:Moony Metatag

Average Percentage

Correct

lAngoll Rating).

Expert t Name

Expert 2 Name

Expert 3 Name

Expert 4 Name

Experts Name

Expert 6 Name

Expert 7 Name

E pert 8 Name

Expert 9 Name

Expert tO Name

Standard Deviation

Topic

. Cut 6000

Score

Moderate

Difficulty

2.0 El 75.00 70 80 75 75 80 85 60 8.16 Approximate Difficulty Rating

2.0 E2 72.86 65 80 75 65 80 85 60 9.51 25 - 48.3 Hard

2.0 M1 1=221 6000 50 65 70 50 50 65 70 9.57 48.4 - 71.; Moderate

2.0 ES 73.57 50 70 so 75 70 90 70 9.45 71.8 - 45 Easy

2.0 E4 77.14 70 80 85 75 80 90 60 9.44

2.0 ES 74.29 50 85 65 75 80 75 80 8.66 Standard !Deviation

2.0 M2 r=1 69.29 50 80 75 60 70 so 60 9.32

2.0 E6 73.57 50 70 so 65 75 85 80 9.00 A standard deviation of more than

2.0E7 80.00 70 80 85 85 80 90 70 7.54 10 will trigger an alert. Discuss the

2.0E8 81.43 70 85 95 70 80 90 80 9.45 outliers with the judges who set

20 MS MEI 7143 65 75 85 65 70 00 60 9.00 them to determine why. Change as

2.0 E9 77.14 70 80 85 65 80 90 70 9.05. necessary.

20 M4 = 67.14 65 75 80 50 75 75 50 12.1

2.0 E10 84.29 75 85 90 45 go 85 ao 673 24 in this section 46:;

2.0 Ell 82.14 70 85 90 45 go 85 70 9051 27 Moderate In this section 52:;

2.0 E12 74.29 65 70 85 85 75 70 70 7.87 1 1111,. In thissection 351

2.0 US 7786 70 80 85 60 75 85 70 6.36 52 TOTAL 10051

2.0 E14 75.00 65 65 80 80 75 80 80 7.C7

Copyright © 2020 by James R. Parry

33

Experiment #3A – Random Selection – Real Client #2 Data Using cut score (Parry Method) results from real client data (Tables 28, 29, and 30 – partial views),

the items were divided into three topics within the main topic and further subdivided into difficulty

sub-sub topics based upon the results of the cut score rating tool. The cut score/difficulty for the

entire database (134 items) was determined to be 64.37% by averaging all three topic cut scores.

Table 28 - Experiment #3A - Difficulty Calculations - Real Data - Topic 1

Table 29 - Experiment #3A - Difficulty Calculations - Real Data - Topic 2

Page 35: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

CUT SCORE CALCULATION TOOL

Course/Certification Namml FAIRNESS RESEARCH 3 I Test Name:1 TEST NAME

RVaSS

ix

Facilitator Name/Date: Far Hater Name/Date Revision 1 Facilitator Name/Date Datermmiddivyyy I

IICE

Enter Topk/TP0/Subject ID: Topic 3 Revision 2 Facilitator Name/Date

TM1 c cpreatl:M1eettool ie tM1e intellectual property of a d Copyright .2010 2019 by Compacc Concultant., LLC, Uce is limited to the terms of rM1e End Lf er Licence Pqr ement (EL,A) Thic copy ic licenced to: 30 DRY DEMO ONLY

Test kering/EP Enter • If New. R if Retired

Difficulty Metatag

Average

Percentage

Mgt rect

lAn Rating)

Expert 1 Name

Expert 2 Name

Expert 3 Name

Expert 4 Name

Expert 5 Name

Expert 6 Name

Expert 7 Name

Expert 8 Name

Expert 9 Name

Expert 10 Name

Standard Deviation

Topic

Cut 55.00

Score

Moderate

Diffilty.cu

50 El I 74.29 50 70 90 75 75 so 70 9.32 Approximate Difficulty Rating

30E2 77.14 70 70 90 75 75 SO 70 9.06 25-483 Hard

30 ES 77.14 70 70 90 75 75 90 70 9.06 48.4- 71; Moderate

3.0 E4 78.57 70 80 90 75 75 so 70 8.52 718-95 Easy

311 Ml Berate 67.14 60 65 70 65 70 80 60 6.99

30 M2 ir oderate 63.57 55 70 75 65 65 65 40 1127 Standard Deviation

32 M3 Moderate 61.43 65 65 75 50 50 65 50 9.00

30 ES 78.57 70 80 90 75 75 90 70 852 A standard deviation of more than

30E6 79.29 70 80 90 75 70 so SO 838 10 will trigger an alert. Discuss the

3.0 E7 73.57 65 70 85 70 70 85 70 802 outliers with the judges who set

30 ER 7357 70 75 90 70 75 75 50 9.00 them to determine why_ Change as

30 E9 77.14 75 75 90 65 75 so 70 9.51 necessary.

30 E10 7857 75 so 90 75 70 90 70 8.52

30 Ell 77.86 65 80 90 70 80 90 70 9.94 14 In this section 35%

30 M4 .derate 6286 55 65 70 50 55 65 70 7.56 24 Moderate In t his section 60%

3.0 M5 :Aerate 71.43 50 70 so 65 65 80 SO 8.52 2 IIIIIIIIIIIIIIII, In this section 5%

30 E12 77.86 70 so 90 75 75 so 55 9.51 40 TOTAL 100%

3.0 M6 Moderate 58.57 50 50 75 50 50 55 70 Mil

Topic

Topic Cut

Score g.

Difficulty

Items in

Topic

% of Total

Items

Avaiable

Hard

% From

Topic

Available

Mod

% From

Topic

Available

Easy

% From

Topic

Topic 1 58 42 31.34% 7 17% 28 67n 7 _7-

Topic 2 59 52 38.81% 1 2% 27 52% 24 =l

Topic 3 66 40 29.85% 2 5% 24 60% 14 E.5':

Total #

Needed

From

Topic

Use Hard

(Calculated)

Use Hard

(Actual)

Use Mod

(Calculated)

Use Mod

(Actual)

Use Easy

(Calculated)

Use Easy

(Actual). Topic

6.21 1.04 1 4.18 4 1.04 1 Topic 1

7.76 0.15 1 4.03 4 3.58 3 Topic 2

5.97 0.30 1 3.5-8 3 2.09 2 Topk 3

Copyright © 2020 by James R. Parry

34

Table 30 - Experiment #3A - Difficulty Calculations - Real Data - Topic 3

Topic 1 consisted of 42 test-items with a difficulty rating of 58 (Moderate). Topic 2 consisted of

52 test-items with a difficulty rating of 69 (Moderate). Topic 3 consisted of 40 test-items with a

difficulty rating of 66 (Moderate). Each topic had a mix of hard, moderate, and easy items. (Table

31)

Table 31 - Experiment #3 - Item Difficulty Distribution by Topic

Referring to the design philosophy of the Compass Consultants spreadsheet tool as to the number

of items drawn from each section, it appears that topic 2 was considered to be the ‘most’ important

with 38.81% of the available items, topic 1 was the next ‘most’ important with 31.34% and topic

3 was the least important with 29.85%.

The final test design to maintain fairness in both content and difficulty is shown in table 32.

Table 32 - Experiment #3 - Recommended Test Design

Page 36: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt

QID

1

SCORE

Attempt

4310

Experiment

2

SCORE

3A

Attempt

1311)

- Random

3

SCORE

Selection

Attempt

QID

of 20

4

SCORE

items from

Attempt

QID

all 3

5

SCORE

topics.

Attempt

Q1D

Real Client

6

SCORE

Data.

Attempt

COD

Desired target

7

SCORE

Attempt

QID

difficulty is

8

SCORE

64.37

Attempt

131D

9

SCORE

Atlen

QM

pt 10

SCORE

10 E7 7285 10 HA 40013 1 0 E5 72 86 1 0 E5 7286 10 El 77 14 10 117 42 14 10 E4 84 24 10 M11 57 86 10 E4 8429 10 E5 7286

10(1 41 43 10 M15 58 57 10 M10 4857 133 M1 4857 10 M16 5500 10 M12 5500 10 ES 7286 10 M12 65 00 10 M2 5214 10 E7 72 86

1 0 115 4786 10 M17 6071 10 Km 5785 10 M19 5000 10 M20 5929 10 M13 55 43 10111 41 43 10 MIS 50 71 10 M22 5143 10112 44 29

10 H5 3286 101A2 62 14 10 M13 56 43 10 M2 6214 20 E14 7500 10 M25 54 29 10 114 4060 10 M23 71 43 20 E16 7571 10115 47 86

1.0 M12 52.14 1.0 M3 57.14 1.0 M14 53.57 1.0 MS 60.01 2.0 615 7531 1.0 M4 52.85 1.0 M12 65.00 2.0 El 75.01 2.0 E19 75.013 1.0 M2 52.14

1.0 M13 56.43 2.0 613 7785 1.0 M28 55.71 10 M7 6387 2.0 617 77.14 1.0 M6 62.14 1.0 M13 56.43 2.0 Eli 77.86 2.0 Ell 73.57 1.0 M3 57.14

1.0 M15 55.00 2.0 E23 76.43 10 M5 62.14 2.0 El0 84.29 2.0 E24 72.14 2.0 613 77.85 1.0 M16 55.00 2.0 08 80.71 2.0 Ml 6000 2.0 610 84.29

1.0 M18 50.71 2.0 ES 74.29 2.0 E21 73.57 2.0 E21 73.57 2.0 E5 74.29 2.0 EIS 75.71 1.0 M18 50.71 2.0 El9 75.00 2.0 M14 6785 2.0 E14 75.00

2.0 E21 73.57 2.0 M12 60.71 2.0 E9 77.14 2.0 E22 74.29 2.0 ES 73.57 2.0 617 77.14 1.0 M21 5785 2.0 M1 60.00 2.0 M24 50.71 2.0 616 75.71

2.0E22 74.29 2.0 M13 64.29 2.0 M10 64.29 2.0E24 72.14 2.0E7 80.013 2.0E19 75.00 10 M4 6285 2.0 M10 64.29 2.0 M27 57.85 2.0E9 77.14

20E7 80.00 20 M15 6143 211M2 6929 10M10 6429 20111 4643 20 M10 6429 2001 73 57 211M2 6929 20 M3 7143 2 OH1 4643

20 hin 57 14 20 M17 67 85 20 M21 5929 20 M14 6786 20 M12 5071 20 M16 6429 20E6 73 57 20 M21 5929 29 M5 5071 20 M15 5143

20 M12 6071 20 M23 5786 29 M23 5786 20 M15 6143 233 M16 6429 20 M19 60 71 20E7 8000 2 OM22 5857 20 M5 7000 20 M26 57 14

20 M6 7000 20 M24 6071 20 M25 57 14 20M18 5929 30 E2 77 14 20 M25 5000 20E8 81 43 20 M25 6000 20 M8 5786 20 M27 57 86

3.0 Ell 77.85 2.0 M5 60.71 3.0 E7 73.57 2.0 M25 60.01 3.0 E3 77.14 2.0 M3 71.43 2.0 M14 5785 2.0 M5 60.71 3.0 [10 7857 3.0 E10 78.57

3.0 En 72.85 3.0 H2 46.43 3.0 68 73.57 3.0112 45.43 3.0 E4 78.57 3.0 E14 72.14 2.0 M21 59.29 2.0 M7 65.00 3.0 ES 7857 3.0 M1 57.14

3.0 M19 70.00 3.0 M22 57.85 3.0 M1 67.14 3.0 M12 55.00 3.0E9 77.14 3.0 0 73.57 3.00 77.14 2.0 M9 65.00 3.0 M17 50.00 3.0 M18 5000

3.0 M2 53.57 3.0 M24 54.29 3.0 M12 55.01 3.0 M13 60.00 3.0 M12 55.00 3.0 M15 49.29 3.0 0 73.57 3.0 Ell 77.85 3.0 M21 52.86 3.0 M21 62.86

30 M3 6143 30 MS 71 43 30 M15 49 29 30M14 6071 30 M3 5143 30 M22 5786 30 M13 5000 3 0 M13 6000 313 M22 5786 30 M5 71 43

30 MS 5857 3 0 M7 57 14 30 M15 68 57 30 M17 6000 30 M7 5714 30 M9 55 43 30 M8 55 00 30 MI7 6000 30 M9 55 43 30 M6 5857

Difficulty 6206 Difficulty 6139 Difficulty 62.64 Difficulty 62.82 Difficulty 69.21 Difficulty 63.93 Difficulty 6539 Difficulty 65.69 Difficulty 67.14 Difficulty 64.54

Easy 5 Easy 3 Easy 5 Easy 5 Easy 12 Easy 5 Easy 8 Easy 5 Easy 6 Easy 7

Moderate 11 Moderate 15 Moderate 15 Moderate 14 Moderate 7 Moderate 13 Moderate 10 Moderate 15 Moderate 14 Moderate 10

Hard 3 Hard 2 Hard 0 Hard 1 Hard 1 Hard 1 Hard 2 Hard 0 Hard 0 Hard 3

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 8 Topic 1 5 Topic 1 7 Topic 1 6 Topic 1 3 Topic 1 5 Topic 1 10 Topic 1 4 Topic 1 3 Topic 1 5

Topic 2 5 Topic 2 10 Topic 2 7 Topic 2 9 Topic 2 10 Topic 2 9 Topic 2 6 Topic 2 13 Topic 2 11 Topic 2 8.

Topic3 5 Topic3 5 Topic3 6 Topic3 5 Topic 3 7 Topic3 5 Topic3 4 Topic3 3 Topic3 6 Topic 3 5.

Attempt 11 Attempt 12 Attempt 13 Attempt 14 Attempt 15 Attempt 16 Attempt 17 Attempt IS Attempt 19 Attempt 20

OID SCORE OID SCORE OID SCORE OlD SCORE OID SCORE OID SCORE OID SCORE OID SCORE 01D SCORE OID SCORE

1.0 um 56.43 1.0 M10 48.57 1.0 M14 5387 1.0117 4214 1.0111 41.43 1.0 0 72.66 1.0 E4 84.29 1.0 E2 75.00 1.0 E7 7285 1.0 E9 84.29

1.0 M14 53.57 1.0 M11 97.86 1.06117 60.71 1.06111 57.88 10114 40.00 1.0 M14 53.57 1.061 41.43 1.0111 41.43 1.06115 58.57 1.0114 4000

1.0 M22 61.43 1.0 M15 65.013 1.0 M18 5031 1.06112 65330 1.085 47.86 1.0 M18 50.71 1.0115 3226 1.0 M10 48.57 1.06127 53.57 1.0612 52.14

1.0 M27 53.57 1.0 M18 50.71 1.0 M20 59.29 1.06119 50.03 1.0 M15 58.57 1.0 M18 48.57 1.0 M12 65.00 1.0 M15 58.57 1.0 M7 6857 1.0 M22 51.43

1.0619 57.14 1.0 M20 59.29 1.0 M21 5785 1.06126 5235 1.0 M2 52.14 1.0 M27 53.57 1.0 M17 60.71 1.014119 50.60 1.0 M9 57.14 1.0 M25 54.29

2.0 610 84.29 1.0 M21 57.85 1.0 M22 5143 1.0 M27 5387 1.0 M3 57.14 2.0 615 75.71 1.0 M25 54.29 1.0 M2 62.14 2.0 Ell 82.14 1.0 MS 50.00

2.0 E2 72.86 2.0 E13 77.86 1.0 MS 60.00 1.0 6127 62.14 2.0 613 77.36 2.0 E17 77.14 1.0 M6 62.14 1.0 M6 62.14 2.0 614 7500 1.0 MS 55.00

29611 60 00 20 [15 7571 20 E14 7500 10 MS 6214 10 El 7286 20 E23 75 43 20 El 75 on 10 M9 57 14 20 618 8071 10 M9 57 14

2.0 M13 64.29 2.0 E23 75.43 2.0 E6 7387 2.0 610 84.29 2.0 E22 74.29 2.0 E4 77.14 2.0 E23 75.43 2.0 E22 74.29 2.0 E19 75.00 2.0 03 77.86

20 M15 61 43 20 E4 7714 20 E9 7714 10 612 7429 20 MIS 5929 20 ES 73 57 20 E9 77 14 20 ES 7429 2014111 5714 10 ES 7357

2.0 M27 57.86 2.0 E9 77.14 2.0 M14 5785 2.0E4 77.14 2.0 M18 50.00 2.0 M12 60.71 2.0 M12 60.71 2.06112 60.71 206114 5785 2.0 E8 81.43

2.0 M5 70.00 2.0 M23 57.85 2.0 M26 57.14 2.0 M8 57.85 2.0 M2 59.29 2.0 M15 61.43 2.0 M15 61.43 2.06121 59.29 2.0 M23 57.85 2.0 M13 54.29

3.0 Ell 77.86 2.0 M4 67.14 2.0 MS 6031 3.0 612 7785 2.0 M20 57.14 2.0 M20 67.14 2.0 M25 57.14 2.01022 58.57 2.0 M26 57.14 2.0 M14 57.86

30 E3 7714 30E11 7785 20 M7 5500 30E3 7714 20 M23 5786 20 M24 6071 20 M3 7143 20 M24 6071 29 MS 50_71 20 M26 57 14

30 Ea 73 57 30 Ell 7429 30 E10 7857 30111 4500 29 M27 57 86 20 M25 57 14 20 M7 6500 20 MS 6071 20 M7 65330 20 MS 5786

3.0 62 46.43 3.0 E4 78.57 3.0 E6 79.29 3.014412 5500 2.0 M3 71.43 3.0 E2 77.14 3.0 Ell 7786 2.0 M7 65.00 3.0 El 7429 3.0 M14 80.71

3.0 M1 67.14 3.0 ES 79.29 3.0 E7 7387 3.014415 49.29 2.0 M5 50.71 3.0 E7 73.57 3.0 E5 78.57 3.0 E7 73.57 3.0 E4 7887 3.0 M14 57.14

3 °M16 6897 30M14 6071 30E9 7714 30 6118 5000 30 M13 5000 30111 4501 30M14 6071 3.0E8 73 57 30E6 7929 30 M21 62.86

3.0 M20 60.00 3.0 M17 60.00 3.0 M12 5800 3.0 M20 6000 3.0 M24 54.29 3.0 M17 60.00 3.0 M15 68.57 3.0 M20 60.00 3.0 H2 46.43 3.0 M22 57.86

3.0 M24 54.29 3.0 M5 71.43 3.0 M13 6100 3.0 M9 5643 3.0 MG 58.57 3.0 M2 63.57 3.0 M18 60.00 3.0 M24 54.29 3.0 M4 62.85 3.0 M3 51.43

Difficulty 63.89 Difficulty 6784 Difficulty 65.18 Difficulty 61.50 Difficulty 6043 Difficulty 64.25 Difficulty 6454 Difficulty 6180 Difficulty 66.29 Difficulty 53.72

Easy 5 Easy 9 Easy 7 Easy 5 Easy 3 Easy 8 Easy 5 Easy 5 Easy 8 Easy 4

Moderate 14 Moderate 11 Moderate 13 Moderate 13 Moderate 14 Moderate 11 Moderate 12 Moderate 14 Moderate 11 Moderate 15

Hard 1 Hard 0 Hard 0 Hard 2 Hard 3 Hard 1 Hard 2 Hard 1 Hard I Hard 1

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total Fr rri Topic

Topic 1 5 Topic 1 5 Topic 1 7 Topic 1 8 Topic 1 G Topic 1 5 Topic 1 7 Topic 1 8 Topic 1 5 Topic 1 8

Topic 2 7 Topic 2 7 Topic 2 7 Topic 2 4 Topic 2 11 Topic 2 10 Topic 2 8 Topic 2 8 Topic 2 10 Topic 2 7

Topic3 8 Topic3 7 Topic3 6 Topic3 S Topic3 3 Topic3 5 Topic3 5 Topic3 4 Topic3 5 Topic3 5

Copyright © 2020 by James R. Parry

35

Ignoring the recommendation of item distribution to maintain the cut score, the assessment design

function of Questionmark OnDemand was instructed to select 20 items at random from the single

topic containing all of the test-items. Tables 33, 34, and 35 present the results.

Table 33 - Experiment #3A - Item Selection - Attempts 1 - 10

Table 34 - Experiment #3A - Item Selection - Attempts 11 - 20

Page 37: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt 21 Attempt 22 Attempt 23 Attempt 24 Attempt 25 Attempt 26 Attempt 27 Attempt 28 Attempt 29 Attempt 30

DID SCORE illID SCORE CtID SCORE illID SCORE CtID SCORE CilD SCORE CilD SCORE CilD SCORE illID SCORE CtID SCORE

1.0 H7 42.14 10E5 47.86 1.0 E5 8285 1.0 M1 48.57 1.0 M14 5337 10 M10 48.57 1.0 E7 72.85 1.0 El 77.14 111E1 77.14 1.0 M15 58.57

1 0 M 11 5785 10117 42.14 1.0 M12 65.00 10 M17 60.71 1.0 M1.8 5031 1.0 M17 60.71 1.063 4786 LO M12 65.00 1.003 47.86 1.0 M16 6510

1.0 M19 50.00 1.0 kin 5386 1.0 M14 53.57 10 M25 54.29 1.0 M20 59.29 1.0 M27 53.57 1.064 40.00 LO M17 60.71 10 05 47.86 1.0 M27 5357

1.0 M27 5357 10 M14 5357 10 M16 6510 101027 5357 20 Ell 8214 10 ME. 5500 10 M18 5071 10 M22 6143 1065 3286 10 M3 5714

1.0 M26 55.71 1.0 M37 60.71 1.0 M17 6031 2.0 E12 74.29 2.0 Eli 7531 2.0 El 75.00 1.0 M22 51.43 1.0 M26 52.85 1.0 M18 50.71 2.0 E16 7531

1.0 ME. 55.00 1.0 M18. 50.71 1.0 M22 5113 2.0 114 75.00 2.0 En 7531 21 [12 74.29 1.0 M4 52.85 1.0 M3 57.14 1.0 M28 55.71 2.0 E2 7285

2.0 El 7510 1.0 M20 59.29 10 M3 57.14 2.0 Eli 75.71 2.0 E2 72.85 2.0 E13 77.86 1.0 M9 57.14 1.0 M7 63.57 1.0 M3 57.14 2.0 M10 64.29

2.0130 84.29 1.0 M25 52.86 2.0 El 75.00 2.0 E22 74.29 2.0 E21 73.57 2.0 E15 75.71 2.0 E19 75.00 2.0 El 75.10 2.0 E21 73.57 2.0 Mll 57.14

2.0 E23 7643 20 E16 7571 20 E13 7785 2063 7357 20 E4 7714 20 E23 7543 20 E20 7571 20 EIS 7571 20 M1 50.00 20 M13 64_29

2.0 E6 73.57 2.0 E4 77.14 2.0 E5 7357 2.0 ES 74.29 2.0 M15 5143 2.0 E24 72.14 2.0 E22 74.29 2.0 E7 60.00 2.0 Mll 57.14 2.0 M15 6113

2.0 E8 81.43 2.0 M25 50.00 2.0 E8 8113 2.0 E8 81.43 2.0 M17 5785 2.0 E7 80.00 2.0 E5 74.29 2.0 M19 60.71 2.0 M15 51.43 2.0 M17 5785

2.0 M1 60.00 2.0 M7 55.00 3.0113 7285 2.0 M10 64.29 2.0 M21 59.29 2.0 E9 77.14 2.0 M1 50.00 21 M23 57.85 2.0 M15 54.29 2.0 M18 59.29

2.0 M21 59.29 2.0 M9 65.00 3.0 E5 79.29 2.0 M18 59.29 2.0 M22 5837 2.0 M10 64.29 2.0 M23 57.86 2.0 M26 57.14 2.0 M25 50.00 3.0 [10 78.57

2.0 M23 57.85 3.0 E5 78.57 3.0 M1 6714 2.0 M25 60.00 2.0 M23 57.85 2.0 Mll 57.14 3.0 [12 77E6 2.0 M7 65.00 2.0 M27 57.86 3.0 [14 7214

2.0 M24 60.71 3.0 MIS 57.86 3.0 M10 59.29 2.0 M9 65.00 2.0 M24 5031 2.0 M13 64.29 3.0 E3 77.14 3.0112 77.85 2.0 M3 714.3 31 E5 78.57

2.0 M3 7143 3.0 M12 55.00 3.0 Mll 57.85 3.0113 72.86 2.0 M5 5031 3.0 El0 78.57 3.0 M13 50.011 3.0 E7 73.57 2.0 M4 57.14 3.0 M1 5714

3.0 M19 7010 3.0 M13 60.00 3.0 M20 60110 3.013 77.14 2.0 M5 70.00 3.061 45.00 3.0 M18 50.011 3.0 M11 5785 2.0 M7 5300 3.0 M14 6031

3.0 M21 6285 30 M18 6000 30 M21 6285 30 M13 6000 30 El 7129 30 M13 6000 30 M2 5357 30 M17 60.00 30 E2 7714 30 M19 7010

3.0 M7 57.14 3.0 M2 63.57 3.0 M23 67.85 3.0 M19 70.00 3.0 Ell 77.85 3.0 M19 70.00 3.0 M20 50.00 3.0 M20 60.00 3.061 45.00 3.0 M22 57.85

3.0 M8 65.00 3.0 M20 60.00 3.0 M7 57.14 3.0 M20 60.00 3.0E3 77.14 3.0 M21 62.85 31 M5 58.57 3.0 M6 58.57 31 M8 55.00 3.0 M4 5225

Difficulty 63.45 Difficulty 60.14 Difficulty 66.89 Difficulty 6632 Difficulty 6732 Difficulty 65.43 Difficulty 6336 Difficulty 64.85 Difficulty 59.71 Difficulty 65.25

Easy 5 Easy 3 Easy 7 Easy 9 Easy 9 Easy 9 Easy 7 Easy 5 Easy 3 Easy 5

Moderate 14 Moderate 15 Moderate 13 Moderate 11 Moderate 11 Moderate 10 Moderate 11 Moderate 14 Moderate 13 Moderate 15

Hard 1 Hard 2 Hard 0 Hard 0 Hard 0 Hard 1 Hard 2 Hard 0 Hard 4 Hard 0

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 6 Topic 1 8 Topic 1 7 Topic 1 4 Topic 1 3 Topic 1 4 Topic 1 7 Topic 1 7 Topic 1 7 Topic 1 4

Topic 2 10 Topic 2 5 Topic 2 4 Topic 2 11 Topic 2 14 Topic 2 11 Topic 2 6 Topic 2 7 Topic 2 10 Topic 2 8

Topic 3 4 Topic 3 7 Topic 3 9 Topic 3 5 Topic 3 3 Topic 3 5 Topic 3 7 Topic 3 6 Topic 3 3 Topic 3 8

Sample Difficulty Statistics

Target Cut Score 6437

Mean difficulty 64.15

Median 64.41

Minimum 59.71

Maximum 67.54

Variance Target vs. Mean 0.02

Standard Deviation all Averages 2.44

95% Confidence Score D.S737S4614

Kurtosis -0 858833055

Skewness -0.403006617

Copyright © 2020 by James R. Parry

36

Table 35 - Experiment #3A - Item Selection - Attempts 21 - 30

The item selection difficulty varied widely with each iteration (see ‘Difficulty’ columns in tables

33, 34, and 35). The target cut score/difficulty was 64.37. The randomization produced a difficulty

range between 59.71 and 67.54 with average of 64.15. The standard deviation of the scores was

2.44 with a 95% confidence interval of 0.878 which means that the true population mean is be-

tween 63.27 and 65.03 of the 30 samples. The kurtosis of the average difficulty is -0.859 and the

skewness is -0.403. The number of items at each difficulty level from each topic varied with each

iteration. Table 36 provides a summary of the statistics for the sample. Figure 10 illustrates the

standard distribution curve of the sample.

Table 36 - Difficulty Statistics for Experiment #3A

Page 38: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Standard Distribution of Experiment #3 Samples Random Data

50 55 GO 65 70

Average Test Difficulty

75 30

Cluestion selections

1 random question (s) from topic 'FAIRNESS RESEARCH 311.0 TOPIC 111.0 HARD' excluding subtopics (Avoid previously delivered)

4 random question(s) from topic 'FAIRNESS RESEARCH 311.0 TOPIC 111.0 MODERATE' excluding subtopics (Avoid previously delivered)

1 random question(s) from topic 'FAIRNESS RESEARCH 311.0 TOPIC 111.0 EASY' excluding subtopics (Avoid previously delivered)

1 random question(s) from topic 'FAIRNESS RESEARCH 312.0 TOPIC 212.0 HARD' excluding subtopics (Avoid previously delivered}

4 random question (s} from topic 'FAIRNESS RESEARCH 312.0 TOPIC 212.0 MODERATE' excluding subtopics (Avoid previously delivered)

3 random question(s) from topic 'FAIRNESS RESEARCH 312.0 TOPIC 212.0 EASY' excluding subtopics (Avail previously delivered}

1 random question Cs) from topic 'FAIRNESS RESEARCH 313.0 TOPIC 3/3.0 HARD' excluding subtopics (Avoid previously delivered)

a random question (s} from topic 'FAIRNESS RESEARCH 313.0 TOPIC 313.0 MODERATE' excluding subtopics (Avoid previously delivered)

2 random question(s) from topic 'FAIRNESS RESEARCH 313.0 TOPIC 3/3.13 EASY' excluding subtopics (Avoid previously delivered)

Copyright © 2020 by James R. Parry

37

Figure 10 - Standard Distribution of Test Difficulty - Experiment #3A - Random Selection

The content (topic) coverage was erratic as illustrated by the three-color (delineated by sections)

display and ‘Total From Topic’ columns in tables 33, 34, and 35.

Conclusion: All topics were not covered equally in either difficulty or content. Although the av-

erage (mean) difficulty of the sample is within an acceptable range of the desired difficulty/cut

score the standard deviation indicates a significant variation among attempts.

Experiment #3B – Stratified Randomization – Real Client #2 Data Using the same data from experiment #3A, following the recommended test design, the assessment

design function of Questionmark OnDemand was instructed to select 20 items in a stratified ran-

dom fashion from the sub-topics, following the recommended design shown in table 32 (figure

11).

Figure 11 - Item Selection Criteria - Experiment #3B

Page 39: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Attempt

QID

1

SCORE

Attempt

1/10

Experiment

2

SCORE

3B - Directed

Attempt

QID

3

SCORE

Random

Attempt

QID

Selection

4

SCORE

of 20 items

1/10

Attempts

from

SCORE

all 3 topics.

Attempt

QID

Real

6

SCORE

Client Data

QID

Attempt?

Desired

SCORE

target

Attempt

QID

difficu

8

SCORE

ty is 64.37

Attempt

QID

9

SCORE

Attempt

QID

10

SCORE

1.0 114 40.00 1.0113 47.86 1.0 15 32.85 1.0 H2 44.29 1.0113 47.86 1.0 H7 4214 1.0 H3 47.86 1.0117 42.14 10 H2 44.29 1.0 H2 44.29

1.0 M7 63.57 1.0 M28 55.71 10 ivas 50.71 1.0 M25 54.29 1.0 M15 65.00 1.0 MS 55.00 1.0 M5 62.14 1.0 M27 53.57 1.0 M20 59.29 10 M5 62.14

1.0 M20 59.29 1.0 pna 57.86 10 M23 71.43 1.0 M28 55.71 1.0 M10 48.57 10 M26 5285 1.0 M3 57.14 1.0 M10 48.57 1.0 M14 53.57 1.0 M17 60.71

19 M28 5571 1O M13 5643 1 0 M1 4857 10 M14 5357 1O M25 5286 10 M13 5643 10 M4 5286 1O M14 53.57 1O M8 5500 10 M27 53.57

1.0 M22 61.43 1.0 M14 53.57 1.0 M3 57.14 1.0 M17 60.71 1.0 M20 59,86 1.0 M4 6235 1.0 M20 59.29 1.0 M13 55.43 10 M23 7143 1.0 M21 57.86

10 ES 72.86 1.0 E2 75.013 1.012 75110 10 El 77.14 1.0 E7 72,86 1.0 E7 72.85 10 65 72.86 1.0 E7 72.86 1.016 8285 10 E5 82.86

2 0 111 4643 20111 4643 29(11 4643 20111 4643 20111 4643 29(11 4643 20111 4643 20111 4643 29111 4643 20E11 4643

2.0 M8 67.86 2.0 M12 60.71 2.0 M18 59.29 2.0 M26 57.14 2.O M11 57.14 2.0 M1 6010 2.0 M24 60.71 2.0 M25 60.00 2.0 M17 67.85 2.0 M2 59.24

2.0 M21 59.29 2.0 M21 59.29 29 M13 64.29 2.0M4 67.14 2.0 M12 60.71 2.0 M5 6031 2.0M2 59.29 2.0 M17 67.86 2.0 M27 57,85 2.0 M10 64.29

2.0 M26 57.14 2.0 M25 57.14 2.0 M19 60.71 2.0 M15 61.43 2.0 M25 60.00 2.0 M20 67.14 2.0 M14 67.86 2.0 M8 67.86 2.0 M1 60.00 2.0 M12 60.71

2.0 M25 50.00 2.0 M18 59.29 2.0M9 6510 2.0M9 65.00 2.0 M24 60.71 2.0 M25 6010 2.0 M15 51.43 2.0 M4 67.14 2.0 M10 64.29 2.0 M26 57.14

2.0 E20 75.71 2.0 E15 75.71 2.0 E3 73.57 2.0 E23 76.43 2.0 E3 73.57 2.0 ES 81.43 2.0 619 75.013 2.0 E24 72.14 2.0E20 75.71 2.0 65 73.57

2.0 E10 84.29 2.0 E16 75.71 2.0 Ell 82.14 2.0615 75.71 2.0 119 75.00 2.0 E21 73.57 2.0 65 74.29 2.0 E17 77.14 2.0 Ell 82.14 2.0E4 77.14

2.0 ES 74.24 2.0 E9 77.14 2.0113 77.85 2.0 E2 72.86 2.0114 75.00 2.0 El 74.29 2.0 613 77.86 2.0E16 75.71 2.0 E22 74.29 2.0 E9 77.14

30112 46 43 30111 4500 30 112 4643 30E11 4500 30112 4643 30 H2 4643 30E11 4500 30111 4500 30112 4643 30E11 4500

3.0 M23 67.86 3.0 M24 54.29 3.0 M16 68.57 3.0 M10 54.29 3.0 M7 57.14 3.0 M4 6285 3.0 M5 58.57 3.0 M19 70.00 3.0 M22 57.85 3.0 M13 60.00

30 M2 6357 3O M18 6000 3.0M13 6010 30 M14 60.71 3O M12 5500 30 M17 6000 3O M16 68.57 3O M14 6071 3O M7 5714 3O M23 6786

3.0 MG 58.57 3.0 M21 62.86 3.0 M21 6285 3.0 M1 67.14 3.0 M21 62.86 3.0 M8 6510 3.0 M8 55.00 3.0 M5 71.43 3.0 M17 60.00 3.0 M9 56.43

3.0 E7 73.57 3.0 E7 73.57 3.0 E9 77.14 3.0 E2 77.14 3.0 E3 77.14 3.0 E6 79.29 3.0 E3 77.14 3.0 E4 78.57 3.0 E14 7214 3.0 15 79.29

3.012 77.14 3.016 79.29 3.0 E4 70.57 3.0 112 77.86 3.0 E2 77.14 3.014 78.57 3.0 E7 73.57 3.0 Ell 77.86 3.016 79.29 3.0110 78.57

Difficulty 63.25 Difficulty 61.64 Difficulty 62.93 Difficulty 62.75 Difficulty 61.56 Difficulty 62.89 Difficulty 64.14 Difficulty 63.25 Difficulty 6339 Difficulty 63.71

Easy 6 Easy Easy 6 Easy 6 Easy 6 6 6 Easy 6 Easy 6 Easy 6 Easy Easy 6

Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11

Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 6 Topic 1 5 Topic 1 6 Topic 1 6 Topic 1 5 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6

Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 6 Topic 2 8 Topic 2 8

Topic3 6 Topic 3 5 Topic 3 6 Topic3 6 Topic 3 5 Topic 3 6 Topic 3 6 Topic 3 5 Topic3 6 Topic 3 6

Attempt

QID

11

SCORE

Attempt

0413

12

SCORE

Attempt

QID

13

SCORE

Attempt

(HD

14

SCORE

Attempt

(HD

15

SCORE

Attempt

(HD

16

SCORE

Attempt

CND

17

SCORE

Attempt

CUD

18

SCORE

Attempt

CND

19

SCORE

Attempt

CID

20

SCORE

10 116 3286 10 814 40 00 10 65 47 86 10115 47 86 10115 47 86 10111 41 43 10117 42 14 10 113 4786 10 61 4143 10 113 47.86

1.0 M17 60.71 1.0 M5 60.00 1.0 M14 53.57 1.0 M20 54.24 1.0 M4 62.86 1.0 M1.1 57.86 1.0 M21 57364 1.0 M20 59.29 1.0 M12 65.181 1.0 M10 48.57

10 M19 50 013 10 kin 57 846 10 M17 60 71 10 M26 52 86 10 M8 55 00 1O M20 5929 1O M22 6143 10 M22 6143 10 M14 5357 10 M28 5571

1.0 M1 48.57 1.0 M19 50.00 1.0 M7 63.57 10 MG 62.14 10 M5 60.00 1.0 M16 65.00 1.0 M9 57.14 1.0 M18 50.71 1.0 M25 54.29 10 M20 59.29

10 M23 71 43 10 M24 56 43 10 M23 71 43 10 M24 5643 10 M20 5929 10 M17 5071 1 0 Mll 5786 10 M4 6286 10 M20 5929 10 M5 6000

10 E3 77 14 10 15 72 86 10 E5 72 86 10 11 77 14 10 E4 84 29 10 E5 82 86 10 15 7286 10 12 7510 10 E7 7286 10 E7 7285

2.0 H1 46.43 2.0111 46.43 2.0111 46.43 2.0111 46.43 2.061 46.43 2.061 46.43 2.061 46.43 2.061 46.43 2.061 45.43 2.061 45.43

20 M17 67 86 20 M2 69 29 2O M24 6071 2O M25 57 14 20 M3 71 43 20 M27 57 86 20 M24 50_71 20 M2 6329 20 M3 7143 20 M25 6000

2.0 M10 64.29 2.0 M12 60.71 2.0 M12 60.71 2.0 Mb 70.00 2.0 M24 6021 2.0 M12 60.71 2.0 M20 6714 2.0 M23 57.85 2.0 M6 70.00 2.0 M16 64.29

20 M6 70 013 20 M10 64 29 2 0 Mll 57 14 2O M10 6429 20 M17 5736 20 M4 6714 20 M22 5857 20 M18 5929 20 M5 6071 2O M1 6000

2.0 M26 57.14 2.0 M16 64.29 2.0 M4 67.14 2.0 M12 60.71 2.0 M12 60.71 2.0 Mb 70.0141 2.0 M9 65,112 2.0 M6 7010 2.0 M26 57.14 2.0 M22 50.57

2O El 75 013 2 0 120 7571 20 E5 74 29 20 E5 74 29 20 E4 77 14 20 E5 7429 20 E24 7214 20 E21 7357 20 E4 7714 20 E8 8143

2.0 619 75.00 2.0 615 75.71 2.0 E21 73.57 2.0 Ell 82.14 2.0 E5 74.29 2.0 E22 74.29 2.0 E21 73.57 2.0 614 75.161 2.0 Ell 82.14 2.0 E7 80.00

2.0 E13 77.86 2.0 E23 76.43 2.0 E19 75.00 2.0 E3 73.57 2.0 E5 73.57 2.0 E16 75.71 2.0 E7 80.00 2.0 E9 77.34 2.0 612 74.29 2.0 121 73.57

30 H2 46 43 30111 45 00 30112 46 43 30112 4643 3 0 112 4643 3062 4643 3061 4500 3061 45110 3061 4510 3061 4500

3.0 M8 65.00 3.0 M8 65.00 3.0 M21 52.86 3.0 M2 63.57 3.0 M14 60.71 3.0 M22 57.86 3.0 M12 55.00 3.0M16 68.57 3.0 M13 60.00 3.0 M1 67.14

3.0 M5 71.43 3.0 M21 62.66 3.0 M8 65.013 3.0 M9 56.43 3.0 M4 5236 3.0 M15 49.29 3.0 M5 71.43 3.0 M4 62.85 3.0 M14 60.71 3.0 M14 60.71

3.0 M14 5371 3.0 M19 70.00 3.0 M3 5113 3.0 M12 55.00 3.0 M23 5736 3.0 M4 52.85 3.0 M15 49.29 3.O M11 57.85 3.0 M 11 57.85 3.0 M17 60.00

3.0 E6 79.29 3.0 E14 72.14 3.0 E13 72.86 3.0 E9 77.14 3.0 E4 78.57 3.0 E2 77.14 3.0 65 79.29 3.0 EIO 78.57 3.0 E3 77.14 3.0 El 74.29

3 0E14 7214 3O Ell 77 86 3018 73 57 3 0 [13 72 86 30 E7 7357 3015 7857 3013 7714 30 19 7714 30 El 7429 30 E6 7929

Difficulty 63.46 Difficulty 63.14 Difficulty 63.36 Difficulty 62.79 Difficulty 64.57 Difficulty 63.29 Difficulty 62.50 Difficulty 63.79 Difficulty 63.04 Difficulty 62.75

Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6

Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11

Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 8

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 5 Topic 1 5 Topic 1 5 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6

Topic 2 S. Topic 2 8 Topic 2 0 Topic 2 0 Topic 2 0 Topic 2 S. Topic 2 S. Topic 2 S. Topic 2 S. Topic 2 5

Topic3 6 Topic3 5 Topic3 5 Topic3 6 Topic3 6 Topic3 6 Topic3 6 Topic 3 6 Topic3 6 Topic 3 6

Copyright © 2020 by James R. Parry

38

Thirty (n=30) iterations of a 20-question test were generated as illustrated in tables 37, 38, and 39.

Table 37 - Experiment #3B Item Selections - Attempts 1 - 10

Table 38 - Experiment #3B Item Selections - Attempts 11 - 20

Page 40: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

I W

Attempt 21 Attempt 22 Attempt 23 Attempt 24 Attempt 25 Attempt 26 Attempt 27 Attempt 23 Attempt 29 Attempt 30

QID SCORE QID SCORE 8/10 SCORE 8/10 SCORE QID SCORE QID SCORE QID SCORE QID SCORE QID SCORE 8/10 SCORE

1063 4786 1065 3286 10 H4 4000 10 H1 4143 10117 42 14 10 65 4786 1003 4786 1067 42 14 10 64 4000 10 H5 4786

10M15 58 57 10 M4 6286 10 M24 5643 10 M5 6000 10 M1 48 57 10 M21 5786 10 M19 5000 10 M27 5357 10 M3 5714 10 M15 6500

10 M12 6500 10 M2 6214 10 M8 5300 10 M12 5500 10 M24 56 43 10 M27 53 57 10 M21 5786 10 MI1 5386 10 MS 5500 10 M20 5929

10 M17 6071 10 M22 6143 10 M18 5071 10 M15 5857 10 M3 57 14 111 Mfi 60 00 10 Mfi 62 14 10 M5 6000 10 M12 6500 10 M1 4857

1.0 M2 62.14 1.0 M19 50.00 1.0M2 6214 1.0 M19 50.00 10 M9 57.14 1.0 M15 58.57 1.0 M23 71.43 1.0 M24 56.43 1.0 M19 50.00 1.0 M27 5337

1 0 E4 84 29 10 El 77 14 10 E7 72 86 10 E5 8.2 86 10 E5 8285 10 E7 72 85 10 E4 84 29 10 E2 7500 10 El 75 00 10 E7 7286

2061 4643 20E1 4543 20E11 4543 20E11 4343 20 01 4643 20E11 46 43 20E11 4643 20E1 4643 20E1 4543 20E11 4543

20 MS 6500 20 M16 6429 2 0 M9 6500 2 0 M1 6000 2 0 M25 5000 2 0 M27 57 85 20 M17 6386 20 M5 6071 20 M22 5857 20 M10 6429

2.0 M27 57.86 2.0 M27 57.86 2.0 M20 6714 2.0 MS 67.86 2.0 M5 60.71 2.0 M14 67.86 2.0 M22 58.57 2.0 M24 60.71 2.0 M1 60.00 2.0 M25 6080

2.0 M1 60.00 2.0 M14 67.86 2.0 M2 69.29 20M11 5714 2.0 M15 64.29 2.0 M25 60.00 2.0M5 60.71 2.0M7 65.00 2.0 M14 6736 2.0 M23 5386

2.0 M6 70.00 2.0 M13 64.29 2.0 M5 60.71 2.0 M19 60.71 2.0 M15 51.43 2.0 M7 65.00 2.0 M24 60.71 2.0 M10 64.29 2.0 M5 60.71 2.0 M22 58.57

2.0 E10 84.29 2.0 E1.5 75.71 2.0 E23 7343 2.0 E9 7714 2.0 El 81.43 2.0 E8 81.43 2.0 E21 73.57 2.0 E19 75.00 2.0 Ell 8214 2.0 El5 7531

2.0 E3 73.57 2.0 El 75.00 2.0 E14 75.00 2.0 E2 7286 2.0 E14 7500 2.0 E19 75.00 2.0 E20 75.71 2.0 E24 72.14 2.0 E9 7714 2.0 E9 77.14

2.0 ES 81.43 2.0 E24 72.14 2.0 Ell 7423 2.0 El3 77.86 2.0 E15 75.71 2.0 E5 74.29 2.0 E10 84.29 2.0 E21 73.57 2.0 ES 81.43 2.0 E21 73.57

3.0 1-11 45.00 3.0 1-11 45.00 3.0111 45.00 3.0111 45.00 3.0 02 46.43 3.0 Ill 45810 3.0112 4643 3.0112 46.43 3.0112 45.43 3.0112 45.43

3.0 M14 60.71 3.0 M6 58.57 30M11 57.86 3.0 M23 67.86 3.0 M1 67.14 3.0 M11 5786 3.0 M2 63.57 3.O M11 57.86 3.0 M3 6143 3.0 M17 6000

3.0 M9 56.43 3.0 M5 71.43 3.0 M3 61.43 3.0 M22 57.86 3.0 M2 53.57 3.0 M8 5300 3.0 M3 61.43 3.0 M10 56.29 3.0 M10 59.29 3.0 M7 57.14

3.0M11 5786 3.0 M17 60.00 3.0 M24 54.24 3.0 M9 55.43 3.0 M8 55.00 3.0 M23 57.85 3.0 M24 54.29 3.0 M9 56.43 3.0 M4 62.86 3.0 M18 6001

3.0 E4 78.57 3.0 E9 77.14 3.0 Ell 7 796 3.0 El4 72.14 3.0 El° 78.57 3.0 E12 77.85 3.0 El 74.29 3.0 E9 77.14 3.0 ES 77.14 3.0 E2 77.14

3.0 E6 79.29 3.0 E3 77.14 3.0 El 74.29 3.0 E7 73.57 3.0 E9 77.14 3.0 E9 77.14 3.0 E2 77.14 3.0 E4 78.57 3.0 E2 7714 3.0 E3 7714

❑ifficulty 64.75 Difficulty 62.96 Difficulty 62.11 Difficulty 62.54 Difficulty 63.36 Difficulty 63.47 Difficulty 63.93 Diffficulty 61.93 Difficulty 63.04 Difficulty 61.93

Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6 Easy 6

Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate 11 Moderate /1

Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3 Hard 3

Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic Total From Topic

Topic 1 6 Tapir 1 6 Topic 1 6 Topic 1 6 Topic 1 5 Topic 1 5 Topic 1 6 Topic 1 6 Topic 1 6 Topic 1 6

Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 8 Topic 2 B. Topic 2 B. Tapir 2 8 Topic 2 8 Topic 2 8 Topic 2 8

Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6 Topic 3 6

Sample Difficulty Statistics

Target Cut Score 16437

Mean difficulty 63.13

Median 63.09

Minimum 61.93

Maximum 64.75

Variance Target vs. Mean 0.76

Standard Deviation all averages 0.76

95% Confidence Score 0273697889

Kurtosis 0127180112

Skewness 0351166789

Copyright © 2020 by James R. Parry

39

Table 39 - Experiment #3B Item Selections - Attempts 21 - 30

The item selection difficulty remained constant as stratified with each iteration (see ‘Difficulty’

columns in tables 37, 38, and 39). The target cut score/difficulty was 64.37. The stratified random-

ization produced a difficulty range between 61.93 and 64.75 with average (mean) of 63.13. The

standard deviation of the scores was 0.76 with a 95% confidence interval of 0.274 which means

that the true population mean is between 62.86 and 63.40 of the 30 samples. The kurtosis of the

average difficulty is 0.127 and the skewness is 0.351. The number of items at each difficulty level

from each topic varied with each iteration. Table 28 provides a summary of the statistics for the

sample. Figure 12 illustrates the standard distribution curve of the sample.

Table 40 - Difficulty Statistics for Experiment #3B

Page 41: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

55 60 65 70 75 SO

Average Test Difficulty

Standard Distribution of Experiment #3 Samples Stratified-Random Data

Copyright © 2020 by James R. Parry

40

Figure 12 - Standard Distribution of Test Difficulty - Experiment #3B - Stratified Random Selection

The content (topic) coverage was equal as stratified for each iteration as illustrated by the four-

color (delineated by sections) display and ‘Total From Topic’ columns in tables 37, 38, and 39.

Conclusion: All topics were covered equally as desired in both difficulty and content. Comparing

the spread of test difficulty scores in figures 10 and 12 shows that the stratified randomization

consistently produced tests well within an acceptable range to meet the desired cut score of 64.37.

Although the simple randomization in experiment 3A produced tests with an average (mean) dif-

ficulty (64.15) closer to the desired difficulty (64.37) than the stratified randomization in experi-

ment 3B (63.13), the standard deviation of the results in experiment 3B (0.76) indicated signifi-

cantly less variance from attempt to attempt than the standard deviation produced by experiment

3A (2.44).

CONCLUSIONS Key factors in maintaining defensibility and fairness of assessments are being able to justify the

cut or passing score as well as maintain validity. In that regard, each iteration of a parallel assess-

ment or test must be parallel in both difficulty and topic coverage otherwise there is unfair bias

each time an assessment is generated. Conclusions drawn from the experiments described in this

paper are as follow:

1. If test items are selected at random from a test-item database, without regard to either dif-

ficulty or subject matter, the resulting assessments will be inconsistent in coverage of topics

and difficulty. This is evidenced by the results of experiments 1A, 2A, and 3A.

a. Selection of the number of test-items from each topic was inconsistent with each

iteration of the randomly generated assessments.

i. Experiment 1A tables 6, 7, and 8

ii. Experiment 2A tables 20, 21, and 22

iii. Experiment 3A tables 33, 34, and 35

b. Each item in the test-item database was empirically assigned a difficulty score rang-

ing from .25 (25%) to .95 (95%) using the Parry Method (a variation of the Angoff

Method). Each item was then classified as Easy, Moderate, or Hard based upon the

score range determined by dividing the entire score range into thirds (see table A-

1) The “cut” score for the entire database was determined by averaging the results

of item score. This cut score was the target for each iteration of the assessment.

Page 42: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

41

i. The mean (average) difficulty or cut score for each experiment of randomly

generated assessments was within one point of the target cut score, and ac-

tually closer to the desired cut score than the mean (average) using strati-

fied-random selection (table 41).

ii. The standard deviation among each iteration was higher than those gener-

ated using stratified-random selection which translates to a wider spread or

variance of actual difficulties among iterations. This is evidenced by the

comparison of the standard score distribution plots (Figure 13).

2. If test items are selected from a test-item database, using stratified-random selection by

difficulty and subject matter, the resulting assessments will be consistent in coverage of

topics and difficulty. This is evidenced by the results of experiments 1B, 2B, and 3B.

a. Selection of the number of test-items from each topic was consistent with each it-

eration of the randomly generated assessments. The selection of the number of

items from each topic was a forced selection based upon the number of items avail-

able in each topic vs. the desired test length. The philosophy behind this selection

is described in Appendix A.

i. Experiment 1B tables 6, 7, and 8

ii. Experiment 2B tables 20, 21, and 22

iii. Experiment 3B tables 33, 34, and 35

b. As previously stated, each item in the test-item database was empirically assigned

a difficulty score and placed into appropriate topic and difficulty folders.

i. Stratified-random selection, based upon difficulty as well as topic, consist-

ently produced assessments within several points of the target cut score. The

assessment difficulty in relation to the target cut score varied slightly with

each iteration.

1. Experiment 1B tables 6, 7, and 8

2. Experiment 2B tables 20, 21, and 22

3. Experiment 3B tables 33, 34, and 35

Page 43: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Standard Distribution

la%

1_2%

10%

Random of Experiment #1 Samples

Data

Standard Distribution of Experiment #1 Samples Stratified-Random Data

3$.%

Ax

ZI.E.

20%

I IS% '3,- a- 103E

5%

0% 40 4 50 55 60 65 70 75 5%

Average Ten Difficu by

g

S IN

a o 695 a_

..,,

2%

0% 40 45 SO 55

F

_

60 65 7C .2.0

Average Ten Drfficulty

Standard Distribution of Experiment #2 Samples Random Data

20% 15% 16% 14%

.2 12% L15, 1.1. s . ' 6%

4% 2%

Standard Distribution of Experiment #2 Samples Stratified-Random Data

60%

50%

40% 5

30% E g

20%

10%

55 67 09 71 73 75 77 79 81 83 35 Average Ten Difficulty

0% 05 67 09 71 73 75 77 ra ss as 55

Average Test Difficulty

Standard Distribution

MN

is%

LIN

8 12%

.. 10%

06

of Experiment #3 Samples Random Data

Standard Distribution of Experiment #3 Samples Stratified-Random Data

6016

50%

V% 2 1 3066

SO 55 60 65 70 7, 07

Average Test Difficulty

2% -.41111111

111%

0% 0% SO SS 63 40 70 75 07

Average Test Difficulty

Copyright © 2020 by James R. Parry

42

ii. Although the difference between the mean (average) difficulty or cut score

and the target cut score for each experiment of stratified-random generated

assessments was slightly more than that of the randomly generated assess-

ments, (table 41), the standard deviation among iterations was lower which

translates to a smaller spread or variance of actual difficulties among itera-

tions. This is evidenced by the comparison of the standard score distribution

plots (Figure 13).

Target Difficulty (Cut-Score) vs. Actual Results

Experiment # Target Random Selection Stratified-Random

Selection

1 58.21 58.54 58.84

2 76.32 75.87 74.11

3 64.73 64.15 63.13

Table 41 - Target Difficulty (Cut-Score) vs. Actual Results

Figure 13 - Comparison of Standard Distribution of Difficulty - Random Selection vs. Stratified-Random Selection

Page 44: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

43

RECOMMENDATIONS In order to maintain fairness and ensure tests are valid, reliable, without bias and defensible when

generated from a test-item database:

• Test-items must be constructed using universally recognized standards

• Cut scores should be established using a recognized test-centered method or, if appropriate,

a test-taker centered method, because arbitrary methods are not defensible

• Each item in a test-item database should be evaluated by a panel of expert judges and a

difficulty score or rating established based upon the agreed upon MAC level of the target

test-taker

• Test items should be selected using stratified randomization based upon both topic cover-

age as well as item difficulty to ensure equitable parallel assessments are generated

• Tests should not be generated in a pure random fashion from a test-item database without

regard to content because content coverage will be erratic

• Tests should not be generated in a pure random fashion from test-item database without

regard to difficulty of test-items because difficulty among tests will be erratic

Page 45: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

44

LIST OF FIGURES Figure 1 – Designing Criterion-Referenced Tests .......................................................................................... 8

Figure 2 - Questionmark OnDemand Item Selection Criteria for Hypothetical Data ................................. 18

Figure 3 - Standard Distribution of Test Difficulty - Experiment #1A - Random Selection ......................... 20

Figure 4 - Topic, Sub-topic Structure Used in Experiments ........................................................................ 21

Figure 5 - Item Selection Criteria - Experiment #1B .................................................................................... 22

Figure 6 - Standard Distribution of Test Difficulty - Experiment #1B - Stratified Random Selection ......... 24

Figure 7 - Standard Distribution of Test Difficulty - Experiment #2A - Random Selection ......................... 29

Figure 8 - Item Selection Criteria - Experiment #2B .................................................................................... 30

Figure 9 - Standard Distribution of Test Difficulty - Experiment #2B - Stratified Random Selection ......... 32

Figure 10 - Standard Distribution of Test Difficulty - Experiment #3A - Random Selection ....................... 37

Figure 11 - Item Selection Criteria - Experiment #3B.................................................................................. 37

Figure 12 - Standard Distribution of Test Difficulty - Experiment #3B - Stratified Random Selection ....... 40

Figure 13 - Comparison of Standard Distribution of Difficulty - Random Selection vs. Stratified-Random

Selection ...................................................................................................................................................... 42

Page 46: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

45

LIST OF TABLES Table 1 - Number of Test-Items per Objective ............................................................................................. 9

Table 2 - Topic Weights for Electrical Safety Assessment .......................................................................... 10

Table 3 - Test Time Tool .............................................................................................................................. 11

Table 4 -Experiment # 1- Cut Score Calculation Tool (partial view) showing hypothetical data ................ 17

Table 5 – Experiment #1 - Recommended Test Difficulty Distribution for Hypothetical Data – Random

Selection ...................................................................................................................................................... 17

Table 6 - Experiment #1A - Item Selection - Attempts 1 - 10 ..................................................................... 18

Table 7 - Experiment #1A - Item Selection - Attempts 11 - 20 ................................................................... 19

Table 8 - Experiment #1A - Item Selection - Attempts 21 - 30 ................................................................... 19

Table 9 - Difficulty Statistics for Experiment #1A ....................................................................................... 20

Table 10 - Recommended Test Design - Stratified Randomization - Experiment #1B ................................ 21

Table 11 - Experiment #1B Item Selections - Attempts 1 - 10 .................................................................... 22

Table 12 - Experiment #1B Item Selections - Attempts 11 - 20 .................................................................. 23

Table 13 - Experiment #1B Item Selections - Attempts 21 - 30 .................................................................. 23

Table 14 - Difficulty Statistics for Experiment #1B ...................................................................................... 24

Table 15 - Experiment #2A - Difficulty Calculations - Real Data - Topic 1 ................................................... 25

Table 16 - Experiment #2A - Difficulty Calculations - Real Data - Topic 2 ................................................... 25

Table 17 - Experiment #2A - Difficulty Calculations - Real Data - Topic 3 ................................................... 26

Table 18 - Experiment #2 - Item Difficulty Distribution by Topic ................................................................ 26

Table 19 - Experiment #2 - Recommended Test Design ............................................................................. 26

Table 20 - Experiment #2A - Item Selection - Attempts 1 - 10 ................................................................... 27

Table 21 - Experiment #2A - Item Selection - Attempts 11 - 20 ................................................................. 27

Table 22 - Experiment #2A - Item Selection - Attempts 21 - 30 ................................................................. 28

Table 23 - Difficulty Statistics for Experiment #2A ..................................................................................... 28

Table 24 - Experiment #2B Item Selections - Attempts 1 - 10 .................................................................... 30

Table 25 - Experiment #2B Item Selections - Attempts 11 - 20 .................................................................. 31

Table 26 - Experiment #2B Item Selections - Attempts 21 - 30 .................................................................. 31

Table 27 - Difficulty Statistics for Experiment #2B ...................................................................................... 32

Table 28 - Experiment #3A - Difficulty Calculations - Real Data - Topic 1 ................................................... 33

Table 29 - Experiment #3A - Difficulty Calculations - Real Data - Topic 2 ................................................... 33

Table 30 - Experiment #3A - Difficulty Calculations - Real Data - Topic 3 ................................................... 34

Table 31 - Experiment #3 - Item Difficulty Distribution by Topic ................................................................ 34

Table 32 - Experiment #3 - Recommended Test Design ............................................................................. 34

Table 33 - Experiment #3A - Item Selection - Attempts 1 - 10 ................................................................... 35

Table 34 - Experiment #3A - Item Selection - Attempts 11 - 20 ................................................................. 35

Table 35 - Experiment #3A - Item Selection - Attempts 21 - 30 ................................................................. 36

Table 36 - Difficulty Statistics for Experiment #3A ..................................................................................... 36

Table 37 - Experiment #3B Item Selections - Attempts 1 - 10 .................................................................... 38

Table 38 - Experiment #3B Item Selections - Attempts 11 - 20 .................................................................. 38

Table 39 - Experiment #3B Item Selections - Attempts 21 - 30 .................................................................. 39

Table 40 - Difficulty Statistics for Experiment #3B ...................................................................................... 39

Table 41 - Target Difficulty (Cut-Score) vs. Actual Results .......................................................................... 42

Page 47: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Difficulty Values Based on Number of Alternatives

ALES Y 4 MODERATE 55.1- 80

L 2 80.1- 95 50- 55

3 74.51-95 53.8-74.5 33-53.8

4 71.8-95 48.4-71.7 25-48.3

5 70.2-95 45.1 - 70.1 20-45

6 69.1- 95 42.9- 69 16.7-42.8

7 68.4-95 41.3-68.3 14.2-41.2

Copyright © 2020 by James R. Parry

A-1

APPENDIX A

Description of the spreadsheet tool: The spreadsheet tool is designed to be used in conjunction with most any test-centered cut score

setting method that produces a difficulty rating for individual test-items. It has been used very

successfully in conjunction with the Modified Angoff method25 and is designed primarily for 4-

alternative multiple-choice26 test-items with allowable rating scores between .25 and .95. A ver-

sion has also been developed for use with 2 through 7-alternative items with allowable rating

scores adjusted to suit the number of alternatives with a minimum rating score of 14.2 for a 7-

alternative item (table A-1). There are two trains of thought for the maximum rating ceiling; 1.00

or .95 (100% or 95%). The original Angoff Method allowed any value from .25 through 1.00 and

only asked the judges to consider one individual. I chose a variation of the Modified Angoff

Method (Parry Method) which asks the judges to consider 100 test-takers who are considered to

be minimally competent (in the subject being tested) and restricts the judges to preset values al-

lowed for simplicity. (.25, .30, .35, .40, .45, .50, .55, .60, .65, .70, .75, .80, .85, .90, .95). 1.0 (100%)

is eliminated as an item that is responded to correctly by a minimally competent examinee 100%

of the time is considered an unnecessary test-item in most cases.

Table A- 1 - Difficulty Values Based on Number of Alternatives

The description of the spreadsheet function will use examples from the 4-alternative tool. Both

tools function the same with the exception of the 2 through 7-alternative tool requires an entry by

the user to identify the number of alternatives to ensure proper difficulty calculations. The tool has

not been optimized for items that use multiple-response27 items types but does allow for the mixing

of multiple-choice items with different numbers of alternatives. It is recommended that mixing

test-items with varying numbers of alternatives within the database be limited and used with ex-

treme caution as the results of stratified random selection may generate assessments outside of the

desired difficulty range or present an unbalanced mix of items with varying numbers of alterna-

tives.

25 A well-established method involving three basic steps: conceptualizing the borderline examinee, identifying spe-cific test-items, and using expert judges to estimate what percentage of borderline examinees should respond cor-rectly. 26 A multiple-choice test-item only allows the test-taker to select one alternative as the correct/incorrect choice 27 A multiple-response test-item allows the test-taker to select more than one alternative as correct/incorrect re-sponses.

Page 48: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

A-2

Design philosophy of the Compass Consultants, LLC spreadsheet tool: The tool is designed

to assist in setting a cut score for an assessment based on the results of a test-centered cut-score

rating session. Additionally, it will determine the number of items from each section at each level

of difficulty (Hard, Moderate or Easy) as set by the cut-score rating of each item. This assumption

is made for 4-choice, multiple choice items with a floor of 25% and a ceiling of 95%. The diffi-

culty is then assigned based on dividing the difference between 25% and 95% by 3 to arrive at the

three difficulty levels. The workbook is designed to accommodate up to ten (10) reviewers on the

rating panel. The ratings assigned to each test-item by the individual judges are averaged and a

difficulty score is assigned to the test-item. If the judge’s individual ratings produce a standard

deviation28 of 10 or greater the item is flagged for discussion among the judges to either come to

a consensus by modifying their ratings, retire the item as a ‘bad’ item or leave it as written and

rated. A final test design is proposed after all items in all topics have been rated. The totals required

from each section are based upon the numbers of each level of difficulty available in each section

as well as the total number of items available. An assumption is made that if there are fewer items

available in any particular section(s) than in other section(s), then that section is of less importance

or has significantly fewer objectives. As data for each item is entered in each section, the final test

design worksheet will be updated automatically.

Function of the Spreadsheet Tool The tool is designed to accommodate 20 topics with 200 test-items per topic. The number of topics

and test-items can be expanded upon request to Compass Consultants, LLC. The tool will allow

for the input from up to 10 judges. Typically, 6 – 8 judges are sufficient with less than three not

producing reliable ratings and more than 10 as being difficult to arrive at a consensus. Refer to

table A-2 as the spreadsheet tool is described.

• As all of the heading information is entered by the facilitator, each of the individual work-

sheets populate automatically

• Generally, the Enter * If New, R if Retired block is left blank initially. As items are

reviewed during the rating session, some may be omitted. If an item has NEVER been

presented on an assessment and is omitted during the review, all data can be omitted and

the row left blank. If additional items are added to the database in the future they should

be indicated with an ‘*’ for easy identification. It is not usually necessary to re-evaluate all

items but be aware the section/topic difficulty as well as the cut score may change. If an

item that is currently in use on an assessment is subsequently retired from use the worksheet

row should be updated with an ‘R’. This will remove any cut score/difficulty calculations

from the tool and the section/topic difficulty as well as the cut score may change. To main-

tain defensibility for possible future challenges this data should remain in the permanent

assessment documentation.

• Each test-item is given a Question Identifier (Test-item QID) in the test-item database

before a rating session begins.

• As each expert (judge) ‘score’ is entered for each item the tool is updated in real time:

28 Standard Deviation (σ) is a measure of how spread out numbers are.

Page 49: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

. . , . . CUT SCORE CALCULATION TOOL

Course/Certifiotion Nairieil FAIRNESS RESEARCH 3 Ted Named TEST NAME

Comm Facil itator Name/Date: Facilator Name/Date Revision 1 Facilitator Narne/Date D atedrnmiddityyyy COSSilhaStSALC

Enter Topk/TPO/Subject ID: Topic 1 Revision 2 Facilitator Name/Date

Thic cpre,d. ool intelle nal prop, of NA Copyright .2010 2015 by Comp, Concultant... LLC. Lf. ic d to the of nd Lf..er Licence Agr nt (EL11..4) Thic copy I: licenced to: 30 DAY 10EMO ONLY

Ted Item DID Enter` If New. R if Retired

Difficulty

ME.tat'g

Average Percentage

Coment

Rawl

Expert 1 Name

0Name

Expert 2 Name

Expert 3 Name

Expert 4 Name

Experts Name

Expert 6 Name

Expert 7 Name

NMI 8 Name

Expert9 Name

Expert 10 Name

II Standard Deviation

Topic Moderate

Cut 58.00 Difroulty

Score

1.0 M1 Moderate 48.57 60 60 40 45 40 55 40 Approximate Difficulty Rating

10 M2 Moderate 62.14 55 70 60 75 65 60 50 25 - 48.3 Hard

1.0.M3 I - 57.14 50 60 70 60 60 EU 40 48.4 - 71di Moderate

It 50 65 50 65 60 50 40 71_8 - 95 Easy

10 M4 Moderate 62.85 70 70 70 60 60 50 50 756

10 M5 Moderate 60.00 60 70 70 50 50 70 50 /000 Standard Deviation

1.0 El 77.14 70 85 30 75 70 90 70 gDgi

1.0 E2 75.00 70 60 90 75 70 30 50 9.57 A standard deviation of more than

1.0 E3 77.14 70 80 90 75 75 80 70 5.99 10 will trigger an alert. Disc uss the

1.0 E4 84.29 75 90 95 90 75 90 75 8.85 outliers with the judges who set

1.0 M6 Moderate 62.14 55 70 50 50 70 70 70 9.94 them to determine why. Change as

1_0111 41.43 30 50 35 35 50 50 40 852 necessary_

1.0 M7 Moderate 63.57 60 65 70 50 65 55 70 6.90

10 M8 Moderate 5500 55 60 55 50 50 55 40 '.7..10 7 In tills section 17%

R 70 80 75 8.0 70 90 40 28 Moderate In this section 67%

10 M9 Moderate 57.14 50 50 70 50 50 70 50 7 In 'Lis section 17%

1.0 M10 Moderate 48.57 50 45 50 35 55 55 50 7:3:: 42 TOTAL 10095

1.0 Mll Moderate 57.86 50 70 50 50 65 60 60 8.1'9

Copyright © 2020 by James R. Parry

A-3

o Difficulty Metatag presents the difficulty as easy, moderate, or hard and the color

changes as appropriate with green indicating ‘easy, yellow indicating ‘moderate’,

and red indicating ‘hard’ based upon the judges average score

o Average Percentage Correct (Angoff Rating) for the item is updated to reflect

the running average score

o Standard Deviation is updated. If the standard deviation is 10 or greater the block

displays a red flag to facilitate location during the discussion phase of the rating

session

o The Topic Cut Score block updates to reflect the current cut score/difficulty for

the topic

o A running total of the number and percentage of items at each difficulty level is

displayed in a block to the lower right.

Table A- 2 - Cut Score Calculation Tool Data Entry and Display

Page 50: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

SECTION 1 DESCRIPTIVE STATISTICS

Judges Judge's

Minimum Maximum Standard Mean

Rating Rating Error of Rating

'the Mean

Judge

Number

of Items

Counted

Judge's

MI n/ Max

SD

SD of

Judge

FromCut

Score

1 39 25 75 55.67 6.32 35.36 1. g9

146 2 39 35 90 60.56 0.23 38.89

3 39 35 95 6133 032 4243 7 m

4 39 25 90 53.89 0.52 45.96 325

5 39 30 as 59.11 007 38.89 044

6 39 40 90 64.78 0.71 35.36 4A5

7 39 25 75 53.67 0.55 35.36 3A1

8 39 0 0 0.00

9 39 0 CI 0.00

III 39 0 0 0.00

of t he

Section i

Calculated Cut Score

(Mean) 58.49

Standard Error

Mean for Entire 1.36

mean, also

of the

the

error,

fr6u rage Standard

Deviation for

Number of

Retired (not

Unft 814 The stand and error of the

. lied the standard deviation

mean, is used to esti mate

standard deviation of a sampl ing

distribution. The smallerthe

the more re liable the measurement.

Items

counted} 3

SECTION 1 DESCRIPI1VE STATISTICS

SD of Judge's

Judge

Minna' From Cut SD

Score

SECI1ON 2 DESCRIPTIVE STATISTICS.

Judges

Min/Max

SD

SD of I

judge.

From Cut

Score

6.76

Judge

Number

of Items

Counted

Minimum

Rating

Maximum

Rating

Judge's

Mean

Rating

Judge's

Standard

Error of

the Mean

Judge

1

2

Number

of Items

Counted

Panimum

Rating

Maximum

Rating

fudge's

Mean

Rating

Judge's

Standard

Error of

the Mean

1 39 25 75 55 67 032 3536 1.99 49 30 75 59.36 097 3182

2 39 35 so 60.56 023 38.89 1.46 49 45 90 69.55 0.06 31,82 0.44

3 39 35 95 6133 032 42.43 2.01 3 49 45 95 75.64 0.68 3536 4.74

4 39 25 90 53.89 0.52 4596 3.25 4 49 35 95 64.82 0.42 42.43 2.91

5 39 30 85 5911 01/7 3889 0.44 5 49 50 80 6986 01)4 2121 0.31

6 39 40 90 6478 071 3536 4.45 6 49 55 90 7636 075 2475 5.26

7 39 25 75 53.67 0.55 3536 3.41 7 49 40 80 6591 031 2828 2.14

8 39 o o ODD 8 49 O o 0.00

9 39 o o 0.00 9 49 O o 0.00

10 39 o o 0 CC 10 49 o o 0.00

Calculated Cut

(Mean)

Average Standard

Deviation for

Mean for EntireSectkm (Mean)

Calculated Cut

Average Standard

Deviation for

Score 58.49

Standard Error of the 1.38

mean, also

of the

the

error,

Score 68.93

Standard Error of the 123

Mean for Entire Sectkm

The standard error of the mean, also

called the standard deviation of the

mean, is used to estim ate the

standard deviation of a sampling

distribution. The sm alter the error,

the more reliable the measurement.

Unit 8.64

The standard error of the

called the standard deviation

mean, is used to estim ate

standard deviation of a sampling

distribution. Thesm alter the

the more reliable the measurement.

Unit 8.64

Number of Items

Retired (not counted l 3 Number of Items

Retired (not counted! 3

Copyright © 2020 by James R. Parry

A-4

As data is entered, descriptive statistics are calculated for each judges’ ratings as illustrated by

table A-3.

Table A- 3 - Judges' Descriptive Statistics

The descriptive statistics are copied for each topic/section to a separate worksheet “CONSOLI-

DATED DESCRIPTIVE STATISTICS” that displays all of the descriptive statistics for the entire

workbook on one worksheet (Figure A-4 – partial view).

Table A- 4 - Consolidated Descriptive Statistics Sample

Page 51: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

JUDGES COM PARITIVE RATINGS

Judge's Mean

Rating For All

Sections

Judge's Standard

Deviation From

Assessment Cut

Score

Judge's

Standard

Error of the

Mean

57.94 4.29 0.37

65.24 0.87 0.08

70.98 4.94 0.43

59.69 3.05 C, 26

63.11 0.63 0.05

71.72 5.45 0.47

60.87 2.21 0.19 =Ir NOTE: These calculations assume thatthe judge has

contributed input to all topics being analyzed. If the judge

did not contribute, their average rating will not be

included in this overall rating table- refer to the

individual section descriptive statistics for their

information.

Judge

1

2

3

4

5

5

7

9

10

Copyright © 2020 by James R. Parry

A-5

In addition to the consolidated view of the descriptive statistics, a table comparing the judges’

ratings for the entire database is presented (Table A-5). This table is useful to determine if indi-

vidual judges have typically scored higher or lower than the rest of the group and whether consid-

eration should be given to disregard that judge’s scores.

Table A- 5 - Judges' Comparative Ratings

The results of each or the worksheets are consolidated on the “FINAL TEST DESIGN” worksheet

(Table A-6) which displays totals and percentages of items at each difficulty level for each topic

as well as presents a recommended test design to maintain the projected cut score as well as ensure

every iteration of the assessment that is generated in a stratified-random format is equal in content

and content difficulty. Each column is explained below unless it is self-explanatory.

Page 52: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Final Directed-Randomized Test Design Blueprint for: TEST NAME nuniddirrirr

Topic To Cut pic Score& Diffim,ity

Items in Topic

%of-final Items

Avaiable Herd

%Frorn Topic

Available Mod

%From Topic

Available Easy

% From Topic

Total ir Needed Fran Topic

Use Hard (Calculated)

Use Hard {Actual}

Use Mad 'Calculated'.

Use Mod IA Et. IF

Use Easy {Calculated}

Use Easy {Actual)

Topic

Topic 1 56 42 31.54% 7 17 ?..• 28 67?.. 7 6.27 1.04 1 418 4 1 Topic 1 Topic 2 64 5.2 38.9' ' 1 2% 27 5236 24 7.76 0.15 1 4.03 4 3 Topic 2

Topic 3 36 40 29. 2 5 5 24 6:-% 14 5.97 0.30 1 1.5R 2 Topic 3

4.1 r

0 0 C. 0 r 0.00 ... r , lir r

r 4.1

5,1 r 0 C.: 0 0 r 5.1. 6.1 r 0 C.: 0 r 0 0.00 6.1

7.1 0 C. 0 C 0 0.00 7.1

8.1 0 C. 0.00 8.1

9,1 0 C. - - - - ., 0.00 9.1

10.1 0 C. - C 0.00 10.1

11.1 0 C. - - 0.00 11.1

12.1 0 C. - - - 0.00 12.1

13.1 0 C. 0 C 0 0.00 13,1 14.1 r 0 C. 0 r r

0 I 0.00 r r r

14.1

15.1 r 0 C. 0 r C r

0 I 0.00 r r r

15.1

16.1 r 0 C. 0 r r 0 r 0.00 r

r r 16.1

17.1 r 0 C. 0 P C r

0 r 0.00 r r r

17,1 18.1 r o C.: 0 r 0 0.00 18.1

19.1 0 0 C 0 0.00 19.1

20.1 r 0 O. 0 0 DAC 20.1

TOTAL 131 100.00% 10 79 45 2:- .CC 3 11 6.72 6

:: appears in the "Total* Needed From Topic" block - you do not have sufficient items NOTE: If

in the topic indicated to design a fa i test.

C

c

n

o

n

n

i

rilli:r1S, LUC mDifrrier:ty

Test

tScoCreut

Set

Desired

Test Size

After all cut-score session data has been entered on section worksheets, set the desired test size in the block

to the left. Based upon the number of available items, the quantity of Hard, Moderate and Easy from each

section will populate automatically. Use these results to design the test in your test item database using

64.00 20 established difficulty Metatags or sub-topic Approximate Difficulty Ratings. Note:- Due to rounding errors in

Checksum Excel, the unit/item difficulty totals may require you to round up or down manually to achieve desired test

Approximate Test Time in Minutes Based 17.43

20 size. Set the actual number desired bsed upon the calculated results in the columns labeled "Actual" above.

The Checksum to the left will alert you if the selected value does not match the desired test size.

on Item Difficult/

Copyright © 2020 by James R. Parry

A-6

Table A- 6 - Final Test Design Display

• Column D – displays the percentage of items from each topic in relation to the total items

available.

• Columns E through J – display the total number and percentage of items at each level of

difficulty in each topic. NOTE: Percentages are rounded for display.

• Column K – presents the total number of items required from each topic to maintain the

percentages shown in column D for the desired test size entered in the “Set Desired Test

Size” block at the bottom of column F.

• Columns L, N, and P – presents the recommended number of items from each topic at each

level of difficulty to be used to maintain the desired cut score as well as topic coverage.

These numbers were not rounded to allow the test designer to make informed decisions as

to how many items to enter in columns M, O, and Q.

Page 53: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Set

Desired

Test Size

I 20

ChteelcSum

22

Copyright © 2020 by James R. Parry

A-7

• Columns M, O, and Q – The test designer enters the whole number of items as close to the

recommended numbers as possible while referring to the “Checksum” block below the

desired test size block at the bottom of column F. If there are more items selected than the

test size, the checksum block will alert the designer by changing color (Figure A2). The

designer must then make an informed decision as to which topic(s) to subtract or add items

to ensure topic equity.

Figure A- 1 - Test Design Sheet Checksum Warning

• The note at the bottom of columns J through R explains a warning that will appear in the

“Total needed from topic” block(s) if there is not a sufficient quantity of items available

to generate a fair test of the size desired.

• The box labeled “Approximate Test Time in Minutes Based on Item Difficulty” is calcu-

lated using rule of thumb values for the time it takes a test-taker to respond to a test item

based upon its difficulty (see table 3).

Note: The spreadsheet tool is available for licensing. Send request to: info@gocompassconsult-

ants.com

Page 54: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

A-8

LIST OF FIGURES IN APPENDIX A Figure A- 1 - Test Design Sheet Checksum Warning ................................................................................. A-7

LIST OF TABLES IN APPENDIX A Table A- 1 - Difficulty Values Based on Number of Alternatives .............................................................. A-1

Table A- 2 - Cut Score Calculation Tool Data Entry and Display ............................................................... A-3

Table A- 3 - Judges' Descriptive Statistics................................................................................................ A-4

Table A- 4 - Consolidated Descriptive Statistics Sample ........................................................................... A-4

Table A- 5 - Judges' Comparative Ratings ................................................................................................. A-5

Table A- 6 - Final Test Design Display ....................................................................................................... A-6

Page 55: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Copyright © 2020 by James R. Parry

R-1

REFERENCES AERA; APA; NCME. (2014). Standards for Educational and Psychological Testing. American Educational

Research Association (AERA); American Psychological Association (APA); National Council on

Measurement in Education (NCME). Washington, DC: American Educational Research

Association.

Cizek, G. J. (2006). Standard Setting. (S. Downing, & T. Haldyna, Eds.) Handbook of test development, 225

- 258.

Coscarelli, W., Barrett, A., Kleeman, J., & Shrock, S. (2005). The problem of saltatory cut-score: some

issues and recommendations for. Proceedings of the 9th CAA Conference. Loughborough:

Loughborough University. Retrieved from https://hdl.handle.net/2134/1984.

Dictionary.com Unabridged. (2020, March 6). empirical. Retrieved March 6, 2020, from Dictionary.com:

https://www.dictionary.com/browse/empirical

Downing, S. M. (2006). Twelve Steps for Effective Test Development. (S. M. Downing, & T. M. Haladyna,

Eds.) Handbook of Test Development, 3-25.

Great Schools Partnership. (2013, August 29). Measurement Error. Retrieved from The Glossary of

Education Reform for Journalists, Parents, and Community Members:

https://www.edglossary.org/measurement-error/

Higgins, P. (2009, May). Item Difficulty and Time Usage. Retrieved March 2020, from Measurement

Research Associates, Inc.: https://www.rasch.org/mra/mra-05-09.htm

Livingston, S. A., & Zieky, M. J. (1982). Passing scores - A manual for setting standards on educational

and occupational tests. Princeton, NJ: Educational Testing Service.

Shrock, S. A., & Coscarelli, W. C. (2007). Criterion-Referenced Test Development - Technical and Legal

Guidelines for Corporate Training (3 ed.). San Francisco, CA, USA: Pfeiffer.

United States Coast Guard. (2015). Training system standard operation procedure 10 - Testing (SOP-10).

Washington, DC, US: United States Coast Guard Forces Command.

Vale, C. D. (2006). Computerized Item Banking. (S. Downing, & T. Haladyna, Eds.) Handbook of test

development, 261 - 285.

Page 56: Ensuring Fairness in Difficulty and Content Among Parallel ...gocompassconsultants.com/uploads/8/5/5/0/85508724/fairness_in_t… · time the assessment is presented. To maintain fairness

Ack-1

ACKNOWLEDGEMENTS The author gratefully acknowledges:

John Kleeman for his friendship, review and recommendations.

John Kleeman, is Executive Director and Founder of Questionmark. John has a first-class degree

from Trinity College, Cambridge, and is a Chartered Engineer. John wrote the first version of the

Questionmark assessment software system and then founded Questionmark in 1988 to market,

develop and support it. John has been heavily involved in assessment software development for

over 20 years and has also participated in several standards initiatives: he was on the original team

that created IMS QTI and was the instigator and chairman of the panel that produced the Standard

BS 7988, which has now become ISO 23988.

Eric Shepherd for his friendship, review, encouragement and publishing guidance.

Eric Shepherd has led international businesses and associations focused on talent, assessments,

and success. Eric currently leads Talent Transformation Guild which is focused on helping leaders,

talent management professionals, and consultants manage the changes that are now being driven

by the pandemic and the technology driven-revolutions. The Guild uses the vocabulary of the Tal-

ent Transformation Pyramid to help members understand, learn, and take action, to prepare for the

new world of work. Eric stepped away from a CEO role in which he built a SaaS company into a

multi-million-dollar international assessment software business. Eric has also led industry and

standards initiatives to promote best practices for competency development, assessments, learning,

and interoperability.

Sandra Parry, my loving wife, for putting up with my long hours spent tucked away in my office

working on this paper.

ABOUT THE AUTHOR Jim Parry, a 22+ year veteran of the United States Coast Guard, is the Owner and Chief Executive

Manager of Compass Consultants, LLC with over 38 years’ experience in course design, develop-

ment, presentation and assessment design and analysis. He holds a Master of Education degree

from the University of West Florida and is a Certified Performance Technologist (CPT), awarded

by the International Society of Performance Improvement (ISPI). Jim has been a presenter of pre-

conference workshops and educational sessions at various professional conferences for many

years. He is an internationally recognized consultant providing services concerning test design,

development, establishment of cut scores, and analysis. Some of Jim’s recent major accomplish-

ments include the development of a comprehensive standard operating procedure (SOP) manual

on testing for the U.S. Coast Guard, a spreadsheet tool to assist in recording and analyzing the

results of test-centered cut score setting meetings and using the spreadsheet tool to recommend the

design of assessments using directed randomization in order to maintain fairness. Jim is a consult-

ing partner of Questionmark Corporation.