crt development

17
CRT Development Item specifications and analysis

Upload: garrett-tillman

Post on 31-Dec-2015

26 views

Category:

Documents


5 download

DESCRIPTION

CRT Development. Item specifications and analysis. Considerations for CRTs. Unlike NRTs, individual CRT items are not ‘expendable’ because they have been written to assess specific areas of interest - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: CRT Development

CRT Development

Item specifications and analysis

Page 2: CRT Development

Considerations for CRTs

• Unlike NRTs, individual CRT items are not ‘expendable’ because they have been written to assess specific areas of interest

• “If a criterion-referenced test doesn’t unambiguously describe just what it’s measuring, it offers no advantage over norm-referenced measures.” (Popham, 1984, p. 29)

Popham, W. J. (1984). Specifying the domain of content or behaviors. In R. A. Berk (Ed.), A guide to criterion-referenced test construction (pp. 29-48). Baltimore, MD: The Johns Hopkins University Press.

Page 3: CRT Development

CRT score interpretation

(Popham, 1984, p. 31)

‘Good’ CRT

‘Bad’ CRT

Page 4: CRT Development

Test Specifications

• ‘Blueprints’ for creating test items• Ensure that item content matches

objectives (or criteria) to be assessed• Though usually associated with CRTs, can

also be useful in NRT development (Davidson & Lynch, 2002)

• Recent criticism: Many CRT specs (and resulting tests) are too tied to specific item types and lead to ‘narrow’ learning

Page 5: CRT Development

Specification components

• General Description (GD) – brief statement of the focus of the assessment

• Prompt Attributes (PA) – details what will be given to the test taker

• Response Attributes (RA) – describes what should happen when the test-taker responds to the prompt

• Sample Item (SI) • Specification Supplement (SS) – other useful

information regarding the item or scoring

(Davidson & Lynch, 2002)

Page 6: CRT Development

Item – specification congruence

(Brown, 1996, p. 78)

Page 7: CRT Development

CRT Statistical Item Analysis

• Based on criterion groups• To select groups, ask: Who should be able to

master the objectives and who should not?• Logical group comparisons

– Pre-instruction / post-instruction– Uninstructed / instructed– Contrasting groups

• The interpretation of the analysis will depend in part on the groups chosen

Page 8: CRT Development

Pre-instruction / post-instruction

Advantages • Individual as well as

group gains can be measured

• Can give diagnostic information about progress and program

Disadvantages• Requires post-test• Potential for test

effect

Berk, R. A. (1984). Conducting the item analysis. In R. A. Berk (Ed.), A guide to criterion-referenced test construction (pp. 97-143). Baltimore, MD: The Johns Hopkins University Press.

Page 9: CRT Development

Uninstructed / instructed

Advantages• Analysis can be

conducted at one point in time

• Test can be used immediately for mastery / non-mastery decisions

Disadvantages• Group identification

might be difficult • Group performance

might be affected by a variety of factors (i.e., age, background, etc.)

Page 10: CRT Development

Contrasting groups

Advantages• Does not equate

instruction with mastery

• Sample of masters is proportional to population

Disadvantages• Defining ‘mastery’ can

be difficult• Individually creating

each group is time consuming

• Extraneous variables

Page 11: CRT Development

Guidelines for selecting CRT items

Item Characteristic

Criterion Index value

Item-spec congruence

Matches objective being tested

IF (difficulty) Hard for UG

Easy for IG

IF less than .5

IF greater than .7

Discrimination Positively discriminates between criterion groups

High positive

Page 12: CRT Development

Item discrimination for groups

Uninstructed

Non-masters

Instructed

Masters

DI = IF (‘master’) – IF (‘non-master’)

Sometimes called DIFF (difference score)

(Berk, 1984, p. 194)

Page 13: CRT Development

(Brown, 1996, p. 81)

Page 14: CRT Development

Item analysis interpretation

(Berk, 1984, p. 125)

Page 15: CRT Development

Distractor efficiency analysis

• Each distractor should be selected by more students in the uninstructed (or incompetent) group than in the instructed (or competent) group.

• At least a few uninstructed (or incompetent) students (5 – 10%) should choose each distractor.

• No distractor should receive as many responses by the instructed (or competent) group as the correct answer.

(Berk, 1984, p. 127)

Page 16: CRT Development

The B-index

• Difficulty index calculated from one test administration

• The criterion groups are defined by their passing or failing the test

• Failing is defined as falling below a predetermined cut score

• The validity of the cut score decision will affect the validity of the B-index

Page 17: CRT Development

(Brown, 1996, p. 83)