1 psy 6430 unit 4 reliability and validity lecture: monday, 2/23 and wednesday, 2/25 exam: monday,...

53
1 PSY 6430 Unit 4 Reliability and Validity Lecture: Monday, 2/23 and Wednesday, 2/25 Exam: Monday, 3/02 SOs ME1: Wednesday, 3/04 ME1: Monday, 3/16, after semester break

Upload: emerald-cain

Post on 28-Dec-2015

221 views

Category:

Documents


2 download

TRANSCRIPT

1

PSY 6430 Unit 4

Reliability and Validity

Lecture: Monday, 2/23 and Wednesday, 2/25 Exam: Monday, 3/02

SOs ME1: Wednesday, 3/04ME1: Monday, 3/16, after semester break

Schedule Make-up exam: Monday, 3/16, after break It will cover Units 1-4 I will hand out the study objectives Wednesday, 3/04

Schedule is awkward because of the cancellation of class on 2/18: OBM Net and BAAM conferences

I can’t give you the ME1 study objectives until I grade and hand back E4 on Wed., 3/04

I want to give you your grades on ME1 before the last day to withdraw with academic penalty

But that means that Unit 5 lecture will be split, with lectures on 3/04 and 3/18, and the exam on 3/23

I just couldn’t make the calendar work well

2

Make-up Exam If you have missed an exam, you need to take

this exam or your missing exam score will turn into a zero

If you have taken Exams 1-4, but would like to try to replace one of your scores with ME1, you should take the exam If you get a lower grade on ME1, I will throw out the

ME1 grade – ME1 can only help your grade, it cannot hurt it

If you are satisfied with your exam scores, you do not have to take the exam – you get the day off!

3

4

Unit introduction I have found that students cannot understand reliability and validity

unless they first understand correlation Thus, I am first going to review correlation and statistical

significance before dealing with reliability and validity In traditional I/O psychology programs, students would be required

to take a generic tests and measurements course before taking a course in personnel selection, but since our program does not emphasize testing, we don’t have that type of course

Unfortunately, Gatewood, Feild, & Barrick discuss correlation in some detail as it relates to validity, but don’t talk about it much before they discuss reliability; yet correlation is the primary way to determine reliability as well

I could not find relevant supplemental material that dealt with this topic the way I wanted to deal with it in this course, so bear with me a bit

5

SO1 (NFE): Correlation, validity, and reliability

of selection instruments A correlation coefficient indicates

whether two variables are related and the extent to which they are related

Correlation is typically used in selection to determine whether the scores on a selection instrument are really measuring what you want

to measure (do the scores actually reflect the KSAs you want to measure and the person’s competence)

your selection instruments are related to how well a person performs on the job

Reliability refers to whether the selection instrument is accurately measuring the knowledge, skill and/or ability it is supposed to be measuring

Validity refers to whether your selection instruments are related to the job

6

SO1 (NFE): Correlation and reliability

With respect to reliability, correlation is used to answer the following questions: Is the selection instrument accurately measuring the ability, skill,

or knowledge it is supposed to be measuring Does the person’s score accurately reflect his/her competence

with respect to what is being measured Reliability does not indicate whether the selection

procedure is related to performance on the job

7

SO1 (NFE): Correlation and reliability

One measure of reliability is the stability/consistency with respect to how a person scores when he/she takes the test two different times

In order to be useful for selection, the score a person receives must be reasonably the same each time he/she takes the testExample: Assume that math is required to perform well on the job. A company administers a math test, and a person gets a 75. If the same person took the test the next day and only scored a 20, the test would not be useful for selection purposes. Why? Because you would not know whether the 75 or 20 represented what his/her math skills really were.

A high correlation between test scores indicates that the test is “reliable”

8

SO1(NFE): Correlation and validity

With respect to validity, correlation is used to answer the following two questions: Is the score that a person receives on a personnel selection

instrument related to a measure of his or her job performance? If so, to what degree are the two related?

If scores on the selection instrument and the measures of job performance are highly correlated, then the selection instruments are considered to be related to the job and can be used to select individuals for the job in the future

9

SO2: Some basic terms SO2: Terms related to correlation

r = correlation coefficient x = selection test/instrument y = measure of job performance rxy = validity correlation coefficient; that is, the correlation

between a selection test and measure of job performance rxx = reliability correlation coefficient; that is, the correlation

between two administrations of the same test or two tests that measure the same thing (alternate forms of the same test)

10

SO3: Some basic terms, validity SO3: Terms related to validity

Predictor = selection test/instrument; you use the score on the selection test to predict job performance

Criterion = measure of job performance

11

SO4A: Elements of a correlation 4A. Two elements of a correlation coefficient

Magnitude: how strong the relationship is Sign, + or -: whether the relationship is positive or negative

4A. Magnitude and sign Correlations go from -1 to +1 -1 indicates a strong negative relationship +1 indicates a strong positive relationship 0 indicates there is no relationship

How would you rank order the following correlations in terms of magnitude? -.20, +.05, +.15

12

SO4B: Inverse relationship 4B. If there was an negative or inverse relationship

between the scores on a social skills test and performance measures for computer programmers, what would that mean?

(next slide for diagrams of positive/negative relationships)

13

SO5: Fairly high positive, fairly high negative and zero relationship between test scores and measures of performance

High positive relationship People with good test scores perform well People with poor test scores don’t perform well Thus, if you knew a person’s test score but you

didn’t know what his performance score is, you could make a good guess what his performance is

High negative relationship People with good test scores don’t perform well People with poor test scores perform well Once again if you knew a person’s test score,

you could guess what his performance was Zero relationship

Some people with good test scores perform well but just about as many do not perform well

Some people with poor test scores perform well but just about as many do not perform well

If you know a person’s test score, but don’t know the person’s performance score, you could not guess what his performance was

Test

Per

form

ance

Low

Low

High

High

Test

Per

form

ance

Low

Low

High

High

TestP

erfo

rman

ce

Low

Low

High

High

14

SO6: NFE, but possible confusion You determine the validity of a test using current

employees Administer the test to them and then collect measures of

performance and correlate them If the correlation coefficient is statistically significant, we

conclude that the test is job related You then administer the test to a group of job

applicants You now have scores from the test for the applicants but you

do not have measures for job performance (you haven’t hired them yet)

You use the scores from the test to predict how well the person will do on the job, based on the validity coefficient from your current employees

15

SO7: Statistical significance

The correlation between the test scores and the performance measures must be statistically significant at the .05 level in order for the selection test to be considered a valid predictor of job performance.

If it is not, then the selection test is not considered to be a valid predictor and you should not use it to select applicants.

16

SO8: Statistical significance Descriptive vs. inferential statistics

Assume you have ten current employees.

You administer a test to them and correlate the test scores with a measure of job performance.The resulting correlation is .50.If we are concerned only with the performance of these particular 10 employees, we can accept this correlation as a completely accurate description of the degree to which the test scores are related to their job performance measures. (descriptive statistics)However, in selection we are not just interested in these particular 10 employees. Rather, we want to know if we can use the test scores to predict the job performance of others (future applicants). (inferential statistics)

(for those of you who just had 6340, this should be easy – the book is a little misleading- not wrong, but misleading)

17

SO8: Statistical significanceThe question becomes: Is the test related to job performance for all potential employees (the entire population of employees), not just for your particular 10 employees (the sample). Your ten employees constitute only a very small sample of that whole “population” of potential employees. Clearly if we took another 10 employees, administered the test to them and correlated the scores with their job performance measures, the correlation would not be the same - it might be higher, it might be lower.Given that the correlation would not be the same for another group of employees, how do we know that the test is actually valid? That is, is actually related to performance?That is what statistical significance tells us.The question asked is rather simple: Given the correlation (.50) we obtained with our particular sample (our 10 employees), what are the chances that the real correlation between the test and performance measure is actually zero?

18

SO8: Statistical signficance What we mean when we say that a correlation is significant at the

.05 level (three critical parts): The chances are not greater than 5 out of 100 that the correlation for the

whole population of employees is zero given that We obtained the correlation we did (in my example, .50) or larger For our sample which contained a specific number of individuals (in my

example, 10 individuals)

In other words, what are the chances we are wrong? What are the chances that the validity coefficient for the entire population of employees is really zero, given that we obtained a correlation coefficient of .50 based on our 10 employees?

If our correlation of .50 was significant at the .01 level,what would that mean?

(click for question)

19

SO8: Statistical significance To determine whether a correlation is statistically significant for

the number of employees in your sample, you consult a statistical significance table (I have provided a sample at the end of the study objectives)

In order for a correlation coefficient to be statistically significant at the .05 level with a sample size of 10, the correlation must be at least .63

Thus, my correlation is not statistically significant The chances are greater than 5 out of 100 that we are wrong; that

is, the chances are greater than 5 out of 100 that the actual correlation between the test and the performance measure for the population of employees is actually zero

Thus, we must conclude that the test is not job related and will not predict the job performance of applicants

It is NOT valid

20

SO9: What statistical significance does not mean9A Statistical significance tells us nothing about the real magnitude or

size of the correlation

It does not mean that the true correlation between the test and performance scores is the correlation you obtained with your sample or even approximates that correlation

It simply means that there is a 95% probability that the correlation is not zero.

9B It does not mean that if you correlated the test scores and performance measures for different samples, there is a 95% probability that you would obtain the same correlation (in my example, .50)

It simply means that there is a 95% probability that the correlation is not zero.

(Assume, .50 correlation that was statistically significant at .05)

21

SO11: Sample size and reliability of the correlation11A A correlation coefficient is less reliable with small

sample sizes. What does this mean?

The size of the correlation is going to vary more if your sample size is small; it will be less stable from sample to sample

That is, if you correlated the test scores with performance measures for four groups of 10 employees each, the size of the correlation is likely to be quite different for the four groups, and differ more in size than if you correlated the test scores with performance scores for four groups of 50 employees each.

22

SO11: Sample size and reliability of the correlation11B Why are correlations less reliable with small sample

sizes?

A larger sample means the correlation you obtain is going to be more reliable because you are sampling a greater number of individuals from the population. With smaller samples, the correlation is going to differ more from sample to sample because of sampling errors - you may have one or two “unusual” cases.

For example, assume that your total population is 100 (not theoretically possible or correct).

If you correlate the test scores with the performance scores for 90 of those individuals, you would expect a more reliable correlation than if you correlated them with a sample of 5, 10, or even 50.

23

SO12: Statistical significance and size of the sample

As the sample size decreases, the correlation required to achieve significance increases. Why?

Because correlations based on small sample sizes are unreliable. The size of the correlation is going to vary more across samples if you use a small sample size.

Because of that variation, the magnitude of any one correlation coefficient from any one sample must be larger to be statistically significant to compensate for the fact that the correlation from that sample may, indeed, be wrong.

More technically, the correlation may not be representative of the true correlation for the entire population.

(highly related to the preceding material; first sentence is not adequate for the exam)

24

NFE: Statistical significance and sample size

While reliability coefficientsoften range from .80 to themid .90s, validity coefficientsrarely exceed .50. They often range from .30-.50, but can even be much lower than that.

25

SO13: Sample size and validity coefficients Regardless of the reason, what is wrong with a small

sample size when correlating test scores with performance measures?As the sample size decreases, the probability of not finding a statistically significant relationship between the test/predictor and the criterion (performance measure) increases.Thus, you are much more likely to conclude that your test is not valid and hence not useful, when in fact it may well be.

26

SO14: Study by Schmidt Frank Schmidt correlated scores from a clerical test

with performance measures for 1,500 post office letter sorters The correlation for the entire sample was .22 The correlation was statistically significant

He and his colleagues then divided this sample up into 63 groups of 68 individuals each (68 = most common size of group for a validation study) Validity coefficients ranged from -.03 to .48! Less than a third were statistically significant!

(terrific study! Demonstrates how size of the correlation can vary from sample to sample; Frank Schmidt is one of THE names in selection; click, implications; valid when it is not: ~.25 correlation, sig at .05 level for 68; next slide - reliability)

Validity coefficients may be very misleading with small (?) sample sizes and lead to the conclusion that your test is not valid when in fact it is or vice versa!!

27

Reliability

28

SO15: Reliability Fundamental definition

The degree of stability, dependability, or consistency, scores on a measure (either the test or the performance measure)

(NFE) Essence of ReliabilityTo what extent does the score reflect the person’s ability vs. the extent to which the score reflects measurement error Is the instrument accurately measuring the KSA it is supposed to be measuring? Does the person’s score accurately reflect his/her competence with respect to

what is being measured?

29

SO15: NFE but confusion about reliability

Reliability is a theoretical concept that must be operationally defined

Because of that, there are different ways to assess itIn behavior analysis, for example, interobserver agreement is a form of reliability: are you consistently and accurately measuring the behavior you say you are measuring? Are your definitions of behavior adequate? Are your observers accurately measuring the behavior? Are you using the right sampling procedure?

Frequency count, whole interval, partial interval, time sampling? The data you obtain consists of the “true” measure of

behaviors and the “errors” that creep in because of measurement error due to the above (related to SO16) Just as in selection you can conceive of your data having two

“parts”: True measure of behavior + the error

30

SO15: Reliability With respect to selection instruments, there are three

primary ways to operationalize “reliability” Stability Dependability Consistency

31

SO15: Reliability Stability

Does the person get approximately the same score if he/she takes the test several times?

DependabilityDoes the test accurately sample the relevant content? That is, is it measuring what it is supposed to be measuring?For example, does a math test give an accurate indication of a person’s mathematical ability or is there something wrong with some of the items on the test?

ConsistencyAre the items on the test measuring the same thing?Do all of the items on a mechanical ability test measure mechanical ability?

32

Introduction: NFE Four basic ways to assess reliability

Test-retest, with a time delay in between Parallel forms, no time delay Parallel forms, with a time delay in between Internal consistency, split half reliability

33

SO17: Test-retest reliability

17A: Test-retest reliability, what is it? The same test is administered twice to the same individuals,

with a time interval in between The scores are then correlated

17B: Resulting coefficient is called what, and why? coefficient of stability It measures how stable the scores are on that test over time A KSA should remain stable, given that no learning has taken

place 17C: What does it indicate?

How stable the score is over time

34

SO18: Test interval for test-retest method

18A: Why is an interval that is too short inappropriate? Memory - the person can remember the items and how

he/she responded the first time

18B: Will an interval that is too short underestimate or overestimate reliability? Why? Overestimates it A person is likely to get the same or a similar score because

he/she remembers the items, not because the test shows good stability over time

35

SO19: Test interval, for test-retest method SO19: In general how long should the interval be?

Several weeks (3-4 weeks) to several months However, long intervals (6 months or so) can also get you

into trouble

36

SO20: Test interval, for test-retest method

20A: Why is an interval that is too long inappropriate? Learning may occur during the interval - the person’s KSA may

actually change during that time period 20B: Will an interval that is too long underestimate or

overestimate reliability? Why? Underestimates it A person is going to score differently on the test because his/her

competency on the KSA has changed, not because the score on the test is not stable over time If the person hadn’t acquired more competency, the person may

have gotten the same score Also relevant to the alternate or parallel form method of

reliability if an interval is used

(math ability - may have had a class in math)

37

SO21: Test-retest reliability Test-retest reliability is appropriate if you are interested in

whether a measure is stable over time If a measure has high test-retest reliability (.85 or above),

you can conclude that the test is free from error associated with passage of time

*If a measure has low test-retest reliability (below .85), however, you would not know whether The test actually has low reliability - test suffers from

error due to passage of time The low correlation is due to the fact that the KSA

being measured has actually changed (and hence your test may actually be reliable)

*this part, NFE

38

SO22: Parallel forms reliability Parallel/alternate/equivalent forms reliability, what it is?

Two different tests that measure the same thing are administered to the same individuals with no (or a very short) time interval or a time interval in between

Two arithmetic tests that are designed to measure the same thing but have different problems

Two clerical proofreading tests that are designed to measure the same thing but have different items

How is the reliability determined? Correlate the test scores from the two tests

39

SO22, cont: Parallel forms reliability

If no time interval, or a short interval, what is the reliability coefficient called? Why? Coefficient of equivalence It indicates the consistency with which the KSA is measured by the two

instruments Conceptually, it tells you whether your test is actually measuring what it is

supposed to be measuring - the underlying KSA being assessed by the two measures

If the coefficient is high (.85 or higher): add this for the exam You can conclude that the two tests are consistently measuring what they

are supposed to be measuring

40

SO23: Parallel forms with a time interval in between reliability

What is the reliability coefficient called? Why? Coefficient of equivalence and stability It indicates the consistency with which the KSA is measured by

the two instruments It also indicates whether the scores are stable over time

(small warning – students often miss this when I ask it on the exam; another slide on this)

41

SO23: Parallel forms with a time interval in between reliability

If the coefficient is high (.85 or higher): You can conclude that the two tests are consistently measuring

what they are supposed to be measuring AND The scores are stable over time

If the coefficient is low, however, you don’t know whether: The two tests are not equivalent - they are not measuring the

same thing but again you don’t know which test is not measuring what it is supposed to be measuring (or whether neither is measuring what it is supposed to be measuring)

The scores are not stable over time Some combination of the above

(if things work out, you know more than just test-retest or parallel forms w/o interval, but if not, then you are left wondering what the problem is)

42

SO25: Parallel forms vs. Test-retest In general, does parallel form method tend to

underestimate or overestimate reliability? Tends to underestimate it

Why? In practice, it is VERY difficult to develop two identical tests

Which method is better? If you can obtain equivalent forms, parallel form is almost

always preferred Why?

Because scores would be the same if individuals took an equivalent test at a different time That is, the test is measuring what you think it is, and the

scores are stable over time

43

SO26: Internal consistency What is internal consistency and what does it show ?

It shows the extent to which items on the same are measuring the same thing

Let’s say you have an arithmetic test with 10 itemsIf each item is truly measuring a person’s arithmetic ability, and the person gets one of the problems right, he/she should, theoretically, get of the other nine right as wellOn the other hand, if he/she misses one of the problems, he/she should miss the other nine as well

(next slide on this as well)

44

SO26: Internal consistency

Internal consistency is only good for unidimensional tests - that is, for a test in which all of the items are supposed to be measuring the same thing

It is not appropriate for multidimensional tests - tests that measure different KSAs in one test Why? A person might do well on one KSA, but not the

other because of his/her different competencies on the two KSAs

(last slide on this)

45

SO27: Statistical interpretation of a reliability coefficient

Let’s assume you administered the same exam to the same individuals with an interval in between and correlated the scores

The resulting correlation coefficient is .90 How is that statistically interpreted?

90% of the differences in the scores between the individuals who took the test is due to “true” differences in ability, while 10% is due to measurement error

46

SO28: Minimum and preferred reliability correlation coefficients

Minimum = .85 Preferred = at least .90 Why?

You are correlating a measure with itself If the measure does not correlate with itself, it cannot

correlate with something else (job performance) As you will see next unit, if a test is not reliable it cannot be

valid (although it can be reliable and not be valid) That is, if the test is not reliable it cannot be related to the

job and you cannot use it to select applicants

(authors don’t give a figure; depends on the situation – rule of thumb)

47

SO29: Generally, how do differences between individuals affect reliability estimates

In general, the greater the differences between individuals on the KSA being measured, the higher the correlation

This may seem counterintuitive, but remember in order to have a high positive correlation: High performers must perform well on both tests Middle performers must perform middling on both tests Low performers must perform low on both tests

Thus, you need to have a range of scores (high, medium, and low) in order to get a strong correlation

Anything that restricts/reduces the range of scores on either test will, in general, decrease the magnitude of the correlation

(example on the next screen)

48

You administer a math test to high school students, community college students, and college engineering students

You re-administer the same math test to the same individuals

The high school students score relatively poorly on both administrations of the test, the cc students middling, while the college engineering students score much better on both administrations of the test

When you plot the scores you get the diagram on the right, which represents a high positive correlation

Now, let’s give the test to only college engineering students and redraw the diagram (assume the circled data points represent the scores of the engineering students in both diagrams)

You still have a low positive correlation between the two test administrations, but it is not as strong or nearly as high of a correlation

Test, Time 1

Te

st,

Tim

e 2

Low

High

Low High

Te

st,

Tim

e 2

Low

High

Test, Time 1

Low High

29B: Why from a statistical perspective?

Correction to the study objective: the answer is in 130,1 not 131,1.

Variation between individuals represents differences in the ability

Variation within an individual represents error Therefore, if variation increases between

individuals while variation within an individual stays the same, the correlation/reliability will be higher

49

50

SO30: Length of the test and reliability estimates In general, as the length of the test increases, so too will

the reliability. Why?Think of a test that is designed to measure mathematical ability. The items on the test are only a sample of all possible items. If you have 5 math problems, a person may miss one just because of error (i.e., misread a 2 as a 5, or made a “stupid” error because he/she was hurrying, etc.). The more problems you have, the more likely it is that the person’s score will actually represent his/her “true” ability; he/she can make one or two errors “by mistake” without having it affect the person’s overall score on the exam as much.

Behavior analysis analogyWith within-subject data, the more data points you have for an individual during each phase, the more confident you are that the data actually represent the person’s true performance under that condition, not simply momentary fluctuations due to unknown factors in the environment

51

SO31: Difficulty of test items and reliability estimates Test questions of moderate difficulty (about 50% of

test takers answer them correctly) will result in higher reliability estimates

Why? Basically the exact same issue we have been dealing with If the test items are too easy, most people will answer them

correctly (no low scores) If the test items are too difficult, most people will answer

them incorrectly (no high scores) Thus, you will not have a range of scores on the test

GREs, SATs are designed so VERY few individuals get all of the items correct

Again, the diagrams from SO29 are relevant

(diagrams on next slide)

52

Top diagram represents a situation where the test items are of moderate difficulty

Thus, you get a range of low, medium, and high scores

Bottom diagram represents a situation where the test items are too easy

Everyone gets a very high score Could actually end up with a zero

correlation, or close to zero

Test, Time 1

Te

st,

Tim

e 2

Low

High

Low High

Te

st,

Tim

e 2

Low

High

Test, Time 1

Low High

(last slide)

53

THE END!!

QUESTIONS??