validity & realibility

Upload: aidilubaidillah

Post on 06-Apr-2018

225 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/3/2019 Validity & Realibility

    1/13

    ReliabilityThe consistency of scores obtained from oneinstrument to another or from the sameinstrument over different groups.

    Test retestEquivalency

    Inter rater

    Internal consistency

    Split half

  • 8/3/2019 Validity & Realibility

    2/13

    Test retest reliability Method for testing the same subject at

    a later date, ensuring that there is acorrelation between the results.

    An educational test taken after amonth should yield the same resultsas the original.

  • 8/3/2019 Validity & Realibility

    3/13

    Equivalency reliability method which two items measure

    identical concepts at an identical levelof difficulty.

    is determined by relating two sets oftest scores to one another to highlightthe degree of relationship.

  • 8/3/2019 Validity & Realibility

    4/13

    Inter-rater reliability Method use two or more independentjudges score the test.

    The scores are then compared to

    determine the consistency of the raterestimates.

    One way to test inter rater reliability is

    to have each rater assign each testitem a score.

  • 8/3/2019 Validity & Realibility

    5/13

    Internal consistencyreliability Method used to judges the consistency of

    result across items on the same test.

    Compare test items that measure the same

    construct to determine the tests internalconsistency.

    Example, when a question is very similar toanother test question, it may indicate that

    the two questions are being used togauge(measure) reliability.

  • 8/3/2019 Validity & Realibility

    6/13

    Split-half reliability A measure of consistency where a test

    is split into two and the scores foreach half of the test is compared with

    one another. If the test is consistent, it leads the

    experimental to believe that it is most

    likely measuring the same thing.

  • 8/3/2019 Validity & Realibility

    7/13

    Factors affecting

    reliability Test length - the longer a test is, the more

    reliable Speed - when a test is a speed test, reliability

    can be problematic. Group homogeneity - the more heterogeneous

    the group of students who take the test, the morereliable the measure will be.

    Item difficulty - reliability will be low if a test is soeasy or difficult.

    Objectivity - objectively scored tests show a

    higher reliability than subjectively scored tests. Test-retest interval -the shorter the time interval

    between two administrations of a test, the higherthe reliability will be.

    Variation with the testing situation. Errors in the

    testing situation can cause test scores to vary.

  • 8/3/2019 Validity & Realibility

    8/13

    Relationship between

    validity and reliability Reliability is an essential component

    of validity but on its own, is notsufficient measured of validity.

    A test can be reliable but not valid,whereas a test cannot be valid yetunreliable.

  • 8/3/2019 Validity & Realibility

    9/13

    Validity Validityrefers to the accuracy of an assessment --

    whether or not it measures what it is supposed tomeasure. Even if a test is reliable, it may not provide avalid measure. Lets imagine a bathroom scale thatconsistently tells you that you weigh 130 pounds.

    The reliability (consistency) of this scale is very good,but it is not accurate (valid) because you actually weigh145 pounds (perhaps you re-set the scale in a weakmoment)

    Since teachers, parents, and school districts makedecisions about students based on assessments (suchas grades, promotions, and graduation), the validityinferred from the assessments is essential -- even morecrucial than the reliability.

  • 8/3/2019 Validity & Realibility

    10/13

    Example 1 :

  • 8/3/2019 Validity & Realibility

    11/13

    Example 2:

  • 8/3/2019 Validity & Realibility

    12/13

    Type of Validity Definition Example

    Content The extent to which the

    content of the test matches

    the instructional objectives.

    A semester or quarter exam

    that only includes content

    covered during the last six

    weeks is not a valid measureof the course's overall

    objectives -- it has very low

    content validity.

    Criterion The extent to which scores on

    the test are in agreement with(concurrent validity) or predict

    (predictive validity) an external

    criterion.

    If the end-of-year math tests

    in 4th grade correlate highlywith the statewide math tests,

    they would have high

    concurrent validity.

    Construct The extent to which an

    assessment corresponds toother variables, as predicted

    by some rationale or theory.

    If you can correctly

    hypothesize that the studentswill perform differently on a

    reading test than English-

    speaking students (because

    of theory), the assessment

    may have construct validity.

  • 8/3/2019 Validity & Realibility

    13/13

    Factors thatinfluences

    validityassessment

    Nature ofthe group

    Moderatorvariables

    Criterion-predictor

    relationship

    Consistency of the validitycoefficient for subgroups whichdiffer in any characteristic (e. g.age, gender, educational level,

    etc, )

    There must be a linearrelationship between predictor

    and criterion.

    Variables like age, gender,personality characteristics mayhelp to predict performance forparticular variables only