Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

17
Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Transcript of Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Page 1: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Emily H. Wughalter, Ed.D.

Measurement & Evaluation

Spring 2010

Page 2: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Reliability means the consistency in the measurement from one testing to the next trial.

Page 3: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Validity means does the measure or the score measure what it purports (is supposed) to measure.

Page 4: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

For a measure to be valid it must be reliable; however, a measure can be reliable without it being valid.

Page 5: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Xobserved

XTrue

XError

Page 6: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Xobserved = Xtrue+XError

Page 7: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Testing environment should be favorable Testing environment should be well-

organized Administrator should be competent to run a

favorable testing environment Range of talent Motivation of the performers

Page 8: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Good day vs bad day Learning, forgetting, and fatigue Length of the test Test difficulty Ability of the test to discriminate Number of performers

Page 9: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Nature of measurement device Selection of scoring unit Precision Errors in measurement Number of trials Recording errors

Page 10: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Classroom management Warm-up opportunity

Page 11: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Test-Retest (Stability Reliability) Internal Consistency

Page 12: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Difference or change scores should be avoided because of ceiling and floor effects.

Difference scores are highly unreliable scores

Page 13: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Objectivity means interrater reliability or consistency from one person (rater) to another.

Page 14: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

A criterion score represents the score that will be used to represent an individual’s performance The criterion score may be the mean of a series

of trials. The criterion score may be the best score for a

series of trials.

Page 15: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

When selecting a criterion measure, whether it is the best score, the mean of all the trials score, or the mean of the best 2 or 3 trials score, a researcher must determine which of the criterion measures represents the most reliable and most valid score.

Page 16: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.

Inconsistent scorers Inconsistent performance Inconsistent measures Failure to develop a good set of instructions Failure to follow testing procedures

Page 17: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.