emily h. wughalter, ed.d. measurement & evaluation spring 2010
TRANSCRIPT
![Page 1: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/1.jpg)
Emily H. Wughalter, Ed.D.
Measurement & Evaluation
Spring 2010
![Page 2: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/2.jpg)
Reliability means the consistency in the measurement from one testing to the next trial.
![Page 3: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/3.jpg)
Validity means does the measure or the score measure what it purports (is supposed) to measure.
![Page 4: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/4.jpg)
For a measure to be valid it must be reliable; however, a measure can be reliable without it being valid.
![Page 5: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/5.jpg)
Xobserved
XTrue
XError
![Page 6: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/6.jpg)
Xobserved = Xtrue+XError
![Page 7: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/7.jpg)
Testing environment should be favorable Testing environment should be well-
organized Administrator should be competent to run a
favorable testing environment Range of talent Motivation of the performers
![Page 8: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/8.jpg)
Good day vs bad day Learning, forgetting, and fatigue Length of the test Test difficulty Ability of the test to discriminate Number of performers
![Page 9: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/9.jpg)
Nature of measurement device Selection of scoring unit Precision Errors in measurement Number of trials Recording errors
![Page 10: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/10.jpg)
Classroom management Warm-up opportunity
![Page 11: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/11.jpg)
Test-Retest (Stability Reliability) Internal Consistency
![Page 12: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/12.jpg)
Difference or change scores should be avoided because of ceiling and floor effects.
Difference scores are highly unreliable scores
![Page 13: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/13.jpg)
Objectivity means interrater reliability or consistency from one person (rater) to another.
![Page 14: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/14.jpg)
A criterion score represents the score that will be used to represent an individual’s performance The criterion score may be the mean of a series
of trials. The criterion score may be the best score for a
series of trials.
![Page 15: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/15.jpg)
When selecting a criterion measure, whether it is the best score, the mean of all the trials score, or the mean of the best 2 or 3 trials score, a researcher must determine which of the criterion measures represents the most reliable and most valid score.
![Page 16: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/16.jpg)
Inconsistent scorers Inconsistent performance Inconsistent measures Failure to develop a good set of instructions Failure to follow testing procedures
![Page 17: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010](https://reader035.vdocument.in/reader035/viewer/2022062304/56649e555503460f94b4cf93/html5/thumbnails/17.jpg)