developing evaluation instruments. evaluations formative – how are we doing? summative – how did...
TRANSCRIPT
Developing Evaluation Instruments
Evaluations
• Formative – How are we doing?
• Summative – How did we do?
• Confirmative – How are we still doing?
Formative Evaluation
• Even excellent plans and concepts can be improved– Gauge success by classroom program
presentation– Informs the instructor how well instructional
package is serving the objectives as it progresses
Formative Evaluation
• Program of instruction is evaluated through– Test results– Learner reactions/comments– General observation of learners during class– Subject-matter expert reviews– Colleague suggestions
• Quality control of the development process
Formative Evaluation
• Questions that may be asked– Level of learning acceptable compared to
objectives?– Learners using knowledge to perform skills?– Time element acceptable?– Activities appropriate?– Tests measuring objectives?
Summative Evaluation
• Measures success of major outcomes
• Usually follows all instruction, projects and testing in a program
Summative Evaluation
• Measures– Learning efficiency (material mastered/time)– Program cost (Development and delivery)– Continuing expenses– Student comments/reactions/evaluations– Long-term benefit of program
Confirmative Evaluation
• Although instruction initially effective, some problems appear over time
• Tracks learner experiences over time to judge program’s validity
• Relies upon numerous data-collection instruments (interviews, performance assessments, etc.)
Confirmative Evaluation
• Questions to investigate– Do learners continue to perform correctly over
time?– Do materials still meet original objectives?– How has technology and/or resources, trends,
and attitudes changed since inception? – How to best meet clients’ needs over time?
Confirmative Evaluation
– Changes needed in materials? What are the costs?
– If instruction not working as well as before:• Should instruction continue as is?
• Should it be revised?
• Should it be terminated?
• What might replace it?
Validity
• The test assesses what it is supposed to measure – For example:
• Performance tests assess processes and outcomes relating to skills or competencies
• Course attitude surveys need to measure reactions to the course
Reliability
• The test will produce consistent results whenever used
• More questions relating to specific objectives = more reliable test
• Standardized delivery of test
• Scoring methods – the less subjective, the better
Pretesting
• Judges the learner’s preparation to study the material (course or topic)
• Determines competencies already mastered
Testing for Prerequisites
• Standardized paper-and-pencil test
• Performance observations
• Questionnaire (Looks at learner background, training, and experiences)
• Review of past work experience
• Interview past supervisors/managers
Objective tests
• Multiple choice
• True or False?
• Matching items
Constructed-Response
• Make learner plan answers and express them in their own words
• Short-answer
• Essay questions
• Problem solving
Skills and Behavior
• Actions observed
• Process, product, or both?
• Constraints/limitations?
• Testing conditions simulated or realistic?
The Rubric
• General assessment of overall product
• Certain elements attain points
• More objective tool
From Designing Effective Instruction 4th Ed., Morrison, Ross, & Kemp, 2004.