series. institutionresearch in washington, d.c. formerly maryland's state director of student...

49
ED 429 989 AUTHOR TITLE INSTITUTION ISBN PUB DATE NOTE AVAILABLE FROM PUB TYPE EDRS PRICE DESCRIPTORS ABSTRACT DOCUMENT RESUME TM 029 692 McTighe, Jay; Ferrara, Steven Assessing Learning in the Classroom. Student Assessment Series. National Education Association, Washington, DC. ISBN-0-8106-2070-7 1998-00-00 48p.; For an earlier related document, see ED 393 870. NEA Professional Library, Distribution Center, P.O. Box 2035, Annapolis Junction, MD 20701-2035; Tel: 800-299-4200 (Toll Free). Guides Non-Classroom (055) EDRS Price MF01 Plus Postage. PC Not Available from EDRS. *Educational Assessment; *Evaluation Methods; Feedback; *Student Evaluation; *Teacher Made Tests; *Test Construction; Test Use The premise of this book is that the primary purpose of classroom assessment is to inform teaching and improve learning, not to sort and select students or to justify a grade. The common principles of effective assessment are covered, and the strengths and limitations of a variety of effective approaches are reviewed. Vignettes illustrate classroom assessment in action, and guiding questions and a framework for planning classroom assessments supplement each discussion. The following chapters are included: (1) "Ongoing Assessment of Student Learning"; (2) "Large-Scale versus Classroom Assessment"; (3) "Effective Classroom Assessment"; (4) "Assessment Approaches and Methods"; (5) "Evaluation Methods and Rules"; and (6) "Communication and Feedback Methods." A glossary is included. (Contains 23 references.) (SLD) ******************************************************************************** Reproductions supplied by EDRS are the best that can be made from the original document. ********************************************************************************

Upload: others

Post on 04-Jun-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

ED 429 989

AUTHORTITLE

INSTITUTIONISBNPUB DATENOTEAVAILABLE FROM

PUB TYPEEDRS PRICEDESCRIPTORS

ABSTRACT

DOCUMENT RESUME

TM 029 692

McTighe, Jay; Ferrara, StevenAssessing Learning in the Classroom. Student AssessmentSeries.National Education Association, Washington, DC.ISBN-0-8106-2070-71998-00-0048p.; For an earlier related document, see ED 393 870.NEA Professional Library, Distribution Center, P.O. Box2035, Annapolis Junction, MD 20701-2035; Tel: 800-299-4200(Toll Free).Guides Non-Classroom (055)EDRS Price MF01 Plus Postage. PC Not Available from EDRS.*Educational Assessment; *Evaluation Methods; Feedback;*Student Evaluation; *Teacher Made Tests; *TestConstruction; Test Use

The premise of this book is that the primary purpose ofclassroom assessment is to inform teaching and improve learning, not to sortand select students or to justify a grade. The common principles of effectiveassessment are covered, and the strengths and limitations of a variety ofeffective approaches are reviewed. Vignettes illustrate classroom assessmentin action, and guiding questions and a framework for planning classroomassessments supplement each discussion. The following chapters are included:(1) "Ongoing Assessment of Student Learning"; (2) "Large-Scale versusClassroom Assessment"; (3) "Effective Classroom Assessment"; (4) "AssessmentApproaches and Methods"; (5) "Evaluation Methods and Rules"; and (6)"Communication and Feedback Methods." A glossary is included. (Contains 23references.) (SLD)

********************************************************************************

Reproductions supplied by EDRS are the best that can be madefrom the original document.

********************************************************************************

Page 2: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

C7100

"Fr vgsf

A

LI 2 r I O.

PERMISSION TO REPRODUCE ANDDISSEMINATE THIS MATERIAL IN MICROFICHE,

AND IN ELECTRONIC MEDIA FOR ERICCOLLECTION SUBSCRIBERS ONLY,

HAS BEEN GRANTED BY

"..1 TO THE EDUCATIONAL RESOURCES2 A INFORMATION CENTER (ERIC)

LI

U S DEPARTMENT OF EDUCATIONOffice of Educational Research and Improvement

EDUCATIONAL RESOURCES INFORMATIONCENTER (ERIC)

Iff"Tri's document has been reproduced asreceived from the person or organizationoriginating it

0 Minor changes have been made toimprove reproduction quality

Points of view or opinions stated in thisdocument do not necessarily representofficial OERI position or policy.

BEST COPY AVAILABLE

1 2

neaNAliONAL EDUCATION ASSOCIATION

Page 3: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

AS7La/

C LA S-,C00h\

Page 4: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

The Authors

Jay Mc Tighe directs the Maryland Assessment Consortium, a collaboration amongMaryland's school systems to develop and share performance assessments with teachersthroughout the state. He is well known for his work with "thinking skills," having coordinat-ed statewide efforts to develop instructional strategies, curriculum models, and assessmentprocedures for improving the quality of student thinking.

Steven Ferrara is director of research and psychometrics at the American Institutes forResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, hehas taught at the preschool, middle school, high school, and graduate levels. He has spent thelast 12 years conducting research, writing, teaching, and speaking on the effects of large-scaleand classroom assessment on teaching and learning.

Authors' AcknowledgmentsThe authors wish to thank the following individuals for their comments on and critique of

various drafts of this document: Gail Goldberg, Joan Heiss, Mary Kimmins, HillaryMichaels, Ruth Mitchell, and Barbara Reeves. They also wish to thank the many users of theFramework of Assessment Approaches and Methods in an earlier version of this book fortheir helpful comments on using the Framework.

J.M. and S.F.

Page 5: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Student Assessment Series

Jay McTighe and Steven Ferrara

Glen W. CutlipSeries Editor

neaNA110NAL EDUCATION ASSOCIATION

Page 6: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Copyright © 1998National Education Association of the United States

NoteThe opinions expressed in this publication should not be construed as representing the

policy or position of the National Education Association. These materials are intended to bediscussion documents for educators who are concerned with specialized interests of the pro-fession.

Reproduction of any part of this book must include the usual credit line and copyrightnotice. Address communications to Editor, NEA Teaching and Learning.

Library of Congress Cataloguing-in-Publication DataMcTighe, Jay.Assessing learning in the classroom / Jay Mc Tighe and Steven Ferrara.p. cm. (Student assessment series)Includes bibliographical references (p. ).ISBN 0-8106-2070-71. Achievement testsUnited States. 2. Educational tests and measurementsUnited

States. I. Ferrara, Steven. II. Title. III. Series.LB3060.3.M38 1998 98-28469371.27'1'0973dc21 CIP

s

Page 7: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

CONIEN.TS

Introduction vii

I. Ongoing Assessment of Student Learning 1

II. Large-Scale versus Classroom Assessment 4

III. Effective Classroom Assessment 6Content Standards, Purpose, and Audience 9

IV. Assessment Approaches and Methods 11

Selected-Response Format 11

Constructed-Response Format 14

V. Evaluation Methods and Roles 21

VI. Communication and Feedback Methods 29

Afterword 32

Glossary 33

References 36

Page 8: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

INTRODUMONThe field of student assessmentfrom methodology and techniques to the use of

resultsis changing, and these changes are dramatically affecting the work of edu-cation employees.

On one hand, these changes have created new options. For example, classroomassessment instruments have expanded to include assessments based on portfolios,projects, and performances. Teachers now assess a student's performance based onpredetermined criteria more closely aligned with the instructional objectives of thelesson and tailor instruction more specifically to individual students. Studentsbecome partners with the teacher in assessment by having access to these criteria atthe beginning of the lesson. Classroom assessment is truly becoming an integralpart of the instructional program as more and more teachers add these assessmenttechniques to their repertoire.

On the other hand, changes in student assessment have created new concerns,especially in the use of assessment results. Today, assessment results are being usedfor more than comparing an individual student's performance against a state ornational norm, and for more than providing data for making program improvementdecisions. They are being used to determine the success or failure of teachers andschools. Policy makers and others are using large-scale assessments to decidewhether teachers and schools are providing an adequate education to all studentsand attaching consequences, positive and negative, on the basis of student assess-ment results. The use of student test scores has raised the stakes for all educationemployees.

Consequently, student assessment is part of every teacher's work. In fact, near-ly one-third of a classroom teacher's time is spent assessing and evaluating students.Many influential groups have identified competence in student assessment as essen-tial for the training and licensing of new teachers and the upgrading of the skills ofpracticing teachers (National Board for Professional Teaching Standards, InterstateNew Teacher Assessment Consortium, National Council for Accreditation ofTeacher Education, Educational Testing Service, and the National Association ofState Directors of Teacher Education and Certification). These groups estimate thatless than one-half of currently practicing teachers have received adequate trainingin student assessment.

To help members and other educators keep abreast of the ever-changing field ofstudent assessment, the National Education Association (NEA) commissioned lead-ing assessment experts to write about student assessment from their perspectives.Experts Jay Mc Tighe and Steven Ferrara, the authors of this book on classroom-based assessment of learning, believe that "the primary purpose of classroomassessment is to inform teaching and improve learning, not to sort and select stu-dents or to justify a grade." In this book, a revised edition of an NEA publicationprinted in 1994, they discuss principles of effective classroom assessment, illustratea variety of assessment approaches and methods, and provide a framework for plan-

VI I

Page 9: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

ning. Their readable, practical approach to an evolving and sometimes complex sub-ject is intended to be of use to teachers at all levels, preschool through graduatestudies, as well as to other education employees.

The NEA developed the Student Assessment Series to help teachers and othereducation employees improve their knowledge and skills in student assessment andhopes readers will find the series a valuable resource for current and future studentassessment practices.

Y111

Glen W. Cut lipSeries Editor

Page 10: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

I.ONGOING ASSESSMENTOF STUDENT LEARNING

Ongoing assessment of student learning in the classroom is an essential aspect ofeffective teaching. Teachers can use a variety of assessment methods to diagnosestudents' strengths and needs, plan and adjust instruction, and provide feedback tostudents and parents regarding progress and achievement. The basic premise of thisbook is that the primary purpose of classroom assessment is to inform teaching andimprove learning, not to sort and select students or to justify a grade.

The book is intended for teachers from the preschool to graduate school levels touse in examining a variety of methods for effectively and fairly assessing their stu-dents. While the choice of particular assessment methods will vary according to thepurpose of the assessment, the con-tent of the curriculum, and the agelevels of students, a set of commonprinciples underlies effective class- The primary purpose ofroom assessment. This book covers classroom assessment isthese principles, provides thestrengths and limitations of a vari- to inform teaching andety of assessment approaches, pre- improve learning, not tosents a series of vignettes to illus-trate classroom assessment in sort and select studentsaction, and offers a set of guiding Or tO justify a grade.questions and a framework forplanning classroom assessments toimprove teaching and learning.

Teachers frequently begin new units of study by introducing or reviewing keyvocabulary with the recognition that an understanding of certain basic concepts willenhance subsequent learning of important principles and procedures in the unit.Likewise, this book begins with a review of basic terminology commonly associat-ed with classroom assessment. (Additional assessment terminology is providedthroughout the book and definitions of related terms important to the topic of class-room assessment are provided in the Glossary.)

10

Page 11: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Assessment refers to "any systematic basis for making inferences about charac-teristics of people, usually based on various sources of evidence; the global processof synthesizing information about individuals in order to understand and describethem better" (Brown 1983). While considering this textbook definition, it is inter-esting to note that the term assessment is derived from the Latin root assidere mean-ing "to sit beside." Although this original meaning may seem at odds with present-day images of large-scale standardized testingnumber 2 pencils, "bubble" sheets,rigid time limits, silent work, and so onit conforms more closely with the arrayof assessment methods routinely used by teachers for assessing their students.Assidere suggests that, in addition to tests and projects, classroom assessmentsinclude informal methods of "sitting beside," observing, and conversing with stu-dents as a means of understanding and describing what they know and can do.

The terms assessment, testing,and evaluation are frequently usedinterchangeably, but they have dis-

It is interesting to notetinct meanings. Assessment is abroad term referring to the process that the term assessmentof gathering and synthesizing infor- is derived from the Latinmation to better understand anddescribe characteristics of people. root assidere meaning "toTesting is one type of assessment. sit beside."Tests generally utilize a paper-and-pencil format, are administered andtaken within established time limits, restrict test takers' access to resources (e.g.,reference materials), and yield a limited range of acceptable responses. Evaluationinvolves making a judgment regarding quality, value, or worth, based on set crite-ria. Teacher questioning, reviews of student work folders, and paper-and-penciltests are commonly used assessment methods for gathering information about stu-dent learning. Scoring a student essay and assigning report card grades are exam-ples of evaluation.

Another pair of widely used terms, summative assessment and formative assess-ment, pertain to the purpose and timing of classroom assessments. Summativeassessment refers to any culminating assessment that provides a summary report onthe degree of knowledge or proficiency attained at the conclusion of a unit, course,or program of study. A final exam, senior exhibition, or dissertation defense areexamples of summative assessments. Formative assessment refers to any ongoingdiagnostic assessment that provides information to help teachers adjust instructionand improve student performance. For instance, prior to the start of a unit on theCivil War, a teacher might ask students to make a "web" or an outline to show whatthey already know about this period of history as a means of obtaining informationabout students' prior knowledge. The teacher might also randomly select and inter-view several students to check their perceptions and awareness of the Civil War.Formative assessment also can be used during instruction to check on studentunderstandings and misconceptions. Teachers often use brief written and oral

11

Page 12: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

quizzes and classroom discussions to determine if students have learned coursematerial and can apply the skills they have been taught. Such activities provideteachers with valuable information that allows them to adjust instruction to improvestudent learning.

Although the term alternative assessment appears widely in the recent educationliterature (Herman, Aschbacher, and Winters 1992), there is no universally agreed-upon definition for the term alternative. Generally, alternative assessment is used torefer to those assessments that differ from the multiple-choice, timed, "one-shot"approaches that characterize most standardized and some classroom assessments.The term should be avoided as it is imprecise and open to various interpretations.

1 2

Page 13: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

H.LARGE-SCALE VERSUSCLASSROOM ASSESSMENT

Different types of assessments address different information needs. The purpos-es and audiences for assessment information influence what is assessed, how it isassessed, and how the results are communicated and used. Large-scale assessmentshave very specific purposes. For example, standardized tests, such as the Iowa Testsof Basic Skills (ITBS), California Achievement Tests (CAT), and the StanfordAchievement Test, are used primarily to satisfy the demands for educationalaccountability. The results of assessments such as these are reported to legislatures,boards of education, school administrators, parents, and the general public.Standardized tests are generally norm referenced to allow for easy interpretation.They are designed to determine how well students have learned particular conceptsand skills as compared to other students in a norming group. The results of norm-referenced assessments may be conveniently displayed so that observers can readi-ly distinguish achievement above or below the norm.

Not all large-scale standardized tests are norm referenced. Some, such as theCollege Board's Advanced Placement Examinations, the National Assessment ofEducational Progress (NAEP), and certain state-level competency tests, are criteri-on referenced. These tests evaluate and report student performance compared to pre-established standards.

Furthermore, not all standardized tests are multiple-choice in nature. Severalstates currently use standardized performance assessments, featuring open-endedtasks, for "high stakes" accountability purposes.

Standardized tests are considered "high stakes tests" if their results are used for con-sequential decisions such as promotion, graduation, admission, certification, evalua-tion, or where rewards and sanctions are involved. For example, a districtwide mini-mum competency exam would be "high stakes" for students if passing the exam is arequirement for a high school diploma. For an extended discussion of standardized test-ing, see The Role of High-Stakes Testing in School Reform (Smith 1993). Because theyare intended to provide accountability information, "one-shot" large-scale standardizedtests typically do not provide sufficiently detailed or timely information regarding stu-dent achievement of specific curriculum goals.

13

Page 14: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Classroom assessments serve other purposes and audiences. At the classroomlevel, teachers have different assessment needsdiagnosing student strengths andweaknesses, informing students and parents about progress, planning and adjustinginstruction, and motivating students to focus on valued knowledge and skills. Withthese purposes in mind, classroom assessments may be tailored directly to the cur-riculum and to the information needs of individual teachers, students, and parents.Unlike "one-shot" standardized tests, assessments designed to promote learning inthe classroom are more likely to be used over time, include an array of methods,focus on elements of quality, offer a more personalized picture of student achieve-ment, and provide timely and specific feedback.

Page 15: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

EFFECTIVE CLASSROOM ASSESSMENT

A large variety of methods is available to teachers for assessing student learning(Airasian 1991; Cross and Angelo 1988; Ferrara and McTighe 1992; Stiggins 1994).Regardless of the particular methods employed, effective classroom assessment isguided by three fundamental principles. Classroom assessment should: (1) informteaching and improve learning; (2) use multiple sources of information; and (3) pro-vide valid, reliable, and fair measurements.

The first principle is based on the premise that the primary purpose of classroomassessment is to inform teaching and improve learning (Mitchell and Neill 1992).Thus, effective classroom assessment must be an ongoing process instead of a sin-gle event at the conclusion of instruction. Rather than waiting until the end of a unitof study or course to assess students, effective teachers employ formative assess-ments at the beginning of instruc-tion to determine students' priorknowledge, and they assess regular-ly throughout the unit or course of Effective teachers employstudy to obtain information to helP formative assessments atthem adjust their teaching based on the beginning of instruc-the learning needs of students. Theyrecognize that assessment results tion .. and they assesscan inform them about the effective- regularly throughout theness of their teaching as well as thedegree of student learning, unit or course of study.

When using performance-basedassessments, teachers can maketheir evaluative criteria explicit in advance to serve as a focus for both instructionand evaluation. Effective teachers help their students understand that the criteriadescribe the desired elements of quality. They provide regular feedback to studentsbased on the identified criteria and allow students to revise their work based uponthis feedback. They also involve students in peer- and self-evaluation using the cri-teria in order to engage students more actively in improving their performance.

15

Page 16: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Assessment for learning recognizes the mutually supportive relationship betweeninstruction and assessment. Like a Möbius strip where one side appears to seam-lessly blend into the other, classroom assessment should reflect and promote goodinstruction. For example, teachers following a process approach to teaching writingwould allow their students to develop drafts, receive feedback, and make revisionsas part of the assessment. Likewise, if teachers teach science through a hands-on,experimental approach, their assessment should include hands-on investigations.

The second principle of sound classroom assessment calls for a synthesis ofinformation from several sources. The importance of using multiple sources ofinformation when assessing learning in the classroom can be illustrated through theanalogy of taking photographs. A single assessment, such as a written test, is like asnapshot of student learning. While a snapshot is informative, it is generally incom-plete since it portrays an individual at a single moment in time within a particularcontext. It is inappropriate to use one snapshot of student performance as the solebasis for drawing conclusions about how well a student has achieved desired learn-ing outcomes. Classroom assessment offers a distinct advantage over a large-scaleassessment in that it allows teachers to take frequent samplings of student learningusing an array of methods. To continue the analogy of taking photographs, class-room assessment provides an opportunity to construct a "photo album" containinga variety of pictures taken at differ-ent times with different lenses,backgrounds, and compositions.

The principle of multipleThe photo album reveals a richer,more complete picture of each stu- sources is especiallydent than any single snapshot can important when theprovide. Applying the principle ofmultiple sources is especially assessment information isimportant when the assessment used as a basis for mak-information is used as a basis for iimaking critical summative deci- ng crtical summativesions, such as assigning report card decisions, such as assign-grades or determining promotion. ing report card grades orThe third principle of classroomassessment concerns validity, relia- determining promotion.bility, and fairness. Validity refersto the degree to which an assess-ment measures what it was intended to measure. For example, to assess students'abilities to conduct research using primary and secondary sources, a media special-ist should observe students' use of these sources directly as they work on theirresearch projects. For this learning outcome, a paper-and-pencil test of studentknowledge of library references would be an indirect and less valid assessmentsince it does not reveal the ability to actually use the references purposefully.

Reliability refers to the dependability and consistency of assessment results. Ifthe same assessment yielded markedly different results with the same students

1 6 7.

Page 17: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

(without intervening variables such as extra instruction or practice time), one wouldquestion its reliability. Performance assessments present a special challenge sincethey call for judgment-based evaluation of student products and performances. Areliable evaluation would result in equivalent ratings by the same rater on differentoccasions. For instance, an observation checklist can be used reliably as long asteachers are careful to ensure that their ratings would not differ substantially fromoccasion to occasion (e.g., Monday morning versus Friday afternoon). When teach-ers are involved in school- or district-level evaluations based on a set of criteria usedthroughout the school or district, inter-rater reliability must also be considered. Inthis case, scores on a writing assessment would be considered reliable if differentraters assign similar scores to the same essays.

Fairness in classroom assessment refers to giving all students an equal chance toshow what they know and can do. Fairness is compromised when teachers assesssomething that has not been taught or use assessment methods that are incongruentwith instruction (e.g., asking for recall of facts when the emphasis has been on rea-soning and problem solving). The fairness of teacher judgments is also challengedby the "halo" and "pitchfork" effects, where expectations based on a student's pastattitude, behavior, or previous performance influence the evaluation of his or hercurrent performance.

Subtle, unintended racial, ethnic,religious, or gender biases also pre-sent roadblocks to the fair assess- Fairness is compromisedment of students. Such biases may when teachers assessnegatively influence students' atti-tudes toward, and performances on, something that has notclassroom assessments. For exam- been taught or useple, the junior high mathematics assessment methodsteacher who routinely uses sportsstatistics as a main source for prob- that are incongruentlem-solving tasks could turn off with instruction.those students who are not sportsfans. Likewise, insensitivity todiverse religious beliefs (e.g.,choosing reading passages involving only Christian holidays), gender/racial images(e.g., depicting all doctors as white males), or socioeconomic status (e.g., assumingthat all kids have access to a telephone or home computer) may result in unfair eval-uation of individuals or groups. Teachers must be on guard so that biases do notinfluence their evaluations of a student's performance.

After teachers consider these three general principles, they should address somefundamental questions related to planning classroom assessments.

1 7

Page 18: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Conqent Standarrds Pupose and Audtenceg if

Just as teachers have numerous instructional techniques and strategies fromwhich to choose, they also have a variety of methods available for assessing learn-ing. Teachers can determine which assessment methods to use by responding to sev-eral key questions (see Figure 1).

The first question, under standards/benchmarks, concerns content standards, orthe intended results of the teaching. Teachers should ask: "What do we want stu-dents to understand and be able to do?" Content standards typically fall into threecategories: (1) declarative knowledgewhat we want students to understand (facts,concepts, principles, generalizations); (2) procedural knowledgewhat we wantstudents to be able to do (skills, processes, strategies); and (3) attitudes, values, orhabits of mindhow we would like students to be disposed to act (e.g., appreciatethe arts, treat people with respect, avoid impulsive behavior). The choice of specif-ic assessment methods should be determined in large part by the nature of the con-tent standards being assessed (Marzano, Pickering, and McTighe 1993). For exam-ple, to assess students' ability to write an effective persuasive essay, the assessmentshould involve gathering samples of students' persuasive writing and evaluatingthem against specified criteria. In this case, a multiple-choice test would be ill-suit-ed to measure the intended outcome. Likewise, to assess students' ability to workcooperatively on a research project, the assessment should assess group processesand products as well as individual performance.

In addition to considering content standards, teachers need to raise questionsabout the purpose(s) and audience(s) for classroom assessments. They should ask:"Why are we assessing and how will the assessment information be used? For whomare the assessment results intended?" Purpose and audience influence not only theassessment methods selected, but also the ways in which the results of classroomassessments are communicated. For example, to provide parents of a primary-gradestudent with an interim report of progress in language arts, the teacher might arrangea conference to describe the child's reading skills in terms of a developmental pro-file and review a work folder containing samples of the child's writing.

18

Page 19: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Fig

ure

1

Cla

ssro

om A

sses

smen

t Pla

nnin

g: K

eyQ

uest

ions

taric

iarO

s/B

erkt

imik

sP

urpO

se(0

) :fo

r,A

Sse

ssrh

ent

.W

hat-

do

Ave

wor

k itd

'ele

ntt t

o --

unde

titan

d-ah

4-13

6- a

ble

to-

de?

hi: a

re:w

e ,d

sses

sing

- an

d- h

ew w

ill-

the:

:gik

esS

nier

it iii

fonn

etio

n, h

e ,L

ised

?-

SuO

i-010

1(s)

,10e

,A's

_tes

tffig

Ot

,or:

who

m'a

re-7

-the

:-as

tesi

then

tlesu

lts_

ihte

nded

?-

LI LI

0 D

iagn

ose

stud

ent s

tren

gths

/ nee

dsC

I Pro

vide

feed

back

on

stud

ent

lear

ning

LI P

rovi

de a

bas

is fo

r in

stru

ctio

nal

plac

emen

tLI

Info

rm a

nd g

uide

inst

ruct

ion

CI C

omm

unic

ate

lear

ning

expe

ctat

ions

CI M

otiv

ate;

focu

s st

uden

tat

tent

ion

and

effo

rtLI

Pro

vide

pra

ctic

e ap

plyi

ngkn

owle

dge

and

skill

sC

I Pro

vide

a b

asis

for

eval

uatio

n:gr

adin

gpr

omot

ion

prog

ram

sel

ectio

n/ad

mis

sion

sLI

Pro

vide

acc

ount

abili

ty d

ata

CI G

auge

pro

gram

effe

ctiv

enes

s

Tea

cher

/ ins

truc

tor

Stu

dent

sP

aren

ts

Gra

de-le

vel/

depa

rtm

ent t

eam

sO

ther

facu

ltyS

choo

l adm

inis

trat

ors

Cur

ricul

um s

uper

viso

rsB

usin

ess

com

mun

ity/

empl

oyer

s11

1 C

olle

ge a

dmis

sion

s of

ficer

s0

Pol

icy

mak

ers

(e.g

., bo

ard

ofed

ucat

ion,

sta

te le

gisl

atur

e)C

I Gen

eral

pub

licD

Oth

er:

1

Page 20: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

ASSESSMENT APPROACHESAND METHODS

After teachers identify content standards, purposes, and audiences, they need toselect assessment approaches and methods (see Figure 2, which provides a system-atic guide to selecting assessment methods). Teachers will choose from two for-matsselected-response format and constructed-response formatand the fiveassessment approaches within the two formats. They will select approaches accord-ing to whether they want students to:

1. select a response2. construct a response3. create a product4. provide an observable performance or5. describe their thinking/learning processEach of these formats and approaches has its strengths and limitations, and it is

instructive to look at classroom examples of how teachers use the various assess-ment methods.

SeJected-Response FameSelected-response formatswhich include multiple-choice, true-false, matching,

and enhanced multiple-choice itemsare widely known and used in educationaltesting, especially at the secondary and post-secondary levels (Stiggins and Conklin1992). Multiple-choice items are the most common type of selected responses,appearing on most commercially produced tests, as well as on many classroomassessments. The selected-response format presents students with a question, prob-lem, or statement followed by a set of alternative responses. Students make a selec-tion from among the given alternatives rather than generate their own response.While most selected-response items have a single correct or best response, it is pos-sible to create "enhanced multiple-choice" items that have more than one acceptableanswer contained among the alternatives.

Selected-response formats have a number of advantages. They allow teachers to effi-ciently and objectively assess students' knowledge of factual information, concepts and

Page 21: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Figu

re 2

Fra

mew

ork

of A

sses

smen

t App

roac

hes

and

Met

hods

How

mig

ht w

eas

sess

stu

dent

lear

ning

in th

e cl

assr

oom

?

Sel

ecte

d-_R

tspo

ilse

For

inat

O M

ultip

le-c

hoic

eO

Tru

e-fa

lse

LI M

atch

ing

O E

nhan

ced

mul

tiple

choi

ce

Con

stru

cted

-Res

pons

elor

mat

-

Brie

fC

onst

ruct

ed R

espo

nse

0 F

ill in

the

blan

kw

ord(

s)ph

rase

s(s)

LI S

hort

ans

wer

sent

ence

s(s)

para

grap

hs

0 La

bel a

dia

gram

D "

Sho

w y

our

wor

k"O

Vis

ual r

epre

sent

atio

nw

ebco

ncep

t map

flow

cha

rtgr

aph/

tabl

eill

ustr

atio

n

LI E

ssay

D R

esea

rch

pape

r

0 S

tory

/pla

yLI

Poe

mLI

Por

tfolio

LI A

rt e

xhib

itLI

Sci

ence

pro

ject

LI M

odel

LI V

ideo

/aud

iota

peLI

Spr

eads

heet

11:1

Lab

rep

ort

21

LI O

ral p

rese

ntat

ion

LI D

ance

/mov

emen

tD

Sci

ence

lab

dem

onst

ratio

nLI

Ath

letic

ski

llspe

rfor

man

ceD

Dra

mat

ic r

eadi

ngO

Ena

ctm

ent

LI D

ebat

eD

Mus

ical

rec

ital

O K

eybo

ardi

ngO

Tea

ch-a

-less

on

LI O

ral q

uest

ioni

ngO

Obs

erva

tion

("ki

d w

atch

ing"

)O

Inte

rvie

wLI

Con

fere

nce

LI P

roce

ss d

escr

iptio

n

CI "

Thi

nk a

loud

"O

Lea

rnin

g lo

g

Page 22: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

principles, and the application of basicskills. Because assessments usingthese formats can accommodate alarge number of items, they enable ateacher to sample a broad range ofknowledge and skills in a limitedamount of time. Because selected-response items contain correct oracceptable responses, they are easilyand objectively scored as correct orincorrect using an answer key.Machine-scorable answer sheets andhand-scoring templates simplify thescoring process, allowing teachers toquickly obtain results fortimely feedback to students.

Despite their advantages,assessments using selected-response formats have limi-tations. Instead of assessingthe application of knowledgeand higher-order skills inmeaningful "real world" sit-uations, they tend to assessknowledge and skills in iso-lation and out of context.Selected-response items can-not adequately measure cer-tain learning outcomes, suchas critical thinking, creativi-ty, oral communication, andsocial skills. While real-world issues and problemsrarely have single correctanswers, the widespread useof assessments with select-ed-response formats maycommunicate to students anunintended message aboutthe nature of knowledge andlearningthat recognizingthe "right answer" is the pri-mary goal of education.Critics also express concernthat multiple-choice tests

The widespread use ofassessments with selected-response formats maycommunicate to studentsan unintended message ...that recognizing the "rightanswer" is the primary goalof education.

, CLASROOM ,E, ; , PLES

, prior :to introducirig' phOtosynthesiS, a highschool biolOgy tedcher 'gives yier,,,SnideritsPle-chOiee;pieteSt to aSSesS,their IcnoWledge` of basic ,

Concepts;"related ',to, the topic. Usiiii;the',ehOpt%scanning, inachine'to score ,the ,pretest, the teacher

: qUiCklY detennines, the prior knowledge 'of 'studentslicher elaSsei, and PlAns'her inStinctiOnaceordinglY.

A',c011ege',ec,OnoinicS,OgeSSOr includes ; a hfulti-,'PleChOice ketiOn'a her "inidtera;and, nationslcf assess 'students' ',knowledge 'iof key con-c9M:::4P0,''PrIr,19plesc presented 'in'reSulti,,:are'tised; in-COrijunctiOn',;ith, Other factors:(responses to essay qiiestions,, research,paper,,,class

articipatiOn) *deriVe the OOP kra0'1,At, the, beginning of 'an; AlDS,aWareness unit; A

middle:: school' health, teacher',givesr' his; students atrue-false; quiz to determine their knowledge of -

basic',facts about ,the, disease.,psing, a ',scoring key,the teacher, quickly:realiZeS: ttiactke:rOjority-off stu-

dents ,hold;seVerak COinm7oi-I,inis:cOneeptiOnS'regard=:,,ing the, Spread 9f Aps.Tiiig,irifdr,r6tiOn'help's the ,

;teacher iii:addresSihese- ini'ScoficePtiOns` during 'the

Vpper eleirientary,,sindents: take a practice quizin vthich.:'theY maiCh'siateS-ith their CaPital Chies

3hey then-exchange`papers,,witha,partner and check'each Other's resPonseS byeconsulting a:United States'map. Incorrect;,,,matches, are, identified, for, later,reViei,V-with'A,"study "-s

,

22 LEST COPY AVAILABLE

Page 23: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

lead to "multiple-choice" teaching, that is, a focus on acquisition of facts rather thanan emphasis on understanding and the thoughtful application of knowledge (Mitchell1992; Perrone 1991; Wiggins 1992). With a recognition of their advantages and limi-tations, teachers may appropriately incorporate selected-response formats as part of abalanced menu of assessment approaches.

The development of fair and valid tests using selected-response items is a chal-lenging and time-consuming process. While a complete treatment of this topic isbeyond the scope of this publication, several excellent resources are available toassist teachers in designing assessments using selected-response formats. For moreinformation, see Carlson (1985), Haladyna (1994), and Nitko (1983).

ConzirracUd-Respornse FCRITiAleConstructed-response format refers

students to construct aresponse, create a product,or perform a demonstrationto show what they knowand can do. In this book,constructed-response for-mats include brief con-structed responses, studentproducts, student perfor-mances, and process-focused assessments (seeFigure 2, second throughfifth columns). Other writ-ers refer to such assess-ments as "performanceassessments."

Brief ConstructedResponse

Unlike selected-responseitems that call for a selectionfrom given alternatives,brief constructed-responseitems ask students to gener-ate brief responses to open-ended questions, problems,or prompts. Short writtenanswers and visual represen-tations (e.g., concept map,flow chart, graph) are exam-

to those assessment activities that call upon

,CLASSR00m 'EXAMpLES,A.mid,d1e qchool,'Scienae teather involves her

students',in an'investigation of the ,absorbencyrates,::,of'different brancis,'OflPaPer toWels,F011OWingithinVeStigatiori, theY recOOreSUltS,'"of their.ata,e0l,leetion on. a thart:lind statetheik`i'eOnclusien's in Writ-ing:;StUdents are evaluated on the 'effectiVenesS WithWhiCh their chart§ Coniniunicate the results and on:the acciffaey Of theiir*ritten 0661444s ,

As part of a ctilininating.aSsesSinen(forabeialstudies unit; fourth graders arc gii,,oii'a:iAarik map of:,their state.,They,are aSked tOtheir and the three',Other:larieSecities,.t',and labelffie two'niajor rivers.

As part:ofia CdurSe rev iewia,high school ,p'sy-: chologY teacher:prOvides his Studenti 'With a ce,ac-parisOnvinatrix,'COntaining different: pychnlogicaV,theorieS on bit 0' 4.i( is and :,arious persdnality disor-ders en the ottier. They are asked tOof the matrik to explain the':personalitYlaiserders

. according to each theorY.The' teacher, then reVieWSthe matriceS:arid dikeiisSes thein,,With the Class toclear uPniisunderStandings.

. ., .

,

A sikthgrade .inatheinaticS teacher aSks her gni.=dents to shoW, their Work as 'thei atteniPitO' 'solveMultisterword problems. 'In addition',to examiningtheir solUtiOns,.she looksat their ,Ii/60:;fpr,eyidence,'of, apPropriate'uSe, of algorithms and'pioblern-SolV:-.---'ing strategies.' $he,Provides', feedback throughbrief,

. . written eominents.'

2 3

BEST COPY AVAILABLE

Page 24: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

ples of widely used brief constructed-response methods. While brief constructed-response items may seek a correct or acceptable response (e.g., fill in the blank), theyare more likely to yield a range of responses. Thus, the evaluation of student respons-es requires judgment, guided by criteria. This approach may be used for assessingdeclarative knowledge and procedural proficiency. In addition, assessments using briefconstructed response items can provide insight into understanding and reasoning whenstudents are requested to show their work and explain or defend their answers inwriting.

Assessments using brief constructed-response items offer several advantages. Theyrequire less time to administer than other types of assessments using constructed-response formats. Since they elicit short responses, several brief constructed-responseitems may be used to assess multiple content standards. Evaluation of student responsesis straightforward, guided by criteria and model responses.

Brief constructed-response items are limited in their ability to adequately assess atti-tudes, values, or habits of mind. In addition, they require judgment-based evaluation,which takes time and introduces potential problems of scoring reliability and fairness.Teachers are cautioned against regularly re-using brief constructed-response items forsummative assessments so that students cannot give memorized responses to knownquestions and tasks.

Performance-Based AssessmentPerformance-based assessments include student products, student performances,

and process-focused assessments. Performance-based assessments require studentsto apply knowledge and skills rather than simply to recall and recognize. Thus, per-formance-based assessmentsare more likely to reveal stu-dent understanding. They arewell suited to assessing appli-cation of content-specificknowledge, integration ofknowledge across subjectareas, and life-long learningcompetencies such as effectivedecision making, communica-tion, and cooperation (Shepard1989).

The current interest in per-formance-based methods haspopularized additional assess-ment terms, such as authentic assessment, rubric, anchors, and standards. The termauthentic assessment, popularized by Grant Wiggins (Wiggins 1989), is used todescribe performance-based assessments that engage students in applying knowl-edge and skills in ways that they are used in the "real world." According to Wiggins,authentic assessments should also reflect good instructional practice in ways that

The term authentic assessmentpopularized by GrantWiggins ... is used todescribe performance-basedassessments that engage stu-dents in applying knowledgeand skills in ways that they areused in the "real world."

0

Page 25: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

make "teaching to the test" legitimate and worthwhile.Performance-based assessments generally do not yield a single correct answer or

solution but allow for a wide range of responses. Thus, evaluations of studentresponses, products, and performances must be based on judgments. The evaluativejudgments are guided by criteria that define the desired elements of quality (Ferrara,Goldberg, and Mc Tighe 1995). One widely used scoring tool is a rubric, a scoringtool used to evaluate the quality of constructed-response products and perfor-mances. Rubrics consist of a fixed measurement scale (e.g., four-point) and a list ofcriteria that describe the characteristics for each score point. Rubrics are frequent-ly accompanied by representative examples of student products or performancesthat illustrate each of the points on the scale. These examples are called anchors.

The term standards is frequently used in conjunction with performance-basedassessments. There are three distinct ways in which the term is used: (1) contentstandards, which specify what students should know and be able to do; (2) perfor-mance standards, which set expectations about how well students should perform;and (3) opportunity-to-learn standards, having to do with the necessary resourcesand conditions for effective teaching and learning. Performance-based assessmentscall for decisions about content standards as well as expected standards for perfor-mance (Diez 1993). Three primary types of performance-based assessments areproducts, performances, and process-focused assessments.

Product. Student products provide tangible indicators of the application of knowl-edge and skills. Many educators believe that product assessment is especially "authen-tic" because it closely resembles real work outside of school. Teachers may evaluatewritten products (e.g., essays, research papers, laboratory reports), visual products(e.g., two- and three-dimensional models, displays, videotapes), aural products (e.g.,an audiotape of an oral presentation), and other types of products to determine degreesof proficiency or levels of quality.

Product assessment calls for the selection or development of criteria for evaluation.The criteria are incorporated into a scoring rubric, rating scale, or checklist. Manyteachers recognize that evalua-tion criteria also serve aninstructional purpose: providing Many teachers recognize thatstudents with a clear focus onelements of quality to guide evaluation criteria also servetheir work. When the criteria are an instructional purpose: pro-made public, students may be

viding students with a clearinvolved in using them in peer-and self-evaluation of products. focus on elements of quality

One application of product to guide their work.assessment is systematicallycollecting representative samples of student work over timein portfolios. Portfolios allow teachers, students, parents, and others to observedevelopment and growth in learning. Portfolio assessment has been widely used

2 5

Page 26: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

6.A.S'SROOM :EXAMPLES.

Students deVelORacOmputer program for'a1:1":4: YanCedig4:§0:1091 co:41:pUter.,blaSS.Their teacher e'4114iastildentS".PrOgrarniningprogram'S written Code for accuracy and efficiencYin addition; StUdentS MuSt run theprograM: to demonstrate that,. it_ performs' the .SpeCifiedfunetionS;. unk*ii*fill: re-

-graMs Milk, be "debugged"tintil they.satisfaCtOrily,UlfilLtheyeilUirementS:.se.COnd gradeleaCher collects bi Week l k.e*arnPleS,Of, representative student :Wor

in a lanitiage'artS,POrtfOlie:- The:';'epliected snident:SaniPleS Are.reVieWed with: parentSduring Mid-year conferences. parents with tangibleillustratiOn4of their child's literacy deVelopment. Theleacher uSeShthe actual prOducts',aleng with

develOpMental scaleof reading and Writing for the prithary.grades, 16.disfcns's.ii-te stu-dents skill Stiengths,and Point out areas 'needing special attention:

college engineering profesOr assigns his'students.tO#Ork in teams tOdeSign ariSelfprope*d 'hOVering Vehicle that CeriespondS tO:',certain specified parame

ters, In:"additionA0 the model; students must, individUallprepare a techniCal reportrelated totheir design. The models.and technical tei*iit's are'evalUatedio deterinine sttidents' under§tanding pf; and 'ithility: to.apply, prindiPleS of aerodynamics: 'A'Culminat-

:inglioVering" contest is=iised io determine the ituisteffectiVe deSigns:Fifth grade art :stUderitS Create ;a:landsdape uSing teinpera paints. ljSing a skills

.checklist,,their art teacher assesseS their paintings' to. determine .their proficiency inusing theinedium. He also asSesses their Understanding of the use of CoMPOSitionaleleinents Tor creating, an illusion of depth:, indiviatial student conferences are arrangedto provide :feecibaCk:',.

A Middle school!.SCieriee teacher reYieWs, no students': latiOratory,TeportS .6 deter-Mine their effectiVenes's':in applyintexpeinnental procedures arid the'aeciiracy of theirdata eollection, Her ivrittenWintnents inthe Margins point, ,oitt errors:arid:Offer speeif-,ic SuggestionS 'for ininroverneriCThe report§ are returned, diSCUSSedi, and:filed in 'theStudentS.'.Scienea folder§ fOr future' reference. ;

over the years in the visual arts, architecture, and technical areas.In recent years teachers have increasingly used portfolios to document learning

in other subject areas, especially the language arts. For additional information on theuse of portfolios in the classroom, see Student Portfolios (National EducationAssociation 1993).

The use of products and portfolios can be appealing. When students are givenopportunities to produce authentic products, they often become more engaged in,and committed to, their learning. Unlike standardized assessments that require uni-form student responses, performance-based assessments in which students create aproduct allow students to express their individuality. Product assessments also indi-cate what students can do, while revealing what they need to learn or improve.When teachers share the criteria used to evaluate products, students know the ele-ments of quality that will serve as a guide for peer- and self-evaluation. Previouslydeveloped products can serve an instructional purpose when they are presented asmodels of excellence for students (McTighe 1997; Wiggins 1992).

01 7

Page 27: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Despite their benefits, product assessments have their drawbacks. Criteria for judg-ing the products must be identified, and product evaluation can be a time-consumingprocess. In addition, teachers must be careful when evaluating student products thattheir judgments are not unduly influenced by extraneous variables, such as neatness orspelling. Practicality must also be considered. The time required to develop qualityproducts may compete with other instructional priorities. Product assessments requireresources, including funds for materials and space for display and storage.

Performance. Using performance assessments, teachers are able to observedirectly the application of desired skills and knowledge. Performance assessmentsare among the most authentic types of student assessments because they can repli-cate the kinds of actual performances occurring in the world outside of school.Performances have been used widely to assess learning in certain disciplines, suchas vocal and instrumental music, physical education, speech, and theater, whereperformance is the natural focus of instruction. However, teachers in other subjectscan include performances, such as oral presentations, demonstrations, and debates,as part of an array of assessment methods.

As with product assessments, teachers must develop criteria and scoring tools,such as rubrics, rating scales, or checklists, to evaluate student performances.Students gain additional instructional value when they apply the scoring tools for

C LASS ROOM EXAMPLES

Students in the, schoOl. Orchestra .PartiCipate in' a diess.rehearSalltwO,Weeks beforethe- public performance'. The niusie. teacher, Works ,with;the,:,StudentS':to"0,a1 Uate their . .

. perfOrniance during -the rehearsal, and identify,nreas Of weakneSs.'-During the ensuing,

..PractiteS; the orchestra' members cohcentrate bri making impro'vernents' in.theSe:'neeitS,.prior to the'actual performance before 'a live andieriee.

'.A high, 'school Social studiei-teacher .sets,,up .an 'in-claSS debate as W'tntininliiing'activitY for' a contern'porary iSsues. unii: Students wcirk as part of a-teatn' tO debate theisStie 'of gim control: Theteather Will rate stUdentS" perforniances in the plebateS on Sey-

-erg dimensions including.their tiriOrStanding of the Bill of:RightSi,persuasiveneis oftheir: arguments, iiSe2,'Of ;SUOPOrting:,factlial3.infortinatien,' :affectiVeness in-:countering. ,

rebuttalS",'and'obse0anee Of ritlek of debating: ,

'An elementariiibYsienleducationieaCher tises'a skills Chealigi dtiting the unit On ,_

intrOciuctory gymrinStiCs to aSseSs.studentS'proficiency....EaCh'StUdent reeeiVes',a cOPy,:of the checklist .anO Works, With, n'partner:to try :to sUccesSfully,PerfOrM the. identified7skills. The' Completed .checkliStS nre ,USed aS one COMpOnent'of the' culminating'grade.'

. , .for the. unit. ,

A- high sehOol speech teaCher,:workS, With a hOine eCOnoMiCs teaCher ,fripreparingstudents to make oral' preSentatieris to' cOrtimUnieate-thereSults Of a'nutritiOn research, .

project., ,Using a, rating .SCale, Abe-hi:4nd. econothics tedefier evalitateS the Students on ,.

. . . .

aceUracY and; conipleteness'..of their...knowledge ,of. the ."food.:pyramid.'," the Speech

:teacher uSeS :a scoring rubriefOr'dclivery'cifan inforMatiVe speech tO,evaltiate'the oral..presentations.

. .

7DEST COPY AVAILABLE

Page 28: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

f 7

SSROOM EXAMPLES

iteiatuieteacherlegnlarly,:.`poseerpretation und6r,sfail:cti,:ti, that.Reed

cA kindergarten teacher iiterviews each of her 'ehildreifiri the heiinning-Of the year..,

This orMat'AgsessMent::. rovi es useful information ,about:eognitiyeandclingnis,tic.eVelopinent,;.'g6Cial skills, and areas, ofpersOrial Were*,z A Middle sehOorscieiat Oidies',,teabfierlcarefully,'obgefves ';§tudents,lo Osess. their:

recioperatiye.skills,asAiey*ik.On 'aso6iailstndies'projeetithaents'id seryer,s,,, giving , obserxable

,Mdicatoi*Of,coiip4atiVe skills .:Tlle'teaelieil and sttident,.ob'serVefseedbiek tb the'clas.'dn'theeffectiVenes s, , r, C9,0,pqr4P4e grPu

,chOOf YinatheMaties, ,teaeher,

m 1,b,students::4,nyr:p...06ulikpof.h9iigt0; the teacher can identify fallacious rea-soning d giVefeedback:oif:thelaiikopiiatenes's of 'str'ateiiest.heY'are rovi in& hee 7

peer- and self-evaluation. Such involvement helps students to internalize the ele-ments of quality embedded in the criteria. Many teachers have observed that stu-dents are motivated to put forth greater effort when they perform before "real" audi-ences of other students, staff, parents, or expert judges. Schools also benefit frompositive public relations when students perform for the community.

Many teachers have observed that students are motivated to put forth greatereffort when they perform before "real" audiences of other students, staff, parents, orexpert judges.

Despite their positive features,performance assessments can betime- and labor-intensive for stu-dents and teachers. Time must beallocated for rehearsal as well as forthe actual performances. The evalu-ation of performances is particularlysusceptible to evaluator biases, mak-ing fair, valid, and reliable assess-ment a challenge.

Process-focused Assessment.Process-focused assessments pro-vide information on students' learn-

Rather than focusingon tangible products orperformances, [process-focused assessment]focuses on gaining insightsinto the underlyingcognitive processes usedby students.

ing strategies and thinking process-es. Rather than focusing on tangibleproducts or performances, this approach focuses on gaining insights into the under-lying cognitive processes used by students. A variety of process-focused assess-

28

Page 29: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

ments are routinely used as a natural part of teaching. For example, teachers mayelicit students' thinking processes using oral questions such as: "How are these twothings alike and different?" or by asking students to "think out loud" as they solvea problem or make a decision. Teachers may ask students to document their think-ing over time by keeping a learning log. Also, teachers can learn about students'thinking processes by observing students as they function in the classroom. This"kid watching" method is especially well suited to assessing the development ofattitudes or habits of mind, such as persistence.

Process-focused assessments are formative in that they provide diagnostic infor-mation to teachers and feedback to students. They may also support the develoP-ment of students' metacognition by heightening their awareness of cognitiveprocesses and worthwhile strategies. Process-focused assessment methods are typi-cally used over time, rather than on single occasions. Thus, they are rarely used instandardized, high stakes evaluations of students.

202 9

Page 30: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

EVALUATION METHODS AND ROLESIn addition to making choices about classroom assessment methods, teachers

should consider options for evaluating student work (see Figure 3).One question teachers must ask is: "How will we evaluate student knowledge

and proficiency?" They should determine evaluation methods largely by the assess-ment approach and the nature of the student responses to the assessment item ortask. Selected-response format items and some brief constructed response items(e.g., fill in the blank) yield a single correct or best answer. Most often teachersscore such items using a key with the answers. Sometimes they ask students to"bubble in" their answers on an answer sheet that can be scanned by machine orhand-scored by overlaying a scoring template. Scoring of selected-response formatitems is relatively quick, easy, and objective.

Assessments using constructed-response formats elicit a range of responses,products, or performances that reflect varying degrees of quality and different lev-els of proficiency. Because such assessments typically do not have a single correctanswer, teachers must rely on judgment-based methods to evaluate responses tothese open-ended assessments. Four primary types of evaluation methods are usedwith constructed-response formats: scoring rubrics, rating scales, checklists, andwritten and oral comments.

A scoring rubric consists of evaluative criteria, a fixed scale (e.g., four or sixpoints), and descriptive terms for discriminating among different degrees of under-standing, quality, or proficiency. The term rubric has its origins in the Latin wordrubrica, meaning "red earth used to mark something of significance." Today, educa-tors use rubric to communicate the important qualities in a product or performance.

Scoring rubrics can be holistic (providing an overall impression of the elementsof quality and levels of performance in a student's work) or analytic (indicating thelevel of performance of a student's work on two or more separate traits). For exam-ple, the reading rubric in Figure 4 presents an example of a holistic rubric for evalu-ating reading comprehension. The oral presentation rubric in Figure 5 shows an ana-lytic rubric for oral presentations. Notice that in the analytic rubric, four traits (con-tent, organization, delivery, language conventions) are evaluated independently.

21

Page 31: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Fig

ure

3

Eva

luat

ion

and

Com

mun

icat

ion

Met

hods

valu

atio

n M

eth

ow .w

il-W

e ev

olua

te-

stud

ent

owle

dge,

cind

pro

fic

ienc

Sel

ecte

d-R

espo

nse

For

mat

*O

ans

wer

key

O s

corin

g te

mpl

ate

O m

achi

ne s

corin

g

Con

stru

cted

-Res

pons

e F

orm

atO

sco

ring

rubr

ic

gene

ric

task

-spe

cific

O b

ipol

ar r

atin

g sc

ale

O c

heck

list

O w

ritte

n/or

al c

omm

ents

Eva

luat

orO

teac

her(

s)/ i

nstr

ucto

r(s)

O p

eers

O e

xper

t jud

ges

(ext

erna

l rat

ers)

O s

tude

nt (

self-

eval

uatio

n)O

par

ents

/ com

mun

ity m

embe

rsO

oth

er:

31

Pre

sent

atio

n of

Res

ults

O n

umer

ical

sco

re

perc

enta

ge s

core

s

poin

t tot

als

O le

tter

grad

eO

par

ents

O d

evel

opm

enta

l/pro

ficie

ncy

scal

eO

rub

ricO

che

cklis

tO

writ

ten

com

men

tsO

nar

rativ

e re

port

(w

ritte

n)ve

rbal

rep

ort/

conf

eren

ce

*The

se e

valu

atio

n m

etho

ds m

ay b

e us

ed fo

r so

me

brie

f con

stru

cted

res

pons

e ite

ms

(e.g

., fil

l in

the

blan

k, s

hort

ans

wer

).

Page 32: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Figure 4

Reading RubricRating Scale '. , valuatrie,,Critena,i -

4 Reader displays a sophisticated understanding of the text withsubstantial evidence of constructing meaning. Multiple connectionsare made between the text and the reader's ideas/experiences.Interpretations are sophisticated and directly supported by appro-priate text references. Reader explicitly takes a critical stance (e.g.,analyzes the author's style, questions the text, provides alternateinterpretations, views the text from multiple perspectives).

3 Reader displays a solid understanding of the text with clear evi-dence of constructing meaning. Connections are made betweenthe text and the reader's ideas/experiences. Interpretations aremade and generally supported by appropriate text references.Reader may reveal a critical stance toward the text.

2 Reader displays a partial understanding of the text with some evi-dence of constructing meaning.A connection may be madebetween the text and the reader's ideas/experiences, but it is notdeveloped. Interpretations are not made and/or not supported byappropriate text references. Reader shows no evidence of a criti-cal stance toward the text.

I Reader displays a superficial understanding of the text with limitedevidence of constructing meaning. No connections are madebetween the text and the reader's ideas/experiences. Reader pro-vides no interpretations or evidence ofa critical stance.

0 Reader displays no evidence of text comprehension or construct-ing meaning.

Holistic rubrics are most appropriately used for summative purposes (such as theevaluation provided at the conclusion of unit or a course) where the goal is to pro-vide an overall picture of student performance. Most report card grades representholistic evaluation, since a variety of "subscores" (tests, quizzes, performance tasks,homework, classwork, etc.) are collapsed into a single symbolthe letter gradefor each subject.

GEST COPY AVAILABLE

3 2

Page 33: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Figure 5

Oral Presentation Rubric

Content OrganizationLanguage

Delivery Conventions

4 (Varies byassign-

ment)

Coherent organi-zation through-out; logicalsequence; smoothtransitions; effec-tive introductionand conclusion

Excellent volume;fluent deliverywith varied into-nation; effectivebody languageand eye contact

Highly effectiveuse of language

enhances themessage; few, ifany, grammaticalmistakes

3 (Varies byassign-

ment)

Good organiza-tion generally butwith some breakin the logical flowof ideas; cleartransitions; identi-fiable introduc-tion and conclu-sion

Adequate volumeand intonation;generally fluent;generally effectivebody language

and eye contact

Generally effec-tive use of lan-guage supportsthe message;

minor grammati-cal errors do notinterfere withmessage

2 (Varies byassign-

ment)

Flawed organiza-tion; ideas notdeveloped; weaktransitions; inef-fective conclusion

Volume is toolow or too loud;delivery is notfluent; body lan-guage and eyecontact do notenhance message

Use of languagenot always alignedwith the message;grammaticalerrors may inter-fere with message

(Varies byassign-

ment)

Lack of organiza-tion; flow of ideasdifficult to follow;no evidence oftransitions; nointroduction orconclusion

Message cannotbe understooddue to low vol-ume; straineddelivery; ineffec-tive body lan-guage; lack of eyecontact

Major grammati-cal errors makethe message verydifficult or impos-sible to follow

Holistic rubrics have their place, but teachers should employ primarily analyticrubrics for day-to-day evaluation in their classrooms. Since they identify and eval-uate particular traits, analytic rubrics provide more detailed and specific feedback tostudents about the strengths of their performance and the areas needing attention. If

%Jo

&_41.BEST COPY AVAILABLE

Page 34: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Figure 6

Art Rubric

ating. cale valUatwe ntena, ,

,

3 Identifies three or more relevant differences between the work ofMatisse and van Allsburg (e.g., use of color, level of detail/simplifica-tion, use of line and shape, materials, process)

Identifies a preference for one artist's style

Supports preference with two or more well-stated reasons citingspecific examples from the artist's work

Uses a variety of art vocabulary terms appropriately

2 Identifies two relevant differences between the work of Matisseand van Allsburg

Identifies a preference for one artist's style

Supports preference with one reason citing an example from theartist's work

Uses one or two art vocabulary terms appropriately

I Does not clearly identify significant differences between the workof Matisse and van Allsburg

Identifies a preference for one artist's style, but does not supportpreference with reasons or examples

Does not use art vocabulary terms appropriately

the goal is to improve student learning, not simply grade it, then such specific feed-back is needed. How can students improve their research skills, for instance, if allthey receive is a "3" on a holistic rubric (or a "B-" on a research report)? Such eval-uations provide little meaningful guidance about how to do a better job in the future.An analytic rubric, on the other hand, offers greater specificity. For example, a stu-dent receiving the following descriptive comments on an analytic rubric "usesseveral appropriate sources to gather information on the topic" and "needs to docu-ment all sources using standard bibliographic notation"is informed about astrength of the research (use of multiple sources) and a weakness (lack of completedocumentation). The intent of such feedback is to encourage the student to becomemore attentive to the importance of careful source documentation on future researchprojects.

Rather than choosing between these two types of rubrics, teachers can use bothduring a course or unit of study. They can use the analytic rubric(s) "along the way"

3 4 BEST COPY AVAILABLE

Page 35: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

to inform teaching and guide student practice and revision, and they can use theholistic rubric(s) at the conclusion of a performance task or unit assessment to pro-vide an overall evaluation of student knowledge and proficiency.

In addition to being analytic or holistic, rubrics also may be generic or task-spe-cific. A generic rubric provides general criteria for evaluating a student's perfor-mance in a given performance area. The rubrics shown in Figures 4 and 5 are gener-ic rubrics since they may be used to evaluate a variety of responses to reading andoral presentations, respectively. In contrast, a task-specific rubric is designed for usewith a particular assessment task. For example, the art rubric in Figure 6 presents arubric used to assess the task of comparing the styles and techniques of two artists(Matisse and van Allsburg) and indicating a preference. Notice that a task-specificrubric, such as this one, cannot be used to evaluate responses to different perfor-mance tasks.

Generic rubrics offer the capability of multiple applications within a given area,such as mathematical problem solving, persuasive writing, and research. Ratherthan creating a new rubric for each and every performance task, the same rubric canbe taught to students, posted in the room, and used throughout the year (and oftenacross grade levels). With repeated use, the criteria contained in the generic rubriccan be internalized by students so that they are better able to consider the qualitiesof effective performance while they are working, as well as to evaluate their ownwork when they are finished.

There are times, however, when a task-specific rubric will be preferable. Forinstance, task-specific rubrics tend to yield greater reliability (consistency) whenused by different teachers. Thus, a department or grade-level team might employ atask-specific rubric for use with a common performance task or final exam given bymore than one teacher. Task-specific rubrics can be customized from genericrubrics.

Rubrics are most effectively used for evaluation or instruction when they areaccompanied by examples of responses for each score point. These examples oranchors provide tangible illustrations of the various points on the rating scale.Perhaps the greatest advantage of rubrics is their clear delineation of the elementsof quality. They provide students with clear performance targets, expectations aboutwhat is most important, and criteria for evaluating and improving their own work.They provide teachers with specific criteria for reliably evaluating student respons-es, products, or performances; a "tool" for increasing the consistency of evaluationamong teachers; and clear targets for instruction.

These evaluation methods require time to collect or develop rubrics, to iden-tify representative anchors, to develop proficiency in applying them reliably, andto use them for evaluating student responses, products, and performances.Nonetheless, some schools and districts have recognized the significant profes-sional development benefits of providing opportunities for teachers to worktogether on scoring student responses, products, and performances and identify-ing anchors.

Rating scales may also be used to evaluate responses to open-ended questions

/0635

Page 36: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

and tasks. Bipolar rating scales (seeFigure 7), for example, are widelyused on questionnaires and can beapplied to educational assessmentsas well. Such a scale might be usedin conjunction with evaluationsrelated to program selection (e.g.,special education placement) or forpeer evaluation of a product or per-formance (e.g., "This oral presenta-tion achieves its stated purpose.").

Checklists contain categories(i.e., specific features or dimensions)for evaluation and rating options foreach category. The rating optionsmay offer a simple "yes" or "no" to

Some schools and districtshave recognized thesignificant professionaldevelopment benefits ofproviding opportunities forteachers to work togetheron scoring studentresponses, products,and performances andidentifying anchors.

indicate the presence or absence ofeach dimension, or a narrow scale,such as "never," "rarely," or "frequently." Checklists are easy-to-use, efficient eval-uation tools. They can be used while teaching a lesson or leading a discussion andare especially useful when observing students at work. Checklists may also be usedas guides by students, individually or in groups, while they engage in performanceactivities.

While rating scales and checklists are simple to apply in the classroom, they gen-erally do not provide the detailed, explicit criteria found in rubrics. Thus, they areopen to differing interpretations and greater subjectivity when used to evaluate stu-dent products and performances.

Written and oral comments can be effective in evaluating student work becausethey enable teachers to communicate clearly and directly with their students aboutelements of quality, expected standards of performance, areas of strengths, andneeded improvements. These methods allow teachers to provide evaluative feedbackto students on a personal level. Written and oral comments can require a great deal

Figure 7

Sample Bipolar Rating Scale

StronglyDisagree

Disagree NotSure

Agree StronglyAgree

-2 0 2

gRST COPY AVAILABLE

Page 37: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

of teacher time and are especially demanding for secondary teachers working withone hundred or more students per day. The effectiveness of personal comments maybe diminished if teachers provide only negative feedback (identifying errors orproblems), make nonspecific positive comments that do not acknowledge particularaspects of student effort and work, or make comments that do not address all impor-tant elements of quality.

Teachers must also ask: "Who will be involved in the evaluation?" As always,this guiding question should be answered with content standards, purposes, audi-ences, and methods in mind. The question also brings to mind the opportunity toinvolve others in the evaluation process. Teachers may involve other staff members,parents, or community experts in the evaluation of student products (e.g., sciencefair projects) and performances (e.g., public-speaking exhibitions). They may alsoinvolve students. When students are engaged in applying criteria for self- and peer-evaluation, they begin to internalize elements of quality and performance standardsin ways that can lead to improvements in the quality of their work and learning.

Page 38: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

COMMUNICATION ANDFEEDBACK METHODS

After evaluations are made, teachers must ask: "How will we communicateassessment results?" A variety of methods can be used, including numerical scores,letter grades, verbal and written reports, scales, and checklists (see Figure 3). Thechoice of communication methods should be determined by assessment purposesand methods, evaluation methods,and especially the audience for theassessment information. The choice of communi-

Numerical scores (e.g., percent- cation methods shouldage correct or number of pointsearned on a classroom quiz) and let- be determined by assess-ter grades are widely used methods ment purposes andfor communicating the results ofclassroom assessments. Both meth- methods, evaluationods are efficient to use and succinct, methods, and especiallybut numerical scores and grades, bythemselves, do not explicitly com- the audience for themunicate the elements of quality and assessment information.standards of performance that theyare meant to reflect. For example,saying that 70 percent correct is a "C" can mean one thing on an easy task and some-thing different on a difficult task and it does not make clear what a student knowsand can do. Likewise, when students are graded "on a curve," their knowledge orperformance level is communicated in relation to other students in the class, not interms of established criteria and standards.

Developmental and proficiency scales are generally more informative thannumerical scores and grades because they contain descriptions of different degreesof quality and levels of performance (see Figure 8 for an example of a develop-mental scale for reading). Information about student learning presented in terms of

38

Page 39: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

developmental or proficiency levels can be especially meaningful to parents.Recognizing this fact, some schools and districts have revised their report cards,especially for the primary grades, to incorporate features of developmental and pro-ficiency scales.

Checklists can also be effective for communicating assessment results becausethey present ratings on identified criteria or elements of quality. They are a quickand efficient method for providing direct and timely feedback to students. However,checklist developers must be careful to avoid poorly defined categories, such as cre-ativity, that are open to diverse interpretations.

Figure 8

Developmental Reading ScaleEmergent Reader0 follows along in the text when adult readsO is aware of relationship of printed text to oral languageO uses picture cues when recalling story0 pretends to read; memorizes favorite stories

Beginner ReaderLi reads word-for-word; struggles with unfamiliar materialLI has limited sight vocabulary of one- and two-syllable words0 attempts to pronounce and figure out meaning of new words0 demonstrates comprehension of simple textO occasionally monitors comprehension and self-corrects

Competent ReaderO reads familiar material comfortably0 has large sight vocabulary

O uses context clues to figure out meaning of unfamiliar words0 actively constructs meaning0 regularly monitors comprehension and self-corrects

Fluent ReaderLI reads fluently with expression0 has extensive sight vocabulary

O readily determines meaning of unfamiliar words using context cluesO reads a wide variety of materials with understandingO independently monitors comprehension; appropriately

applies comprehension strategies

Lao

3 9

Page 40: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Written comments, narrative reports, verbal reports, and conferences can beeffective communication methods because they provide opportunities to clearly anddirectly connect student effort and performance to elements of quality and standardsof performance. They also allow teachers to provide more individualized and per-sonal feedback than the other communication methods. Regrettably, the time-con-suming nature of these methods often limits their use, especially for teachers at thesecondary level because of the greater student-to-teacher ratio.

4 0

Page 41: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

AFTERWORD

Assessment is an essential component of the teaching and learning process.Without effective classroom assessment, it is impossible for teachers to knowwhether students are "hitting the target"that is, learning what is important forthem to learn. However, the significance of classroom assessment extends beyondthe role of measuring learning. What we assess, how we assess and evaluate, andhow we communicate results send a clear message to students about what is worthlearning, how it should be learned, what elements ofquality are most important, andhow well we expect them to perform. By considering the key questions and princi-ples presented here, teachers will be better equipped to develop and use classroomassessments that provide fair, valid, and reliable information that will inform teach-ing and promote learning.

41

Page 42: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

GLOSSARY

analytic scoringscoring procedure in which responses, products, or perfor-mances are evaluated for selected dimensions, with each dimension receiving a sep-arate score. For example, a piece of writing may be evaluated on several categories,such as organization, use of details, attention to audience, and language usage andmechanics. Analytic scores may be weighted and totaled.

anchor(s)representative responses, products, or performances used to illustrateeach point on a scoring scale. They are also referred to as "models" and "range-find-er papers." Anchors for the highest score point are sometimes referred to as exem-plars.

assessmentany systematic basis for making inferences about characteristics ofpeople, usually based on various sources of evidence; the global process of synthe-sizing information about individuals in order to understand and describe them bet-ter (Brown 1983).

authentic assessmentrefers to assessment tasks that evoke demonstrations ofknowledge and skills in ways that they are applied in the "real world." Ideally,authentic assessment tasks also engage students and reflect best instructional activ-ities. Thus, teaching to the task may be desirable.

content standarda goal statement specifying desired knowledge, skills orprocesses, and attitudes to be developed as a result of educational experiences.

criteriaguidelines, rules, or principles by which student responses, products, orperformances are evaluated.

criterion referencedan approach for describing a student's performance on anassessment according to established criteria.

evaluationjudgment regarding the quality, value, or worth of a response, product,or performance based upon established criteria.

formative assessmentongoing, diagnostic assessment providing information(feedback) to guide instruction and improve student performance.

generalizabilitythe extent to which responses, products, or performances sam-pled by a set of assessment activities are representative of the broader domain beingassessed.

holistic scoringa scoring procedure yielding a single score based upon an over-all impression of a response, product, or performance.

indicatora specific description of an outcome in terms of observable and assess-able behaviors. An indicator specifies what a person who possesses the qualities

4 233

Page 43: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

articulated in a content standard knows or can do. Generally, several indicators areneeded to adequately describe each content standard.interdisciplinary or integrated assessmentassessment that uses tasks that teststudents' abilities to apply concepts, principles, skills, and processes from two ormore subject disciplines to a central question, theme, issue, or problem.norm referencedan approach for describing a student's performance on anassessment by comparison to a norm group.

opportunity-to-learn standardsthe conditions and resources necessary forteachers and schools to meet higher standards for students.

performance-based assessment (or performance assessment)an assessmentactivity that requires students to construct a response, create a product, or performa demonstration. Performance-based assessments generally do not yield a singlecorrect answer or solution but allow for a wider range of responses. Thus, evalua-tions of student responses, products, and performances are based on judgmentsguided by criteria.

performance standardan established level of achievement, quality, or proficien-cy. Performance standards set expectations about how much students should knowand how well students should perform.

performance taskan assessment activity, or set of activities, related to one ormore content standards, that elicits one or more responses to a question or problem.portfolioa purposeful, integrated collection of student work showing effort,progress, or achievement in one or more areas (adapted from Paulson, Paulson, andMeyer 1991). Since they feature works selected over time, portfolios are well suit-ed to assess student growth and development.

primary trait(s) scoringa scoring procedure in which responses, products, orperformances are evaluated by limiting attention to a single criterion. These indi-vidual criteria are based upon the trait determined to be essential for a successfulperformance on a given task. For example, persuasiveness might be the primary traitbeing evaluated in a note to a principal urging a change in a school rule. Scorerswould attend only to that trait.

proficiencyhaving or demonstrating a high degree of knowledge or skill in a par-ticular area.

reliabilitythe degree to which an assessment yields dependable and consistentresults.

rubrica scoring tool used to evaluate a student's performance in a content area.Rubrics consist of a fixed measurement scale (e.g., a four-point scale) and a list ofcriteria that describe the characteristics of products or performances for each scorepoint. Rubrics are frequently accompanied by examples (anchors) of studentresponses, products, or performances to illustrate each of the points on the scale.standardized assessmentan assessment that uses a set of consistent proceduresfor constructing, administering, and scoring. The goal of standardization is to ensure

4 3

Page 44: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

that all students are assessed under uniform conditions so that interpretation of theirperformance is comparable and not influenced by differing conditions (Brown1983).

summative assessmentculminating assessment for a unit, grade level, or courseof study providing a status report on mastery or degree of proficiency according toidentified content standards.

testa set of questions or situations designed to elicit responses that permit aninference about what a student knows or can do. Tests generally utilize a paper-and-pencil format, occur within established time limits, restrict access to resources (e.g.,reference materials), and yield a limited range of acceptable responses.

validityrefers to the degree to which an assessment measures what it is intendedto measure.

Page 45: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

REFERENCES

Airasian, P.W. 1997. Classroom assessment. 3rd ed. New York: McGraw-Hill.

Brown, F.G. 1983. Principles of educational and psychological testing. 3rd ed. NewYork: Holt, Rinehart and Winston.

Carlson, S.B. 1985. Creative classroom testing: 10 designs for assessment andinstruction. Princeton, N.J.: Educational Testing Service.

Cross, K.P., and T.A. Angelo. 1988. Classroom assessment techniques: A handbookfor faculty. (Technical Report No. 88-A-004.0). Ann Arbor, Mich.: University ofMichigan, National Center for Research to Improve Postsecondary Teaching andLearning.

Diez, M. 1993. Essays on emerging assessment issues. Washington, D.C.: AmericanAssociation of Colleges for Teacher Education.

Ferrara, S., G. Goldberg, and J. McTighe. 1995. Ways in which teachers communi-cate learning targets, criteria, and standards for performance to their students.Paper presented at the annual meeting of the American Educational ResearchAssociation, San Francisco.

Ferrara, S., and J. McTighe. 1992. Assessment: a thoughtful process. In If mindsmatter: A foreword to the future. Vol. 2. Palatine, Ill.: Skylight Publishing.

Haladyna, T.M. 1994. Developing and validating multiple-choice test items.Hillsdale, N.J.: Lawrence Erlbaum Associates.

Herman, J.L., P.R. Aschbacher, and L. Winters. 1992. A practical guide to alterna-tive assessment. Alexandria, Va.: Association for Supervision and CurriculumDevelopment.

Marzano, R., D. Pickering, and J. McTighe. 1993. Assessing outcomes: Performanceassessment using dimensions of learning. Alexandria, Va.: Association forSupervision and Curriculum Development.

McTighe, J. 1997. What happens between assessments? Educational Leadership54(4): 6-12.

/36

4 5

Page 46: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Mitchell, R. 1992. Testing for learning: How new approaches to evaluation canimprove America's schools. New York: The Free Press.

Mitchell, R., and M. Neill. 1992. Criteria for evaluation of student assessment sys-tems. Washington, D.C.: National Forum on Assessment.

National Education Association. 1993. Student porYblios. Washington, D.C.: NEAProfessional Library.

Nitko, A.J. 1983. Educational tests and measurement: An introduction. New York:Harcourt Brace Jovanovich.

Paulson, F.L., P.R. Paulson, and C.A. Meyer. 1991. What makes a portfolio a port-folio? Educational Leadership 48(5): 60-63.

Perrone, V., ed. 1991. Expanding student assessment. Alexandria, Va.: Associationfor Supervision and Curriculum Development.

Shepard, L. 1989. Why we need better assessments. Educational Leadership 46(7):4-9.

Smith, M.L. 1993. The role of high-stakes testing in school reform. Washington,D.C.: National Education Association.

Stiggins, R.J. 1997. Student-centered classroom assessment. 2nd ed. New York:Macmillan College Publishing Company.

Stiggins, R.J., and N.F. Conklin. 1992. In teachers' hands: Investigating the prac-tices of classroom assessment. Albany, N.Y.: State University of New York Press.

Wiggins, G.P. 1989. Teaching to the (authentic) test. Educational Leadership 46(7):41-47.

. 1992. Standards, not standardization: Evoking quality student work.Educational Leadership 4(5): 18-25.

Page 47: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

For more information about the Student Assessment Series,contact the National Education Association at

202-822-7350.

Copies of this book may be purchased from theNEA Professional Library

Distribution Center, P.O. Box 2035Annapolis Junction, MD 20701-2035

1-800-229-4200

4 7

Page 48: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

Student Assessment Series

Glen W. Cut lipSeries Editor

ISBN 0-8106-2070-7

neaNATIONAL EDUCATION ASSOCIATION

1201 16th Street, N.W.Washington, DC 20036-3290

eAlb$'5

Page 49: Series. INSTITUTIONResearch in Washington, D.C. Formerly Maryland's State Director of Student Assessment, he has taught at the preschool, middle school, high school, and graduate levels

LI

(9/92)

U.S. DEPARTMENT OF EDUCATIONOffice of Educational Research and improvement (OERI)

Educational Resources information Center (ERIC)

NOTICE

REPRODUCTION BASIS

[ERIC

This document is covered by a signed "Reproduction Release(Blanket)" form (on file within the ERIC system), encompassing allor classes of documents from its source organization and, therefore,does not require a "Specific Document" Release form.

This document is Federally-funded, or carries its own permission toreproduce, or is otherwise in the public domain and, therefore, maybe reproduced by ERIC without a signed Reproduction Releaseform (either "Specific Document" or "Blanket").