docoment besohe.' - wordpress.com€¦ · in the context of soci-al science research and...

28
•1 . ?- DOCOMENT BESOHE".' ED 20U 351» , TH 810 295 ^ -RUTHOP Schwandt, Thomas A. TITLE . ' Defining Bigor and Pelevance in y,ocatipnal Education Evaluation. , PHB DATE 17 Apr*81 ' NOTE ' 23P.; Paper presented .at the Annua'l Meeting of the American Educational Research Association (65th,. Los - • Anaeles, CA, April 13-17, 1981). , - EDPS PFICE MF01/PC02 Plus Postage.. * ' , ' DE-SCRIPTOPS Definitions: *Epistemology: Ethnograph'y; *Evaluation Methods: *Proqram_ Evaluation: Research Design: Vocational Education ' ^ IDENTIFIERS *Relevance (Evaluation): *Rigor (Ev.aluation) ABSTRACT \ , ' The terms "rigor" and "relevance" most "often surface in discussions of methodological adequacy. Assessing epistemological •relevance is equivalent to answering the question, ."Is this ^particular research question worth asking at all?" Epistemological riaor refers to the properties of a "researchable" problem. If^one accepts ^he proposition that different kinds of questions teguire different ki^ids of methodologies, then the question of methodological relej^ance can be asked as, "Is this particular fletliod ,(or model) appropriate to the questions that I am trying: to answer?" Methodological rigor is a determination of whether the method selected me^ts certain standards for a "good" or "trustworthy" method^ a "sound" design, or an""approp]|^a't'e" type of data analysis. It is freguently ar-gued tliat one must cnoose between rigorous a-nd relevant methods. .This argument demonstrates a confounding of the notions of methodological rigor and methodological relevence. Rigor and relevence are no* necessarily inversely^related. These issues are being discussed' i^ vocational/ evaluators, and warrant additional " ' investigation. (BW) -\ ***^ci^**^c^|t*i^*i^tik****•^l•^l^ei^*7^c^t**J^^ik***i^* ******** *********** * 'Peproductions supplied by EDItS are the best that <jan be made * from the original document. * *t!******c***-lnti****^t**_**M* ********.********************************* ******

Upload: others

Post on 19-Apr-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•1 . ?-

DOCOMENT BESOHE".'

ED 20U 351» • , TH 810 295 • ^

-RUTHOP Schwandt, Thomas A. TITLE . ' Defining Bigor and Pelevance in y,ocatipnal Education

Evaluation. , PHB DATE 17 Apr*81 ' NOTE ' 23P.; Paper presented .at the Annua'l Meeting of the

American Educational Research Association (65th,. Los • - • Anaeles, CA, April 13-17, 1981). , -

EDPS PFICE MF01/PC02 Plus Postage.. * ' , ' • DE-SCRIPTOPS Definitions: *Epistemology: Ethnograph'y; *Evaluation

Methods: *Proqram_ Evaluation: Research Design: Vocational Education '

IDENTIFIERS *Relevance (Evaluation): *Rigor (Ev.aluation)

ABSTRACT \ , ' The terms " r igor" and "re levance" most "often surface

in d i scuss ions of methodological adequacy. Assessing epis temological •relevance is equivalent to answering the ques t ion , . " I s t h i s ^par t icu la r research question worth asking at a l l ? " Epistemological r i ao r r e fe r s to the p rope r t i e s of a " researchable" problem. If^one accepts ^he proposi t ion t ha t d i f fe ren t kinds of quest ions t e g u i r e d i f fe ren t ki^ids of methodologies, then the question of methodological relej^ance can be asked a s , " I s t h i s p a r t i c u l a r fletliod ,(or model) appropr ia te t o the ques t ions tha t I am t ry ing: t o answer?" Methodological r i go r i s a determination of whether the method se lec ted me^ts c e r t a i n s tandards for a "good" or " t rus tworthy" method^ a "sound" design, or an""approp]|^a't'e" type of data a n a l y s i s . I t i s freguently ar-gued tliat one must cnoose between r igorous a-nd re levan t methods. .This argument demonstrates a confounding of the notions of methodological r igo r and methodological r e levence . Rigor and relevence are no* necessa r i ly i n v e r s e l y ^ r e l a t e d . These i s sues are being discussed' i^ vocational/ eva lua to r s , and warrant addi t iona l " ' i n v e s t i g a t i o n . (BW)

-\

***^ci^**^c^|t*i^*i^tik****•^l•^l^ei^*7^c^t**J^^ik***i^* ******** ***********

* 'Peproductions supplied by EDItS a re the best t ha t <jan be made • * from the o r i g i n a l document. * *t!******c***-lnti****^t**_**M* ********.********************************* ******

Page 2: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•J . M' \

y

o

LLI Defining Rigor and Relevance in

Vocational Education Evaluation

U ^ DEPAffTMENT OF EDUCATION

NATIONAL INSTITUTE OF EDUCATION

EDUCATIONAL RESOURCES INFORMATION

rfNTER (ERIC* '

^ The document ha& been (epJCxloced a^

(ecerved f'O'n The pefK>n of org&nudtton

Minor changes have been made 10 improve

leprodLiction quality

• Points ot view o ' opinions state*! in ih i j docu

ment 00 not necesian'Y teptcsent oHiCiai Nl£

pOS't'On 01 policy

,«?

Thomas A. Schwandt "* Department of Vocational Education

Sniith Research Center Indiana University

Bloomington, Indiana 47405

•PERMISSION TO REPRODUCE THIS

MATERIAL HAS BEEN GRANTEC? BY

TO THE EDUCATIONAL RESOURCES

INFORMATION CENTER (ERIC) "

Q. \ ^

1. • Paper presented at the 1981 annual meeting of the,

American Educational Research Association L<?s Angeles, California, April 17, 1981

s!' \

Page 3: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

J ^

DEFINlf^G RIGOR AND RELEVANCE IN^ VOCATIONAL EDUCATION EVALUATION

The term^ 'rigor' and 'relevance' most often surface in discussions of

methodological adequacy.* If they have a, 'classical' meaning, then, m6st

likely, the phrase 'rigor vereus relevance' refers to the tradeoffs involved

in designing an experiment that has both high' internal validity (rigor) and

hi§h external validity (relevance) (Campbell and Stahley, 1966), Within the

context of the current methodological "debate, these two terms have been

used, in a rather general way to characterize the differences between 'hard'

data and traditional, scientific, or" quantitative methodology which is'

ri^rous and 'soft' data antj, less cfonventional, naturalistic, or qualitative

methodology which is relevant-

'I initially intended to investigate vocational education evaluation "

models, methods, and frameworks in view of their treatment of. these two

dimensions of methodological adequacy. Yet, attempts to clarify the.meaning

of the terms 'rigor/ and 'relevance' revealed that they have an epistemological

as well as a metho^jidlogical meaning.

Hence, in what follovfs, I propose a more expanded analysis of rigor and

relevance. -Pour distinct, though interrelated, notions of rigor and relevance

are identified: epistemological relevance, epistemological rigor, method-

. ological relevande, and methodologicai rigor. Each is explained and an attempt

is made to illustrate the treatment; of each in the'literature on Vocational

education evaluation. The paper concludes with a discussion of the need to

further investigate each dimension. \

K

* •

o KJ

Page 4: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

\

Rigor and Relevance

2

Defining Rigor and Relevance

Lbgical -and conceptual ana lys i s of the not ions of r i go r and relevance

in the context of soci-al science research and evaluat ion r evea l s an

epis temological and a methodolo'gical usage of each term.. These four aspec ts

of r igor and relevance are explained below.

Epistemological Relevance

\ I t i s commonplace to charac te r i ze the major c r i t e r i a for adequate "researoh

problems as relevance and f ru i t f u lne s s . . For example, S te ine r (X978) expla ins .

t ha t for an educat ional problem to be r e l evan t i t must be genera t ive of

s c i e n t i f i c , phi lo 'sophical , or p r a x i o l o g i c a l knowledge about educatiort^ The

mark of a f r u i t f u l problem i s t h a t i t must be capable of leading to the

"^ extension of knowledge. Assessing epis temological relevance i s thus equivalent

• to ansverin.g the ques t ion , " I s t h i s p a r t i c u l a r research quest ion worth aisking

a t a l l ? " , • , • • •

^ We mxijht extend t h i s notioft of epis temolpgical relevance in r e s e a r t h to

the domain of evaluat ion in the following way. To determine whether a * . " ". • '

particular evaluation question is worth asking (or whether a particular f-

evaluation is worth pursuing) , we mu^. first apswer twa questions: (1) What

is to be evaluated?, an^ (2) Why is it to be evaluated? These qifestions must ' . " V /' . - •

be answered to the sat is!fact ion of stcikeholdeirs in any given evaluatif)n before

quest ions of how to proceed are proposed. Fa i lu re tq adequateTy speci fy- the ~ • ' - '

evaluand—the e n t i t y being evaluated (Scriveri, 1^79)--an*, to iden t i fy the

intended uses and users of eva lua t ive •information w i l l ' l i k e l y r e s u l t ii> an

inaccura te evaluat ion of l i t t l e use to anyone.' ^

•• • { :

Page 5: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•TFT- •

«^

^

, ^j. • Rigpr ' ^ d . JfeiB.^^c^e ,

- •" ' ' % ' ,3--. . / : '

Determining epistemological relevancib need no.t be''na;^owly conceived ,as

a task uni<5ue to positivi-stic science; 'f'he following two approaches to

assessing epistemological relevance orig/inate j.n' quite "different points of

view regarding the nature of evaluation/' ' The f i r s t ajsproach-, 'proposed by" . ' ' • / • • ' I ^ ,

Joseph Wholey .and his colleagues at thd Urban Insiutute -(Wholly, 1975,^ 197,7;

Horst, et al. ' , 1974) i s conroatible with thfe- ti:adj.tional view of evaluation. ' / ' ' c ••' ' '

as 'evaluation-research' . The approach, known as "evaluability assessment," ' : * . ' ' ' • ' '

requJXH's clarifying and defining the evaluaijd frdn the |)erspectj.ves of both^

' the "user and the evaluator. The major elements'.of" an -evalutoility as'sessmen't

• . • ' . / ' " ' ' - ' ' ' - • ' : . " ' " " - •

are characterized by Wholey (1977) as:, • • ' • ' ' ' • ' / • . • ' ' , '

' ' , / . • • • • \

1. Determining-the bo\indaries/of the. problem/priograin, ^ i . e . , •>. / ' what i s i t that i s to be c^alyzed?, wlhat'ar«_,the progsairi objectives?

2. Gathering information^ t h ^ defines-program^bjex^tiyes,^ ' a c t i v i t i e s , and underlying, a'ssumptityis. '.'^

' ' ' * / ' " ' r i • ' ' ' • ,

/ ' . ** , ' " ** * 3. Developing a model of pip^ram acti\?,itias and obj_e<J3fi'*/es

from the point of view <j»f the intended^'Tiser''of thef^ evaluation informa'tion. i

t J

4. Determining to what e x t ^ t the ^efih'i'tdph ^f the program,' ' • • as containedin the model, ' i f . suff ic ientry unambiguous to permit a useful evaluq|tion_. • , i t •' ' .•

• * • / -• • . . - . ' "

' 5 . Presentation of the above infbrmatxon to ' the in-tended use-r' of the Valua t ion and determination-of next steps to'X>^ .taken.

'' • . ' • • ' " . r • .

Though quite 'anti thetical to-the'general approach .of' evaluation r e s e a r c h ^ ' ' ' ' . •• • : ' -^ .-• • I

Guba's (1978) commentary on the methodology'of na tu ra l i s t i c inqui/7. in ' ' ' • •• ' - \ ' . . • • '.•

evaluation also addresses the que'sftipn 6t ep;Ls'temological relevance.- In r > • • •

surfacing the concerns andissues of "relevant par t ies to the evaluation,

^the na tura l i s t ic investiga^tor i s efgaged^ih a process of determining what i s

to be evaluated and why i t i s ' to be evaluated. Having c y c l ^ through repeated*

^ !

\ \

*s, > i

A

5

Page 6: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

"• - i^' • . . Rigor and Relevance • - - v ,

I* ,

: • ^

phases of discovery and verif icat ion, the evaluator possesses a preliminary

set of categories of information., Gulia suggests that considerations of salience,,

credibi l i ty , unicjueness, heuris t ic value, f eas ib i l i ty , and materiality be

employed to p r io r i t i ze the categories and thereby focus the inquiry.' As a

resul t ' c f this process,* the evaluator i s aJble to pursue those-categories o^

concerns and issues- "most worthy of further exploration."

It ,should be apparent from these two examples that , regardless of one's

philosophical orientation regarding the nature of evaluation, assessment of

epistemological relevance iS a . c r i t i c a l f i r s t step in evaluation. Though

the tec^ iques of the evaluation researcher and the na tu ra l i s t i c investigator

are quite-different, both aim a t clarifying the nature of the evaluand and-

surfa-cing the concerns of potential users of evaluation information. " '

Epistemological Rigor . . .

• When we discuss, the properties of a 'researchable' problem, we are

. ' ' ' ' '• , " »

speaking- oi epistemological r igor. As was the case with epistemological

- <elevance, th is notion oi rigor i s important regardless of the pa r^cu la r

philosophical orientation of the researcher or evaluator. However, ' s c i en t i f i c '

and '^naturalist ic ' inq'uirers assess the dimension-of epistemological r igor in - '- •" ^ • . - • I • • '

quite, different wa ys- ' ' , - • - 1*'

Vftiezje evaluation i s viewed- as an extension of sc ien t i f ic research, the - \ ' •• " ' _ ;

,-asse'ssmej/it'of epi*stemolo^cal rigor is quite straightforward. Here, rigor r •?effers -to the extent -tb which evaluation questions are cast in a form that i s

m'easu:f'.mle or .testable. Assessment of 'rigor i s largely context-independent

'.and £ griow^.^ Jt'iiivp'Lves Adequately specifying the e;npirical ^referents for f . ' / ;• V - ' . • /

/ . . • ' - • -a> ,._ -

Page 7: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

Rigor and Relevance

' 5 '

-V • -?'

and'the connections between variables contained in the statement of the research--. " • * .

or evaluation problem. Darcy (1979, 1980) i l l u s t r a t e s th is process in the , •'

'context of vocational"education outcomes evaluation. Beginning with a

meaningful 6utcqme statement, he explains that (1) th i s statement "must Ke

translated into a, hypothesis with careful specification of dependent and '

independent variables, (2) the ent i ty on which the outcome, i s to be observed

must be delineated, and (3) empirical indicators of the outcome niust b?

specified. Outcome statements so formulated meet the t e s t of epistemological

r igor.

wHere evaluation is placed more in the tradit ion of ethnography (e .g . ,

Cuba, 1978; Pdtton, 1980; Filstead, 1979) t ^ determination of epistertological \

rigor i s no less important, yet the process i s less apparent because problems

are not so carefully circumscribed prior to the investigation. Here, tfhe

assessment of epistemological rigor i s la rge ly context-dependent and a "

pos ter ior i . That i s , epistemologifcal. rigor i s not assessed at the outset of

a^ inyestigation by determining whether or not a problem is :ijesearchable/ rather -

rigor i s assessed near the conclusion of the investigation by determining

whether or hot a problem has been' adequately researched (investigated). r • " '. . ' '

Natural is t ic or ethnographic Evaluation approaches begin with problems . ^ , " . -

that a^e not largely delimited. Hence,, the na tu ra l i s t i c evaluator defines

epistemological rigor in terms of whether the l im i t s ' t o an-investig^tiion qf .•,;'"

a problem have been reached^ Cuba (1978) de^jCxibe's th is process as on^ pf P\

reaching "closure" by applying the c r i t e r i a of "exhaustlpn of resourCesV' . ' ..

"saturation," "emergence of r egu la r i t i e s , " and "overextension" to the activity""

^

i.

Page 8: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

- . * * , J - •<- -'" > Rigor and Relevance

' ^ f co l l e c t i ng information. If t^ese s i g n a l s . f o r closure appear, then the

' n a t u r a l i s t i c evaluator i s reasonahjly c e r t a i n t h a t epis temoldgical r igor has

been a t t a i n e d . , ' " , •

Both approaches focus on s e t t i n g the boundaries t h a t define a researchable

problem- In the case of evaluat ion resea rch , these boundaries are defined in

te rms 'of condi t ions f o r s t a t i n g the problem. In n a t u r a l i s t i c eva lua t ion , these

boundaries are defined in terms of outer l i m i t s t h a t s ignal completion of an

i n v e s t i g a t i o n . ' '

Methodological Relevance

I f we accept the propos i t ion t h a t .different -kinds of evaluat ion ques t ions

requi re d i f f e r e n t kinds of methodologies, then the quest ion of methodological

relevance can be asked as , " Is t h i s par . t icular 'method (or model) appropria te

to the ques t ions t h a t I am t ry ing to answer?" Assessir/g methodological

relevance i s l a r g e l y a. mat te r of determining the t r a d e o f f s , in terms of

s t r eng ths and weaJcnesses, of methods t h a t are ava i l ab le to the evalu^torf ^

Questions of methodological relevance are l a rge ly means-end ques t ions .

I t i s only a f t e r knowing what we are t ry ing to discover t h a t we can decide

how to proceed. For example, i f we wish t o t e s t causal hypotheses, then we-

might choose an experimental design. If we wish t o ac t as the surogate eyes

and ea rs of decisifon makers who des i r e information about what r e a l l y t akes

place in a program, then we might choose a case study approach.

Determining me^odologica l re'levance r equ i r e s (1) a review of the

condi t ions which must be p resen t to f a c i l i t a t e the use of a given method

and (2) a careful considera t ion of the i n t r i n s i c s t r eng ths and weaJcnesses

Page 9: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•A

Rigor and Relevance

7

{

1 of the method. ^ For ,e'xample, in order to make use of an eJJ^erimental or

quasi-expeximental evaluation design, conditions such,as the following must

obtain: a clear, precise statement of intended prograp resu l t s ; a reasonably 1 * . ' •

'controlled' program set t ing; a reasonably 'uniform' treatment across

p a r t i c i p ^ t s and over time; a large enough sample; and an abi l i ty to select

apd assign individuals randomly to treatment and control groups, or the

avai labi l i ty of a comparison group. Likewise, the measurement of program

effects by means of objective, standardized instruments such as aptitude t e s t s ,

achievement" t e s t s , or at t i tude scales requires that there be a'program logic

exhibiting valid linkages between the program's goals, the tireatment delivered,

and the instrvunents used to measure outcomes.

To understand the second objective—the process of weighing the in t r ins ic

merits of a given technique—consider the followirfg review of the technique of

documentary analysis- This method involves the a:nalysis of written progifaii\

materials—e.g., interim reports-, ' internal memoranda, ac t iv i ty logs', e tc .—

to gain a clearer ' insight of program planning and operation. I t s s t r 0 ^ t h s

are that i t i s ent irely unolDtrusive and nonreactive, that thf'documents

themselves are unch^ging and express the perspectives of the i r authors in the / . - -

authors' own nati;/al language. On the oth^r hand, ^mong i t s weaknesses are

that the docum 'nt may not be representative, i t i s usually uniperspectival,

represents ^ ique events, ahd may be temporally and spat ia l ly specific. In

a si;nilar way, every method available to the evaluator-can be scrutinized for

i t s in t r ins ic adequacy or merit. , ,

The act ivi ty Of determining methodological relevance i s clearly not a

simple process. The choice of one method over another involves tije evaluator'

Vt.

„ 0.'

Page 10: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

Rigor and' Relevarice

• 8 '

I and other par t ies to the evaluation in a series of trade-offs 'for which no

set of rules will suffice. I t may be tempting to say .that the determination

of relevance should be based on the principle of maximizing u t i l i t y . But

that ra ises the question of how we are to measure u t i l i t y and for whom. A

more plausible approach, may be to argue that instead of seeking optim.al or

maximal solutions to, the problem of methodological relevance, we .should adopt

a strategy of "sat isf icing" (Simon, 1976)—choosing methods which are not

necessarily the best but 'good enough' given the goals of the evaluation, * - - • ' .

the l imitat ions of the metijods themselves, the problems inherent in the

part icular evaluation s i tuat ion, and the needs of relevant par t ies to the

evaluation.

Questions of methodological relevance naturally raige the poss ib i l i ty

of combinibg methods in a single study. The rationale for the use of multiple

methods i s captured nicely by Webb, et_ al_. (1966) : ' '

• ' Once a proposition has been confirmed by €wo or more

' measurement processes, the uncertainty of i t s interpretation

i s greatly xe^uced. The most persuasive evidence canes

through a triangulation of measurement processes. (p, 3, ,

emphasis added)

De.nzin (1978) further suggests that., there are four, types.of triangulation

available: (1) data triangplation--u9d.ng a variety of data sources, . (2)

investigator triangulation—using several different evaluators, (3) methodological

triangulation—using several different methods to examine the same questions so

that the flaws of one metHod can be compensated for by-the strengths of other '

methods, and (4) theory triangulation—using multiple perspectives to in terpre t •

/

Page 11: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

T mm

\

\ ' - ' . Rigor and Relevance

• • 9

the same se t of o-bjectives. However, though the r a t i o n a l e may be v e l l -

\ '^ . established^ ac tua l ly e f fec t ing \a methodological mix in a given study i s a

complicated mat ter (cf. Trend, 19Y9) r e q u i r i n g a g r ea t deal more i n v e s t i g a t i o n .

(

I \ ' • Methodological Rigor

' Unlike methodological relevance \which is an assessment of instrumental

worth, methodological rigor is a deteAnination of intrinsic merit. When

we )ask, "Is this method/evaluation-desJ^gn/type oi. dafa analysis rigorous?"

we are asking whe'ther it meets certain agreed upon standards for a 'good'_ or

'trustworthy' method, a 'sound' design-, or an 'appropriate! type of data

analysis.

Until veri' recently, ttfe only avaiJable ahd^^gree'd upon standards for

methodological rigor v/ere the canons'for what constituted rigorous Scientific

inquiry. For example, in their review pf federal evaluation studies^

Bernstein and Freeman (1975) developed a composite index of scientific

standard^ for measuring the quality (rigor) of ev.aluation-. research. Their it ''• ' • *•

r a t i ng s'cheme i s shown belov in I'able 1.

'i.

Insert Table 1 about here .

The Bernstein and Freeman index is fairly representative of the types of

• \ standards currentli' in use for judging the methodological rigor 'of both .

• research .and evaluation studies. §imilar sets of standards are commonly

employed in assessing the 'internal arid ex'temal validity of 'research designs

(Campbell'and Stanley, 1966; Cook and Campbell, 1979), the psychome'fric'

firoperties of-measurement devices (Guilford, 1954; Nunnally-, 1978), etc.

• . • ; ^ . , , , • •

Page 12: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

' ., ' , " . Rigor and Relevance' . - • , . ' , . ' •

Jt . • _ • ID ' * , •* •• • •• J- '

-standards for-assessing the .methodological rigor of evali^tion stvidies •

conducted in a na tu ra l i s t i c or ethnographic mode are far^iess well developed-s

-' A recent paper by Cuba (In press) represents one of ' the first attempts to c

spe-cify standards 'for naturalistic, studies- Table 2 below displays the

A , * .

criteria which Cuba proposes for assessing the methodological rigor - * • ' ' , " • • • , • ' ' ' > ; ' . -

("trustworthiness"') of« ni^turalis£ic inquiries. Cuba defines the naturalistic

investigator's analog for criteria'such as objectivity ("confirmability"),

reliability ("dependcUoility"), generalizability ("transferabij.ity"), and^

internal validity ("credibility") and lists and'briefly explains mfethods - .

that might be used to determine whether these priteria have been met. ' • •

, _.. -^_J-_4. \- ^ _ — ^ ' ' •' i " • •-

Insert Table 2 about here ' " • • . •

Efforts svich asr this to specify the standards for judging not only"tlie desiqn

but the product of na tu ra l i s t i c inquiries are indispensable in view-of the e

•' growing in teres t in. the use of na tu ra l i s t i c and ethnographic;-methods. * r»

> * # >

Relationships Between Rigor apd Relevance

The'preceding four categories of rigor arid rfelevance—epistemological

rigor, epistemolCgical relevance, methodological relevance, and. ifiethodological

rigor—have been presented in the i r most logical sequence. I t should be

apparent that efforts to frame a question in a rigorous way should commence

only after i t i s determiaed what i t i s that we are askipa-^P^iJ^ewise (assuming .^^-'^^ • . r &> ^ -the existence of standards for judging the rigor of both quanti tat ive and

'• - * jf • . «

quali tat ive methods), i t i s reasonable to believe that 'questions of which

12

Page 13: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•i

Rigor -and Relevirice

. " - 11

• method to use can be s e t t l e d before examining how these methods might be

used (or whether they have been used) in the most r igorous fashion. * ' » , -

This ana lys i s of r i g o r and relfevance has a l so demonstrated t h a t these

not ions are not necessa r i ly inverse ly r e l a t e d . In o ther words) an increase > • •

. in re levance 'need not r e s u l t in a decrease in r i g o r and"Vice-versa. To be

'—sure, the foUr dimensions o f . r i g o r and irelevance a re not or thogonal . For.

• • • • • • • ' A • • . • •

example, an assessment of epis}:emological relevancey.informs the assessment

of methodologica l re levance . Never the less , one need "^lot always t radeoff - , « . ' -

r i go r for re levance . _ . ' , •" ^ Tradeoffs between r4.gor and- relevance •frequently (and qu i t e inappropr ia te ly)

4 • ' " • ' I

. ', ' » . ' characterize the chqice of evaluation and research methods. It-is argued that

, . . A •

<, iine jnust chopse', between r igorous and r e l evan t methods. This demonstrates a ,

confounding of the no t ions of methodologioarl r igor ' and methodological relev,ance.

As was discussed above, the relevance o£ any method can Be assessed with

respec t to the goals of the research-A^w evaluat ion. Methods are i n s t r u m p h t a l i t i e s ; • f ^ , . • .

' the s u i t a b i l i t y df a method for meeting the goals of inqui ry determine i t s -, re levance. Rigor i s another mat te r . Once a ijglevant method ha^^een chosen,

. • ' * t > ' ' ' ' • ' -

' s teps can be taken to ensure the r igorous use of t h ^ t method. The only we l l -• / • ' \ ' ' -

developed and agreed upon standards, for r i g o r kpply t o the «se o f jguan t i t a t i ve

-methods. We have ohly r ecen t ly begun to i n v e s t i g a t e s tandards for r i g o r t h a t

govern the use of q u a l i t a t i v e methods. I t i s not the case t h a t < ^ a l i t a t i v e •

methods are inheren t ly non-r igorous .(and herffce, somehow morfe re levant ) ' , but "

t h a t , a t p re sen t , we are uncer ta in of bow to judge whether they have beeri - . f >• '' . . -

used in. a*, r igorous fashion. ,• '

«.

1 '-,

Page 14: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

r

X

, •' Rigor and Relevance H * . . .

^ ' . . ,12 -

Rigor and Relevance in Vocatioaal Education Evaluation

• »

Given the diversity of approaches to and methods for evaluating

vocational education programs, it"is hazardous to offer general statements

regarding the extent to which the enterprise of vocational education *

evaluation is," addressing these questions of rigor and-relevance.. However, it

is commonplace to find questions of rigor artS relevan'ce addressed to varying

degrees within the context of particular methods or approaches. From these

discussions there emerge several central tendencies wh'ich are discussed below.

Epistemologic&l relevance is emerging as a primary concern after several

years of evaluation efforts. For example, following a two-year study of

vocational education outcomes by the National Center for Research in

^•ational Education (Darcy, 1979, 1980) it was recoimiended that:

Iri planning evaluation studies-, care should be taken to ,«

• . ' determine clearly what is to be -evaluated and 'i7hat . »

criteria, data, and evaluation standards dc^ 'to .be used.

(Darcj^ 1980, p. 70)

This recommendation stems frcStn several findings of this study which point to

i

shortcomings in assessing epistemological relevance: (1) Terms such a§

'outcomes,' 'outcome measuresj' 'program goals,' and 'program benefits' lack

precise definition,. (2) It is not "clear what is being evaluated—outcomes,

groups of students, programs, etc., (3) There is^little app;reciation of the

range, diversity, and complexity of possible outcomes, and (4) The relative importpuice of outcomes vis-a-vis other types of evaluation has not been well

• < • , • "

addressed.

'- *

1 •

a. "at

Page 15: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

/ -:

Rigor and JlelevaAce

13

Ihe issue of epistemological relevancy has-been alluded to in other

ways as well. In discussing the cpllection of evaluatrve" data by/means o'f .^ ^ > ' - y >.\ - ' " • .

a standardized vo9ational, education d^ta system,j |^wes (1978) no,teci/tha-€^

answering Questions of why d.ata'-are to be collec?ted (a dimension of assessing"*"

epistemological relevance) will determine the use and u t i l i t y of such a.

system. Kievit (1978) sought to lay out a rat ionale for linking kinds of y^—:.T"^'

evaluative data to the values perspectives of potent ial usters, thereby^

" • . . r^ . . . •. '

addressing the question of why vocational education programs are in need

of evaluation, > . . • j

Determining epistemclogical rigor has always been, and will l ikely . •.•:'-, '''

remain, a majojr concern of vooational evaluators. For example, Lecht's

(1974) discussion of indicators of "vocational program success qan b^ viewed

largely as an attempt-to address questions of. epiatemological "rigdr in the

definition and ^measur^ent of, those indicators. Mdst ifecehtly, the link

between episjemoj^gical rigor and relevance has been demonstrated in the

vocational education outcomes study noted ea r l i e r . The study attempted to If' document epistemological relevance for outcomes evalu«ition ty reqtiiring

. (1)' a clea;tr ra t ionale for the choice of art outcome, (2) evxi| n<?e of the

appropriateness of-a'-g4ven outcome as a basis for program evaluation,

(3) i l lus t ra t ion of the potential impact of r esu l t s , .and (4) identif ication

of relevant audiences for evaluative information. As noted ea r l i e r , the

Study then addressed epistemological rigor by indicating how outcome

state'ments are to be translated into empirical measures of outcomes. Owing

to the re la t ively recent importation of qual i ta t ive techniques to vocational

.1 r -t-O

/s

Page 16: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

^ " . • 9 " , • Rigor andxiRelevance

'- ' ' ^ ' ^ • > • • • ' -

• • • - " \ ' ' • " , . . ^ ' • f

eva lua t ion ; there are no commentaries on procedures for es tab l i s t i ing

epis temolo^ical r i go r in ethnographic o r n a t u r a l i s t i c .vocational evaluat ion

s t u d i e s . ' ' *

Hbwever, as a l t e r n a t i v e metho'dplogies for eva lua t ing soc ia l programs

have' found t h e i r way in to eva lua t ioos of -O-ocational programs, d i scuss ions

of metJteadological relevance have emerged. Bollahd (1979), for example,

b r i e f l y addresses the question of methodological relevance by l i s t i n g the • -x •

r e l a t i v e s t r eng ths and weaJcnesses of var ious da ta ga ther ing techniques in

. her review of vocat ional education outcomes s t u d i e s . Grasso (1-979) < ;. ' . '

discusses the s u i t a b i l i t y of impact ' '£valuation for meeting the evalua t ion ' .- • ^ •• • • I

reqilirements spe l led out in the 1-976 vocat ional edud^tlon l e g i s l a t i o n . -

Spirer-(1980) po in t s - ' to i the u t i l i t y of the case study method in voca t ional

education eva lua t ion . "Bonnet (1979) d i scusses a l t e r n a t i v e "methods for

mea^Bijring the outcomeajot carreer education in view of the outcome go&ls

s e t by the Office of Car'aer Educa t ion . Final ly^ Rif^el (1980) r ecen t ly

of fered a very re f lex ive p resen ta t ion on- the u t i l i t y of the pase study

approach in the Vocational Education Study, and Peaxsol (1980)' commented on

' the ' impl ica t ions of coijdjining q u a n t i t a t i v e and q u a l i t a t i v e methods.

Methodological r i g q r h^s perhaps been the mast f requent ly addressed

aspect of r i go r and relevance in voca t iona l educatioh eva lua t ion . Most, i f

not £L1, of these d i scuss ions are concerned with specifying s tandards for

s c i e n t i f i c r i go r as i t i s commonly^-perceived in the research community.

Hence, Bolland (1979) spec i f i e s e igh t b a s i c components of a sound research

r epo r t . Morell (1979) and Franchak.and^ Sp i re r (197^) address design aqd

IG-• 1 , .

• \ ,

Page 17: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

t.

Rigor" and Relevance

15

. . . f

s t a t i s t i c a l ' i s sues i ^ the app l i ca t ion of follow-up research to voca t ional

eva lua t ion . Borgatfea "(1979) d i scusses the requirements for 'good',

experimental and quasi-experimental degighs. V , .

Several papers address quest ions of methodological r i g o r and relevance

sifhxiltane'ously in reviewing.a p a r t i c u l a r research technique. ' Pucel (1979)

addresses question's of methddological relevance by po in t ing to the types of ,

quest ions t h a t can be^answered'through long i tud ina l s t u d i e s . He a l so focuses

on aspect^ of methodological r i g o r in the use of the method ( e . g . , solving ^

problems in implementation, specifying types of da ta to be co l l ec ted , e j t c . ) .

Likewise, Franchak, e t a l . (1980) seek to demonstrate methodological re lavance^

by, l inking*the use of long i tud ina l methods to c r i t i c a l data needs in voca t ional

educatipn, and they address problems of r i go r in reviewing bas ic s t r a t e g i e s

and procedures' far Ibngi tud ina l s t u d i e s . S imi la r ly , severa l pub l i ca t ions in

the Career Education Measurement Ser ies ( e . g . , McCaslin, et^ a l • / 1979;

Mctaslin and- Walker, 1979) address both methodological r i g o r and re levance i^

d iscuss ing the s e l e c t i o n , eva lua t ion , and design, of instruments to evaluate

I career education". . ' ' k ,,

In summary, the importance of addressing the i s sues of epis temological

''and methodOjlogical r i g o r and relevanc* can be seen in Lee ' s (1979) discuss ion

of the f ac to r s governing the use of evaluat ion da ta . Lea! i d e n t i f i e s the

jEollowing f ive f a c t o r s : (1) avai la±) i l i ty (making, evalua t ion data ava i l ab le

^ to users i M ^ way t h a t can.be rea'dil^' understood) , (2) r e l i a b i l i t y , (3)

c r e d i b i l i t y , (4) u t i l i t y ( c o l l e c t i n g , analyzing, and i n t e r p r e t i n g evalua t ion

data in view of t h e i r p o t e n t i a l u ses ) , and (5) c o n s i s t e n c / ( c o l l e c t i n g ,

Page 18: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

•"..

Rigor and' ReleVanae . • • > . ' '

1 6 ' , '•

analyzing, and making available data within the boiindaries of possible

action). I t i s possible to recast these factors as functions of addressing . • -* ri^or and relevance in evaluation studies• Thus, failure 'to address ^^ ' ' ^ .

qiibstions of episi^emological fe lev^ce may lead to problems in consistency » ** *

and utility; failure to address questions of qpistemological ri^or and questions

of methodological rigor may result in problems with reliability and credibility;

finally, failure to address questions of methodological relevance may lead to

p'roblems with utility and. availability.

Avenues for Future Study >

r ' y • '

All four dimensions of rigor and relevance discussed in this ^aper

warrant further attention by "the community of vocational education evaluators

and researchers. Epistemological relevance—determining what i-s to be

evaluated and why—must clearly be our- foremost concern. Premature' focus on \ fe-the selection, of appropriate methods will l ikely Encourage th,e approach of

'solutions in. search of problems.' That i s , we may" attempt to f i t existing

(and new auid developing)' evaluation s t ra tegies to par t icular vocatioiikal

education evaluation problems without f i r s t understanding what i t is we wish

to know and why. There should be no equivocating^: arguing that we have the

' r igh t ' solution but the 'wrong' problem is simply an argument for the'wrong

solutioa. We should not hesitate to re t reat , from solutions to make a more

careful diagnosis of the problem, ' , . _

Attending ta epistemological rigor presents us with two different types " ! •

of problems. It appears that we are fully in possession o^ the knowledge of

\ .

t r

1

Page 19: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

Rigor and Relevance

17

what constitutes/ a testable or measurable problem from a pos i t iv i s t i c

perspective. What we heed are morfe attempts, such as that demonstrated in

1;he vocational/education outcomes jstudy, to ap|)ly th is knowledge to par t icu la r

evaluation questions. On the other hand, as na tu ra l i s t i c inquiry becomes

increasingly/relevant to evaluations of vocational, education progtams, we

will need to/devote our efforts-tp specifying procedures for determining the

boundaries of such investigations'- The lack of a p r io r i constraints ,

character is t ic of»this Approach, Idoes not imply a to ta l lack or regard for

• J ' ' . ' • '

constraints which demonstrate r:^ifor in the investigation of problems; .

Methodological relevance—including both an assessment of the in t r ih s i c

merits ojf methods and an investigation of the poss ib i l i t i e s for combining • / . • ^

methodsf-demands our most careful at tention, les t the choice of methods becomes sunply a matter of what i s 'current ly in vogue. We must guard against

nqi the nqrmative appeal of certain established methods as being the most,(or

/ • • • - ' • ' ' ' '

the only) 'rational' strategies and investigate the contextual limils •/ " " , : • '

governing the .scope of these s t r a teg ies . We must be careful not to mistake

evidence of the inapplicabil i ty of certain methods as simply problems with

implementation. , / • •

Finally,^ in the area of methodological r igor, we lack l i t ' t l e knowledge

of t radi t ional .standards for assessing the sc ient i f ic ac^equacy of quanti tat ive

methods and experimental designs. Yet, we are largely ignorant of how to -^

judge the merit of case studies, emergent designs) and similar methods and

tools associated with na tura l i s t i c or ethnographic inquiries. .

In generaj., we need to become more open and public about our discussions

or rigor and relevance. There are re la t ively few-accounts of the conduct of

f

1 I \J

Page 20: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

T - ^

/'" •. T Rigor and Rel,evance

18

vocational educdi:ion. inquiries that are reflexive. Reflsxivity refers to the [ . ' ' . . ' *' ^

capacity of thought to bend back upon i t se l f , to become an object to i t s e l f

(Ruby, 1980). TO'be rfeflexive i s hot the same as being self-conscious or

ref lect ive. Most evaluators and researchers are.probably self-conscious, yet

that ;kind of awarenes^ remains private knowledge for the inquirer, detached

from the product of his or her inquiry. There are re la t ively few accounts of

inquiry in which inquirers reveal the epistemological "and ajciological

assumptions which caused them to choose a par t icular set of questiions to «

investigate, to seek answers to those questions in a par t icular way, and, , :i?acular way. By' finally, to present their findings in a pa r^cu la r way. By' engaging in this

kind of. ref lexivi ty about our research cCnd evaluation, we are more l ikely

to address c r i t i c a l issues in rigor and relevance.

^

/

'•V

2'.

Page 21: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

1 ' /•

TABLE 1

Criteria for Assessing the Quality

of ^.valuation ,Resea;rcU.*

Criterion

Sampling

Rating

systematic nonraiidom, cluster, or nonsysbematic

D^ta Analysis 2 - q t i an t i t a t ive 1 - q u a l i t a t i v e and q u a n t i t a t i v e -0 - q u a l i t a t i v e ' '

S t a t i s t i c a l Procedures

4 - mu l t i va r i a t e 3 - descr<iptive 2 - r a t i n g s from q u a l i t a t i v e da ta 1 - n a r r a t i v e da ta o n l y 0 - no systemat ic ma te r i a l

Design

.'-

4

.•

-' i '

* • 1 "

SajnpUing'- •"'*

/ < \ : ' ' • • ' •

< • ' , ,

• ^5easyrement • Proce'cjjxres <

3 - experimental or quasi-experimental with randomization and cont ro l grdu^s

2 - experimental or quasi-experimental withou-t both randomization and con t ro l groups

1 - l ong i tud ina l or c ross ' - sec t ional without cont ro l or comparison

0 - d e s c r i p t i v e , n a r r a t i v e f

"2 - r ep re sen ta t i ve 1 - poss ib ly r ep resen ta t ive \ 0 - haphazard

1 - judged adequate in face validity 0 - judged less than adequate in face,validity

'i-

/ •Berns te in , -Freeman,' 1^74/pp . , 100-101

> >, y

/ ' *'* -\

2. 9

Page 22: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

• >.

TABLE 2

• C r i t e r i a f o r A s s e s s i n g t h e T r u s t w o r t h i n e s s

of N a t u r a l i s t i c I n q u i r i e s *

Aspect' of Method, S tudy, Procedure N a t u r a l i s t i c - Term

•Methods f o r De te rmin ing Whether C r i t e r i a Are Met

Tru th Value

Applicability

Credibility

T r a n s f e r a b i l i t y .

• #

Pro longed engagement a t i ^ s ' i t e r ' P e e r d e b r i e f i n g , . T r i a n g u l a t i o n , .Member c h e c k s , C o l l e c t i o n ,of r e f e r e n t i a l a d e ^ a c y m a t e r i a l s *,

Theore t ica l /puj rpQS i v e sai t5>JingV-Collect ion of"-' " th i ck" . "desc r ip t ive" d a b i ,

C o n s i s t e n c y

N e u t r a l i t y

D e p e n d a b i l i t y

C o n f i r m a b i i i t y

-. ^ -Overlag? methods, S t ^ w i s e ' ' . ^ e i p l i c a t i o n , E s t a b l i s h \

^ _ ' "aua l t " : t r a i j l , -

'. • • Triahgul'ation^"'. C o n f i r A a b i l i ' t y aud i ' t -*' '

• i' . - , • , . - - .

J •

» ft

"Cuba, Inptess, p a s s i m

ic ^ ' S - ' W I

•- ^ r ( . '•

» •

$ . •

Page 23: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

• • References

i • . • • - •

Bernstein,, I .N . , s Freeman,-H.E;-* Academic and Ent repreneur ia l Research. > ',

; - New York: The RusSell Sage 'Foundation, ^975.•

" ^ - ' : ' " • ' . • " ' : ' . • • . •

Bollahd, K.A. Vocatipnal .Education Outcoities: ^ Evaluat ive Bib l iography

of Empirical Studies'.>, Columbus,^',OH:'r The Nat ional Center for Research

Tn Vocational•Education, i979 , , , , "

Bonn'et, D.G. Measuring Sti/dent Learning in Career Education. In T.

• • * ••' '1'' • - " ' 7 Abramson, C.K". "Tittle, L. Cohen (Eds.), Handbook of Vocational

i- T f— •• Education E y ^ u a t i o n . Beverly -Hil ls , CA: Sage? 1979.

Borgattar E.F,' ^Methodological Considetations•• Experimental 'and

Nc^n-experimental Designs and Causal Inference . In T. Abramson,

et' a l . , (Eds.) , Handbook of Vocational Education Evaluat ion.

, ^Beverly H i l l s , CA: '^Sage", 1979.

Campbell, D.T., '& Stanley , J .C. Exp.erimental and Quasi-experimental

Design^ 'fo-r. Research. Chicago: Rand McNally, 1966.'

Cook, T.D., & tampbeir , D.T. Quasi-experimentat ion: Oesign and Analysis

Issues for Fie ld S e t t i n g s . Chicago: Rand McNally,•1979.

'».«

Page 24: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

References, continued

Darcy,-R.L. Some Key Outc6mes of Vocational Education. ' Coltunbus, OH:

. . . . •, . -• , The I4ational Center fo^ Research in Vocat ional Education, 1980. s.

Darcy, R:L. Vocational Educfation Outcomes:' Perspective for Evaluation.

Columbus, OH: The National Center for Research in Vocatioi;ial

, Education; 1979. ' •• ' .

* * * '* *

Denzin, N.K. The Research Act. New york: McGraw Hill, 1978. '

<• . : . • '

Drewes, D.W. Outcome Standardization -for'Compliance or Direction;

. 'The Critical Distinction. Paper presented at the National

Conference on Oatcome Measures, Louisville, Kentucky, August, 1978.

. Filstead, W.J. Qualitative Methods: ' A Needed Perspective in Evaluation

Research, ;tn T.D. dook, C.S. Reichardt, (Eds.), Qualitative and

' Quantitative Methods in Evaluation Research. ' Beverly Hills, CA:

Sage, 1979. , ' ' ' '

Franchak, S.j. , Frank-en, .M*E. , S Subisak, J. Specifications fot ^ , , _ _

Longitudinal Studies. Columbus, OH: The National Center for .

Research in Vocational Education,-1980. ' •

Page 25: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

References,** continued • •

> • ">

-Franchak, S.J. , & Spirer, J.E. Evaluation Handbook, Vol. 1; Gqidelines

and'Practices for Follow-up Studies of Former Vocational Education

Students. Golumbus, OH: The National Center for Research in «

VocaN^ional Education, J.978.

Grasso, J.T. Impact Evaluation; .The State of the Art. ColumbusV ,OH:

:nter The National C^nter^ for Research in Vocational Education, 1979.

Guba, E.G. Criteria for Assessing the Trustworthiness of Naturalistic

Inquiries. In press.

Guba, E.G. Toward a Methodology of Naturalistic Inquiry in Educational

Evaluation. Los Angeles, CA; Center, for the Study of Evaluation, »>

UCLA Graduate School of Education, University of California^ 1978.

Guilford, J.P. Psychometric Methods. New York: McGraw Hill, 1954.

'Horst, P., et al. Program m.anagement and the federal evaluato-r. Public

Administration Review, July/August 1974, 300-308. •

Kievit, M.B. Perspectivism in Choosing and Interpreting Outcome Measures

in Vocational Education. Paper presented at the National Conference .

on Oufecome. Measiires, Louisville, Kentucky, August, 1978.

or-

Page 26: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

References, continued ' . »- '

— - V V

Lecht, L.A. Evaluating Vocational Educa t ion-Pol ic ies and Plans for the

1970s. New York: Praeger , .J.974« • -

*'<Lee, A.M. Use of Evaluat ive Data-by Vocational Educators , ^ lurabus , OH:'

'The National Center for Research iri Vocational Education, 1979<

•• Mc'Caslin, N.L. , Gross, C . J . , &-Walker, J . P . Career Education Measures: '

A Compendium of Evaluat ion Instruments . Columbus,. OH: The Nat ional

Center for Research in Vocational Educa_tion, 1979. - . . , . _

"• • • • . c

'McCaslin,. N.L. , s Walker, J . P . A Guide for Imprp.ving Locally Developed V

Career Education Measures; Columbus, OH: The National Center'for

Research in Vocational Education, 1979.

Morell, J. Followrup-Research as Evalua'tion Strategy: T?heory and

Methodologies. In T. Abramson, £t al. (Eds.), Handbook of Vocational -s

Education. Evaluation. Beverly Hills, CA: Sage, 1979.

i

Nunnally, J .C. Psychometric Theqry (2nd e d . ) . New York: McGra'ii H i l l ,

1978. _ " ^ ' , _

T ' ' ' r . y

Patton, M.Q.^ Qualitative Evaluation Methods. Beverly Hills, CA: Sage, '

1980: ,m

o G

Page 27: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

Rife rences, continued " ^ ^—- ' •' \ ht •

" . ^ <"

Pearsoi'; J . A. ''Combilning Quant i ta t ive and. Qua l i t a t ive Educational Research

Methods: A Pe r spec t ive . .Pape r presented a t the JIhnual Meeting of

the American Vocational Associat ion, New Orleans , •Louis iana , December, ,

> 1-980. . - ' ..

Pucel , D.J. Longi tudinal Methods as Tools for Evaluating Vocational Education.

Columbus, OH: The National Center for Reseai-ch in Vocational Education, ' •4-

' 19.79. . •

R i f fe l , R. The Case Stuc^y and I t s Use on the National ]8 i s t i tu te of Education

V o c a t i ^ a l .Education Study.' .Paper presented a t the Annual Meeting of

the American Vocational Associa t ion, '^ew Orleans, Louisiana, December,

1980. " * -'

Ruby, ' J . Exposing yourse l f : Ref l ex iv i ty , anthropology, ' and f i lm. Semiotica,

. " 1980,' 30, i ja-?-?. ^ .

I

Scriven, M. Pyoduct Evaluation. Research and Evaluation Program Paper >

and Report Series No. 29. Portland*, OR: The Program, Northwest .

^Regional Eduoitional Laboratory, 979.

Simon, H.A. Admihis,trative Behavior (3rd edjj . New York.: 'The Free P^ess,

1976." » ^ i

Page 28: DOCOMENT BESOHE.' - WordPress.com€¦ · in the context of soci-al science research and evaluation reveals an epistemological and a methodolo'gical usage of each term.. These four

. i

V

J

**

References, continued _ . • ,

Sp i re r , J .E . The Case Study Method; Guidel ines , P r a c t i c e s and Appl ica t ions

for Vocational Education Evaluat ion. Columbus, OH:" The Nat ional

Center for Research in Vocational Education, 1980.

S te ine r , E.S. Logical and Conceptual Analyt ic T^^hniques for E<^cational

Researchers. Washington/ DC: Univers i ty Press of America, 1978.

Trend, M.G. On the Reconci l ia t ion of Qua l i t a t ive and Quan t i t a t ive f , *

Analyses: A Case Study. In T.D. Cook, C.S. Reichardt (Eds . ) ,

Q u a l i t a t i v e and Quant i ta t ive Methods in Evaluation Research.

Beverly H i l l s , CA: Sage, 1979.

Webb, E . J . , e t - a l . Unobtrusive Measures. Chicago: RawJMcNally, 1966.

Wholey, J . S . Eva luab i l i t y Assessment'. In L. Rutman (Ed. ) , Evaluat ion

Research Methods; A Basic Guide. • Beverly H i l l s , CA: .SageJ 1977. ,

Wholey, J . S . Evaluat ion: When i s i t r e a l l y needed? Evaluat ion, 1975, 2_,

89-93. . • " • '«»

\

/