on evaluating the secondary impact of leadership programs: a … · 2016. 10. 28. · when...

20
Drawing above used with permission of the artist, Rebecca Grace Jones. On Evaluating the Secondary Impact of Leadership Programs: A White Paper Justus Baird and Aaron Dorfman Processes and methods to evaluate the impact of leadership programs run by nonprofit agencies, particularly of programs designed to leverage leaders for impact on their communities and society. Produced with the support of a Wexner Graduate Fellowship Alumni Collaboration Grant

Upload: others

Post on 01-Oct-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

Drawingaboveusedwithpermissionoftheartist,RebeccaGraceJones.

OnEvaluatingtheSecondaryImpactof

LeadershipPrograms:AWhitePaper

JustusBairdandAaronDorfman

Processesandmethodstoevaluatetheimpact

ofleadershipprogramsrunbynonprofitagencies,

particularlyofprogramsdesignedtoleverageleaders

forimpactontheircommunitiesandsociety.

ProducedwiththesupportofaWexnerGraduateFellowshipAlumniCollaborationGrant

Page 2: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

1

Summary

Thisreportoutlinesanapproachandsomespecificmethodsforevaluatingaparticular,andoften

neglected,aspectofnonprofit leadershiptraininganddevelopmentprograms.Whiletheprimary

(or direct) impacts of these programs on the participating leaders are often evaluated and

understood, it’s less clear how leadership programs ought to think about and evaluate their

secondary impact on the communities inwhich their programalumniwork. This report offers a

conceptualframeworkandtoolsforstaffofleadershiptraininganddevelopmentprogramstouse

inevaluatingthatsecondaryimpact.

INTRODUCTION

There’sanapocryphalstoryinwhichaHarvardBusinessSchoolstudentvisitsthecoursematerials

officeatHBSandisdisgruntledbythelackadaisicalservice.Fedupwithwaiting,hefinallyrapshis

knucklesonthecounterandbarks,“Hey!Canyouhurryitup?I’mthecustomer!”Inresponse,the

clerkcasuallylowersthecrosswordpuzzleinwhichshe’sbeenengrossedandremarks,“You’renot

thecustomer.You’retheproduct.”Thisjokeencapsulatesaperceptionthatisoftenheldamong

those who manage programs focused on leadership training and development: that the

participantsarethe“product”ofthelearningprocess.

Many nonprofits run programs to invest in future

skilled leaders and professionals. While we are

deeply committed to the intrinsic value that these

programsoffertothosewhoparticipateinthem,we

should be equally committed to understanding the

impactthattheleadersinwhomwe’veinvestedhave

on their communities and networks, what are

referredto in this reportas“secondary impacts.” In

fact, the real measure of our success is not the

leader’s experiencealone, but also the change they

catalyze in their communities, their networks, and

thelargerworld.Inthiscontext,ourparticipantsare

alsoameanstosupportourlargervisionsforchange.

Our implied theory of change is that investing in

leaders translates to better-equipped leaders who

canmoreeffectivelypursue thechangeweallneed

to see in theworld. Themeasure of our success is

howmuchmoreeffectivelytheyrealizethatchange.

This report was born out of a shared need of its

authors, staff leaders at two different nonprofits,

American Jewish World Service and Auburn

Seminary.Wewerebothoverseeingmajorprograms

Case study from the Talmud: Training

leaderstoachievesecondaryimpacts

(Thisstoryabout leadership inamomentof

crisis comes from the Babylonian Talmud,

BavaMetzia85b.)

OnedayRabbiHaninaandRabbiHiyyawere

arguing, and Rabbi Hanina said to Rabbi

Hiyya: ‘How can you argue with me? If,

Heaven forbid, the Torah were forgotten in

Israel, Iwould [be able to] restore it bymy

argumentative powers.' To which Rabbi

Hiyyarejoined:'Howcanyouarguewithme,

whoactuallypreventedtheTorahfrombeing

forgotten in Israel? How did I do it? I went

and sowed flax, made nets [from the flax],

trappeddeerandgavethemeattoorphans,

andpreparedscrolls[fromtheirskins],upon

whichIwrotetheFiveBooks[ofTorah].Then

I went to a town [where there were no

teachers] and taught the Five Books to five

children,andthesixorders[oftheMishnah]

to six children. And I told them: "Until I

return, teacheachother theFiveBooksand

the Mishnah;" and thus I preserved the

Torahfrombeingforgottenin Israel.'This is

whatRabbi[YehudahaNasimeantwhenhe]

said,'HowgreataretheworksofHiyya!'

Page 3: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

Background

2

thataspiretostrengthentheleadershipofacohortofleaders.Forbothofus,theinvestmentin

leaderswasclearlyameanstoanend:therealdesiredoutcomewasmakinganimpactonsociety,

andtheleadersweweretrainingservedasvehiclesforachievingthatimpact.Wewereoperating

under a hypothesis about social change that presumes that great leaders make great things

happen, so investing inpeople tohelp thembecomegreat leaders—inspiring,empowering,and

trainingthemtocatalyzecommunaltransformation—wouldleadtogreatchange.

Ourbare-bonestheoryofchangelookedsomethinglikethis:

Whilewehadmanytoolsforevaluatingourprogram’simpactontheleadersthemselves(A->Bor

“primaryimpact”),whenitcametounderstandingormeasuringhoweachleader’sexperiencein

ourprogramcontributedtoanimpactonthecommunitiestheyled(A->Cor“secondaryimpact”),

wewereataloss.

Stirredbythisfrustration,weproducedatwo-dayseminarfornonprofitstafftostudyapproaches

to evaluating the secondary impacts of leadership programs. We are grateful to the Wexner

Foundation for underwriting the costs associated with the seminar (through an alumni

collaborationgrant),heldinApril2015attheNewYorkofficesofAmericanJewishWorldService.

Twenty-two colleagues who lead similar leadership programs (mostly in the Jewish nonprofit

sector)participatedintheseminar.Wehiredtwoexpertsinevaluationtoleadusinourlearning

andreflection:Dr.JewlyaLynnfromtheSparkPolicyInstituteinDenverandTanyaBeerfromthe

CenterforEvaluationInnovationinWashington,D.C.Thereportsharesthetoolsandinsightswe

gained through the seminar, in the hopes that other nonprofit staff and funders designing and

supportingleadershipeffortscanbetterassesstheimpacttheirleadershaveonthecommunities

theyserve, inorderto learnandimprovethoseprogramstoeffectyetmorepositivechangefor

ourworld.

AOrganizabonrecruitspeoplewithgreatleadership

potenbal,andinspires,

empowers,andtrainsthemtoadvancetheorganizabon's

mission.

BLeadersemergefromleadershipprogramandbegintouse

theirnewskillstocatalyze

changeamongpeopleintheirnetworksandcommunibes.

CChangesinthe

adtudes,knowledge,andbehaviorwithin

leaders'communibessupporttheimpactthe

organizabonisseekingtomake.

Page 4: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

Background

3

EVALUATINGIMPACTINACOMPLEXSYSTEM

Theleadershipprogramsatissueaddresswhatareoftencalledcomplexproblemsorsystems.As

opposed to simple problems (baking a cake) and complicated problems (sending a rocket into

space), complex problems or systems—such as raising a child, changing immigration policy, or

solving climate change—are characterized by a multiplicity of interconnected factors, some of

whichareunknownorinvisible;aninabilitytopredictconsequences;andunstablecontexts.

When evaluating impact in the context of a complex system, experts encourage evaluators to

develop a plausible case for how a program contributed to the outcomes, as opposed to

overstatingthecase.TheSparkPolicyInstituteencouragesevaluatorsto:

• Aimforcontributionratherthanattribution(i.e.don’tassumethatanyparticularoutcome

isentirelyattributabletoasingleintervention);

• Lookforemergentoutcomes,especiallythoseyoudidn’tanticipateorexpect;

• Keepoutcomesrealistic;

• Differentiatebetweenoutcomesthatyou“expecttosee,”“wouldliketosee,”and“would

lovetosee”;and

• Use a “golden spike”1to link to other evidence: connect your impact to already proven

causalrelationships,focusingyourevaluationonthegapsornewlearning.

Perhapsthemostimportantapproachtoevaluatingcomplexsystemsistofocusmoreon“learning

andadaptation”thanontraditional“accountabilityforimpact.”

Evaluations that focus on learning and adaptation encourage developmental learning and risk

taking. The complex problems our programs are addressing are never entirely solved; what

worked last timemaynotwork in the future; andwhatworks in one contextmaynotwork in

Page 5: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

4

another.Developmentalevaluationmethodstrackeffectsastheyunfold,notexpectingmethods

orevenintendedoutcomestoremainstable.Inthiscontext,evaluationencouragesstafftoadapt

towhatislearned,acknowledgewhatisn’tworking,andusethenewknowledgetoinformfuture

decisions.IntheirbookGettingtoMaybe:HowtheWorldisChanged,Westley,Zimmerman,and

Pattoncall thisa“safe-fail”approach,asopposedtoa“fail-safe”approach2.Thepoint is that in

thecontextof complexsystems, spending lotsof timeand resourcesonmeasuring theoriginal,

predictedimpactmaynotbeaneffectivewaytogettotheultimateoutcome.

THREE-STEPCYCLEFORBETTEREVALUATION

Movingfromtheideaofmoreeffectivelyevaluatingleadershipprogramsinacomplexsystemto

actuallydoingitcanbedaunting.Wehavefoundinourownworkthatittakessignificanttimeand

energytogetinstitutionalbuy-inandinvolvecolleaguesinnewbehaviors.WhetheryoutryaDIY

(“doityourself”)approachorinvolveaconsultant,thesethreestepsformthecoreelementsofan

evaluationandlearningcycle.

Thefirststepistoreflectonwhatyouaretryingtoachieve(outcomes)andhowyouaretrying

toachieveit(changepathways).Itisimpossibletoeffectivelymeasureorlearnaboutimpactuntil

youhavefirstarticulatedasclearlyaspossiblewhatthedesiredoutcomesareandhowyouthink

theywillcomeabout.Thesecondstepisto identifyandutilizemethods to collectandanalyze

datathatcanmeasurethedesiredoutcomes.Evaluationexpertshavedevelopedawiderangeof

creativemethodsand tools for this step,wellbeyond the surveys, interviews, case studies, and

focusgroupsthatwearemostfamiliarwith.Andevenforsomeofthesetried-and-truemethods,

expertsproposerefinementsthatmakethemmoreeffective inthecontextofcomplexsystems.

Thethirdstepistofeedthelearningbackintotheongoingprogramdesignandimplementation.

Step1:ArticulateMeaningfulOutcomesandChangePathways

Mostprogramstreatevaluationasanexerciseincountingheads(e.g.,Howmanypeopleapplied

to/participatedin/graduatedfrommyprogram?)and/orcustomersatisfaction(e.g.,Howwould

you rateyourexperience inmyprogramona scaleof1-5?).Wedesign leadershipprograms to

leverage leaders to effect meaningful change in the world. The first step in evaluating these

programsistoarticulatethechangeweseektoproduceandthewaywethinkitwillmaterialize:

What dowewant to see happen in the communities our leaders serve?Willmore people join

thosecommunities?Willcommunitymembersengageinmore—orbetter—politicalactivism?Will

they engagemore productively in civil discourse around contentious issues?Will theymanage

theirinstitutionsmoresustainably?

Onewaytobeginarticulatingtheoutcomesandchangepathwaysofa leadershipprogram is to

developapersona,aniconic/archetypalstoryofaprogramparticipantwhoexemplifieswhatthe

program is trying to accomplish. Using personas can help concretize program design in the

observedinterests,experiences,andperspectivesofrealpeople.Developingapersonaconsistsof

thefollowingkeysteps:

Page 6: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

5

1. Identifytargetusers

2. Developaresearchprocess(interviewandobservationprotocol)

3. Learnfromtargetusers(throughinterviews,focusgroups,observation,etc.)

4. Analyzeandsynthesizelearningintooneormorepersonas

5. Produceawrittenprofileforeachpersona

ShlomoGoltzhaswrittenahelpfulguide fordeveloping robustprogrampersonas.Below,you’ll

also find a case study from one of the program directors who participated in our evaluation

seminar. While this is a description of an actual program participant and not a synthesized

persona, it conveys the value of having an individual story to capture the kind of change your

programisdesignedtoproduce.

Anothervaluabletoolforarticulatingoutcomesandchangepathwaysisthelogicmodel.(SeeW.K.

Kellogg Foundation’s logic model development guide.) A logic model is a flow-chart diagram

depictingthecausalrelationshipsamongthevariouselementsofprogramdesign.Whilethereare

manylogic-modelformats,mostincludeanddescriberelationshipsamongthefollowingelements.

• Plannedwork:

§ Inputs: resources invested in the program (e.g., funding, staff time, materials,

politicalcapital,etc.)

§ Activities: activities undertaken using inputs (e.g., training programs, political

advocacy,directservice,etc.)

• Intendedresults:

§ Outputs:directandusuallyquantifiableproductsresultingfromactivities(e.g.,#of

people served, # of lobbying visits or petition signatures, customer satisfaction

targets,etc.)

§ Outcomes: specific changes in program participants’ attitudes, behaviors,

knowledge, skills, etc., that result from the program (e.g., reduction in child

malnutrition,passageofspecificlegislation,increasein#ofskilledvolunteers,etc.)

§ Impact: systemic transformation supported by the program’s outcomes (e.g.,

healthier familiesandchildren, reduction incrimerates/violence, increase incivic

engagement,etc.)

Page 7: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

6

Some logicmodels also include articulations of the problem the program is designed to solve,

underlying assumptions that inform the logic model design, and critical factors in the external

operatingenvironment.

Logicmodelscanbeusedformanypurposesinbothprogramdesignandevaluation,buttheircore

utilityisinmakingexplicitandtransparentthecausalrelationshipsamongthevariouselementsof

aprogram.

3

Page 8: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

7

AnExampleinPractice:BronfmanYouthFellowships(BYFI)

Julie(notherrealname)grewupinapartoftheU.S.withasmallJewishcommunity.Shebelongedto

a Reform Jewish synagogue, attended Hebrew School, and participated in Reform Jewish summer

programsgrowingup.Her limitedexposure to Jewishdiversity resulted invery few friendshipswith

JewsoutsideoftheReformmovement.

JulietraveledtoIsraelwithBYFIthesummerbeforehersenioryearofhighschoolandcontinuedwith

BYFIseminarsduringher senioryear.Theprogramexposedher to Jewishpluralism,text study,and

Israel in a multifaceted way. She built close relationships with Jews different from her (e.g., she

emergedfromtheprogrambestfriendswithanobservantOrthodoxyoungwoman).

Juliewentontoattendan IvyLeagueschool.Becauseofherexperience inBYFI, shedecided toget

involved and joined a prayer community, but came to see that the Hillel—dominated by Orthodox

students on her campus—seemed to alienate Jews frommore secular backgrounds and thosewith

left-leaningperspectivesonIsrael.

As a junior, Julie decided to run forHillel president against a politically conservative and religiously

Orthodoxfellowstudent.Shewastheunderdog,butspenta lotoftimereachingouttostudentson

thefringesandthoseinthecenter.HerBYFIexperiencegavehervaluabletoolsforhowtospeakwith

Jewsfromdiversebackgroundsandhelpedherhonoreachperson’spointofviewasequallyvalid.She

gainedpeople’strustandwontheelection.

AmonthpriortotheelectionshewasfeelingdisheartenedaboutpolarizedattitudestowardIsraelon

campus.SheattendedaBYFIseminarforcollegealumnidesignedtocreateasafespacetoexplorethe

Israel discourse on campus, provide discussion tools, and build community for students who care

about Israelacrossdifferentpoliticalperspectives.Shebrokedown in tearsat theseminar,buthad

severalmentoringconversationswithstaffandpeersthatbolsteredher.Aftertheseminarshewrote

to staff saying the retreathad rejuvenated her and given her increased confidenceand strength to

keep her campaign going. The sense of being part of a community that believed in her and the

reminderthatthereareJewishspacesthatcanenableheartfelt-yet-civildiscussionaboutIsraelgave

herhopeandenergy.

AsHillel president she focused onmaking those on themargins feelwelcome.She hosted the first

everpre-Passoverbagelbrunch,withover250studentsattending(mostofwhomwerenot“theusual

suspects”).ShehasalsoconvincedtheHillelprofessionalstafftodevelopprogramsthatprovidesmall-

groupsettingsforIsraelconversationtobringpeopletogether.

Onelastpoint:Juliewenttocollegeintendingtomajorinpoliticalscience.Becauseofherexposureto

arangeofJewishtextsandideasthroughBYFI,shedecidedtotryYiddishandisnowaYiddishmajor.

~BeckyVoorwinde,ExecutiveDirectorofTheBronfmanYouthFellowships

Page 9: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

8

Step2:OpentheEvaluationToolbox:UtilizeaRangeofMethodstoCollectand

AnalyzeData

Ifthefirststepisarticulatingmeaningfuloutcomesandthemeansbywhichweintendtoachieve

them,thenextstepistoopentheevaluationtoolboxtofindmethodsandtoolsthatcaptureand

analyzedatarelevanttothoseoutcomes.

Thecentraltaskofthisstepistomatchoneormoreevaluationtoolstoyourdesiredoutcome(s).

Todothat,youneedtoexploretheevaluationtoolbox,andthenreflectonhowyouwouldusethe

datathataspecifictoolwouldgenerate.Sofirstwe’llintroduceyoutothetoolbox,andthenwe’ll

offerasetofquestionstoreflectonasyouconsiderpotentialevaluationmethods.

Oneofthemostexcitingmomentsoftheseminarwaswhenourtwoevaluationexpertswalkedus

throughasetof15evaluationtoolstheyandotherexpertshavedevelopedforprogramproviders.

Whileitisdifficulttorecreatethatexperienceinawrittenreport,webelievethattastingthe

varietyofevaluationmethodsthatareavailableisanimportantexperienceintheprocessof

improvingourapproachtoevaluating/assessingourprograms.

Belowarebriefdescriptionsofahandfuloftheseevaluationtools.Attheendofthisreport,you’ll

findabriefsummaryofall15differentapproaches,achartyoucanusetoidentifyadditional

approachesthatyoumaywishtoexplore.Dependingontheevaluationexperienceofyourstaff,

youmaydecidetohireanevaluationconsultanttoidentify—andsupportyouinusing—themost

relevanttoolstocollectdatarelatedtoyourdesiredoutcomes.

IdentityLeadershipInventory

UsetheIdentityLeadershipInventorytomeasurealeader’sabilitytomobilizeanddirectfollowers’

energies.

The Identity Leadership Inventory (ILI) is a simple, validated3survey instrument designed to

measure a leader’s influence on her or his community vis-a-vis four interconnected

relationships:

• Theleader’sabilityto“beoneofus”

• Theleader’sabilityto“doitforus”

• Theleader’sabilityto“craftasenseofus”

Page 10: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

9

• Theleader’sabilityto“embedasenseofus”

By deploying the ILI in a leader’s community before, during, and after a leadership-training

intervention, you can infer some of the intervention’s value in supporting the leader’s

influence in her/his community. You can find the ILI survey instrument and guidance for

administeringitinNiklasK.Steffen,etal’s

“Identity Leadership Inventory (ILI) Instrument and Scoring Guide (ILI Version 1.0),”

downloadablehere:https://www.researchgate.net/publication/271071707_ILI_Guide

OutcomeHarvesting

UseOutcomeHarvestingtoidentifychangesinacomplexsystemandworkbackwardsto

determinewhetherandhowaninterventioncontributedtothosechanges.

Theoutcomeharvestingmethodologyisunusualinthatitdoesnotseektomeasureprogress

towardapredeterminedsetofoutcomes.Instead,itstartsbyidentifyingchanges(e.g.,shifts

in communal behavior, adoption of new regulations/legislation, emergence of new

memes/messages,etc.)withinasystemaffectedbyanintervention(e.g.,acommunitywhose

leaderhasexperiencedaleadership-trainingprogram)andthenworksbackwardtodetermine

whetherthere’scredibleevidencethattheinterventioncontributedtothechanges.Itseeksto

answerthefollowingquestions:

• Whathappened?

• Whodidit(orcontributedtoit)?

• Howdoweknowthis?Istherecorroboratingevidence?

• Whyisthisimportant?Whatdowedowithwhatwefoundout?

It’sparticularlyapplicableincontextsinwhichtheconnectionsbetweencauseandeffectare

ambiguous. You can find detailed guidance on the outcomes harvesting methodology in

Ricardo Wilson-Grau and Heather Britt’s “Outcome Harvesting,” downloadable here:

http://www.outcomemapping.ca/download/wilsongrau_en_Outome%20Harvesting%20Brief_

revised%20Nov%202013.pdf

Page 11: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

10

MostSignificantChange

UseMostSignificantChangetoassessnotonlywhatoutcomesaprogramhasachieved,butalso

howvariousstakeholdersvaluethoseoutcomesindifferentways.

TheMostSignificantChange(MSC)techniqueisahighlyparticipatorymethodologydesigned

toengageabroadcross-sectionof stakeholders in theevaluationofagiven intervention. In

thecontextofa leadership-trainingprogram,membersof the leader’s communitywouldbe

invitedtosharestoriesofwhattheyconsidertobesignificantchangesresultingfromleader’s

participation.Thisprocessofsharingnotonlyprovidesusefulevidencefortheimpactsofthe

intervention, but also surfaces valuable insights into what community members consider

“significant.” Once these stories have been collected, a group of select stakeholders (staff,

advisors,etc.)discussthestoriesanddeterminewhichtheyconsider“mostsignificant,”along

withdescribingarationale fortheirdeterminations.Theseanalysesarethenfedbacktothe

community to focus its attention on impact, shared with leaders, and used to refine the

trainingprogram.Youcanfinddetailedbackgroundinformationandimplementationguidance

inRickDaviesand JessDart’sThe ‘MostSignificantChange’ (MSC)Technique:AGuide to Its

Use,downloadablehere:http://www.mande.co.uk/docs/MSCGuide.pdf

CollectiveEfficacyScale

UseaCollectiveEfficacyScaletoassessagroup’ssharedbeliefinitsjointcapabilitytoorganize

andexecuteacourseofactioninpursuitofitsdesiredresults.

TheCollective Efficacy Scale emerged from criminal-justice research intohow communities’

sharedvaluesandwillingnesstocooperatecouldexert influenceoncommunitymembersto

Page 12: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

11

reduceviolenceandcrime.Ithasbeenadaptedforuseinschoolsandmanyothercontextsin

whichcommunitymembers’senseofthemselvesasamutuallysupportivecommunitycapable

of collective action is important. By deploying the Collective Efficacy Scale in a leader’s

communitybefore,during,andafteraleadership-trainingprogram,youcaninfersomeofthe

program’svalueinbuildingthecommunity’ssenseofefficacy.Youcanfindmoreinformation

on the Collective Efficacy Scale in John M. Carroll, Mary Beth Rosson, and Jingying Zhou’s

“Collective Efficacy as a Measure of Community,” downloadable here:

https://www.researchgate.net/publication/221519069_Collective_efficacy_as_a_measure_of_

community.

RippleEffectsMapping

UseRippleEffectsMappingtoreflectuponandvisuallymaptheintendedandunintendedchanges

producedbyacomplexprogramorcollaboration.

Ripple Effects Mapping (REM) is a participatory qualitative evaluation method in which

stakeholders collectively and visually map the interconnected outcomes resulting from a

programorintervention.InanREMprocess,participantscometogethertoreflectontheways

in which a program has affected their lives and the lives of others. Individual impacts are

captured using mind-mapping software or on a large white board, and the group works

together to identify and map connections among these impacts to generate a collective

representationoftherippleeffectsoftheprogramorintervention.Formoreinformationand

a practical guide for usingREM, seeWashington StateUniversity-Extension’s “Ripple Effects

Mapping for Evaluation,” downloadable at: http://extension.wsu.edu/stevens/wp-

content/uploads/sites/19/2013/12/REM.Complete.pdf.

More details and more tools are included in the section titled “15 Evaluation Tools to Consider”

near the end of the report.

Page 13: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

12

TESTTHETOOLSONYOUROUTCOMES

Onceyouhaveidentifiedoneortwoevaluationtoolsof interest,youcantestthemoutonyour

outcomes.Dothisbyreflectingonthefollowingquestions:

1. Whatoutcome(s)doyouwanttomeasureusingthisevaluationmethod?

2. How are you planning to use the approach?Whowill be the informants? Howwill you

collectthedata?

3. Imaginea setofdatayoumight receive fromthismethod.Whatwouldyoudowith the

data?Howwoulditaffectthedecisionsyoumakeabouttheprogram?

Notethateachofthesemethodsalsointroducesaframeyoucanapplytoyourevaluationevenif

youaren’tusingthefullmethod.Forexample,youcanlearnaboutthetypesofquestionsyou

mightasktocaptureemergentoutcomesbybecomingfamiliarwithoutcomeharvesting,evenif

youdon’tplantodothefullverificationprocessthatispartofthattechnique.

Evaluation experts offer various guidelines for selecting appropriate evaluation tools.No tool is

perfect, comprehensive, or appropriate for all contexts, so deciding upfront what is most

important about the tool can help. Some of the key criteria to consider in choosing evaluation

toolsinclude:

• Accuracy: The accuracy of your evaluation data is highly dependent on the constraints

underwhich you are operating (e.g., having access to relevant respondents at the right

time).

• Timeliness:Datathatwon’tbeavailableintimetomeaningfullyinformdecisionswon’tbe

thathelpful.

• Actionability:Weoftencollectdatathatareinterestinginsteadofdatathatareactionable.

Consideradressrehearsaltotestouttheusefulnessofyourexpecteddata.Imagineresults

fromamethod, and then imaginewhat youwoulddodifferentlybasedonhaving those

data.Ifyoucan’tfigureoutwhatyouwouldchange,thosedatamaynotbeuseful.

• Impartiality:Avoidaskingleadingquestionsorquestionsthatassumepositiveorplanned

outcomes.

• Representative of multiple perspectives: Identify key perspectives you need to further

develop theprogram (e.g.,endusers, funders, community stakeholders,etc.),and finda

toolthatwilldeliverthoseperspectives.

Notethatthesearecriteriatobalance—yourtoolsneedtomeetallofthemtoacertaindegree,

butyouwillalsohavetobalancehowfartogo(e.g.themosttimelydatacollectionmaynotbethe

most accurate or represent themost perspectives, but you don’t want to go so far into being

accurateandrepresentativethatthedatadoesn’tshowupuntilit’stoolatetouseit).

STEP3:APPLYLEARNINGSFORCONTINUOUSPROGRAMIMPROVEMENT

Thisisperhapsthesimplestandyetmostoverlookedstepintheprogramevaluationprocess.The

learninggeneratedthroughevaluationprocessesshouldbefedbackintotheprogramdesignand

processtoensureacontinuouscycleofprogramimprovement.There’sno“magicbullet”

methodologyforthis—it’smostlyafunctionofsettingasidetimeforaplanningteamtoreflecton

Page 14: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

13

theevaluationandidentifywaysinwhichtheprogramcanberefinedtoimproveeffectiveness.

Thisshouldbeanongoingcycleinwhichtheprogramdesign(outcomesandchangepathways),

andlearningandevaluationmethodsareregularlyinterrogatedandrefinedtoensurethatthe

overallinterventioniscontinuallyimproving.

Wehopethisreporthasstimulatedyourthinking,andyourpractices,aroundevaluatingyourown

leadershipprogram.

As your evaluation practices evolve, you may wish to deepen your learning about evaluation

methods. In the addenda you’ll find a short list of resources to consider. Youmay alsowish to

partnerwithotherorganizationsandpoolresourcesorexplorejointevaluationefforts.Couldyou

design or use an instrument together?Workwith a consultant together? Share data you have

gatheredandthelearningithasoffered?

Ultimately,evaluationisacorepartoflearningandimprovement.Goodevaluationgivesusdata

toreflecton,helpsusimaginebetterwaystomakeanimpact,andguidesourongoingeffortsat

improvement.

Developoutcomesandchangepathways

Implementprogram

Evaluateandlearn

Page 15: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

14

NOTETOFUNDERS

Donorsandfundershaveasignificantinfluenceontheevaluationpracticesofnonprofits.

Inmostcases,thisinfluenceisverypositive.Thequestionsembeddedinagrantapplicationand

theinquiriesofaprogramofficeraboutevaluationpracticesforceprogramdesignerstoreflecton

theoutcomestheyhopetoachieveandarticulatehowtheyhopetomeasureprogresstoward

thoseoutcomes.Butsometimesafunder’srequestsforevaluationleadtoacultureinwhich

programdirectorsfeelpressuretoshowhowapre-determinedtargetwasachieved.Whena

programissituatedinacomplexsystem,suchalinearapproachtoevaluationcanshutdown

importantlearningandreflectionthatwillleadtobreakthroughslateron.Weencouragefunders

andnonprofitstafftoengageinopenconversationsaboutapproachestoevaluationthatinvite

curiosityandexcitementandratherthancompliance-drivenfear.Bycommittingtooutcomes,

independentofthesuccessofanyparticularprogrammaticapproach,andencouraginghonesty,

collaboration,andinnovation,funderscanshapealargerfieldofpractice.Somefunderswework

withtakealearning-focusedapproachanduselanguagelike:“Wecaremoreaboutmakingan

impactthantheparticularsuccessorfailureofyourprogram.Wewanttolearnwhatyouare

learningandfindouthowtomakeadifference.Ifthisprogramdoesn’tworkwell,weknowwe

willalllearnalotfromtheexperience,andthenwecantrysomethingelsetogether.”

Page 16: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

15

ACKNOWLEDGEMENTS

ThisreportwasproducedwiththesupportofaWexnerGraduateFellowshipAlumniCollaboration

Grant.

TheauthorsexpressgratitudetotheWexnerFoundationfortheirinvestmentinleadership

developmentandforthefinancialsupportthatmadethisinitiativepossible.

Authorsofthisreport:

• RabbiJustusBairdisDeanofAuburnSeminary.

• AaronDorfmanisPresidentofLippmanKanferFoundationforLivingTorah.HewasVice

PresidentforNationalProgramsatAmericanJewishWorldServiceduringphase1ofthis

project.

Weexpressappreciationtothealmosttwodozenstaffwhooverseeleadershipdevelopment

programswhocametotheApril2015seminarthatledtothedevelopmentofthisreport:

1. SarraAlpert,AVODAH:TheJewishServiceCorps

2. JustusBaird,AuburnSeminary

3. LaurenBerkun,ShalomHartmanInstitute

4. ShifraBroznick,AdvancingWomenProfessionals

5. AaronDorfman,AmericanJewishWorldService

6. MireleGoldsmith,Hazon

7. IritHouvras,AmericanJewishWorldService

8. LeoraIsaacs,SYNERGY

9. StevenJacobson,Dorot

10. JasonKimelman-Block,BendtheArc

11. KateLauzar,SYNERGY

12. SharonMiller,AuburnSeminary

13. LevNelson,T’ruah:TheRabbinicCallforHumanRights

14. HeidiRosbe,Encounter

15. NadiaRoumani,AmericanMuslimCivicLeadershipInitiative

16. ShaunaRuda,JDCEntwine

17. RachelJacobyRosenfield,AmericanJewishWorldService

18. JudithSamuels,SYNERGY

19. ChrisScharen,AuburnSeminary

20. RebeccaSirbu,CLAL:TheNationalJewishCenterforLearningandLeadership

21. EmilySoloff,AmericanJewishCommittee

22. BeckyVoorwinde,TheBronfmanYouthFellowships

Page 17: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

ThismaterialisadaptedfromdatacollectiontoolspresentedaspartofSparkPolicyInstitute’sadvocacyevaluationtoolkit.

15EvaluationToolsToConsiderUsethis2-pagechartofEvaluationToolstoexploremoremethodologiesforevaluatingthe

secondaryimpactofyourleadershipdevelopmentprogram.

Tool Setting #ofInformants TypeofInformationtoCollect

Survey

Onpaper,by

email,orby

phone.

Typicallyalargernumber

ofparticipants(20+).

Close-endedquestionsandopen-

endedquestionstailoredtocapture

datatolearnaboutoneormore

outcomes.

InterviewByphoneor

inperson.

Typicallyasmallnumber

ofparticipants(around5

–15).

Open-endedquestionsthatelicit

moreexplanationandinformation

thansurveysusuallyprovide,

tailoredtocapturedatatolearn

aboutoneormoreoutcomes.

FocusGroup Inperson.Typicallyasmallnumber

ofparticipants(5–10).

Smallnumberofopen-ended

questionsdesignedtoelicit

dialogueamonginformants.

Providesveryin-depthinformation

tolearnaboutoneormore

outcomes.

Document

ReviewInyouroffice!

Noparticipants,butdatais

usuallyformaldocuments,

includingorganization

plans,earnedmedia,

policies,etc.

Canbeanalyzedbycountingkey

piecesofinformationand/or

summarizingthenarrative,tailored

tocapturedatatolearnaboutone

ormoreoutcomes

Observation

Inrealtime,

asanaction

isbeing

taken.

Oneortwoobservers,

collectingdataabout

everyonebeingobserved

(orlistenedto).

Countsandqualitativenarrative

regardingeasilyobservableactions

duringagroupeventoractivityof

anytype,tailoredtocapturedatato

learnaboutoneormoreoutcomes.

CaseStudyMixof

settings.

Typicallyincludesamixof

documentreview,

observation,interviews,

and/orfocusgroups.

Indepthlearningfrommultiple

perspectivesaboutaspecificexample

ofchangehappening(e.g.asubsetof

leaderswhoworkedtogetherora

specificleader),tailoredtocapture

datatolearnaboutoneormore

outcomes.

Outcome

Harvesting

Group

dialogue.

Typicallytheteamof

programstrategistswith

inviteeswhohaveagood

perspectiveontheprogram

Participatorytechniquetoidentify,

formulate,verify,andmakesenseof

outcomeswhenrelationshipsof

causeandeffectareunknown.Good

forsurfacingunplannedoutcomes.

Page 18: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

ThismaterialisadaptedfromdatacollectiontoolspresentedaspartofSparkPolicyInstitute’sadvocacyevaluationtoolkit.

Tool Setting #ofInformants TypeofInformationtoCollect

RippleEffect

Mapping

Paired

interviews

andgroup

dialogue

Typicallyalargegroupof

programparticipantsand

membersofthelarger

network.

Appreciativeinquiry-styleinterviews

toelicitstoriesofsuccessfollowedbya

groupmappingof“ripples”ofchange.

Capturesawiderangeofoutcomesthat

mayhavehappenedindifferent

locations,aswellasgrouphypotheses

aboutwhatlikelysecondaryeffects

thoseoutcomeswillhave.

SocialNetwork

Analysis

Onpaper,by

email,orby

phone.

Canbesurveysor

interviews,dependingon

useofinformation.

Helpstounderstandrelationships

betweenindividualsororganizations,

theresourcesleveragedinthose

relationships,andcanalsoprovide

insightintothewhyandhowbehind

therelationshipswhendone

qualitatively.

360s(e.g.

Servant

Leadership

Questionnaire)

Onpaper,by

email,orby

phone.

Apre-definedinterview

instrumentsdeployedto

multipleaudiencesofthe

leader.

Gathersperspectivesonleader

characterandperformancefromthose

indirectcontactwiththeleaderona

numberofdifferentdimensions,aswell

asfromtheleaderherorhimself.

Identity

Leadership

Inventory

Onpaper,by

email,orby

phone.

Apre-definedinterview

instrumentsdeployedto

multipleaudiencesofthe

leader.

Aleadershipinventorybasedonsocial

identitytheorythatcaptureshowa

leaderaffectsgroupcohesionand

identity,embodiesthegroup’ssenseof

“we”,andcreatesplatformsorvehicles

forthegrouptoactonitsprinciples

andinterests.

Collective

EfficacyScale

Onpaper,by

email,orby

phone.

Apre-definedsurvey

instrumentdeployedto

multipleaudiencesofthe

leader.

Demonstratestheextenttowhicha

grouphasconfidencethatitis

equippedtoachieveitsgoalstogether.

Collectiveefficacyispredictiveof

successbecauseitincreasesagroup’s

abilitytoovercomeadversityand

maintaincommitment.

Appreciative

Inquiry

Interviews

Byphoneor

inperson.

Interviewsconductedwith

audiencesoftheleader.

Focusesonsurfacingpastexperiences,

inthiscasewiththeleader,thatare

importantrepeatinthefuture(very

focusedonstrengths).

Discourse/

FrameAnalysis

Onpaper,by

email,orby

phone.

Interviews,surveysor

focusgroups

Explorestheframesornarratives

peopleusetounderstandanissue,a

group,orthemselves,and/ortheextent

towhichnewframesandnarratives

resonatewiththeirexistingvalues.

Page 19: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

ThismaterialisadaptedfromdatacollectiontoolspresentedaspartofSparkPolicyInstitute’sadvocacyevaluationtoolkit.

RecommendedResources

AmericanEvaluationAssociation(http://www.eval.org/)

TheAmerican EvaluationAssociation’smission is to improve evaluation practices andmethods,

increase evaluation use, promote evaluation as a profession, and support the contribution of

evaluationtothegenerationoftheoryandknowledgeabouteffectivehumanaction.

CenterforEvaluationInnovation(http://www.evaluationinnovation.org/)

Ouraimistopushphilanthropicandnonprofitevaluationpracticeinnewdirectionsandintonew

arenas. We specialize in areas that are challenging to assess, such as advocacy and systems

change.

FreeResourcesforProgramEvaluationandSocialResearchMethods

(http://gsociology.icaap.org/methods/)

ThispagelistsfreeresourcesforProgramEvaluationandSocialResearchMethods.Thefocusison

"how-to"doprogramevaluationandsocialresearch:surveys,focusgroups,sampling,interviews,

andothermethods.Mostoftheselinksaretoresourcesthatcanbereadovertheweb.

SparkPolicyInstitute(http://sparkpolicy.com/)

Spark Policy Institute partnerswith stakeholders throughout the country to develop innovative,

research-based solutions to complex societal problems. Spark combines community and

stakeholder-driven researchwithpractical,hands-onexperienceandbestpractices, allowing for

solutionsthatbridgesectors,issues,beliefs,andvalues.Byintegratingdiversepolicysystems,the

teamatSparkidentifiesanddevelopsthebestsolutionsforallstakeholders.

Page 20: On Evaluating the Secondary Impact of Leadership Programs: A … · 2016. 10. 28. · When evaluating impact in the context of a complex system, experts encourage evaluators to develop

19

Notes

1There’satendencyinprogramevaluationcirclestothinkthat,inordertobecredible,aprogramneedsto

demonstrateandaccountforthefullchainofcausationfromactivitythroughoutputs/outcomestoimpactforeach

discreteprogramorintervention.The“goldenspike”analogy(evokingthefinalconnectionbetweenCentralPacific

andUnionPacificRailroadstocompletetheFirstTranscontinentalRailroadacrosstheUnitedStates)shortcircuitsthis

presumptionbyarguingthatanevaluationcan(andshould)drawonexistingresearchforanalogousproofpointsto

connectactivitiestoultimateimpactwithouthavingtoprovethefullcausalchainforeachdiscreteintervention.Say,

forexample,thatyou’reconcernedaboutchildhoodobesity.Youdesignaninterventionthatreplacesprocessed

snackswithfreshfruitandvegetablesnacksinalocalmiddleschoolcafeteria.Thegoldenspikeconceptsuggeststhat

ifyoucandemonstratethatstudentsareconsumingmorefruitsandvegetablesandfewerprocessedsnacksasa

resultofyourintervention,andyoucancitetoresearchthatdemonstratesthatchildrenwhoincreasetheir

consumptionoffruitsandvegetablesanddecreasetheirconsumptionofprocessedsnacksexperiencereductionsin

obesity,youdon’tneedtoprovethatyourinterventionreducesobesity.Youcansimplyrelyonthe“goldenspike”

connectionbetweenwhatyou’redoingandwhattheresearchindicateswillhappenasaresult.Thewillingnesstouse

the“goldenspike”hasimportantimplicationsforevaluationbudgets,timelines,participantresources,andabilityto

produceresultsthatwillbeseenascredibleanduseful.

2FrancesWestley,BrendaZimmerman,andMichaelQ.Patton,GettingtoMaybe:HowtheWorldisChanged

(Toronto:VintageCanada,2007).