evaluation: building the evidence to demonstrate impacts and outcomes latrice rollins, phd, msw...

45
Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment Calvin McAllister, MPH Research Assistant II Morehouse School of Medicine Prevention Research Center

Upload: kellie-doyle

Post on 29-Dec-2015

220 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Evaluation: Building the Evidence To

Demonstrate Impacts and Outcomes

Latrice Rollins, PhD, MSWAssistant Director of Evaluation and Institutional Assessment

Calvin McAllister, MPHResearch Assistant II

Morehouse School of Medicine Prevention Research Center

Page 2: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Describe evaluation

Describe methods of evaluation

Describe Community-Based Participatory Approach (CBPA) and the systems and infrastructure of community-academic partnerships necessary for CBPA

Detail examples of CBPA to evaluation

Learning Objectives

Page 3: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

What is Evaluation?

Program evaluation is carefully collecting and analyzing information about a program or some aspect of a program in order to make necessary

decisions about the program.

Page 4: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Why Evaluate? Improve the program –

◦ “Balancing the call to prove with the need to improve.” (W.K. Kellogg Foundation)

Determine program effectiveness – ◦ Evaluation supports “accountability and quality control” (W. K. Kellogg Foundation)

◦ Significant influence on program’s futureGenerate new knowledge –

◦ Not just research knowledge◦ Determines not just that a program works, but analyzes how

and why it works With whom is the program most successful? Under what circumstances?

4

Page 5: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

When we seek public funding…

Funding is being provided as an “investment toward the public good.” This isn’t a gift!!

Funding agencies – Federal, State, Public, Private – if they fund you, there is an expectation of results – “Outcomes”or “Impacts”as a result of the funding. As a recipient of the funding – you have the

obligation to do your best to achieve the objectives of the research. Results are expected.

Page 6: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Participatory Approach

A participatory approach to evaluation is an evaluation that involves all the stakeholders in a project - those directly affected by it or by carrying it out - in every phase of evaluating it, and in applying the results of that evaluation to the improvement of the work.

Page 7: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Participatory Approach Participatory monitoring and evaluation is

not just a matter of using participatory techniques within a conventional monitoring and evaluation setting. It is about radically rethinking who initiates and undertakes the process, and who learns or benefits from the findings◦ Institute of Development Studies, 1998

Page 8: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Participatory Traditional

Who drives the evaluation?

Community residents, project staff and other stakeholders

Funders and program managers

Who determines indicators of program progress?

Members of community groups, project staff and other stakeholders; evaluator

Professional evaluators and outside experts

Who is responsible for data collection, analysis and preparing final reports?

Shared responsibility of evaluator and participating stakeholders

Professional evaluators and outside experts

What is the role of the local evaluator?

Coach, facilitator, negotiator, “critical friend”

Expert, leader

Page 9: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Why would you use a participatory approach?

◦ It gives you a better perspective on both the initial needs of the project's beneficiaries, and on its ultimate effects.

◦ It can get you information you wouldn't get otherwise.

◦ It tells you what worked and what didn't from the perspective of those most directly involved - beneficiaries and staff.

◦ It results in a more effective project.

◦ It can provide a voice for those who are often not heard.

◦ It teaches skills that can be used in employment and other areas of life.

◦ It bolsters self-confidence and self-esteem in those who may have little of either. 

◦ It encourages stakeholder ownership of the project.

◦ It can spark creativity in everyone involved.

◦ It encourages working collaboratively.

Page 10: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Why wouldn't you use a participatory approach?

◦ People's lives - illness, child care and relationship problems, etc. - may cause delays or get in the way of the evaluation.                    

◦ You may have to be creative about how you get, record, and report information.

◦ Funders and policy makers may not understand or believein participatory evaluation.

◦ It takes more time than conventional process.

◦ It takes the establishment of trust among all participants in the process.

◦ You have to make sure that everyone's involved, not just "leaders" of various groups.

◦ You have to train people to understand evaluation and how the participatory process works, as well as teaching them basic research skills.

◦ You have to get buy-in and commitment from participants.

Page 11: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

When would you use participatory evaluation?

When you're already committed to a participatory process for your project.

When you have the time, or when results are more important than time.

When you can convince funders that it's a good idea.

When there may be issues in the community or population that outside evaluators (or program providers, for that matter) aren't likely to be aware of.

When you need information that it will be difficult for anyone outside the community or population to get.

When part of the goal of the project is to empower participants and help them develop transferable skills.

When you want to bring the community orpopulation together.

Page 12: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

 Who should be involved in participatory evaluation?

All stakeholders, including:◦ Participants or beneficiaries. ◦ Project line staff and/or volunteers. ◦ Administrators. ◦ Outside evaluators, if they're involved.  ◦ Community officials. ◦ Others whose lives are affected by the

project.

Page 13: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Data Collection Methods Mixed Method Approach including 2 or more

of the following methods:◦ Surveys◦ Interviews◦ Focus Groups◦ Document Review◦ Observations

Page 14: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Types of Data Quantitative

◦ Numbers based on objectives and activities◦ Types of data needed:

Number of participants (process) Grade point averages (outcome) Retention rates (outcome) Survey data (outcome and process)

Qualitative◦Narrative or text from:

Interviews Focus groups Observations

14

Page 15: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

SurveysUse when You want information directly from

a defined group of people to get a general idea of a situation, to generalize about a population, or to get a total count of a particular characteristic

Advantages Many standardized instruments availableCan be anonymousAllows a large sampleStandardized responses easy to analyzeAble to obtain a large amount of data quicklyRelatively low costConvenient for respondents

Disadvantages Sample may not be representativeMay have a low return rateWording can bias responsesClose-ended or brief responses may not provide the whole storyNot suited for all people – e.g., low reading level

Page 16: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

InterviewsUse When You want to understand

impressions and experiences in more detail and be able to expand or clarify responses

Advantages Often better response rate than surveysAllow flexibility in questions/probesAllows more in-depth information to be gathered

Disadvantages Time consumingRequires skilled interviewerLess anonymity for respondentQualitative data more difficult to analyze

Page 17: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Focus GroupsUse When You want to collect in-depth

information from a group of people about their experiences and perceptions related to a specific issue.

Advantages Collect multiple peoples’ input in one sessionAllows in-depth discussionGroup interaction can produce greater insightCan be conducted in short time frameCan be relatively inexpensive compared to interviews

Disadvantages Requires skilled facilitatorLimited number of questions can be askedGroup setting may inhibit or influence opinionsData can be difficult to analyzeNot appropriate for all topics or populations

Page 18: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Document ReviewUse When Program documents or

literature are available and can provide insight into the program or evaluation

Advantages Data already existDoes not interrupt the programLittle or no burden on othersCan provide historical or comparison dataIntroduces little bias

Disadvantages Time consumingData Limited to what exists and is availableData may be incompleteRequires clearly defining the data you’re seeking

Page 19: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

ObservationsUse when You want to learn how the

program actually operates – its processes and activities

Advantages Allows you to learn about the program as it is occurringCan reveal unanticipated information of valueFlexible in the course of collecting data

Disadvantages Time consumingHaving an observer can alter eventsDifficult to observe multiple processes simultaneouslyCan be difficult to interpret observed behaviors

Page 20: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Community-Campus Partnership Example

Page 21: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Theme - Risk Reduction and Early Detection in African American and Other Minority Communities: Coalition for Prevention Research

Morehouse School of Medicine Prevention Research Center (MSM PRC)

Page 22: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

22

MSM Prevention Research Center Overview

Established in 1998

Funded by the Centers for Disease Control and Prevention

Ways of Engaging the community: Research, health promotions, evaluation, and DEFINING THE RESEARCH AGENDA!

Page 23: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

The MSM PRC Community Coalition Board is comprised of:◦ Community Residents◦ Academic Institution Representatives◦ Agencies

within the City of Atlanta NeighborhoodPlanning Units T, V, X, Y and Z.

MSM PRC Community Coalition Board (CCB)

Page 24: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

24

Community Representation

Page 25: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

25

Community DescriptionPRC Service Area City of

AtlantaTotal Population 55,757 420,003

African American/Black 87% 52%

Average Age 34.1 32.9

Average Household Income

$34,389 $49,981

Page 26: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Systems of CCB Engagement in Research and Center Infrastructure

Page 27: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Bi-monthly meetings Scheduled social time and food at each

meeting Formalized Structure and Governance

Ensuring a Shared Community-Campus Experience

Page 28: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

The bylaws permit a maximum of 25 board members, the majority must be community representatives.

The board chair is always a community resident.

All projects and protocols to be implemented by the PRC must be approved by the CCB’s Project Review Committee, which consists of neighborhood representatives.

Example of MSM PRC CCB By-Laws

Page 29: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

MSM PRC CCB: Community Values

1. Research processes and outcomes benefit the community

2. Community partners involved in analysis and interpretation of data and dissemination of results

3. Partnerships to last beyond funded research

4. Community empowered to initiate community-based research

Page 30: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

MSM PRC CCB: Community Values

6. Policies and programs based on mutual respect and justice

7. Right to self-determination8. Community partners at every level9. Enforced principles of informed consent10. Socially, culturally, environmental

sensitive research and application

Blumenthal DS. A community coalition board creates a set of values for community-based research. Preventing Chronic Disease 2006;3(1):A16.

Page 31: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Examples of Community-Based Participatory

Approaches to Evaluation

Page 32: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

The Pittsburgh Community Improvement Association in partnership with MSM PRC CCB, received funding from the DentaQuest Foundation to address the overwhelming need for African American males to increase their awareness of the importance of oral health.

Minority Men’s Oral Health & Dental Access Program

Page 33: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

• Work with advisory board• Conduct Oral Health Needs Assessment• Assess the potential impact and capture important

lessons learned throughout all aspects of health needs assessment process

Evaluation

Page 34: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment
Page 35: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment
Page 36: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Impact oral health care among African American males in Neighborhood Planning Units V, X, Y and Z through educational interventions that demonstrate the importance of oral health and its relationship to the overall quality of life through:

• Partnership development with health and community based organizations to provide oral health education.

• Conduct an oral health community education training and health resource sharing across the identified NPUs.

Actions

Page 37: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

• Policies and interventions to eliminate racial disparities in oral health should be directed at the social, physical and infrastructural characteristics of neighborhoods as well as individuals.  

• Our results could be helpful for policy makers and NPUs citizen advisory councils in assisting the city in developing plans that best meet the needs of their communities.

• Education regarding access to dental care and insurance should be added to the curriculum.

Impact

Page 38: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

United Health Foundation MSM Innovations Learning Laboratory Patient Centered Medical Home and Neighborhood

Page 39: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Demographic Profile -Zip Code 30344 Mean age-34.3 years old

Average annual income - $39,720

28% of households are female-headed

Source: City-data.com. (2010). Zip code detailed profile. http://www.city-data.com/zips/30344.html.

Page 40: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Community Health Needs Assessment Process Was Designed to:

Comprehensively include primary qualitative and quantitative data from community stakeholders and secondary data to identify the demographic profile, health needs, priorities and assets

Engage communities in the review and interpretation of what the data means in real-time

Use community-based recommendations to guide service implementation

Page 41: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Evaluation 5 focus groups were conducted within 30344 to

get community ideas and perceptions related to clinical care quality within 30344.

A total of 47 participants (31 females, 16 males)

All participants resided or were employed in 30344 and were 18 and older

Consisted of community residents and workers, clinicians, and individuals with chronic diseases

Page 42: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Impact The results of these focus groups helped to

identify community preferences and recommendations to clinical care and supports

Established a list of existing clinical assets within 30344 after review of the results from focus groups

Developing an advisory committee

Page 43: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Your Turn!

Page 44: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

How might you apply one or more CBPA concept(s) to your work?

What would be the added value of CBPA to your organization?

Applying CBPA

Page 45: Evaluation: Building the Evidence To Demonstrate Impacts and Outcomes Latrice Rollins, PhD, MSW Assistant Director of Evaluation and Institutional Assessment

Questions