“would you still hang up on this poll if you knew…” presented at the 2009 aapor annual...
TRANSCRIPT
“Would You Still Hang Up on this Poll
If You Knew…”
“Would You Still Hang Up on this Poll
If You Knew…”
Presented at the 2009 AAPOR
Annual ConferenceMay 15, 2009
An Experiment to Improve the Design of Political Message Testing Polls
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
2
Research conducted by:
• Thomas M. Guterbock, Ph.D., Director, Center for Survey Research, University of Virginia, [email protected]
• Deborah L. Rexrode, M.A., Research Analyst, Center for Survey Research, University of Virginia, [email protected]
• Samantha Luks, Ph.D., Polimetrix, Vice-President for Special Projects, YouGovPolimetrix, [email protected]
Sponsored by: University of Virginia Center for Politics
Web Survey Services donated by: YouGovPolimetrix
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
OverviewOverview
• Background—How we got here• Research objectives
– Ideas for improvement
• The experiment• Respondent reactions• Party mismatch• Treatment comparisons• Conclusions & further research
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
4
BackgroundBackground
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
5
Message Testing ≠ Push Polling
Message Testing ≠ Push Polling
• “Push Polls” are not really surveys at all.
• Political advocacy calling conducted under the guise of research.
• Typically very short calls, communicating negative messages about a candidate.– sometimes false messages
• Source of call not clearly identified
• Aimed at altering outcome of the election
• Sometimes live calls, sometimes outbound IVR
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
6
Push Polls are condemnedPush Polls are condemned
• AAPOR issued its first statement in 1997– Renewed statements in 2000, 2004
• National Council of Public Polls issued statement in 1997.
• American Association of Political Consultants issued statement too.– The larger, reputable political consulting firms have
refrained from actual push polling.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
7
BUT: Message Testing Polls are OK?
BUT: Message Testing Polls are OK?
• Political campaigns regularly use surveys to test positive and negative messages for their effect on voters.
• These are recognized as a legitimate form of research.
• These surveys contain ‘push questions.’• The scripts are often unbalanced, treating one
candidate positively and others negatively.• The messages are often strongly partisan in tone.
– Sometimes highly distorted, misleading, even false?
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
9
Why improve Message Test Practice?
Why improve Message Test Practice?
• Message test polls often lead to trouble– For campaigns and candidates– For firms that conduct them– For AAPOR, which must handle complaints
• These polls make respondents angry– Lack of real informed consent– Reaction to partisan, misleading nature of the messages– Feeling of being exploited– Confusion with ‘push polls’
• There is very little published, scientific research that can demonstrate the superiority of any specific technique in message-testing polls.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
10
AAPOR: on Message TestingAAPOR: on Message Testing
“AAPOR recognizes that message tests may need to communicate positive or negative information in strongly political terms, in a tone similar to campaign advertisements. Still, these surveys should be judged by the same ethical standards as any other poll of the public: Do they include any false or misleading statements? Do they treat the respondent with fairness and respect?”
—AAPOR Statement on Push Polls
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
11
BackgroundBackground• AAPOR task force worked on re-writing ‘Push Poll’
statement in 2006.– Recognized message testing as a related issue.
• Guterbock organized a telephone ‘summit’ in December, 2006 with polling practitioners– Participants: Celinda Lake, Daniel Gotoff (LRP/Lake,
Snell, Perry), Mark Mellman (Mellman Group), B. J. Martino (Tarrance Group), Jim Burton (Public Opinion Strategies), John Nienstedt (CERC: Competitive Edge Research and Communication)
• Meeting at AAPOR 2007 (Anaheim)– With academics, AAPOR Standards, practitioners
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
12
Message from the PollstersMessage from the Pollsters
• Message-testing polls are used in nearly all serious campaigns
• Message testing is essential to what the polling firms do.
• Pressures of cost and time limit what practitioners can do in these polls.
• Negative messages are carefully researched– Campaign pollsters are ethically bound not to mislead.– Suggesting that anything in the poll script might not be true
would jeopardize the campaign.
• Where’s the research?
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
13
Proposal and SupportProposal and Support
• Obtained a real message-testing survey conducted by a major political polling firm in 2006– This clarified design issues of concern
– Model for the “generic” questionnaire in our experiment
• Prepared a project description and rationale
• Received a grant from the UVA Center for Politics (Larry Sabato and Kenneth Stroupe)
• Offer of complimentary web survey services from YouGovPolimetrix (Doug Rivers, CEO)
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
14
Research ObjectivesResearch Objectives
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
15
Purpose of the ResearchPurpose of the Research
• Show that improved design can improve respondent experience in a message testing poll.
• If improved designs are used:– Cooperation will improve, costs will be lower
– Data quality will improve
– Fewer complaints, less confusion with ‘push polls’
– Less possibility that public perception of surveys will suffer
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
16
The “Generic” Message Testing Survey
The “Generic” Message Testing Survey
• Closely follows design of a real survey from a large firm
• Starts out as an ordinary pre-election poll• No transitional introductions within the script• No balance of positive and negative statements by
party – Our party is all good; their party is all bad
• Repeated question after each statement: “Would you still vote for …?”
• Strong, increasingly personal, negative statements about the opposing candidate.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
17
Ideas for ImprovementIdeas for Improvement
• Improved transitional introductions– Start like an ordinary poll . . .
– . . . but make the transition to message testing clear
• Improved balance of negative and positive statements by party– No campaign would want to run a fully balanced script
– Test a partially balanced script
• Sprinkle a few positives among the negatives
• And a few negatives among the positives
• Improve the ‘test’ questions
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
18
Improve the transition . . .Improve the transition . . .• Lack of transitional introduction before
negative statements.
– Respondent is caught off guard; strong statements are read to them without any forewarning.
– Expectations of the respondent are violated.• Expects an ‘objective’ or ‘neutral’ survey
– Respondent feels pressured.– These results are strongest when the negative
party statements are mismatched by party of the respondent.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
19
Add a bit of balance . . .Add a bit of balance . . .
• Lack of balance of messages by party creates a stressful experience for respondent.
– Respondent expects polls to be objective and neutral.– Strongly worded statements violate the sense of
fairness for respondent.– Respondents feel pressure to change their opinions.– Respondent loyal to the opposing party will realize that
the results will aid the campaign of politicans they oppose.
– Even partial balancing will symbolize researchers’ intention to be fair, respectful of partisan differences.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
20
Use alternate ‘test’ questions . . .Use alternate ‘test’ questions . . .
• Asking the respondent after each statement if they will change their vote puts unwelcome pressure on the voter– Respondents feel they are being ‘pushed’– Especially if they favor the opposing party, or if they
are independents– Poll fails to measure respondents’ true reaction to some
of the messages.• “That’s distorted, even if somewhat true”• “I don’t believe that”• “That’s not fair”
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
21
The ExperimentThe Experiment
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
22
Fielding the ExperimentFielding the Experiment• Special thanks to Doug Rivers and
YouGovPolimetrix for their support in fielding this experiment.
• YouGovPolimetrix fielded the survey with pre-recruited PollingPoint panelists.
• Main focus: generic 2010 Congressional election• Sample was well balanced between Democrats,
Republicans, and Independents.• Fielded March 2009.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
23
Dependent VariablesDependent Variables
• Self-reported measures of respondent experience– Measure of comfort or stress during the interview
– Fairness of questions
– Believability of partisan statements
– Concern about use of results to aid opposing party
• Were the questions what was expected?• Willingness to participate in the future
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
24
The ExperimentThe Experiment• Full factorial design
• Thirty-six treatments (3 × 2 × 3 × 2)– Three introductions: none vs. two transitional intros
– Unbalanced vs. partially balanced partisan statements
– Three sets of ‘test’ questions (responses to statements)
– Two political party versions
• Random assignment of treatments
• Party versions used equally
• ‘Generic’ treatment assigned a larger number of cases, to facilitate comparison with alternatives.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
25
Transitional IntroductionsTransitional Introductions
• Control: “When you hear the following statements, does knowing about this make you more likely or less likely to vote for this candidate?”
—Wording used on ‘generic’ poll.
• Intro 1: “Here are some statements you might hear from a political candidate running for office. Your thoughts and opinions in response to these statements are an important part of our research.”
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
26
Introduction – Treatment 1Introduction – Treatment 1
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
27
Transitional IntroductionsTransitional Introductions• Intro 2:
“The following statements are the sort that you might hear in a political campaign, or a campaign commercial. These statements are different from what you might expect from a usual poll or survey. You might not agree with these statements, and some are negative. We are testing people’s reactions to these statements and we would like to warn you that these statements could cause some people to react strongly. This is ok. Your thoughts and opinions in response to these questions are an important part of our research.”
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
28
Introduction – Treatment 2Introduction – Treatment 2
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
29
Unbalanced DesignUnbalanced Design
• Follows design of the ‘generic’ poll• Questions were grouped into three blocks.
– Series One: 6 positive statements about members of Congress from favored party
– Series Two: 7 negative statements about opposing party– Series Three: another 7 negative statements about
opposing party• Series 3 included cultural and social issues that are ‘hot-
button’ issues for some voters
Statements were randomized within blocks
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
30
Partially Balanced DesignPartially Balanced Design
• Questions were again grouped into three blocks.– Series One: 4 positive and 2 negative statements about
favored party
– Series Two: 2 positive and 5 negative statements about opposing party
– Series Three: 1 positive and 6 negative statements about opposing party
Statements were randomized within blocks
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
31
Test One QuestionsTest One Questions
• “Does knowing about this make you more likely or less likely to vote for a [D/R] Party candidate, and how strongly do you feel about that?”– Much more likely
– Somewhat more likely
– Somewhat less likely
– Much less likely
Follows wording used on ‘generic’ poll
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
33
Test Two QuestionsTest Two Questions
• “How convincing is this statement as a reason to vote for a [D/R] candidate?” – “Very convincing” to “Not convincing”– Used to evaluate positive statements
• “How serious a doubt does this statement create about your voting for a [D/R] candidate?” – “Serious doubt” to “No doubt”– Used for negative statements
Some firms currently use this wording
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
36
Test Three QuestionsTest Three Questions
• Two evaluative questions asked about each statement:
1) “How believable do you think this statement is?”– “Very believable” to “Not believable”
2) “For you as a voter, how important is it for you to know this information?”
– “Very important” to “Not important”
• Problem: Survey length increased by 2 minutes when Test 3 questions were used
– More mid-survey break-offs with Test 3
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
39
Interview DispositionsInterview Dispositions
• 2,086 complete interviews• 329 incomplete interviews• 300 screenouts
– Low interest in congressional elections
– Industry
• Average interview length: 13 minutes• Response Rate = 30.8%
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
40
Demographic TargetsDemographic Targets
• Based on 2006 American Community survey
• Invitations based on cross-classifications of race, gender, education, and age and expected responsiveness
• No hard quotas employed. All respondents who met screening criteria were permitted to take the survey
Unweighted Respondent DemographicsUnweighted Respondent Demographics
Gender Male 46%
Female 53%
Education HS or less 34%
Some College 34%
College Grad 19%
Post Grad 13%
Age 18-34 21%
Age 40%
55+ 39%
Race White, other 82%
Black 10%
Latino 8%
Party Strong Rep. 16.3%
Moderate Rep. 11.2%
Indep. Rep. 11.2%
Independent 9.7%
Indep. Dem. 9.9%
Moderate Dem. 10.8%
Strong Dem. 24.0%
No Preference 2.9%
Other 2.6%
Unsure/DK 1.4%
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
Respondent ReactionsRespondent Reactions
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
43
Comments on the SurveyComments on the Survey• This is clearly the MOST BIASED survey your firm has presented. You need to
rethink your questions about Republicans and present a moderate position rather than the extremist view you are espousing. Your questions leave no doubt in my mind as to your political position. Not a good thing in a survey.
• This is the most biased survey I ever read! Who wrote these questions? I'm going to take a guess that it was the Democrats and Acorn by putting out such lies about the Republicans. Maybe the Democrats should have inserted their name in where the Republicans were listed. I used to think that polling point was fair and conservative. Now I know it's just a front for those cheating, lying, manipulative phony Democrats.
• This is the single MOST BIASED survey I have taken for your firm. It is garbage!• I refuse to finish this poll. The questions on the Republican congress are far too
critical and on the Democrats far too kind. There is no possible way to do this poll fairly. It's the worst you've ever sent me!!!!!!!!!!!! (note: survey was completed)
• Go into a useful business. • Your questions were left leaning and loaded. I am an independant. I vote for both
parties. I will vote to preserve our God given rights and freedoms, and to seal our Southern border!!!!!!
After launch, a debriefing was added at the end, explaining purpose and design of the experiment.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
44
After the DebriefingAfter the Debriefing• “Be fair and truthful in your questions and you'll get fair and
truthful answers; that statement about this survey not being for a particular political party... yeah right”
• “Don't deliberately give me surveys that you know will stress me out, for one. That would "improve my experience". Seriously now I'm kind of pissed at you.”
• “This poll was designed by the LIBERAL loons at the University of Virginia. How despicable. I'm surprised you went along with this charade.”
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
After the DebriefingAfter the Debriefing• “Be fair and truthful in your questions and you'll get fair and
truthful answers; that statement about this survey not being for a particular political party... yeah right”
• “Don't deliberately give me surveys that you know will stress me out, for one. That would "improve my experience". Seriously now I'm kind of pissed at you.”
• “This poll was designed by the LIBERAL loons at the University of Virginia. How despicable. I'm surprised you went along with this charade.”
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
Results:The importance of party mismatch
Results:The importance of party mismatch
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
47
ResultsResults
• Mismatched: Respondent party is opposite of the party favored in questionnaire version– Includes Independents who ‘lean’ R or D
• Unaffiliated: Respondent is Independent (no leanings) or did not select a political party
• Findings indicate that negative reactions are strongest with mismatched or unaffiliated.
Key to partisan mismatchKey to partisan mismatch
Question-
naire
version
Respondent’s party ID
N
Democrat Independent Republican
Favors
Democrats
Matched
480
Unaffiliated
152
Mismatched
387
Favors Republicans
Mismatched
460
Unaffiliated
137
Matched
413
Percent saying survey was ‘stressful’
Percent saying survey was ‘stressful’
Question-
naire
version
Respondent’s party ID
Democrat Independent Republican
Favors
Democrats
8% 20% 26%
Favors Republicans
26% 17% 7%
Percent saying items were ‘unfair’
Percent saying items were ‘unfair’
Question-
naire
version
Respondent’s party ID
Democrat Independent Republican
Favors
Democrats
8% 24% 47%
Favors Republicans
31% 15% 5%
Remaining analyses exclude ‘matched’
Remaining analyses exclude ‘matched’
Question-
naire
version
Respondent’s party ID
Democrat Independent Republican
Favors
Democrats
Matched
480
Unaffiliated
152
Mismatched
387
Favors Republicans
Mismatched
460
Unaffiliated
137
Matched
413
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
52
Comparison: Generic Treatment vs.
Improved Balance and Improved Test Questions
Comparison: Generic Treatment vs.
Improved Balance and Improved Test Questions
Stress in 18 treatment pairs (D+R)
Stress in 18 treatment pairs (D+R)
Generic Balanced
Balanced + Alternate
tests
Compare:these 12 ‘better’
treatments (N = 503) with the 2 generic
treatments(N=385)
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
60
Regression Results (Mismatched Only)
Regression Results (Mismatched Only)
Design
Improvement:
Stress Misuse Questions Expected
Questions Fair
Questions Believable
Future Participation
Intro 2 -.036 .017 .029 -.006 -.017 -.043
Intro 3 -.026 -.008 .095* -.001 -.025 -.050
Balance -.163* -.125* .163* .270* .260* .064+
Test Two -.174* -.112* .120* .269* .202* .059
Test Three -.102* -.079* .052 .173* .132* .067+
* p<.05; + p<.10. Reference categories: no intro, unbalanced, test
one
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
61
Regression Results (Mismatched and Unaffiliated)
Regression Results (Mismatched and Unaffiliated)
Design
Improvement:
Stress Misuse Questions Expected
Questions Fair
Questions Believable
Future Participation
Intro 2 -.013 -.001 .057+ .022 .016 -.009
Intro 3 -.033 .002 .088* -.003 -.014 .004
Balance -.135* -.133* .129* .234* .228* .042
Test Two -.127* -.083* .090* .255* .180* .059+
Test Three -.079* -.065* .047 .150* .098* .041
* p<.05; + p<.10. Reference categories: no intro, unbalanced, test
one
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
62
ConclusionsConclusions
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
63
ConclusionConclusion• When respondent’s party was a mismatch with the
questionnaire or the respondent was unaffiliated with either party, the experience was stressful for many respondents.
• Respondent assessment of experience is strongly conditioned by the design features of the study.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
64
Further ConclusionFurther Conclusion
• Better design can improve:– Fairness, believability, meeting expectations, concern
about use of results to aid opposing party, future willingness
• Improvements gained through:– better balance of positive and negative statements
about opposing party.– implementing different types of ‘test’ questions. – including a transitional introduction
• makes partisan statements less unexpected for the respondent.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
65
Further analyses: This studyFurther analyses: This study
• Further analysis is needed: – examine the interaction between the various treatment
factors.
– look at content found in the open-ended comments
– effects of the actual messages on vote intention
• We plan to measure future participation of our subjects as YouGov panelists.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
66
Further Research: Future studies Further Research: Future studies
• Test if observed effects operate in a specific, real campaign, with more personal messages
• Replicate experiment in an actual phone survey of cold-called voters
• Explore additional measures of respondent experience.
• We invite political polling firms to incorporate similar tests into their future message-testing polls.
Center for Survey ResearchUniversity of VirginiaCenter for Survey ResearchUniversity of Virginia
In sum . . .Message testing polls
CANbe improved by better design
In sum . . .Message testing polls
CANbe improved by better design
“Would You Still Hang Up on this Poll If You Knew…”“Would You Still Hang Up on this Poll If You Knew…”
Presented at the 2009 AAPOR
Annual ConferenceMay 15, 2009
An Experiment to Improve the Design of Political Message Testing Polls
Thomas M. [email protected]
Deborah L. [email protected]
Samantha [email protected]