how do policymakers update their...

48
How Do Policymakers Update Their Beliefs? Eva Vivalt * The Australian National University Aidan Coville World Bank First version: May 1, 2017 Current version: February 9, 2020 Abstract Evidence-based policymaking requires not only evidence, but also for policymakers to update their beliefs based on that evidence. We examine how policymakers, researchers, and development practitioners update in response to results from academic studies, using a unique opportunity to run an experiment on policymakers. We find evidence of asymmetric up- dating on good news relative to one’s prior beliefs. We also find evidence of “variance neglect”, a bias similar to extension neglect in which con- fidence intervals are neglected relative to how they would be considered under Bayesian updating. Finally, we test whether providing different types of information can mitigate the observed biases. * E-mail: [email protected]. Corresponding author. Research School of Economics and John Mitchell Economics of Poverty Lab, Australian National University, Canberra, ACT 2601. AEARCTR-0001237. E-mail: [email protected]. 1

Upload: others

Post on 22-Jun-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

How Do Policymakers Update Their Beliefs?

Eva Vivalt∗

The Australian National University

Aidan Coville†

World Bank

First version: May 1, 2017

Current version: February 9, 2020

Abstract

Evidence-based policymaking requires not only evidence, but also for

policymakers to update their beliefs based on that evidence. We examine

how policymakers, researchers, and development practitioners update in

response to results from academic studies, using a unique opportunity to

run an experiment on policymakers. We find evidence of asymmetric up-

dating on good news relative to one’s prior beliefs. We also find evidence

of “variance neglect”, a bias similar to extension neglect in which con-

fidence intervals are neglected relative to how they would be considered

under Bayesian updating. Finally, we test whether providing different

types of information can mitigate the observed biases.

∗E-mail: [email protected]. Corresponding author. Research School of Economics andJohn Mitchell Economics of Poverty Lab, Australian National University, Canberra, ACT 2601.AEARCTR-0001237.†E-mail: [email protected].

1

Page 2: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

1 Introduction

The last two decades have seen a transformation in the rigor expected from im-

pact evaluations with the motivation that decisions need to be grounded in the the

most accurate and reliable evidence possible and decisions based on poor-quality ev-

idence may do more harm than good. Yet, in order for policies to be influenced by

evidence, three things must happen. First, there must be reliable evidence on which

policymakers can base their decisions. Second, policymakers must accurately update

their beliefs based on that evidence. Finally, they must have the agency to act on

what the evidence shows. Most research focuses on the crucial first step of generating

new evidence. Some research also focuses on the last question, such as in considering

the political economy issues affecting policy decisions (Persson and Tabellini, 2002;

Krueger, 1993). In this paper, we focus on the intermediate step: how policymakers

and technical advisors interpret new evidence. In particular, are they subject to any

behavioral biases and, if so, can information treatments affect how they interpret

evidence?

There are a number of well-documented behavioral reasons why people may not

accurately update their beliefs when presented with new information (Camerer et al.,

2003; Kahneman, 2003). However, relatively few papers have examined whether and

how policymakers are responsive to evidence. Hjort et al. (2019) test whether inform-

ing policymakers about interventions that proved to be successful in impact eval-

uations translates into policy decisions, while Rogger and Somani (2018) examine

whether bureaucrats learn from basic facts provided to them. Nellis et al. (2019) sur-

vey development practitioners and investigate whether learning about evidence from

meta-analyses of multiple studies relative to individual studies affects beliefs and out-

of-sample prediction accuracy. A number of other papers also focus on prediction

accuracy among experts (Banuri et al., 2015; DellaVigna and Pope, 2018, 2017). We

focus on how policymakers update in response to new information and specifically

on any behavioral biases they might have. We use the term “policymaker” to refer

to mid-level civil servants in charge of a particular program or providing technical

support in a developing country line ministry.

While there are relatively few papers focusing on policymakers, the potential far-

reaching implications of their decisions make them an important population to study.

Further, policymakers may be particularly prone to some behavioural biases. One

2

Page 3: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

focus of the paper is asymmetric optimism, whereby individuals are more responsive

to good news than bad news. The desire to see certain outcomes realized has been

shown to be associated with such optimism (Irwin, 1953; Windschitl et al., 2013;

Stuart et al., 2015), as has one’s perceived sense of control over the event (Langer,

1975). Policymakers, as experts invested in seeing their projects be successful, could

be particularly eager to accept optimistic forecasts. Working in large, bureaucratic

organizations with complex hierarchies that dissuade dissent may compound this

(Easterly, 2002), and government programs are often characterized by a lack of clear

market signals or feedback mechanisms that could enable learning about one’s mis-

predictions. It is also possible that optimism is actually selected for, in that only the

optimistic may through their genuine beliefs convince others as to the merits of their

proposals; there is some related evidence that executives tend to be overconfident in

estimating their expertise, forecasting that they will perform better than average (Liu

et al., 2016; Malmendier and Tate, 2005). Further, policymakers may be routinely

exposed to selected kinds of information that could affect their interpretation of new

evidence. For example, if they are rarely provided with clear information about un-

certainty, such as standard errors or confidence intervals, they may have difficulty

interpreting this information when provided with it.1

We leverage a unique opportunity to run an experiment on policymakers, practi-

tioners and researchers, in collaboration with the World Bank and the Inter-American

Development Bank, that explores (i) whether updating biases exist and (ii) if so,

whether the type of information provided can reduce the impact of these biases. In

particular, we hypothesize that policymakers are biased towards “good news”, over-

weighting positive impact evaluation results compared to negative results (asymmetric

optimism), and that policymakers do not take the variance of impact evaluation re-

sults fully into consideration when updating - a novel bias we term “variance neglect”.

We also leverage the setting to explore how evidence affects real-life decisions, and

we experimentally test whether providing different types of information can help to

mitigate the observed biases.

We focus on the policymakers, practitioners, and researchers who attend World

Bank and Inter-American Development Bank impact evaluation workshops. This

1Indeed, such information is often not provided: the World Bank’s flagship annual publication,the World Development Report, that is widely circulated among policymakers, was reviewed for theperiod 2010-2016. Of thousands of cited papers, information about standard errors or confidenceintervals was only provided 8 times.

3

Page 4: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

focus is intentional: higher-level policymakers will rarely read academic papers them-

selves and might instead rely on briefings from advisors. These attendees are also

demonstrably interested in impact evaluation and thus may be a particularly rele-

vant sample. The workshops are each approximately one week long and designed as

“matchmaking” events between those involved in certain development programs and

researchers: government officials are paired with researchers and tasked with design-

ing a prospective impact evaluation for their program over the course of the week.

Workshop participants include program officers in government agencies; monitoring

and evaluation specialists within government agencies; World Bank or IDB opera-

tional staff; and other international organization operational staff such as technical

advisors at USAID or DFID. The sample includes a group of researchers, both from

academic institutions as well as some international organizations such as the World

Bank. To supplement this sample, we ran the experiment at the World Bank and

the Inter-American Development Bank headquarters in Washington, D.C. Finally, we

separately ran the experiment on a sample of respondents through Amazon’s Me-

chanical Turk (MTurk) for an additional comparison group.

We find evidence of both biases. While we do not observe differences in behavioral

biases across subgroups, we do observe differences in priors. In particular, policymak-

ers have significantly more optimistic priors than practitioners, and both policymakers

and practitioners have significantly more optimistic priors than researchers. We also

find that research results affect a real-life allocation decision. Further, the type of

information that is provided affects updating. This is important because it suggests

that what information is provided and how it is provided can be used as a lever to

improve updating.

The rest of the paper proceeds as follows. First, we discuss the biases that we

are studying. As others have done, we start from the premise that individuals are

Bayesian updating but with an error. In past literature, this has sometimes been

referred to as “quasi-Bayesian” updating (Camerer et al., 2003).2 We then describe

the sample and the experimental design. Finally, we present and discuss results.

2To be distinguished from the quasi-Bayesian models of e.g. Giron and Rios (1980).

4

Page 5: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

2 Model

2.1 Bayesian Updating

To begin, we describe a model of Bayesian updating, which we will then modify to

introduce the behavioral biases of interest. Suppose a policymaker is deciding whether

to implement a program. The program’s effect if it were to be implemented in the

policymaker’s setting, θi, is unknown ex ante. The policymaker’s prior is that θi is

normally distributed across settings, allowing for heterogeneous treatment effects:

θi „ Npµ, τ 2q (1)

where µ is the grand mean and τ 2 is the inter-study variance.

The policymaker has the opportunity to observe a signal about the effect of the

program, Yi with some normally distributed noise, i.e. Yi “ θi ` εi, εi „ Np0, σ2i q. Yi

can be thought of as a point estimate and has variance τ 2 ` σ2i , which we will write

as v2i .

A person who is Bayesian updating will update their estimate of µ according to:

µt “ µt´1 ` kpYi ´ µt´1q (2)

where k “ pv2t´1q{pv2t´1 ` v2i q. In other words, µt is a weighted combination of µt´1

and the new information, Yi, gained in period t. They will also update their estimate

of the variance, so that:

v2t “v2t´1v

2i

v2t´1 ` v2i

(3)

Similar equations could be written for the case of no heterogeneity in treatment

effects, in which case τ 2 “ 0. This can be thought of as the appropriate model for

when one is considering information from replications.

In our experiment, we focus on this latter case, framing the new information

as coming from replications. This is to avoid the estimation challenges posed by

estimating τ 2 ‰ 0. In particular, if we were to introduce heterogeneous treatment

effects, different people could build different mental models of how results depend on

study characteristics and we would not be able to separately estimate the particular

model they have in mind. To avoid this problem, even when we present information

from different settings that might be subject to heterogeneous treatment effects we

5

Page 6: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

will not provide study details that could be used to build a more refined model; all

studies are either described as replications or are otherwise “exchangeable”.3

There are many ways in which individuals could deviate from Bayesian updating.

We focus on two: asymmetric optimism and variance neglect.

2.2 Biases

2.2.1 Asymmetric Optimism

Our definition of asymmetric optimism follows the literature and means updating

more on “good news” - information better than one’s priors - than “bad news” (Eil

and Rao, 2011). Several different kinds of asymmetric optimism with respect to good

news may exist. One hypothesis is that respondents are optimistic with respect to

point estimates, i.e. that they update more on point estimates higher than their

prior mean than they do on point estimates symmetrically lower than their prior

mean. For example, supposing µt´1=3, if we imagine that they alternatively receive

the signal Yi “ 1 or Yi “ 5, we might expect them to update more when they receive

the signal Yi “ 5. It is also possible that, seeing a point estimate with a confidence

interval, someone could believe that the true effect was closer to the upper bound

rather than the lower bound of the confidence interval. In this example, the person

would be optimistic with respect to confidence intervals. We distinguish between these

types of optimism because they have practical significance. Under the second type of

optimism, providing confidence intervals could actually worsen updating relative to

not providing confidence intervals.

To model optimism formally, we adapt Rabin and Schrag’s (1999) seminal paper

on confirmation bias, which models confirmation bias as the misperception of a signal.

In Rabin and Schrag’s model, a person can observe one of two possible signals, but

sometimes mistakes one signal for the other one. To expand this to a world in which

many signals could be perceived, we might say that people observing a signal Yi

perceived that they saw Yi ` γ for some γ ą 0. This would result in a calculated

k greater than the Bayesian k when presented with Yi ą µt´1 and conversely a

calculated k lower than the Bayesian k when presented with Yi ă µt´1. If γ “ 0, the

policymaker would be unbiased.

3For more information on exchangeability, the interested reader is referred to Gelman et al.(2013).

6

Page 7: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure 1: Bayesian Updating vs. Variance Neglect

In this figure, while we do not know exactly where kV N (dashed line) is in relation to kB (solidline), we do know that its slope is less steep; in other words, for a given value of Yi and vt´1,different values of vi result in values of k that are more similar to each other than the k of aBayesian updater.

2.2.2 Variance Neglect

For variance neglect, we only require that respondents pay less attention to the

variance than they would if they were Bayesian updating. For example, consider

the case in which respondents view data with alternatively small or large confidence

intervals. If kBS (kBL ) represents the k that a Bayesian updater would have if receiving

a signal with small (large) confidence intervals, and kV NS (kV N

L ) represents the k that

someone suffering from variance neglect would have upon receiving a signal with small

(large) confidence intervals, kBS ´ kBL ą kV N

S ´ kV NL . Figure A8 illustrates.

Parameterizing this bias is straightforward: we say that individuals pay too little

attention to either v2t´1 or v2i when generating their estimate of µt, so that they update

with:

k “v2t´1

v2t´1 ` pv2i ` λq

(4)

where λ reflects how much they underweight v2i relative to v2t´1.

They could also suffer an error in updating the variance itself:

v2t “v2t´1pv

2i ` λ

1q

v2t´1 ` pv2i ` λ

1q(5)

where λ1 is not necessarily equal to λ. Again, in the case of no heterogeneity in treat-

ment effects, the variance simply comprises the sampling variance. For simplicity,

we focus on how policymakers update their mean rather than how they update their

variance, though a companion paper considers the latter (Vivalt and Coville, 2019).

7

Page 8: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Variance neglect is closely related to, but distinct from, sample size neglect or,

more broadly, extension neglect. In particular, sampling variance depends not just on

the number of observations, but also on the standard deviation, and variance neglect

could also apply to inter-study variation, though we do not test this latter possibility

in this paper.

Prospect theory also bears similarities to variance neglect (Kahneman and Tver-

sky, 1979). Under prospect theory, people overweight small probabilities and un-

derweight large probabilities. They also treat gains and losses differently. The

overweighting of small probabilities and underweighting of large probabilities should

change how respondents treat normally distributed data. Given any normal distri-

bution, respondents should act as though the distribution had larger variance; they

should also act as though the distribution were skewed away from the side represent-

ing a loss. Prospect theory could result in variance neglect. However, there are other

potential causes of variance neglect, and later we will see that prospect theory alone

is inconsistent with some of our results. In particular, respondents do not always act

as though the distribution had larger variance, and this will depend on the width of

the confidence intervals they view.

Variance neglect is also related to the hot hand fallacy and the gambler’s fallacy,

nicely linked elsewhere (Rabin and Vayanos, 2010). Indeed, both the hot hand fallacy

and the gambler’s fallacy result in variance neglect. However, not all cases of variance

neglect stem from the hot hand fallacy or gambler’s fallacy. The situations in which

hot hand fallacy and gambler’s fallacy classically arise are also descriptively different

from the situation we are facing, in which policymakers do not view correlated data

repeatedly but a few data points once.

The next sections provide more information on our data and methods.

3 Sample

This study leverages a unique opportunity to run an experiment on policymakers,

practitioners, and researchers in international development. In particular, we conduct

an enumerated experiment with individuals attending World Bank or Inter-American

Development Bank workshops in which policymakers, practitioners, and researchers

were invited to learn more about impact evaluation and design an impact evaluation.

Data were collected at six World Bank workshops run by the Development Impact

8

Page 9: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Evaluation (DIME) research group. The workshops were conducted in Mexico City

(March 2017), Lagos (May 2017), Washington, DC (May 2017, June 2017), Lisbon

(July 2017), and Dakar (January 2019). The experiment was also conducted at two

Inter-American Development Bank (IDB) workshops in Washington, DC in June,

2017, and May, 2018. Two other World Bank workshops (Mexico City, 2016, and

Nairobi, 2016) were used as pilots to refine the questions and the prior/posterior elic-

itation mechanisms.

These conferences attract participants from around the globe, rather than just

participants from the host country or region. To accommodate more participants,

the survey was translated into Spanish, and respondents had to be fully proficient in

English or Spanish in order to participate. For the workshop in Dakar, French was

also used.

Workshop attendees comprised policymakers, practitioners, and researchers. The

workshops were each approximately one week long and were designed as “matchmak-

ing” events between government staff and researchers. Government counterparts were

paired with researchers and required to design a prospective impact evaluation for

their program over the course of the week. Participants included program officers

in government agencies of various developing countries; monitoring and evaluation

specialists within government agencies; World Bank or IDB operational staff; other

international organization operational staff such as technical advisors at USAID or

DFID; a few staff from NGOs or private sector firms participating in a project; and

academics and other researchers. Those from developing country governments are

considered “policymakers”; international organization operational staff and NGO or

private sector employees are considered “practitioners”; we define “researchers” to be

those in academia or those who either have peer-reviewed publications or else have

“research” or “impact evaluation” in their job title.

Individuals were surveyed by enumerators during breaks in the workshops. Of

526 eligible attendees at the non-pilot workshops, 162 (31%) completed the survey.

The main constraint was that the surveys could only be run during the typically

twice-daily breaks in the workshops and during the lunch period. During the pilots,

individuals were allowed to take the survey by themselves on tablets we provided.

However, we changed approaches after the pilot in favor of one-on-one enumeration

to reduce noise due to participants’ lack of familiarity with operating the tablets and

to increase attentiveness. After making this change, we still had overwhelming inter-

9

Page 10: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

est in the survey among attendees but, being limited to the breaks in the workshops,

only managed to survey an average of 23 participants per workshop. Breaks were

roughly the duration of the survey, and lunch might span 2-3 times the length of the

typical break, depending on workshop timing. Thus, this response rate represents

essentially the maximum number of responses that could be gathered in the allotted

time. We may expect that those who managed to take the survey may have been

particularly interested in taking it or quick to approach the enumerators during a

break, but we have no reason to believe that this represents a substantially different

population than the universe of conference attendees. Response rates by workshop

are detailed in Table 1.

In addition to gathering data at these workshops, past workshop participants were

contacted by e-mail and asked to participate via video conference. The response rate

was much lower in the group contacted by e-mail; of 479 eligible past workshop at-

tendees, 46 (10%) participated in the survey. Finally, participants were recruited at

the World Bank’s headquarters and at the IDB’s headquarters in Washington, D.C.

A table was set up by the cafeteria and passers-by were able to take the survey with

a trained enumerator. 125 responses were collected at the World Bank and 27 at

the IDB over 24 days or 12 lunches, respectively;4 enumerators covered lunch at the

IDB but full or half-days at the World Bank. Summary statistics about the various

recruitment strategies and the breakdown of participants by category (policymaker,

practitioner, researcher) are provided in Table 2.

The experiment was set up to elicit two sets of priors and posteriors, so that with

400 respondents we would expect 800 priors and 800 posteriors. However, some re-

spondents did not complete the entire survey, so we only present results for those who

provided both a prior and a posterior for a given set of questions, resulting in 753

priors and posteriors from 378 respondents, or 94% of the total possible responses.

Finally, a set of responses was elicited on Mechanical Turk to provide a comparison

group. We required a HIT Approval Rate (%) for all Requesters’ HITs greater than

or equal to 95 and Number of HITs Approved greater than or equal to 50.5 1,600 re-

sponses were solicited. In contrast to the policymakers, practitioners and researchers,

who were interviewed one-on-one, the MTurk workers took the survey unsupervised.

4Excluding three responses from support staff at the World Bank and two responses from supportstaff at the IDB. These participants (IT staff, secretaries, lawyers) did not meet our inclusion criteriabut we could not bar them from participating.

5A HIT is a single MTurk task, e.g. a past exercise that they completed.

10

Page 11: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 1: Participants at Workshops

Eligible At Post- ResponseAttendees Workshop Workshop Rate

Mexico, March 2017 86 36 8 0.51Nigeria, May 2017 74 36 1 0.50Washington, DC, May 2017 47 13 2 0.32Washington, DC, June 2017 (IDB) 62 10 0 0.16Washington, DC, June 2017 (WB) 67 21 1 0.33Portugal, July 2017 111 30 8 0.34Washington, DC, May 2018 (IDB) 51 14 0 0.27Senegal, January 2019 54 20 0 0.37

Total 552 180 22 0.37Total (excluding IDB) 439 156 20 0.40

This table shows the number of people surveyed at each workshop and the total number of eligible

attendees. Both values restrict attention to those who could be classified as “policymakers”,

“practitioners” or “researchers”. In addition, to be eligible to take the survey, one had to have not

taken it at a previous workshop (this was primarily a concern for DIME staff) and one had to

speak a survey language fluently (English or Spanish or, for Senegal, English, French or Spanish).

Two of the workshops were held by the Inter-American Development Bank (IDB); all other

workshops were held by the World Bank.

Table 2: Respondents by Recruitment Strategy

Policymakers Practitioners Researchers TotalWorkshops 0.38 0.31 0.32 180Videoconference 0.16 0.29 0.54 68Headquarters surveys 0.04 0.56 0.40 152Total 0.21 0.40 0.39 400

This table shows the percent of respondents who could be classified as policymakers, practitioners

and researchers by each recruitment strategy.

11

Page 12: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Incentives for each of these groups are described in the experimental design section.

4 Experimental Design

4.1 Eliciting Prior and Posterior Beliefs

The overall structure of the experiment is simple: elicit participants’ priors, ran-

domly show them new information, and then elicit posteriors.

To elicit priors, participants were shown a short description of an intervention

(example provided in Figure A.4 in the Appendix) and asked what they thought

the effect of the program was on school enrollment rates in percentage points. We

then asked respondents to place weights on different possible effects the intervention

could have had by having the enumerator drag slider bars next to different ranges

(Figure A.5). This allowed us to elicit a full distribution for their prior belief of the

intervention’s impact. Before using these sliders, participants were shown a video

describing how to use the sliders and were walked through a simple example about

predicting the weather in order to be sure that they understood the exercise. At the

end of this introduction, participants were asked if they understood and were only

allowed to participate further if they stated that they did (Figure A.3). Only one

participant stated that they did not understand the instructions and was prohibited

from continuing.

4.2 Real-Life Data and Decisions

Respondents were first told that they would be shown real data from impact eval-

uations and that they would help decide how a small amount of external funds would

be allocated. In particular, participants received the following text:

You will now be presented with information on two types of programs

that both seek to increase the percent of children enrolled in school: cash

transfers and school meals programs.

We have a small amount of money that you can help to decide how to dis-

tribute. After being shown data on either cash transfers or school meals

12

Page 13: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

programs, you will be asked to decide how you would allocate this set of

funds between cash transfers, school meals programs, or further research.

For example, you could decide to distribute 33% to cash transfers, 33% to

school meals programs, and 34% to further research. At the end of this

study, we will randomly select one person’s responses and distribute the

money according to how that person decided it should be distributed.

Participants were then shown real-life data and asked to allocate funds between the

different options. Some participants were asked to provide their priors and posteriors

before and after seeing the data and before determining their preferred allocations.

The real-life data were taken from AidGrade’s data set of impact evaluation results

(2016). The point estimates and confidence intervals of the data provided for the

conditional cash transfer program were randomized across participants, unbeknownst

to the participants. All respondents received the same data for the school meals pro-

gram. The point estimates and confidence intervals of the different studies shown are

reproduced in Table 3.

Table 3: Real Data Shown

Intervention Point Estimates Confidence Interval WidthsCCT program 1 and 4 2CCT program 1 and 4 5CCT program 2 and 5 2CCT program 2 and 5 5

School meals program 3 4

This table shows the point estimates and confidence interval widths that respondents were shown.Each participant saw the school meals result and one randomly-selected result for conditional cashtransfer programs.

This approach allows us to see how much these differences affect real-world allo-

cations and, in turn, estimate how much allocations could change if respondents were

not biased or if they were to receive information that could help them overcome their

biases.

13

Page 14: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

4.3 Asymmetric Optimism and Variance Neglect

At the conclusion of this section, respondents were told that they were moving on

to a new section of the survey and would not need to use any information that was

previously provided. In this new section, we asked participants to consider effects of

a CCT and a school meals program that had a different, specific description. In the

earlier section, we could not provide many details about the study since we needed

to use real data and needed the description to match across all programs. In this

section, we used hypothetical data rather than real data. The survey merely asked

respondents to “suppose that” they were to guess the effect of a particular program.

The text suggested that enrollment rates were currently at 90 percentage points, so

the most that respondents could reasonably expect the intervention to improve en-

rollment rates was by an additional 10 percentage points. We allowed participants to

guess negative values down to -5.

Participants first provided their prior beliefs as to the effect of a given intervention

by stating their best guess and then placing probability weights in bins as previously

described. They were then randomized into seeing one of several sets of “new data”.

The respondents were asked to imagine that these data represented replications of

studies on the same program. Again, respondents were not informed that they were

presented with randomly-selected results, and audio recordings from surveys where

participants consented to being recorded confirm that no individual questioned why

they were shown those results in a way that suggested they were aware of the ran-

domization.

Participants were presented with hypothetical results from two studies that either

featured a positive or negative outlier relative to their stated prior (“good news” and

“bad news”) and that either provided no confidence intervals, small confidence in-

tervals or large confidence intervals. The positive or negative outlier was randomly

selected but influenced by their stated prior (such that a positive outlier was al-

ways a constant amount above the prior mean, and vice versa for a negative outlier).

The confidence intervals that were provided were randomized independently from the

prior. The “new data” are described in more detail in Table 4, using the example

of someone who previously reported they thought enrollment rates increased by 2

percentage points.

Since the different randomizations could result in confidence intervals stretching

up to 5 above or below the initial mean that respondents provided, we could only

14

Page 15: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 4: Example of Hypothetical Data

Positive outlier: Two data points, one with a mean 1 percentage point belowthe stated value and one with a mean 2 percentage pointsabove the stated value; in the example, they would see themeans 1 and 4.

Negative outlier: Two data points, one with a mean 2 percentage points belowthe stated value and one with a mean 1 percentage point abovethe stated value; in the example, they would see the means 0and 3.

No CIs: No confidence intervals are provided.Small CIs: Confidence intervals are provided that extend 2 percentage

points above/below each disaggregated data point. For themeta-analysis results, these are aggregated as they would bein a fixed effects meta-analysis.

Large CIs: Confidence intervals are provided that extend 3 percentagepoints above/below each disaggregated data point. For themeta-analysis results, these are aggregated as they would bein a fixed effects meta-analysis.

This description of the different types of evidence a participant could view is based on thehypothetical case of someone who previously reported they thought enrollment rates increased by 2percentage points.

15

Page 16: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

follow the above strategy for those who initially stated they expected treatment ef-

fects between 0 and 5 percentage points (otherwise, confidence intervals would be cut

off in the graphical representation). We believe that this is a reasonable range and

most responses indeed fell within 0 to 5 percentage points, especially given that re-

spondents knew that baseline enrollment rates were 90 percentage points. Appendix

Table 14 shows the distribution of prior means, with 82.2% of responses falling in

this range; Figure 2 breaks this down by subgroup, with policymakers notably having

higher prior means than practitioners (pă0.05) and researchers (pă0.001), and with

practitioners also having higher prior means than researchers (pă0.05). To ensure

the graphical representations were properly displayed, those who stated expected val-

ues lower than 0 were shown the same data as those who stated expected values of

0, and those who stated expected values higher than 5 were shown the same data

as those who stated expected values of 5. This poses a slight problem for tests of

asymmetric optimism, since e.g. people who stated expected values greater than 5

will tend to see new data lower than their priors. Since the data are not symmetric,

if we saw them updating less on these values, we would be unable to attribute that

to the values being lower than their priors - it could just be that the people who have

high initial priors are also less likely to shift their priors. For this reason, for tests of

asymmetric optimism we will restrict attention to those whose priors fell between 0

and 5. However, we still use all data for tests of variance neglect.

All data were presented as bar charts, as these represent a common form in which

results are displayed to policymakers. Figure A.6 provides an example. The order in

which the two data points were provided (left to right) was randomized. The data

were also described in the text. After viewing these data, participants were again

presented with a set of slider bars and asked to put weights on different effect ranges,

capturing their posteriors.

By eliciting participants’ µt´1, σ2t´1 and µt in this manner, and given that we

experimentally provide them with Yi and σ2i , we can calculate what their k is using

µt “ µt´1 ` kpYi ´ µt´1q. If we observe that k` ą k´, where k` represents the

calculated k for those receiving the positive outlier treatments and k´ represents the

calculated k for those receiving the negative outlier treatments, that would be evi-

dence of asymmetric optimism. Using the notation of the model, we can estimate γ

such that respondents are Bayesian updating with the correct k based on the wrong

signal.

16

Page 17: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure 2: Distribution of Prior Mean by Subgroup

This figure shows the distribution of the prior mean by each main subgroup, for those passing allthe screenings and tests. Few responses fall below 0, suggesting respondents are paying attentionand averse to giving negative estimates; most fall between 0 and 5.

We can also compare the responses of those who receive a signal with a small

confidence interval and those who receive a signal with a large confidence interval. If

we observe that BkB{Bσ2i ą Bk{Bσ

2i , where kB represents the k of a Bayesian updater,

that would be evidence of variance neglect. We can then estimate λ.

One thing we cannot do is to both estimate optimism with respect to confidence

intervals as well as estimating variance neglect, since confidence intervals derive from

the variance. However, we may still be able to discern whether people are optimistic

with respect to confidence intervals, if this outweighs variance neglect. For example,

imagine that someone is presented with data with a mean that is above their prior

µt´1 and alternatively small or large confidence intervals. Someone who is optimistic

with respect to confidence intervals, and for whom this outweighs any variance ne-

glect, would put more weight on the result with larger confidence intervals. Someone

for whom variance neglect outweighs optimism with respect to confidence intervals

would put less weight on the result with larger confidence intervals, even though they

would put relatively more weight on the result with larger confidence intervals com-

pared to the result with smaller confidence intervals than someone who was Bayesian

updating. In other words, if kS represents the k that someone who saw small confi-

dence intervals gives, kL the k that someone who saw large confidence intervals would

give, and the superscripts V N , O and B represent variance neglect, optimism with

respect to confidence intervals, or Bayesian updating, respectively, we might expect

17

Page 18: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

kBS ´ kBL ą kV N

S ´ kV NL ą 0 ą kOS ´ k

OL .

Since the model depends on having normally distributed priors and data, in all

analyses we restrict attention to those who report normally distributed priors and pos-

teriors. It remains possible, however, that some individuals had seemingly normally

distributed priors and posteriors without believing the new information represented

draws from a normal distribution. As will be described, nearly 80% of the sample

appeared to have normally distributed priors and normally distributed posteriors.

4.4 Information Treatments

If policymakers are biased, are there ways of presenting information to help them

make better decisions? The model makes some predictions as to how providing con-

fidence intervals might affect estimates. If respondents do not suffer from variance

neglect and are not optimistic, they will Bayesian update in the absence of any other

biases, and in this case providing confidence intervals is helpful. If respondents are

not optimistic but do suffer from variance neglect, confidence intervals are less helpful

since respondents somewhat ignore them. If respondents are optimistic and do not

suffer from variance neglect, confidence intervals could help or hurt depending on

whether respondents are optimistic with respect to point estimates or with respect to

confidence intervals. If respondents are optimistic and suffer from variance neglect,

confidence intervals could still help, but less so; confidence intervals alternatively

could still hurt if respondents were optimistic with respect to confidence intervals.

Table 5 summarizes.

The previously described experiment, in which results are presented with or with-

out confidence intervals, can only tell us whether it is better to provide confidence

intervals or not under assumptions about the true study sampling variance. To further

examine how the information provided may affect estimates, we provide respondents

with one of several types of information in a separate treatment. These different

types of information are provided in the context of an introductory question that

asks respondents to estimate temperature. Participants are randomized into receiv-

ing point estimates without confidence intervals; point estimates with confidence in-

tervals; point estimates with confidence intervals and the interquartile range; and

point estimates with confidence intervals, the interquartile range, and maximum and

minimum values. These treatments were constructed so that each subsequent arm

18

Page 19: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 5: Predictions

No variance neglect Variance neglectNo asymmetric Confidence intervals Confidence intervals

optimism are helpful less helpfulAsymmetric Confidence intervals Confidence intervals

optimism are helpful / could harm less helpful / harmful

This table provides theoretical predictions based on whether an individual suffers from asymmetric

optimism and/or variance neglect.

contains the same information as the previous arm plus some additional information;

in other words, the treatments can be thought of as ordered, providing more or less

information about the same distribution. Figure A.2 illustrates.

4.5 Incentives

Policymakers, practitioners and researchers were offered a token gift in the work-

shops (chocolate or coffee costing approximately $5 USD) in exchange for their time.

In addition, participants were informed that at the end of the study, one response

would be drawn at random and awarded an additional prize: a MacBook. We did

not further incentivize responses because we were concerned that policymakers in

particular would fear giving a “wrong” answer, so we did not want to increase the

salience of the possibility of answering “incorrectly” by offering incentives for “cor-

rect” answers. The same incentives were offered to participants at the World Bank

and Inter-American Development Bank headquarters. For those interviews conducted

over videoconference, a $15 Amazon voucher was provided, again without further

conditions, along with entry to the MacBook raffle. Enumerators were trained to

encourage participants who feared giving a wrong answer that we merely wanted to

know what they thought given the information we provided. Despite the lack of for-

mal incentives, respondents appeared to take the survey seriously, with the median

respondent spending 32 minutes on it.6 Respondents also appeared to provide consis-

tent answers, as will be described in the results section.7 The direct interaction with

6This excludes those who, in the data, appear to have taken more than an hour on the survey;we assume these represent individuals who started to take the survey, ran out of time before the endof the workshop break, and continued it in the next break in the workshop.

7For example, those who provided narrower priors updated less on the new information.

19

Page 20: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

an enumerator may have contributed to respondents putting effort into their actions.

MTurk participants were offered $1.50 for the relatively long survey. We were

concerned that without incentivizing thoughtful responses, participants might not

put in the effort to understand and carefully answer the questions. However, not

incentivizing the responses would provide greater comparability with the results from

policymakers, practitioners and researchers. Thus, we continued to not incentivize

responses but chose to implement pre-specified screening questions to filter out inat-

tentive participants. Screening questions are described in the next section.

4.6 Screening

As described, MTurk responses were subject to screening. In particular, the first

question asked what respondents thought the likelihood was that it would rain to-

morrow in their city. They were then asked: “Now suppose that the weather forecast

says there is a 50% chance it will rain tomorrow. Now what do you think is the

likelihood that it will rain tomorrow?” If their new estimate was outside of the range

between their initial answer and 50%, they were excluded, with an exception that

will be described below. This rule excluded those whose responses implied k ă 0

(e.g. they initially answer 10%, then update their answer to 0%, or if they initially

answer 90%, then update their answer to 100%) and those whose responses implied

k ą 1 (e.g. they initially believe the likelihood is 10%, then update their answer

to 60%, or they initially believe the likelihood is 90%, then update their answer to

40%). The second screening question asked what respondents thought the average

monthly temperature would be in Paris this month. Again, they were provided with

new information and those who provided a second answer that implied k ă 0 or k ą 1

were excluded, barring the exception described below. Finally, the third screening

question presented them with pre-populated sliders that put probability weights on

low temperature ranges. They were asked to modify these weights given the new

information that two women who were perfectly informed as to what the weather

would be like decided to wear shorts and a T-shirt; presumably, this would imply a

higher temperature than the low numbers provided. Anyone who modified the sliders

so as to result in a still lower mean temperature was excluded, barring the exception

described below.

The point of these questions was not to screen out people who suffer from gam-

20

Page 21: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

bler’s fallacy or hot hand fallacy or who updated in some other way so as to result in

implied values of k ă 0 or k ą 1. Rather, if an MTurk worker answered in this way

with regards to something as familiar as the weather, it seems likely that they were

simply not paying attention to the question. By repeating the same kind of screening

question three times, we can detect whether someone is consistently answering in a

way that would imply k ă 0, consistently answering in a way that would imply k ą 1,

or answering inconsistently across questions. Hence, we allowed one exception to the

above screening rules: if someone consistently updated in a way that would imply

k ă 0 through all screening questions or in a way that would imply k ą 1 through the

first and second question and not k ă 0 in the third question (as we cannot detect if

k ą 1 in the third question), we included them in the sample. Still, we consider these

participants to be more likely to be inattentive or confused, so they are not part of

our preferred specification. Finally, we also excluded anyone who failed to shift the

pre-populated sliders in the third question to focus on those who put some effort into

answering the questions.

Policymakers, practitioners and researchers were not subject to these constraints

as they went through the survey with an enumerator. Both the policymakers, prac-

titioners and researchers sample and the MTurk sample were asked at the end of the

introductory section if they understood how to use and interpret the slider bars, and

if anyone selected the response “No”, they were excluded from the sample. These

screening questions were described in a pre-analysis plan posted on the Open Science

Framework (OSF) and the study pre-registered at the AEA RCT Registry.8

5 Results

5.1 Descriptive Statistics: Priors

This section describes how respondents’ priors were distributed, with reference to

both the policymakers, practitioners and researchers sample (hereafter referred to as

PPR) and the MTurk sample.

Recalling that our identification strategy for k assumes normally distributed pri-

ors, we test for normality using a Kolmogorov-Smirnov test. This test is typically

8OSF repository: https://osf.io/2da9p/, AEA registration id: AEARCTR-0001237,https://www.socialscienceregistry.org/trials/1237.

21

Page 22: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

conducted for continuous distributions, rather than for data that fall in discrete bins,

as in our case. However, the Kolmogorov-Smirnov test can be applied to discrete

data with modifications. It should also be noted that Kolmogorov-Smirnov tests are

not well-powered for distributions made up of a few discrete bins, as we have. While

15 bins were available, most respondents’ estimates fell into 5 or fewer bins. We first

discard those fairly rare cases in which respondents put weight in only one or two bins

(7.2% of the PPR sample, 3.6% of the MTurk sample). The Kolmogorov-Smirnov

tests of the prior distributions reject 10.0% of the remaining observations in the PPR

sample.

Of 1,675 MTurk respondents9, 1,029 passed the screening questions. Again, one

set of screening questions required respondents to not update very differently across

several questions, i.e. to not answer as though k ą1 for some questions and as though

k ă0 for other questions; 1,183 passed this requirement. There was also one screen-

ing question which provided respondents with a set of pre-filled slider bars and asked

them to adjust the slider bars to reflect their beliefs given new information; since

participants were not required to adjust the bars at all but the question was designed

such that respondents should adjust their estimates upwards, we dropped anyone who

did not adjust any of the bars, considering them too inattentive or uninterested in

putting effort into their responses. After discarding the observations in which weight

was placed in only one or two bins, as previously described, the Kolmogorov-Smirnov

tests then discard an additional 15% of observations.

Finally, k could not be calculated for a subset of observations (11.8% in the PPR

sample, 8.1% for the MTurk sample) for the mechanical reason that the point esti-

mate that respondents were shown, which was based on the first mean value that

they stated, turned out to be exactly equal to the mean that we calculated from

their putting weight in bins.10 It might seem unusual for the new data to be exactly

centered on the midpoint of their prior distribution, but this could happen if a re-

spondent gave a perfectly symmetric distribution. For the MTurk sample, if we could

not calculate k for either of the two sets of questions for which we tried to calculate

9We accidentally gathered slightly more data than initially planned, as a few more people an-swered the survey than filled in a survey code on MTurk within the alloted time, such that the HITwas not counted and was re-offered to other participants.

10Remember that we asked respondents to first state an integer value and then put weight inbins to make the weighting part of the exercise easier. We take these weights as the most accurateestimate of their prior mean, though we use the integer values in a robustness check.

22

Page 23: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

it, we dropped the response and recruited a new participant. For the PPR sample,

incomplete surveys were not discarded due to the relative difficulty of obtaining these

responses.

5.2 Descriptive Statistics: Posteriors

The reported posteriors were broadly similar to the priors in terms of how many

appeared normally distributed. Of those observations in the PPR sample for which

the associated prior passed all the tests, 1.4% were dropped for having a posterior

distributed in 1-2 bins and a further 4.7% of the posterior distributions were rejected

as not normal by a Kolmogorov-Smirnov test. For the MTurk sample, these numbers

were 2.8% and 7.3%, respectively.

Overall, 79% of policymakers, practitioners and researchers and 76% of MTurk re-

spondents reported distributions that are consistent with having normally distributed

priors and normally distributed posteriors using a Kolmogorov-Smirnov test.

5.3 Descriptive Statistics: Distribution of k

Restricting attention to the group that reported normally distributed priors and

posteriors, we calculated k. Figure 2 plots histograms of its distribution, illustrating

a relatively large range, with clusters of estimates around 0 and 1. Table 6 shows

quantiles of k for various types of respondent. Note that k should generally fall be-

tween 0 and 1, with those who take the data mean as their posterior mean having

k “ 1 and those who stick with their initial mean having k “ 0. Surprisingly, only

55% of the PPR estimates and 44% of the MTurk estimates fall within this range,

which will be discussed more later.

Notably, MTurk workers updated more on the data than policymakers, practition-

ers or researchers. This makes intuitive sense: they also reported less familiarity with

the types of interventions discussed. Figure 2 shows that policymakers, who stated

they were less familiar with studies than practitioners or researchers, also had higher

values of k. Somewhat surprisingly, there was a cluster of researchers who appeared

to update too much based on the data.

The bottom plot in Figure 2 distinguishes between responses to a “knowledge”

question asked of all respondents: for each intervention, respondents were asked to

specify whether they had “never heard of it”, “heard of it but never heard of any

23

Page 24: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 6: Quantiles of k

Percentile

Subgroup 10 25 50 75 90PPR All -1.00 0.00 0.50 1.01 2.33

Policymakers -0.29 0.00 0.61 1.27 2.43Practitioners -0.99 0.00 0.33 1.00 1.72Researchers -1.00 0.00 0.60 1.17 3.00

Familiar -0.55 0.00 0.59 1.02 2.43Unfamiliar -2.68 0.00 0.37 1.00 2.11

MTurk All -1.51 -0.01 0.68 1.23 3.16Familiar -2.36 -0.14 0.62 1.15 3.08

Unfamiliar -1.26 0.00 0.72 1.26 3.25

This table shows quantiles of k by subgroup. “PPR” refers to the policymakers, practitioners andresearchers sample. “Unfamiliar” refers to those respondents who noted that they had either neverheard of conditional cash transfers or school meals programs or had heard of them but had neverheard of any studies on them, while “Familiar” refers to those who noted that they had heard ofthem and heard of some studies on them or had heard of them and were very familiar with studieson them.

studies on it”, “heard of it and heard of some studies”, or “heard of it and very famil-

iar with studies”. Those who reported greater familiarity with a type of intervention

typically updated less in response to new information.

Our primary focus was on calculating k from the distributions that respondents

provided. However, each time we asked for a prior or posterior, we also asked respon-

dents to first provide their best guess of the effect of the program. We use these as

estimates of their µt´1 and µt in an alternative specification which produces similar

results.11

5.4 Tests for Biases

The wide dispersion of k values complicates testing for biases. Values of k within

the -0.5 to 1.5 range could simply reflect noise in using the slider bars. However, we

may think that values of k that are much smaller or larger suggest that individuals

are using a different updating heuristic entirely or misinterpreting the exercise. For

example, k could be greater than 1 if respondents have in mind a certain prior based

on older information they hold, are surprised by the results they are shown, and

11Results available upon request.

24

Page 25: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure 3: Distribution of k

This figure plots values of k calculated from respondents’ reported µt´1, µ1t, and the provided Yi

values. Values below -2 or above 3 are not included for legibility. The bottom plot distinguishes

between responses to a “knowledge” question asked of all respondents: for each intervention,

respondents were asked to specify whether they had “never heard of it”, “heard of it but never

heard of any studies on it”, “heard of it and heard of some studies”, or “heard of it and very

familiar with studies”. For visual clarity, the first two categories are collapsed into “Unfamiliar”

and the latter two into “Familiar”.

25

Page 26: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

infer there is a time trend that explains the discrepancy between their older priors

and the new information they were shown. They may then expect that the effects

captured in a subsequent evaluation would also be subject to the same time trend.

We have limited information on the heuristics that respondents are using, based on

descriptive responses to questions about why they answered the way they did.12 As

we cannot fully identify the reasons why some responses imply particularly large or

small values of k, we attempt to separate individuals into different types by restricting

attention to those that provide responses within alternative ranges of k: 0 ď k ď 1

and ´0.5 ď k ď 1.5. The logic is that this may help to distinguish between groups

using very different heuristics. However, we recognize that this division is not precise,

and as an alternative to restricting attention to k within given ranges we also present

results of logistic regressions that use all data but use an indicator variable as the

dependent variable.

Table 7 presents results for regressions of respondents’ calculated values of k on

whether they received the “good news” news treatment. Robust standard errors

are used, clustering observations at the individual level. Receiving the good news

treatment significantly affected k for most specifications in the PPR sample and all

specifications in the MTurk sample.

Recall that to test for variance neglect, we need to consider what someone who

was Bayesian updating would do. Thus, in Table 8, we construct kB ´ k, where kB

is the value that k should have taken if respondents were Bayesian given their stated

priors and the Yi and σ2i they observed. The model implied that kBS ´k

BL ą kV N

S ´kV NL ;

we test this by regressing kB ´ k on whether the respondent saw large or small

confidence intervals. In Table 8, we observe that kBS ´ kS is indeed generally greater

than kBL ´kL, as we predicted. Again, the first column reports the results of a logistic

regression over the full set of responses, with the dependent variable in this case

taking the value 1 if kB ´ k ą 0 and 0 otherwise.

12In particular, after every question, respondents were asked to describe why they answered theway they did, the enumerator summarized this response in a text box, and respondents were givenan opportunity to amend the summary. For a subset of responses, the full audio transcripts of thisexchange were also captured.

26

Page 27: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 7: Tests of Asymmetric Optimism: Regressing k

PPR MTurk

(1) (2) (3) (4) (5) (6)

Good News 2.107*** 0.076 0.172*** 1.766*** 0.091*** 0.118***(0.42) (0.06) (0.06) (0.23) (0.03) (0.04)

Observations 429 229 287 1053 416 594

This table reports the results of regressions of k on an indicator of whether the respondent received

“good news”, i.e. positive new data relative to their priors. Columns (1) - (3) report results using

the policymakers, practitioners and researchers sample; Columns (4) - (6) report results using the

MTurk sample. Columns (1) and (4) present the results of a logistic regression of Ipk ą 0.5q on

whether the respondent received good news, using exponentiated coefficients; Columns (2) and (5)

regress k on whether the respondent received good news, restricting attention to only those

observations for which 0 ď k ď 1; Columns (3) and (6) also regress k on whether the respondent

received good news, considering only those observations for which ´0.5 ď k ď 1.5. Only those who

provided prior means between 0-5 percentage points were included in the tests for asymmetric

optimism, as we were unable to show new data above or below higher or lower priors without going

out of range of what could be displayed.

5.5 Heterogeneity by Profession

We may also be interested in how results vary by sub-sample. We consider het-

erogeneity by profession in Table 9. Depending on the specification, practitioners and

researchers updated less on the data than MTurk workers. This would be consistent

with a story in which they had more background knowledge or narrower priors (as

they do in fact have), and it does not require a difference in updating. Policymakers,

practitioners and researchers do not appear to experience significantly more or less

asymmetric optimism than MTurk workers, the subgroup left out. We also observe

that they were no closer to Bayesian than MTurk workers.13 In the main specifications

which treat the subgroups separately, they do not appear to suffer significantly more

or less variance neglect than the MTurk workers. Collapsing policymakers, practi-

tioners and researchers into a single group, PPR respondents pay more attention to

the variance in some specifications (Appendix Table 15).

13As evident from their having no smaller kB´k, indicated by the regression coefficients on beinga policymaker, practitioner or researcher in Columns (4)-(6).

27

Page 28: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 8: Tests of Variance Neglect: Regressing kB ´ k

PPR MTurk

(1) (2) (3) (4) (5) (6)

Large C.I. 1.342 -0.118** -0.073 0.770** -0.121*** -0.156***(0.29) (0.06) (0.06) (0.10) (0.03) (0.04)

Observations 349 206 256 1025 454 639

This table reports the results of regressions of kB ´ k on an indicator of whether the respondent

saw large confidence intervals as opposed to small confidence intervals. Respondents can be

included regardless of their priors, unlike in testing for asymmetric optimism, but cases in which

respondents were randomized into seeing no confidence intervals are excluded. Columns (1) - (3)

report results using the policymakers, practitioners and researchers sample; Columns (4) - (6)

report results using the MTurk sample. Columns (1) and (4) present the results of a logistic

regression of IpkB ´ k ą 0q on whether the respondent saw a large confidence interval as opposed

to a small confidence interval, using exponentiated coefficients; Columns (2) and (5) regress kB ´ k

on whether the respondent received a large confidence interval as opposed to a small confidence

interval, restricting attention to only those observations for which 0 ď k ď 1; Columns (3) and (6)

also regress kB ´ k on whether the respondent received a large confidence interval or a small

confidence interval, considering only those observations for which ´0.5 ď k ď 1.5. Again,

cluster-robust standard errors are used.

28

Page 29: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 9: Heterogeneity by Profession

Asymmetric Optimism Variance Neglect

(1) (2) (3) (4) (5) (6)

Good News 1.766*** 0.091*** 0.118***(0.23) (0.03) (0.04)

Policymaker 1.277 -0.033 0.060 0.935 0.103 0.064(0.43) (0.09) (0.09) (0.28) (0.09) (0.10)

Policymaker * 1.025 -0.120 -0.027Good News (0.52) (0.15) (0.16)Practitioner 0.540*** -0.163*** -0.164*** 1.291 0.088 0.053

(0.12) (0.05) (0.05) (0.33) (0.07) (0.07)Practitioner * 1.217 -0.030 0.017Good News (0.41) (0.09) (0.10)Researcher 0.811 -0.128** -0.160*** 1.068 0.097 0.013

(0.18) (0.06) (0.06) 0.29 (0.07) (0.08)Researcher * 1.271 0.027 0.114Good News (0.42) (0.10) (0.10)Large C.I. 0.770** -0.121*** -0.156***

(0.10) (0.03) (0.04)Policymaker * 1.967 -0.071 -0.067Large C.I. (0.89) (0.14) (0.13)Practitioner * 1.642 0.036 0.121Large C.I. (0.60) (0.09) (0.10)Researcher * 1.714 -0.016 0.103Large C.I. (0.62) (0.11) (0.11)

Observations 1482 645 881 1374 660 895

This table considers heterogeneity by profession, including both the PPR sample and MTurk

sample and interacting indicator variables for each subgroup. Columns (1) and (4) present

(alternatively) the results of a logistic regression of Ipk ą 0.5q on whether the respondent received

good news or a logistic regression of IpkB ´ k ą 0q on whether the respondent saw large confidence

intervals, using exponentiated coefficients; Columns (2) and (5) present results of regressions of k

and kB ´ k, respectively, on whether the respondent saw large confidence intervals, restricting

attention to those observations for which 0 ď k ď 1; Columns (3) and (6) consider only those

observations for which ´0.5 ď k ď 1.5. The tests for asymmetric optimism and variance neglect

were conducted on slightly different samples. The tests of asymmetric optimism require

respondents to have mean priors between 0 and 5, while the tests for variance neglect require that

respondents be randomized into seeing small or large confidence intervals (as opposed to no

confidence intervals).

29

Page 30: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

5.6 Estimating γ and λ

We can also obtain distributions of γ and λ across individuals and test whether

these distributions are different from 0. We find an average value of γ of 0.36 and

an average value of λ of -2.23 in the PPR data, winsorizing the highest and lowest

1% of the data. γ is significantly different from 0 at p ă 0.05 and λ is significantly

different from 0 at p ă 0.001.14 If γ is positive, it indicates asymmetric optimism;

if λ is negative, it indicates that subjects update too much in response to the new

information, given its confidence interval. In the MTurk data, the equivalent values

for γ and λ are 0.54 and -1.78, significantly different from 0 at p ă 0.0001 and

p ă 0.01, respectively.

5.7 Changes in Allocations

In this section, we present reduced-form estimates of the impact of seeing more

positive results data on allocations. Only policymakers, practitioners and researchers

answered this part of the survey. In some workshops, we did not ask this part of the

survey due to time constraints, so we only have allocations from 308 individuals.

There are several reasons we may expect that information provided may not influ-

ence participants’ allocations. First, in our experiment they only receive information

about a particular outcome variable, and they may care about different outcomes.

Second, they may not feel like they have sufficient information to materially change

their beliefs.15

The average allocations to CCT programs, school meals programs and further

research, respectively, were 35%, 35% and 30%. Recall that in this part of the survey,

respondents randomly viewed a selection of real data on CCT programs, with point

estimates of 1 and 4 or point estimates of 2 and 5 and with confidence intervals rang-

ing 2 or 5 above and below those values. Viewing the larger point estimates resulted

14Without winsorizing, the values are still significant but noisier: the average value of γ is then0.45, the average value of λ, -3.72, and they are significantly different from 0 at p ă 0.05 and p ă 0.1,respectively.

15As the section that asked participants to make allocations used real rather than hypotheticaldata, no study details could be provided, lest they update on characteristics that varied betweenstudies, as previously explained. Updating based on study characteristics would be more realisticbut would prevent us from identifying the effect of observing more positive results. A companionpaper (Coville and Vivalt, 2019) examines how a similar sample values different studies based ontheir characteristics, using a discrete choice experiment.

30

Page 31: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

in an increase in the amount allocated to CCTs of 8 percentage points; viewing re-

sults with larger confidence intervals resulted in a decrease in the amount allocated

to CCTs of 8 percentage points. Interestingly, when respondents saw large confidence

intervals, they allocated 5 percentage points more to further research. Seeing a large

confidence interval did not temper the response to large point estimates. Results are

presented in Table 10.

Table 10: Regressions of Allocations on Evidence Shown

(1) (2) (3) (4) (5) (6)CCT CCT CCT Research Research Research

Funding Funding Funding Funding Funding Funding

Large Point 7.137*** 7.450*** 1.910 -0.347Estimate (1.95) (2.68) (2.48) (3.30)Large Confidence -8.379*** -8.291*** 5.741** 3.132Interval (1.98) (2.58) (2.48) (3.60)Large Point -0.354 4.435Estimate * Large C.I. (3.78) (4.94)

Observations 308 308 308 308 308 308R2 0.04 0.06 0.10 0.00 0.02 0.02

This table reports the results of regressions of allocations to CCTs or to further research onwhether or not large or small point estimates were shown (a mean difference in the point estimatesof 1 percentage point), whether large or small confidence intervals were provided (“large”confidence intervals extended 5 percentage points above or below the point estimate; “small”confidence intervals extended 2 percentage points above or below the point estimate), and theinteraction of the two. We were unable to ask the allocation question at all workshops due to timeconstraints, hence the smaller sample size. Robust standard errors are used.

It is instructive to consider how much allocations would change if respondents were

Bayesian updating. Recall that we estimated γ to be equal to 0.36 and λ to be equal

to -2.23 in the PPR sample. If a policymaker suffering from optimism misperceived

a signal that they thought a certain option had a point estimate that was 0.36 higher

than a Bayesian might think it to be, the estimates in Table 10 would suggest they

might allocate 2.7% more than a Bayesian would to that option. Similarly, someone

suffering from variance neglect might allocate 11.0% more to that option than they

would if they were Bayesian.

In equilibrium, this effect might be partially washed out. For example, if owing

to their biases a policymaker wanted to allocate more to every program, they would

31

Page 32: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

not have enough money to do so, and so we might imagine they would still make the

same allocations across programs that they would if they were Bayesian. However,

this may be unlikely for at least two reasons. First, policymakers are unlikely to

receive information for many projects at the same time. For example, a policymaker

could have priors about a default option and only observe a signal about one other

program and then update too much based on those results. More fundamentally, it

should be emphasized that due to the shape of the normal distribution, thinking that

one perceives Yi ` γ rather than Yi results in more updating for larger values of Yi, a

fact that is obscured in the linear regressions of Table 10. The impact of λ on updat-

ing is also non-linear. Because of this, while we present these back-of-the-envelope

calculations, it should be noted that they are particular to the details of the questions

we asked and the priors respondents had.

5.8 How Much Information Should We Provide?

The type of information that was provided affected updating. Figure 3 shows

results from the PPR and MTurk sample for each of the four treatment arms: pro-

viding results without confidence intervals; with confidence intervals; with confidence

intervals and the interquartile range; and with confidence intervals, the interquartile

range, and maximum and minimum values.

Table 11 shows regression results. Both samples show some increased updat-

ing in response to more information, with the main difference being between seeing

point estimates and/or confidence intervals as opposed to point estimates, confidence

intervals, and the interquartile range, with or without maximum and minimum val-

ues. The significant results imply that some of the previously observed biases may

be mitigated by strategic provision of information.

6 Causes of Variance Neglect

So far, it is not clear what is driving the observed variance neglect. Respondents

could be misinterpreting confidence intervals, or they could be interpreting them cor-

rectly but not putting weight on them for other reasons (e.g. inattentiveness). To

shed light on this issue, we run a set of additional experiments with 600 participants

on MTurk. In the workshop setting, we could not conduct experiments that looked

32

Page 33: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 11: Impact of More Information

PPR MTurk

(1) (2) (3) (4) (5) (6)Ipk ą 0.5q k k Ipk ą 0.5q k k

Point Estimate 0.446* -0.133*** -0.091* 0.662*** -0.096*** -0.065***(0.19) (0.05) (0.05) (0.08) (0.02) (0.02)

Point Estimate and 0.397** -0.103* -0.093* 0.644*** -0.097*** -0.067***Confidence Interval (0.17) (0.05) (0.05) (0.08) (0.02) (0.02)Point Estimate, 0.793 -0.021 -0.019 0.978 0.010 -0.020Confidence Interval, IQR (0.37) (0.04) (0.04) (0.13) (0.02) (0.02)

Observations 388 314 373 3970 2838 3660

This table shows the impact of providing more information on k. Each type of information isrepresented by an indicator variable, with the category left out being shown point estimates,confidence intervals, the interquartile range, and maximum and minimum values. Columns (1)-(3)show results for the PPR sample; Columns (4)-(6), the MTurk sample. Columns (1) and (4) showthe results of a logistic regression on Ipk ą 0.5q with exponentiated coefficients. Columns (2) and(5) restrict attention to 0 ď k ď 1, and Columns (3) and (6) restrict attention to ´0.5 ď k ď 1.5.

like numerical exercises, but we can ask these types of questions of MTurk respon-

dents.

First, we check whether respondents still seem to exhibit variance neglect when

faced with incentivized questions in a value of information game. Respondents are

first asked to provide their full distribution of priors regarding the effect of a particu-

lar program on a particular outcome. Then they are given 100 tokens and asked their

willingness to pay for information from a replication, after which they can modify

their guess. Those who guess correctly within a range will win a bonus. Respondents

are asked their willingness to pay for results with various levels of precision, with the

understanding that after they provide all these estimates one of the estimates will

be selected and the tokens bid on that question taken from them. The more tokens

they bid to receive information, the more likely they would be to receive the infor-

mation, the “cost” of the information being set at some threshold unknown to them.

Remaining tokens that are unspent are also worth a small bonus, so this exercise is

incentive-compatible.

We then compare respondents’ willingness to pay at each level of precision with a

Bayesian’s willingness to pay, given the same priors. We pre-specified that those who

were unwilling to pay for any of the information sets would be excluded. We might

33

Page 34: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

expect that for certain levels of precision, a Bayesian would be willing to pay more for

a study’s results than our respondents, but for other levels of precision, a Bayesian

would be willing to pay less for a study’s results than our respondents. In particular,

for someone exhibiting variance neglect, they should be willing to pay more than a

Bayesian for results with low degrees of precision and less than a Bayesian for results

of high degrees of precision, as in Figure A8.

We also ask respondents to provide their willingness-to-pay for two kinds of confi-

dence intervals: 95% confidence intervals and the equivalent 99% confidence intervals

based on the same mean and standard error. While 99% confidence intervals are

rarely used, under variance neglect wide confidence intervals are treated as somewhat

smaller. It is therefore possible that at least for larger standard errors, 99% confidence

intervals may lead to updating that is closer to Bayesian than 95% confidence intervals

if it helps visually nudge respondents to think of the confidence intervals as particu-

larly large.16 This also allows us to test whether respondents seem to understand the

difference between 95% and 99% confidence intervals by comparing willingness to pay

measures for each 95% confidence interval with their corresponding 99% confidence

interval. To guard against inattentiveness, we impose a requirement that respondents

must successfully pass at least 15 out of 16 attention checks.17

The theoretical willingness-to-pay that respondents should have had given their

priors if they were Bayesian and their observed willingness-to-pay for results with var-

ious confidence intervals are reported in Table 12. Consistent with variance neglect,

subjects are less responsive to differences in confidence interval widths than someone

with the same priors who was Bayesian updating would be.

Interestingly, respondents were willing to pay more for results that provided a 99%

confidence interval than for results with the equivalent 95% confidence interval. For

example, a 95% confidence interval that is 2.12 percentage points wide is equivalent

16The opposite is true if the true result has small confidence intervals: presenting 95% confidenceintervals rather than 99% confidence intervals may help in this case.

17This was pre-specified. The attention checks were as follows. First, some questions allowedrespondents to provide a lower bound that was higher than the upper bound they provided whencombining confidence intervals. Second, some questions allowed respondents to provide a lowerbound that was higher than the upper bound they provided when creating different confidenceintervals, given a 95% confidence interval. Third, some questions allowed respondents to provideanswers that were outside the range of possible values stated in the question. Finally, some questionsallowed respondents to provide larger confidence intervals than the 95% confidence interval whenbeing asked to provide smaller (e.g. 90%) confidence intervals. We count each instance of one ofthese types of errors as a mistake.

34

Page 35: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

to a 99% confidence interval that is 2.78 percentage points wide. Yet, the average

respondent was willing to pay 18.31 tokens for results with a 95% confidence interval

that was 2.12 percentage points wide and willing to pay 20.40 tokens for results with

a 99% confidence interval that was 2.78 percentage points wide. This statistically

significant discrepancy points to a fundamental misunderstanding of confidence in-

tervals. For each 95% confidence interval, respondents were willing to pay more for

the equivalent 99% confidence interval (Table 12). Respondents also appeared to up-

date more on results with 99% confidence intervals than results with equivalent 95%

confidence intervals (Appendix Table 16).

Table 12: Willingness-to-Pay for Equivalent 95% and 99% Confidence Intervals

95% Confidence Intervals

WTP 2.12 3.23 4.35 5.46 6.58 7.70

Bayesian 70.57 47.50 30.04 19.26 12.64 8.5595% Interval 18.31 15.39 13.60 12.89 12.52 12.7399% Interval 20.40 16.81 15.27 14.26 13.95 14.23

p-value 0.0001 0.0025 0.0001 0.0014 0.0007 0.0007

This table shows how the MTurk sample was willing to pay more for every 99% confidence intervalthan the equivalent 95% confidence interval; the p-value is for a t-test of the difference betweenwillingness-to-pay for 95% and 99% intervals. The Bayesian willingness-to-pay figures arestatistically significantly different at pă0.001 for every confidence interval except those in Column(5). Prices are in terms of number of tokens.

We also asked the same subjects to answer a few other questions as part of the same

experiment. In particular, respondents were asked to combine two confidence intervals

framed as coming from replications of the same study. No details were provided

about the studies, so as to focus attention on the confidence intervals, and in the

absence of other information this was purely a mathematical exercise. Close guesses

would be rewarded by a five cent bonus. The top part of Table 13 shows that when

combining relatively small confidence intervals respondents provided estimates that

were too large, while when combining relatively large confidence intervals respondents

provided estimates that were too small.

In a final exercise, the MTurk respondents were provided with 95% confidence

intervals and asked to guess 90%, 75% and 50% confidence intervals. Again, no

35

Page 36: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

other information was provided, to abstract from all issues except variance neglect,

and a five cent bonus was given for every question answered within a window of

the correct answer. When guessing relatively large confidence intervals like the 90%

confidence interval, respondents provided estimates that were too small, and when

guessing relatively small confidence intervals they provided estimates that were too

large (Table 13). These findings are consistent with each other and with at least

partial misinterpretation of confidence intervals. The results in Table 13 are not as

clear as the results in Table 12 as they could also reflect inattentiveness. However,

these results are based on the subsample that passed at least 15 of 16 attention

checks. Further, respondents are not wholly unresponsive to confidence intervals, as

their responses appear correlated with the correct answers. Rather, they do not place

as much weight on confidence intervals as they should.

Table 13: Interpretation of Variance

Combining Confidence Intervals

Widths Correct Response p-value

2 1.41 2.39 ă0.00014 2.83 2.92 0.336 4.24 3.70 ă0.0001

Calculating Confidence Intervals

Interval Correct Response p-value

50% 2.10 2.55 ă0.000175% 3.58 2.91 ă0.000190% 5.12 3.03 ă0.0001

This table shows how the MTurk sample attempted to combine or calculate confidence intervals inan incentivized task. Attention is restricted to those passing at least 15 of 16 attention checks.

7 Conclusion

How policymakers update their beliefs when presented with new information has

important implications for public policy, and this paper provides some first evidence

using a unique opportunity to run an experiment with policymakers, practitioners

36

Page 37: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

and researchers. Respondents recruited through MTurk served as an additional com-

parison group.

We found that many respondents had seemingly normally distributed priors and

posteriors, but the main parameter governing updating, k, often differed substan-

tially from a Bayesian updater’s k. In particular, we found evidence of two behav-

ioral biases: asymmetric optimism and variance neglect. Policymakers, practitioners,

researchers and MTurk workers did not have significantly different biases. However,

they did have different priors, with policymakers being particularly optimistic but

less certain, while researchers were more certain and the most negative.

A model of quasi-Bayesian updating was built to accommodate the biases. The

model’s estimated parameters were shown to be statistically significantly different

from zero, the value they would take if respondents were Bayesian. We also found

that new information affected allocation of resources, suggesting that if policymakers

were Bayesian updaters they might allocate resources differently. We saw that the

amount of information provided matters, with more information leading to increased

updating. This suggests that in cases in which one has to share bad news, providing

more detailed information may improve policy decisions. Finally, in a supplementary

experiment, we found evidence of individuals’ misinterpreting confidence intervals,

being willing to pay different amounts for confidence intervals that were mathemati-

cally equivalent, and updating differently on them as well. This suggests an intriguing

possibility: that strategic choice of unconventional confidence intervals (such as 90%

or 99% confidence intervals) may encourage better updating.

37

Page 38: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

References

Banuri, S., S. Dercon, and V. Gauri (2015). The Biases of Development Professionals.

In World Development Report 2015: Mind, Society, and Behavior, pp. 179–191. The

World Bank.

Camerer, C., G. Loewenstein, and M. Rabin (2003). Advances in Behavioral Eco-

nomics, Volume Roundtable series in behavioral economics, Chapter Behavioural

Economics - Past, Present & Future. New York: Princeton: Princeton University

Press.

Coville, A. and E. Vivalt (2019). Weighing Evidence:Which Attributes Matter?

Working Paper .

DellaVigna, S. and D. Pope (2017). What Motivates Effort? Evidence and Expert

Forecasts. The Review of Economic Studies 85 (2), 1029–1069.

DellaVigna, S. and D. Pope (2018). Predicting Experimental Results: Who Knows

What? Journal of Political Economy 126 (6), 2410–2456.

Easterly, W. (2002). The cartel of good intentions: The problem of bureaucracy in

foreign aid. The Journal of Policy Reform 5 (4), 223–250.

Eil, D. and J. M. Rao (2011). The Good News-Bad News Effect: Asymmetric Pro-

cessing of Objective Information about Yourself. American Economic Journal:

Microeconomics 3 (2), 114–138.

Gelman, A., J. B. Carlin, H. S. Stern, D. B. Dunson, and A. Vehtari (2013). Bayesian

Data Analysis (3 ed.). Taylor & Francis Ltd.

Hjort, J., D. Moreira, G. Rao, and J. F. Santini (2019). How Research Affects Policy:

Experimental Evidence from 2,150 Brazilian Municipalities. NBER Working Paper .

Irwin, F. W. (1953). Stated Expectations as Functions of Probability and Desirability

of Outcomes. Journal of Personality 21 (3), 329–335.

Kahneman, D. (2003). Maps of Bounded Rationality: Psychology for Behavioral

Economics. American Economic Review 93 (5), 1449–1475.

38

Page 39: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Kahneman, D. and A. Tversky (1979). Prospect Theory: An Analysis of Decision

under Risk. Econometrica 47 (2), 263.

Krueger, A. O. (1993). Political Economy of Policy Reform in Developing Countries

(Ohlin Lectures). The MIT Press.

Langer, E. J. (1975). The Illusion of Control. Journal of Personality and Social

Psychology 32 (2), 311–328.

Liu, X., J. Stoutenborough, and A. Vedlitz (2016). Bureaucratic Expertise, Overcon-

fidence, and Policy Choice. Governance 30 (4), 705–725.

Malmendier, U. and G. Tate (2005). CEO Overconfidence and Corporate Investment.

The Journal of Finance 60 (6), 2661–2700.

Nellis, G., T. Dunning, G. Grossman, M. Humphreys, S. D. Hyde, C. McIntosh, and

C. Reardon (2019). Information, Accountability, and Cumulative Learning, Chapter

Learning about Cumulative Learning:An Experiment with Policy Practitioners.

Cambridge University Press.

Persson, T. and G. Tabellini (2002). Political Economics: Explaining Economic Pol-

icy. The MIT Press.

Rabin, M. and J. L. Schrag (1999). First Impressions Matter: A Model of Confirma-

tory Bias. The Quarterly Journal of Economics 114 (1), 37–82.

Rabin, M. and D. Vayanos (2010). The Gamblers and Hot-Hand Fallacies: Theory

and Applications. Review of Economic Studies 77 (2), 730–778.

Rogger, D. O. and R. Somani (2018). Hierarchy and Information. Policy Research

working paper, World Bank Group.

Stuart, J. O. R., P. D. Windschitl, A. R. Smith, and A. M. Scherer (2015). Be-

having Optimistically: How the (Un)Desirability of an Outcome Can Bias Peoples

Preparations for It. Journal of Behavioral Decision Making 30 (1), 54–69.

Vivalt, E. and A. Coville (2019). The Implications of Variance Neglect for the For-

mation and Estimation of Subjective Expectations. Working Paper .

39

Page 40: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Windschitl, P. D., A. M. Scherer, A. R. Smith, and J. P. Rose (2013). Why so confi-

dent? The Influence of Outcome Desirability on Selective Exposure and Likelihood

Judgment. Organizational Behavior and Human Decision Processes 120 (1), 73–86.

40

Page 41: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

For Online Publication: Appendix

Experimental Details

The following diagrams are excerpted from the survey.

Figure A1: Sample Screening Question

Several simple screening questions were used. After this question, respondents were presented withdata and then asked to provide another estimate.

Figure A2: Types of Information Provided for Information Experiment

Four types of information were provided in the information experiment in the introductory sectionof the survey: historical data was presented without confidence intervals, with confidence intervals,with confidence intervals and the interquartile range, and with confidence intervals, theinterquartile range, and maximum and minimum values.

41

Page 42: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure A3: Understanding Check

Respondents were walked through several examples of how they might distribute weights to

different bins. MTurk respondents were provided with the accompanying written text describing

each picture, while policymakers were provided with this information orally. We ran experiments

showing different distributions in the video and understanding check (normal, uniform, and a mix

of normal and uniform) to mitigate concerns that these distributions biased later responses.

Results are available upon request.

42

Page 43: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure A4: Sample Program Description

Respondents were provided with a short description of a conditional cash transfer program and aschool meals program, then asked to provide their best guess as to the effect of the program.

Figure A5: Assigning Likelihoods

Respondents were then asked to use slider bars to place weights on the probability of differentoutcomes.

43

Page 44: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure A6: Sample New Data

Respondents were then randomly shown data and asked to provide another estimate.

44

Page 45: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Figure A7: Assigning Likelihoods after Viewing New Data

Respondents were also asked to provide their posteriors using slider bars.

Figure A8: Allocation Question

Respondents were asked to allocate funds between three options: conditional cash transferprograms, school meals programs, and further research.

45

Page 46: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 14: Distribution of Prior Mean

PPR MTurkPrior mean Frequency Cumulative Frequency Cumulative

percent percent-5 0 0 0 0-4 0 0 0 0-3 0 0 3 0.2-2 0 0 3 0.4-1 1 0.2 4 0.660 15 3.14 25 2.311 48 12.57 120 10.212 108 33.79 249 26.613 110 55.4 210 40.454 59 66.99 155 50.665 89 84.48 294 70.036 30 90.37 134 78.857 13 92.93 130 87.428 19 96.66 101 94.079 12 99.02 62 98.1610 5 100 28 100

Total 509 1,518

This table provides the distribution of prior means for the policymakers, practitioners, andresearchers sample (PPR) as well as the MTurk sample, for those passing all the screenings andtests. Notably, there are few responses below 0; most responses fall between 0 and 5. There is alsosome evidence of rounding: one of the most popular prior means, for both the PPR sample and theMTurk sample, is 5, with large weights also being placed on 2 and 3.

46

Page 47: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 15: Heterogeneity by Sample

Asymmetric Optimism Variance Neglect

(1) (2) (3) (4) (5) (6)Good News 1.766*** 0.091*** 0.118***

(0.23) (0.03) (0.04)PPR Sample 0.739* -0.132*** -0.130*** 1.121 0.094** 0.041

(0.12) (0.04) (0.04) (0.20) (0.05) (0.05)PPR Sample * 1.193 -0.016 0.055Good News (0.28) (0.07) (0.07)Large C.I. 0.770** -0.121*** -0.156***

(0.10) (0.03) (0.04)PPR Sample * 1.742** 0.003 0.083Large C.I. (0.43) (0.07) (0.07)

Observations 1482 645 881 1374 660 895

This table considers heterogeneity by whether the respondent was in the PPR sample or the

MTurk sample. Columns (1) and (4) present (alternatively) the results of a logistic regression of

Ipk ą 0.5q on whether the respondent received good news or a logistic regression of IpkB ´ k ą 0q

on whether the respondent saw large confidence intervals, using exponentiated coefficients;

Columns (2) and (5) present results of regressions of k and kB ´ k, respectively, on whether the

respondent saw large confidence intervals, restricting attention to those observations for which

0 ď k ď 1; Columns (3) and (6) consider only those observations for which ´0.5 ď k ď 1.5. The

tests for asymmetric optimism and variance neglect were conducted on slightly different samples.

The tests of asymmetric optimism require respondents to have mean priors between 0 and 5, while

the tests for variance neglect require that respondents be randomized into seeing small or large

confidence intervals (as opposed to no confidence intervals).

47

Page 48: How Do Policymakers Update Their Beliefs?evavivalt.com/wp-content/uploads/How-Do-Policymakers-Update1.pdf · How Do Policymakers Update Their Beliefs? Eva Vivalt The Australian National

Table 16: Impact of Presenting Information Differently

(1) (2) (3)µt µt µt

µt´1 0.168*** 0.167*** 0.170***(0.04) (0.04) (0.04)

99% C.I.s 0.861*** 0.855***(0.20) (0.19)

No Yi shown 0.335* 0.320*(0.20) (0.19)

Observations 600 600 600R2 0.07 0.04 0.07

This table shows the impact of randomly providing results with 99% confidence intervals rather

than the equivalent 95% confidence intervals, as well as the impact of providing confidence

intervals without a point estimate (Yi), on the new estimate given pµtq. As estimates of mut were

generally low compared to the estimates a Bayesian would provide, a positive coefficient represents

a positive change in this exercise, though for other data these levers may be deleterious.

48