online appendix – not for publication - static.cambridge.org... · agent consists of two...

26
Online Appendix – NOT FOR PUBLICATION “Experimentation and Persuasion in Political Organizations” Alexander V. Hirsch. July 16, 2015. This Online Appendix is divided into six parts. Appendix A provides a general statement of strategies and equilibria, and describes how we do equilibrium selection when there are multiple equilibria. Appendix B proves Proposition 3. Appendix C proves verbal statements in the analysis of experimentation about the principal’s commitment problem and institu- tional solutions. Appendix D contains accessory lemmas used in the other proofs. Appendix E treats the variant of the model with diering preferences rather than beliefs discussed in the analysis of experimentation. Appendix F contains supplemental Mathematica code employed in the proofs. 1

Upload: others

Post on 29-Sep-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

Online Appendix – NOT FOR PUBLICATION“Experimentation and Persuasion in Political Organizations”

Alexander V. Hirsch. July 16, 2015.

This Online Appendix is divided into six parts. Appendix A provides a general statement

of strategies and equilibria, and describes how we do equilibrium selection when there are

multiple equilibria. Appendix B proves Proposition 3. Appendix C proves verbal statements

in the analysis of experimentation about the principal’s commitment problem and institu-

tional solutions. Appendix D contains accessory lemmas used in the other proofs. Appendix

E treats the variant of the model with di↵ering preferences rather than beliefs discussed

in the analysis of experimentation. Appendix F contains supplemental Mathematica code

employed in the proofs.

1

Page 2: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

A Equilibrium Characterization and Selection

We begin by introducing additional notation and providing a general equilibrium charac-

terization for the baseline model. This requires allowing mixed strategies for the principal,

which are omitted from the main text for simplicity. As in the main text, a strategy for the

agent consists of two functions e1 (x1) , e2 (x1

, e

1

, y

1

, x

2) to [0, 1] mapping histories to e↵ort.

To allow for mixed strategies for the principal, we now denote the principal’s strategy as a

probability p

1 2 [0, 1] of initially choosing policy a, and a set of probabilities px1

y

1 2 [0, 1] of

sticking with the initial policy x

1 after outcome y

1 for every (x1

, y

1). Throughout we will

also use P

i

(·) to denote probabilities evaluated with respect to the prior of player i – so

P

2

(! = b) = 1 � ✓

2

denotes the agent’s prior belief that b is the correct policy.

The Agent’s Problem In the second period the agent exerts e↵ort e

2 (x1

, e

1

, y

1

, x

2) =

�P

2

(! = x

2 | x1

, e

1

, y

1) and his expected utility is �

2

[P2

(! = x

2 | x1

, e

1

, y

1)]2. For simplicity,

denote his prior that ! = a as ✓, his initial e↵ort e1 as e, and p

a

s

and p

a

f

as ps

and p

f

. Then

it is straightforward to show that his expected two-period utility when x

1 = a as a function

of first period e↵ort is

U

2

(e, ✓, ps

, p

f

) =

✓e � e

2

2�

+�

2(p

f

· ✓ (e+ (1 � e)h (e, ✓)) + (1 � p

f

) · (1 � ✓) (1 � h (e, ✓)))

+ (ps

� p

f

)�

2✓e (A.1)

By symmetry the agent’s expected utility from exerting e↵ort e

1 on an arbitrary policy

x 2 {a, b} is U2

e

1

, P

2

(! = x) , pxs

, p

x

f

. Also note that U2

(·) is distinct from the expression

U (·) for the agent’s objective function as defined in the main Appendix because the latter

amalgamated utilities from the NLB and the GWL, and did not account for mixed strategies.

The Principal’s Problem The principal’s second period policy choices must be interim-

optimal given e

2 (x1

, e

1

, y

1

, x

2) and her posteriors computed with the agent’s equilibrium

2

Page 3: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

strategy. Thus, she must stay with the initial policy x if it succeeds (pxs

= 1), and can only

stay with the initial policy if it fails�

p

x

f

> 0�

i.f.f. h (e1 (x) , P1

(! = x)) � 1�h (e1 (x) , P2

(! = x)).

If the inequality is strict then p

x

f

= 1.

For simplicity denote e

1 as e and p

a

s

and p

a

f

as p

s

and p

f

. In period 1 the principal’s

expected utility from selecting policy a when she expects first period e↵ort e and future

equilibrium behavior is,

U

1

e

1

, ✓

1

, ✓

2

, p

s

, p

f

= ✓

1

(e+ � (eps

+ (1 � e) pf

· h (e, ✓2

)))+(1 � ✓

1

) (1 � p

f

)� (1 � h (e, ✓2

))

(A.2)

By symmetry her expected utility from some x is U1

e

1

, P

1

(! = x) , P2

(! = x) , pxs

, p

x

f

.

Equilibrium Conditions Strategies�

x

1

, p

a

s

, p

a

f

, p

b

s

, p

b

f

and (e1 (x1) , e2 (x1

, e

1

, y

1

, x

2)) are

an equilibrium if and only if they satisfy the following conditions.

(Agent Optimality)

1. e

2 (x1

, e

1

, y

1

, x

2) = �P

2

(! = x

2 | x1

, e

1

, y

1) (the agent optimizes in the second period)

2. e

1 (x) 2 argmaxe

12[0,1]

U

2

e

1

, P

2

(! = x) , pxs

, p

x

f

8x 2 {a, b} (the agent optimizes in the

first period given the principal’s strategy and expectations about his own future e↵ort)

(Principal Optimality)

1. p

x

s

= 1 8x 2 {a, b} (the principal always stays after success)

2. 8x 2 {a, b} p

x

f

� 0 () h (e1 (x) , P1

(! = x)) � 1 � h (e1 (x) , P2

(! = x)) and = 1

if satisfied with strict inequality (the principal only stays after failure if it is interim-

optimal given on-path posteriors)

3. x

1 2 argmaxx2{a,b}

U

1

e

1 (x) , P1

(! = x) , P2

(! = x) , pxs

, p

x

f

(the principal’s initial policy

choice maximizes her expected continuation value).

3

Page 4: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

Equilibrium Selection and Notation Lemma 2 in the Appendix D proves the following

two statements: (i) whenever experimenting is an equilibrium in a subgame x

1, it is the

optimal strategy for the principal even if she could precommit to her future decisions, and

(ii) if experimenting with x

1 = a is not an equilibrium of the subgame following x

1 = a,

then the unique equilibrium is rigid implementation. Together, these statements imply that

we can select the equilibria that are best for the principal by considering only pure strategy

equilibria, and choosing experimentation for sure in the subgame commencing with policy

x

1 2 {a, b} (i.e., px1

s

= 1 and p

x

1

f

= 0) whenever it is an equilibrium.

With this selection and restriction to pure strategies, we now introduce simplified nota-

tion for the agent’s best responses and principal’s utility. First let es (✓2

) denote the agent’s

first-period best response when the principal’s pure strategy is (x1 = a, s), where s 2 {R,E}

denotes whether the principal (R)igidly implements or (E)xperiments with the initial pol-

icy. It is also helpful to state implicit characterizations of eR (✓2

) and e

E (✓2

) so we can

approximate their values in several proofs. The FOC from the proof of Lemma 1 yields that

e

R (✓2

) = �✓

2

1 +�

2k (e, ✓

2

)

and e

E (✓2

) = �✓

2

1 +�

2(k (e, ✓

2

) + 1)

,

where k (e, ✓) = (1 � h (e, ✓))2.

Second, let V

s

1

(✓1

, ✓

2

) denote the principal’s two-period expected utility when her pure

strategy is (x1 = a, s) and the agent best-responds, so V

R

1

(✓1

, ✓

2

) = U

1

e

R (✓2

) , ✓1

, ✓

2

, 1, 1�

and V

E

1

(✓1

, ✓

2

) = U

1

e

E (✓2

) , ✓1

, ✓

2

, 1, 0�

. Third, let ✓ (✓2

) be the unique solution to

h

e

R (✓2

) , ✓ (✓2

)�

= 1 � h

e

R (✓2

) , ✓2

, (A.3)

By the equilibrium characterization, experimenting with x

1 = a is an equilibrium of that

subgame i.f.f. ✓

1

✓ (✓2

), and it is easy verified that ✓ (✓2

) > 1 � ✓

2

. Finally, by symmetry

the agent’s e↵ort on x

1 = b is es (1 � ✓

2

), the principal’s utility is V s

1

(1 � ✓

1

, 1 � ✓

2

), and the

threshold for experimentation with b is 1 � ✓

1

< ✓ (1 � ✓

2

) () ✓

1

> 1 � ✓ (1 � ✓

2

).

4

Page 5: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

B Proof of Proposition 3

In Lemma 3 in the Supplemental Proofs, we prove the following handy property: fixing the

principal’s experimentation decisions down each path of play, if she prefers x

1 = a given

beliefs ✓

1

(in the sense of ex-ante expected utility) then she also prefers it for all higher

beliefs ✓1

> ✓

1

(of course, by symmetry if she prefers x1 = b at beliefs ✓1

then she also prefers

it for all beliefs ✓1

< ✓

1

). We call this property “preference monotonicity” and employ it in

this and the subsequent proofs.

Part 1 We prove that deference expands in GWL as compared to the NLB both in the

baseline model when � � � ⇡ .23505, and when the principal can commit 8�. When � goes

below � and the principal can’t commit, there opens up a very tiny interval of beliefs where

the principal would have deferred in the No Learning Benchmark, but does not in the Game

with Learning. Using Mathematica we find that the size of this interval is maximized when

(� = .123, ✓2

= .545), and at these values is ✓

1

2 (.5, .508). In this interval the principal is

actually best o↵ deferring by experimenting with b, but can’t credibly commit do so because

the agent is working so (unrealistically) little that almost nothing is learned from failure.

Rather than rigidly implement b, she experiments with a to better motivate the agent.

To prove that deference strictly expands when the principal can commit, we argue that the

following property proved in analytically in Lemma 4 in the Supplemental Proofs is su�cient:

a principal with beliefs ✓

1

= 1 � ✓

2

strictly prefers experimenting with b to experimenting

with a. To see this, observe that for ✓1

1� ✓

2

experimenting with a is strictly better than

any other strategy with a, by 1 � ✓

2

< ✓ (✓2

) and Lemma 2. If experimenting with b is also

strictly better than experimenting with a, then it is strictly better than any other strategy

with a, and x

1 = b must be chosen (either experimenting or rigidly implementing). Thus

within the entire deference region from the NLB the principal defers in the GWL. Because

the preference for deference is there strict, the deference region must strictly expand.

To prove that the deference region strictly expands when the principal can’t commit,

5

Page 6: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

observe that the preceding argument proves the principal defers whenever her beliefs ✓1

are

2⇥

1 � ✓ (1 � ✓

2

) , 1 � ✓

2

, because this condition implies that the principal experiments in

the subgame following x

1 = b without commitment. However, to show that the deference

region in the GWL contains the entire deference region in the NLB, we must also show

that the principal will still choose x

1 = b even when ✓

1

2⇥

1

2

, 1 � ✓ (1 � ✓

2

)⇤

and she rigidly

implements in the subgame following x

1 = b. If ✓2

is such that 1

2

� 1 � ✓ (1 � ✓

2

) then

this region is empty. If 1 � ✓ (1 � ✓

2

) >

1

2

, then we require that a principal with beliefs

1

= 1 � ✓ (1 � ✓

2

) weakly prefers rigidly implementing b to experimenting with a. If this

holds then by preference monotonicity a principal with beliefs ✓

1

2⇥

1

2

, 1 � ✓ (1 � ✓

2

)⇤

who

would rigidly implement b if chosen also prefers that to experimenting with a, and so selects

b initially. If it fails, then for some principal beliefs a little bit below 1� ✓ (1 � ✓

2

) < 1� ✓

2

,

the principal will experiment with a in the GWL when she would have deferred in the NLB.

Finally, we prove that a principal with beliefs ✓1

= 1 � ✓ (1 � ✓

2

) weakly prefers rigidly

implementing b to experimenting with a in Lemma 5 in the Supplemental Proofs with the

aid of Mathematica, if and only if � > � ⇡ .23505.

Part 2 We first argue that following three conditions on ✓

2

are jointly su�cient for the

principal to always defer in the first period regardless of her own beliefs; 1) V E

1

(0, 1 � ✓

2

) �

V

R

1

(1, ✓2

), 2) V

R

1

(1, ✓2

) > V

E

1

(1, ✓2

), 3) V

R

1

✓ (1 � ✓

2

) , 1 � ✓

2

> V

E

1

1 � ✓ (1 � ✓

2

) , ✓2

.

Conditions (1) and (2) jointly imply that experimenting with b is better than both ex-

perimenting with or rigidly implementing a when ✓

1

= 1; by preference monotonicity

(proved in Lemma 3 in the Supplemental Proofs) this also implies that experimenting with

b is better 8✓1

2 [0, 1]. Thus, whenever experimenting with b is an equilibrium strategy

(✓1

� 1� ✓ (1 � ✓

2

)) it is chosen. Now whenever experimenting with b is not an equilibrium

strategy�

1

< 1 � ✓ (1 � ✓

2

)�

, the principal compares rigidly implementing b to experiment-

ing with a; again applying preference monotonicity, condition (3) implies that she prefers

the former 8✓1

1 � ✓ (1 � ✓

2

). All possible principal beliefs are covered, which completes

6

Page 7: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

the argument. We next argue that condition (1) is necessary for the principal to always

defer regardless of her own beliefs; if it fails then V

R

1

(1, ✓2

) > V

E

1

(0, 1 � ✓

2

). For ✓1

> ✓ (✓2

)

the principal would rigidly implement a and experiment with b, and by continuity she also

prefers rigidly implementing a to experimenting with b for ✓

1

su�ciently close to 1. Thus

for such ✓

1

she selects x1 = a in equilibrium and does not defer.

Finally, Lemma 6 in the Supplemental Proofs proves analytically that when � > � (where

� is the unique solution to � (1 + �)�

1 + �

2

= 1 and ⇡ .5214), each of the three conditions

k 2 {1, 2, 3} holds for ✓

2

in a nonempty interval (0, "k

).27 Since they then all hold for

2

2 (0, "k

), � > � is therefore su�cient for existence of a range of ✓2

where the principal

always defers. In addition, the supplemental mathematica code verifies that condition 1 fails

8✓2

> 0 when � �, which is equivalent to

V

E

1

(0, 1 � ✓

2

) � V

R

1

(1, ✓2

)

�✓

2

< 0 8✓2

2

0,1

2

when � < �.

� > � is thus also necessary for existence of a range of ✓2

where the principal always defers.

C Underexperimentation and Commitment

In this section we formally state, and then prove, several verbal statements in the analysis

of experimentation about the principal sometimes “underexperimenting,” and institutional

arrangements that can help the principal solve this commitment problem.

The first result pertains the possibility of “underexperimentation” in equilibrium. For-

mally, we say that the principal underexperiments if she rigidly implements a policy x

1 on

the equilibrium path of play, but a strategy of experimenting with some policy x

1 or ¬x1

would yield higher ex-ante higher expected utility. In other words, she underexperiments if

she implements a policy in equilibrium, but would experiment with some policy if she could

27Note that we have already shown that property (3) holds 8✓2

when � > � in Lemma 5;however, the proof is computational. In Lemma 6 we prove analytically that property (3)holds for su�ciently low ✓

2

for any � 2 [0, 1].

7

Page 8: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

commit to her entire two period strategy ex-ante.

Underexperimentation happens in the model because it is possible that a principal with

relatively strong beliefs in favor of a policy would be better o↵ ex-ante committing to exper-

iment with that policy in order to better motivate the agent, but after actually observing

failure she will want to renege on the experiment and persist with the initial policy. The agent

will anticipate this rigidity, work accordingly, experimentation will collapse in equilibrium.

Formally, we have the following result.

Proposition C.1. For some beliefs (✓1

, ✓

2

) the principal underexperiments. Conversely, the

principal never experiments with a policy in equilibrium when rigidly implementing some

policy would yield higher ex-ante expected utility.

The next result considers an institutional arrangement that can eliminate underexperi-

mentation, and in particular that will result in the principal’s “optimal policy experiment”

becoming the equilibrium outcome. The phrase “optimal policy experiment” refers to the

policy experiment that would yield the highest ex-ante expected utility if the principal could

commit to her strategy ex-ante. The result states that creating exogenous “costs to rigidity”

can induce the optimal policy experiment when she underexperiments.

Proposition C.2. For any beliefs (✓1

, ✓

2

) s.t. the principal underexperiments, there is a cost

c of maintaining policy after failure that makes her optimal policy experiment an equilibrium.

The final result pertains to a model variant in which the principal can first “appoint”

a player with di↵erent beliefs ✓

1

2 [0, 1] to make decisions in her place, and that player

and the agent will then play the equilibrium that is best for the “original” principal with

beliefs ✓1

. In particular, we look for conditions under which the principal is strictly better o↵

appointing somebody with beliefs that di↵er from her own. This yields the following result.

Proposition C.3. Suppose that a principal with beliefs ✓1

could appoint a player with beliefs

1

to make policy decisions in her place. If appointing herself is not optimal, then any optimal

appointee ✓

⇤1

believes less strongly in the resulting policy x

⇤1 than the principal does.

8

Page 9: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

The intuition here is that the principal would appoint somebody with di↵erent beliefs

when she would like to commit ex-ante to experiment with some policy x

⇤1, but her beliefs

are such that she must rigidly implement x⇤1 whenever she selects it, and so in equilibrium

she either rigidly implements x

⇤1 or experiments with ¬x⇤1. The optimal appointee will

be anybody whose beliefs allow her to experiment with x

⇤1, and this must necessarily be

somebody who believes less strongly in it.

Proofs of underexperimentation and commitment

Proof of Proposition C.1 First, there is never overexperimentation in equilibrium, i.e.,

the principal never experiments with x

1 when rigidly implementing either x1 or ¬x1 would

be better. The former is immediately ruled out by Lemma 2. The latter also ruled out

– if rigidly implementing ¬x1 were optimal with commitment then it must be better than

experimenting with ¬x1

, and by implication the unique equilibrium of the subgame following

¬x1 without commitment; thus, the principal failing to choose it would be a contradiction.

Next, we there 9 (✓1

, ✓

2

) s.t. underexperimentation occurs, i.e. the principal rigidly im-

plements x1 when experimenting would be better. In part 3 of Lemma 6 in the Supplemental

Proofs we show there exists a nonempty interval of ✓2

s.t. the principal prefers rigidly im-

plementing b to experimenting with a when ✓

1

= 1 � ✓ (1 � ✓

2

). By continuity, rigidly

implementing b is thus the equilibrium outcome for ✓

1

= 1 � ✓ (1 � ✓

2

) � " when " > 0 is

su�ciently close to 0, and it is also worse than experimenting with b since the agent’s e↵ort

drops discretely. Formally,

V

E

1

✓ (1 � ✓

2

) , 1 � ✓

2

= U

1

e

E (1 � ✓

2

) , ✓ (1 � ✓

2

) , 1 � ✓

2

, 1, 0�

= U

1

e

E (1 � ✓

2

) , ✓ (1 � ✓

2

) , 1 � ✓

2

, 1, 1�

> U

1

e

R (1 � ✓

2

) , ✓ (1 � ✓

2

) , 1 � ✓

2

, 1, 1�

= V

R

1

✓ (1 � ✓

2

) , 1 � ✓

2

The first and last equalities follow from the definitions, the second follows from the definition

9

Page 10: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

of ✓ (·), and the inequality from U

1

(·) increasing in e

1 and e

E (1 � ✓

2

) > e

R (1 � ✓

2

). ⌅

Proof of Proposition C.2 If there were an exogenous cost c > 0 of maintaining policy

after failure, then for experimentation to fail to be an equilibrium of the subgame following

policy x requires that the principal prefer to reselect x given initial e↵ort e

E (P2

(! = x)),

the players’ resulting posterior beliefs, and the cost c. This condition is,

h

e

E (P2

(! = x)) , P1

(! = x)�

� c

> 1 � h

e

E (P2

(! = x)) , P2

(! = x)�

Now suppose the principal underexperiments by rigidly implementing policy x

1⇤ = a

when c = 0. Experimenting must then be the unique equilibrium of the subgame following

b, it will remain so with any higher cost c > 0, and also rigidly implementing a is better

than experimenting with b. So experimenting with a must be the optimal policy experiment,

but not an equilibrium of the subgame following a without commitment. However, it will

become one when

c � �

h

e

E (✓2

) , ✓1

��

1 � h

e

E (✓2

) , ✓2

���

,

and so the principal will select it in equilibrium. A symmetric argument holds when the

principal underexperiments by rigidly implementing b. ⌅

Proof of Proposition C.3 Suppose appointing herself is not optimal; then for a principal

with beliefs ✓1

, the resulting equilibrium (x⇤, s

⇤) is strictly worse than the equilibrium that

would result if an optimal appointee with beliefs ✓

1

were making policy decisions. Denote

this equilibrium (x, s). First note s must not an equilibrium experimentation decision the

subgame following x for ✓1

(since otherwise the principal would choose it). We next argue

that s = E. If s = R then by implication experimenting must be an equilibrium of the

subgame following x

1 = x for ✓1

; but then by Lemma 2 it is also strictly better than rigidly

implementing s and we have a contradiction. Finally, since (x, s = E) is an equilibrium of

10

Page 11: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

the subgame following x for ✓1

but not ✓1

, it follows that

h

e

E (P2

(! = x)) , P1

(! = x)⌘

< 1 � h

e

E (P2

(! = x)) , P2

(! = x)�

< h

e

E (P2

(! = x)) , P1

(! = x)�

,

implying P

1

(! = x) < P

1

(! = x). ⌅

D Supplemental Proofs

In this section we prove a sequence of lemmas that are employed in the previous proofs.

Lemma 1. Say that the agent’s problem is “well-behaved” when e

⇤ (✓, ⌘1

, ⌘

2

) = argmaxe2[0,1]

(U (e, ✓, ⌘1

, ⌘

2

))

is unique and 2 (0, 1). The set of � s.t. the agent’s problem is well-behaved 8 (✓ 2 (0, 1) , ⌘1

, ⌘

2

) is

an interval � 2 [0, �), where � 2⇣p

5�1

2

,

p3 � 1

and is ⇡ .68466.

Proof: First, it is simple to verify that the derivative of the agent’s objective function

U (e, ✓, ⌘1

, ⌘

2

) from the proof of Propositions 1 and 2 in the main appendix is:

@U (e, ✓, ⌘1

, ⌘

2

)

@e

=1

�e+ �✓

1 +�

2⌘

1

[k (e, ✓) + ⌘

2

]

◆◆

,

where k (e, ✓) = (1 � h (e, ✓))2.

Second, observe @U

@e

> 0 at e = 0, and that @U

@e

is convex in e, which follows from the

convexity of k (e, ✓). The set of maximizers is thus either a singleton in the interior (the first

and possibly only point where the FOC is satisfied), a singleton on the boundary e = 1, or a

pair where one of the two elements is e = 1. This further implies that whenever the problem

is not “well behaved,” e = 1 is a maximizer.

Third, observe that @

2U

@e@�

= e

2 + ✓

2

1

(k (e, ✓) + ⌘

2

) > 0. Thus by Milgrom and Shannon

(1994) the set of maximizers of U (·) is weakly increasing in �. This implies that the set of �

s.t. the problem is well behaved for a given (⌘1

, ⌘

2

, ✓) is an interval; if it were well behaved

11

Page 12: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

for �0 but not �00< �

0, then e = 1 would be a maximizer for the former but not the latter,

contradicting weak set increasingness. Lastly, this implies that the set of � that are well

behaved for all feasible parameters 8 (✓ 2 (0, 1) , ⌘1

, ⌘

2

) is also an interval [0, �); if it were

not then it would also not an interval for some specific profile of parameters (⌘1

, ⌘

2

, ✓), a

contradiction.

We can bound � below and above analytically, and computationally compute an estimated

value. We must have � >

p5�1

2

since the problem is strictly concave for all feasible parameters

(and by implication well behaved) when @U

@e

e=1

= � 1

+✓

1 + �

2

1

(1 + ⌘

2

)�

� 1

+(1 + �) <

0, which holds i.f.f. � <

p5�1

2

. We also must have � <

p3 � 1 since best-response e↵ort

at ✓ = 1 when the principal experiments is e

⇤ (1, 1, 1) = �

1 + �

2

< 1 i.f.f. � <

p3 � 1.

In the supplemental mathematica code to this document, we verify that � ⇡ .68466. Since

the set of maximizers is weakly increasing in ⌘

2

(by @

2U

@e@⌘2= �✓ > 0), to find � it su�ces to

check that the problem is well-behaved 8✓ when ⌘

2

= 1. We thus identify � by compute the

highest � s.t. 8✓ 2 [0, 1], the agent’s utility at the lowest solution to the first-order condition

is greater than his utility from e = 1. ⌅

Lemma 2. The following two statements hold when the principal can play mixed strategies.

(i) Whenever experimenting is an equilibrium of the subgame following initial policy x

1 2

{a, b}, then it is also the optimal (pure or mixed) strategy for the principal if she could

precommit to her responses to success and failure.

(ii) Whenever experimenting with x

1 = a is not an equilibrium of the subgame following

x

1 2 {a, b}, then the unique equilibrium is rigid implementation.

(Part 1) Because of symmetry we can restrict attention to the subgame following x1 = a.

We first must characterize the agent’s best response e↵ort argmaxe2[0,1]

{U2

(e, ✓2

, p

s

, p

f

)} to a

general mixed strategy by the principal in the Game with Learning as characterized in

12

Page 13: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

Appendix A. Taking the derivative of equation (A.1) w.r.t. e yields:

@U

2

(e, ✓, ps

, p

f

)

@e

=1

�e+ �✓

1 +�

2[k (e, ✓) + (p

s

� p

f

)]

◆◆

.

Now, it is easily verified that@U2(e,✓,ps,pf)

@e

=@U(e,✓,1,ps�pf)

@e

, where U (·) is the form of the

agent’s objective function that we used in the Main Appendix, which amalgamated payo↵s

from the NLB and the GWL, but did not account for principal mixed strategies. This means

we can “piggyback” o↵ of the analysis of that problem in Lemma 1. First, it remains true that

the agent’s problem is well behaved 8 (✓, ps

, p

f

) when � < �. Second, when � < � we have

that (i) argmaxe2[0,1]

{U2

(e, ✓2

, p

s

, p

f

)} = e

⇤ (✓, 1, ps

� p

f

) where e

⇤ (✓, ⌘1

, ⌘

2

) is the maximizer of

U (✓, ⌘1

, ⌘

2

), and (ii) e⇤ (✓, 1, ps

� p

f

) is strictly increasing in p

s

and strictly decreasing in p

f

by @

2U

@e@⌘2= �✓ > 0 and Theorem 1 of Edlin and Shannon (1998). For notational simplicity,

for the remainder of the proof we will write e⇤ (✓, 1, ps

� p

f

) as e (✓, ps

� p

f

) so as not to carry

around unecessary terms.

(Part 2) We now prove (i). Employing the characterization in Appendix A, the prin-

cipal’s utility from choosing x

1 = a if she could precommit to her responses to success and

failure (ps

, p

f

) would be U

1

(e (✓2

, p

s

� p

f

) , ✓1

, ✓

2

, p

s

, p

f

). It is easily verified that U

1

(·) is

strictly increasing in p

s

holding first period e↵ort e fixed, and increasing in e when p

s

= 1.

Also recall that e (✓2

, p

s

� p

f

) is increasing in p

s

. Hence,

U

1

(e (✓2

, p

s

� p

f

) , ✓1

, ✓

2

, p

s

, p

f

) < U

1

(e (✓2

, p

s

� p

f

) , ✓1

, ✓

2

, 1, pf

) < U

1

(e (✓2

, 1 � p

f

) , ✓1

, ✓

2

, 1, pf

)

and the optimal choice after success is therefore to always stay, i.e. p

⇤s

= 1. This feature

is shared with the baseline model without commitment. Intuitively, the reason is that a

higher probability of staying with the initial policy after success is both interim-better for

the principal, and also better motivates the agent ex-ante.

Given the above analysis, the principal’s optimal choice after policy failure satisfies p⇤f

2

13

Page 14: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

argmaxpf2[0,1] {U1

(e (✓2

, 1 � p

f

) , ✓1

, ✓

2

, 1, pf

)}. Now her utility U

1

(·) can be rewritten as,

1

e · (1 + �p

s

)

+ (1 � ✓

1

e) · � (pf

· h (e, ✓1

)h (e, ✓2

) + (1 � p

f

) · (1 � h (e, ✓1

)) (1 � h (e, ✓2

))) ,

and it is easily verified that this is decreasing in p

f

whenever h (e, ✓1

) 1 � h (e, ✓2

), i.e., if

posteriors after failure are s.t. it is better to switch. If experimenting is an equilibrium of

the subgame x

1 = a then by definition this property holds for e = e (✓2

, 1) (see equilibrium

conditions in Appendix A). In addition, recall that U1

(·) is increasing in e when p

s

= 1 and

that e (✓2

, 1 � p

f

) is decreasing in p

f

. Combining these observations yields,

U

1

(e (✓2

, 1) , ✓1

, ✓

2

, 1, 0) > U

1

(e (✓2

, 1) , ✓1

, ✓

2

, 1, pf

) > U

1

(e (✓2

, 1 � p

f

) , ✓1

, ✓

2

, 1, pf

)

whenever experimenting is an equilibrium. Consequently�

p

⇤s

= 1, p⇤f

= 0�

, i.e. experiment-

ing, is the optimal strategy with commitment.

(Part 3) We now prove (ii). If experimentation is not an equilibrium of the subgame fol-

lowing x1 = a, then by the equilibrium characterization in Appendix A, h�

e

2

, 1 � p

a

f

, ✓

1

>

1 � h

e

2

, 1 � p

a

f

, ✓

2

when p

a

f

= 0. Another equilibrium with p

a

f

> 0 would require that

the l.h.s. r.h.s. – but this cannot be since e�

2

, 1 � p

a

f

is decreasing in p

a

f

, so the l.h.s. is

increasing and the r.h.s. is decreasing. ⌅

Lemma 3 (Preference Monotonicity). For all (s, s0) 2 {R,E}2,

V

s

1

1

, ✓

2

> V

s

0

1

1 � ✓

1

, 1 � ✓

2

! V

s

1

(✓1

, ✓

2

) > V

s

0

1

(1 � ✓

1

, 1 � ✓

2

) for all ✓1

> ✓

1

.

In words, fixing the principal’s experimentation decisions down each path of play, if she

prefers x

1 = a given beliefs ✓

1

then she also prefers it for any higher belief.

Proof: Because the principal’s expected utility for each first period policy is linear in her

14

Page 15: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

prior beliefs (holding her future experimentation decisions fixed), a nonmonotonicity would

imply that x1 = b is better when ! = a and x

1 = a is better when ! = b. The former could

not be true if she is rigidly implementing b since it would always fail, and the latter could

not be true if she is rigidly implementing a. Thus, a nonmonotonicity requires that she be

experimenting down both paths of play. To rule it out, it therefore su�ces to show that

experimenting with b is better than experimenting with a when ! = b (given the agent is

predisposed to b, i.e. ✓2

1

2

). This is,

(1 + �) eE (1 � ✓

2

) > �

1 � h

e

E (✓2

) , ✓2

��

Applying the definition from Appendix A, we know e

E (1 � ✓) > � (1 � ✓) ! the l.h.s. is

> (1 + �)� (1 � ✓

2

). Also, e

E (✓2

) < �✓

2

(1 + �) ! the r.h.s. <

�(1�✓2)

1�✓

22�(1+�)

. The above

inequality will thus hold when

(1 + �)� (1 � ✓

2

) >� (1 � ✓

2

)

1 � ✓

2

2

� (1 + �)() (1 + �)2 <

1

2

2

.

The l.h.s. is < the r.h.s. 8✓2

1

2

when � < 1, which always holds by assumption since

� < � < 1.⌅

Lemma 4. The principal strictly prefers experimenting with b to experimenting with a when

1

= 1 � ✓

2

and ✓

2

<

1

2

.

Proof: Let � (✓2

) = �✓

2

1 + �

2

k

e

E (✓2

) , ✓2

+ 1��

(1 + �)��

1 � h

e

E (1 � ✓

2

) , 1 � ✓

2

��

;

this is the principal’s utility di↵erence between experimenting with a and experimenting with

b when ! = a. Applying symmetry, her expected utility di↵erence between experimenting

with a and experimenting with b with prior ✓1

is ✓1

� (✓2

)� (1 � ✓

1

)� (1 � ✓

2

). We now wish

to show that this is < 0 when ✓

1

= 1 � ✓

2

given that ✓2

<

1

2

, i.e.

(1 � ✓

2

)� (✓2

) � ✓

2

� (1 � ✓

2

) < 0 () � (✓2

)

�✓

2

� � (1 � ✓

2

)

� (1 � ✓

2

)< 0 when ✓

2

<

1

2.

15

Page 16: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

It is simple to verify that

� (✓2

)

�✓

2

=

1 +�

2

k

e

E (✓2

) , ✓2

+ 1�

(1 + �) � 1

1 � e

E (1 � ✓

2

) · (1 � ✓

2

),

and by symmetry it su�ces to show that �(✓2)

�✓2� �(1�✓2)

�(1�✓2)> 0when ✓

2

>

1

2

. Using substitution

and rearranging, we then have that �(✓2)

�✓2� �(1�✓2)

�(1�✓2)> 0 ()

2

e

E (✓2

) � (1 � ✓

2

) eE (1 � ✓

2

)

(1 � ✓

2

e

E (✓2

)) · (1 � (1 � ✓

2

) eE (1 � ✓

2

))>

2(1 + �)

k

e

E (1 � ✓

2

) , 1 � ✓

2

� k

e

E (✓2

) , ✓2

��

(D.1)

Since the denominator of the l.h.s. is < 1, and 1 + � < 2, the above inequality holds if the

following yet stronger inequality holds,

2

e

E (✓2

) � (1 � ✓

2

) eE (1 � ✓

2

) > �

k

e

E (1 � ✓

2

) , 1 � ✓

2

� k

e

E (✓2

) , ✓2

��

. (D.2)

Now again substituting in the definition of eE (·), the l.h.s. can be rewritten

(2✓2

� 1)

1 +�

2

+�

2

2

2

k

e

E (✓2

) , ✓2

� (1 � ✓

2

)2 k�

e

E (1 � ✓

2

) , 1 � ✓

2

��

,

implying that the desired inequality holds i.f.f.

(2✓2

� 1)

1 +�

2

>

1 +�

2(1 � ✓

2

)2◆

k

e

E (1 � ✓

2

) , 1 � ✓

2

�✓

1 +�

2✓

2

2

k

e

E (✓2

) , ✓2

It is easily verified that the r.h.s. <

1 + �

2

� �

k

e

E (1 � ✓

2

) , 1 � ✓

2

� k

e

E (✓2

) , ✓2

��

when

2

>

1

2

, implying that the above inequality holds if the following yet stronger inequality

holds:

2✓2

� 1 > k

e

E (1 � ✓

2

) , 1 � ✓

2

� k

e

E (✓2

) , ✓2

(D.3)

We now show that eqn. (D.3) holds. Substituting in the definition of k (·), observing that

16

Page 17: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

2✓2

� 1 = ✓

2

2

� (1 � ✓

2

2

), and rearranging, we have that the inequality is equivalent to

2

2

1 � 1

(1 � (1 � ✓

2

) eE (1 � ✓

2

))2

> (1 � ✓

2

)2✓

1 � 1

(1 � ✓

2

e

E (✓2

))2

Since e

E (✓2

) is increasing in ✓

2

, this clearly holds because (1 � ✓

2

) eE (1 � ✓

2

) < ✓

2

e

E (✓2

)

when ✓

2

>

1

2

. The desired property is hence shown. ⌅

Lemma 5. For all ✓2

s.t. 1� ✓ (1 � ✓

2

) > 1

2

, the principal weakly prefers rigidly implement-

ing b to experimenting with a when ✓

1

= 1 � ✓ (1 � ✓

2

).

Proof: Observe that 1 � ✓ (1 � ✓

2

) is = 1 at ✓

2

= 0, less than 1

2

at ✓

2

= 1

2

, and strictly

decreasing in ✓

2

. Thus, there exists some unique ✓2

satisfying 1� ✓

1 � ✓

2

= 1

2

s.t. the set

of beliefs⇥

1

2

, 1 � ✓ (1 � ✓

2

)⇤

where the principal would rigidly implement b in that subgame

is nonempty if and only if ✓2

< ✓

2

. Note that ✓

2

is a function of � so we henceforth write

2

(�) for clarity. Now applying the definitions and rearranging, we wish to show that

8� 2⇥

�, �

,

✓ (1 � ✓

2

) ·�

V

R

1

(1, 1 � ✓

2

) � V

E

1

(0, ✓2

)�

>

1 � ✓ (1 � ✓

2

)�

· V E

1

(1, ✓2

) 8✓2

2h

0, ✓2

(�)i

,

where � ⇡ .23505. We verify this step in the supplemental mathematical code. ⌅

Lemma 6. The following three properties hold.

1. When � > �, V E

1

(0, 1 � ✓

2

) � V

R

1

(1, ✓2

) > 0 for ✓

2

in a nonempty interval [0, "1

].

2. When � 2 [0, 1] , V R

1

(1, ✓2

) � V

E

1

(1, ✓2

) > 0 for ✓

2

in a nonempty interval [0, "2

].

3. When � 2 [0, 1] , V

R

1

✓ (1 � ✓

2

) , 1 � ✓

2

� V

E

1

1 � ✓ (1 � ✓

2

) , ✓2

> 0 for ✓

2

in a

nonempty interval [0, "3

].

17

Page 18: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

To show that some function f (✓2

) satisfying f (0) = 0 is > 0 for ✓2

2 (0, ") where " > 0,

it su�ces to show by continuity that f(✓2)

✓2

✓2=0

> 0 provided that this quantity is finite. We

now show this for each of the desired expressions.

Property 1: When ! = a, the principal’s utility from experimenting with b is,

1 � h

e

E (1 � ✓

2

) , 1 � ✓

2

��

=�✓

2

1 � (1 � ✓

2

) eE (1 � ✓

2

),

and from rigidly implementing a is,

e

R (✓2

) (1 + �)+�

1 � e

R (✓2

)�

h

e

R (✓2

) , ✓2

= �✓

2

(1 + �)

1 +�

2k

e

R (✓2

) , ✓2

+

1 � e

R (✓2

)�

2

1 � ✓

2

e

R (✓2

)

!

Since e

R (0) = 0, eE (1) = �

1 + �

2

, and k

e

R (0) , 0�

= 1,

1

�✓

2

V

E

1

(0, 1 � ✓

2

) � V

R

1

(1, ✓2

)�

✓2=0

=

1

1 � �

1 + �

2

!

�✓

(1 + �)

1 +�

2

+ 1

.

Manipulating the above expression demonstrates that it is � 0 i.f.f.

� (1 + �)

1 +�

2

> 1. (D.4)

which holds i.f.f. � > � by definition.

Property 2: When ! = a, the principal’s utility from experimenting with a is,

e

E (✓2

) (1 + �) = �✓

2

1 +�

2

k

e

E (✓2

) , ✓2

+ 1�

(1 + �) ,

Since k

e

E (0) , 0�

= 1 and applying Part 1, the desired condition is equivalent to

(1 + �)

1 +�

2

+ 1 > (1 + �)2

which holds if � (1 + �) < 2 () � < 1.

18

Page 19: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

Property 3: Using the definition in equation (A.3), the threshold function ✓ (✓2

) in closed

form is ✓ (✓2

) = 1�✓2

(1�✓2)+✓2(1�e

E(✓2))

2 . Thus, when ✓

1

= 1 � ✓ (1 � ✓

2

) the di↵erence in the

principal’s utility between rigidly implementing b and experimenting with a is,

✓ (1 � ✓

2

) ·�

V

R

1

(1, 1 � ✓

2

) � V

E

1

(0, ✓2

)�

��

1 � ✓ (1 � ✓

2

)�

· V E

1

(1, ✓2

) (D.5)

The first term is the subjective probability that ! = b times the utility di↵erence between

rigidly implementing b and experimenting with a when ! = b. The second term is the

subjective probability that ! = a times the utility of experimenting with a when ! = a (the

payo↵ from rigidly implementing b when the state is a is 0).

Now to get the desired expression we divide through by ✓

2

and then evaluate at ✓2

= 0

– we do so by dividing ✓ (1 � ✓

2

) by ✓

2

in the first term, then V

E

1

(1, ✓2

) by ✓

2

in the second

term, and then evaluating all expressions at ✓2

= 0. First,

1

2

· ✓ (1 � ✓

2

) =1

2

+ (1 � ✓

2

) (1 � e

E (1 � ✓

2

))2

which is equal to�

1 � �

1 + �

2

���2

at ✓2

= 0. Second, V R

1

(1, 1 � ✓

2

) = 2�, V R

1

(0, ✓2

) = �,

and 1 � ✓ (1 � ✓

2

) = 1 at ✓

2

= 0. Third, from the proof of property 2 we know⇣

1

✓2

·

V

E

1

(1, ✓2

) = �

1 + �

2

k

e

E (✓2

) , ✓2

+ 1��

(1 + �) which is = � (1 + �)2 evaluated at ✓2

= 0.

Assembling these observations, the desired expression is

1 � �

1 +�

2

◆◆�2

· (2� � �) � � (1 + �)2 > 0 ()✓

1 � �

1 +�

2

◆◆

2

· (1 + �)2 < 1

Now the l.h.s. is < (1 � �)2 · (1 + �)2 = (1 � �

2)2 1 8� 2 [0, 1], so the result is shown. ⌅

19

Page 20: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

E Preferences vs. Beliefs

Suppose the state ! 2 {a, b} is payo↵-irrelevant for the agent, and he shares the principal’s

prior ✓1

over P (! = a). Instead, he simply has his own return to e↵ort of ⇡x

e when working

on each policy x 2 {a, b}, and ⇡

b

> ⇡

a

.

In the second period, his objective function from exerting second period e↵ort e on some

policy x 2 {a, b} is simply ⇡

x

e � e

2

2�

regardless of the history. It is straightforward to show

that his optimal level of e↵ort is �⇡x

, and his expected payo↵ is �

2

2

x

.

Now consider the first period. Deriving the agents expected payo↵ from exerting first

period e↵ort e when the principal rigidly implement policy a is immediate. His utility when

the principal experiments with a is:

� e

2

2�+ ⇡

a

e+ ✓

1

e

2⇡

2

a

+�

1 � e

1

2⇡

2

b

◆◆

+ (1 � ✓

2

)

2⇡

2

b

,

which easily reduces to the expression in the main text.

Both objective functions are strictly concave, so the unique solution is given by the first-

order condition. The agent’s e↵ort on a given rigid implementation is simply �⇡

a

. If the

principal experiments then the derivative of the objective function is

� e

+ ⇡

a

� �

2✓

1

2

b

� ⇡

2

a

,

so optimal e↵ort is maxn

�⇡

a

� �

2

2

1

(⇡2

b

� ⇡

2

a

) , 0o

and is strictly lower as stated in the text.

Finally, by symmetry the agent’s e↵ort if the principal rigidly implements b is �⇡b

, and

if she experiments with b is minn

�⇡

b

+ �

2

2

(1 � ✓

1

) (⇡2

b

� ⇡

2

a

) , 1o

. The agent therefore works

harder on a policy experiment with the policy b that he prefers; the “threat” of switching to

his less preferred policy a after failure motivates him, as intuition would suggest.

20

Page 21: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

In[25]:= H*SUPPLEMENTAL MATHEMATICA CODE TO"EXPERIMENTATION AND PERSUASION IN POLITICAL ORGANIZATIONS", 02-21-2015*L

H*------------------------------------BEGINCODE----------------------------------*L

In[26]:=

H*----------Threshold Parameters------------------------*LH*-------------------------------------------------------*L

In[27]:= lLbar := .23505lbar := .684660lfig := .653lhat = l ê. Solve@l * H1 + lL * H1 + Hl ê 2LL ã 1, lD@@1DD

Out[30]= -1 +1

3I27 - 3 78 M1ê3 +

I9 + 78 M1ê3

32ê3

In[31]:=

H*----------Posterior Beliefs After Failure-------------*LH*-------------------------------------------------------*Lh@e_, p_D := H1 - eL * p ê H1 - p * eLk@e_, p_D := H1 - h@e, pDL^2

In[33]:=

In[34]:=

H*---------------The Agent's Problem--------------------*LH*-------------------------------------------------------*L

H*objective function*LU2@e_, t2_, pS_, pF_, l_D := Ht2 * e - e^2 ê H2 * lLL +

Hl ê 2L * HpF * t2 * He + H1 - eL * h@e, t2DL + H1 - pFL * H1 - t2L * H1 - h@e, t2DLL +HpS - pFL * Hl ê 2L * t2 * e

H*optima*LeStar@t2_, d_, l_D := Root@HÒ1 - l * t2 * H1 + Hl ê 2L * Hk@Ò1, t2D + dLLL &, 1DeR@t2_, l_D := eStar@t2, 0, lDeE@t2_, l_D := eStar@t2, 1, lD

F Supplemental Mathematica Code

Page 22: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

In[38]:=

H*---------------The Principal's Problem--------------------*LH*-------------------------------------------------------*L

H*Equation 10*LU1@e_, t1_, t2_, pS_, pF_, l_D :=t1 * He + l * He * pS + H1 - eL * pF * h@e, t2DLL + H1 - t1L * H1 - pFL * l * H1 - h@e, t2DL

H*Pure Strategy Payoffs with agent best-responses - from Proposition 3*LV1R@t1_, t2_, l_D := U1@eR@t2, lD, t1, t2, 1, 1, lDV1E@t1_, t2_, l_D := U1@eE@t2, lD, t1, t2, 1, 0, lD

H*Threshold for experimentation for a given e1*L

thetaBar@t2_, e_D :=1 - t2

1 - 2 e * t2 + e2 * t2

H*Indifference Belief Thresholds for Principal*LbEaR@t2_, l_D := t1 ê. Solve@V1E@1 - t1, 1 - t2, lD == V1R@t1, t2, lD, t1D@@1DDbEaE@t2_, l_D := t1 ê. Solve@V1E@1 - t1, 1 - t2, lD == V1E@t1, t2, lD, t1D@@1DDbRaE@t2_, l_D := t1 ê. Solve@V1R@1 - t1, 1 - t2, lD == V1E@t1, t2, lD, t1D@@1DDaEaR@t2_, l_D := If@V1E@1, t2, lD ¥ V1R@1, t2, lD,

1, t1 ê. Solve@V1E@t1, t2, lD == V1R@t1, t2, lD, t1D@@1DDD

H*------------------------------------ENDCODE----------------------------------*L

H*------------------------------------BEGINFIGURES----------------------------------*L

H*-------------PLOTS FOR FIGURE 2------------------*LManipulate@Plot@8l * t2, l * H1 - t2L, eR@t2, lD, eR@1 - t2, lD, eE@t2, lD, eE@1 - t2, lD<,

8t2, 0, .5<, PlotRange Ø 80, 1<D, 8l, lfig, lfig<D

l

0.1 0.2 0.3 0.4 0.5

0.2

0.4

0.6

0.8

1.0

2 Persuasion_SuppCode_20150221.nb

Page 23: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

In[46]:=

H*-------------PLOT FOR FIGURE 3------------------*L

Manipulate@Show@

Plot@aEaR@t2, lD, 8t2, 0, 1<,PlotRange Ø 80, 1<, PlotStyle Ø Directive@Blue, ThickDD,

Plot@bEaR@t2, lD, 8t2, 0, 1<,PlotRange Ø 80, 1<, PlotStyle Ø Directive@Red, ThickDD,

Plot@thetaBar@t2, eE@t2, lDD, 8t2, 0, 1<, PlotRange Ø 80, 1<,PlotStyle Ø Directive@Black, DashedDD,

Plot@81 - t2, .5, 1, 1000 * Ht2 - .5L<, 8t2, 0, 1<, PlotRange Ø 80, 1<,PlotStyle Ø Directive@Gray, DashedDD,

Plot@bEaE@t2, lD, 8t2, 0, 1<, PlotRange Ø 80, 1<,PlotStyle Ø Directive@Orange, ThickDD,

RegionPlot@t1 < 1 - t2, 8t2, 0, 1<, 8t1, 0, 1<, PlotStyle Ø GrayDD, 8l, lfig, lfig<D

Out[46]=

l

H*------------------------------------ENDFIGURES----------------------------------*L

H*------------------------------------BEGINPROOF SUPPLEMENTS----------------------------------*L

Persuasion_SuppCode_20150221.nb 3

Page 24: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

H*--------For Proof of Lemma 1 HGood BehaviorL-----------------*LH*The following plots verify that the agent's utility for the interior

solution exceeds his utility at boundary solution for all t2 \in @0,1D

when \lambda \approx .68466. The first plot illustrates the entire

parameter space. The second plot focuses on the range t2 \in @.92,.96D*L

Manipulate@Plot@8U2@eE@t2, lD, t2, 1, 0, lD - U2@1, t2, 1, 0, lD<,

8t2, 0, 1<, PlotRange Ø 8-.1, .8<D, 8l, .684660, 0<D

Manipulate@Plot@81000 * HU2@eE@t2, lD, t2, 1, 0, lD - U2@1, t2, 1, 0, lDL<,

8t2, .92, .96<, PlotRange Ø 80, .7<D, 8l, .684660, 0<D

l

0.2 0.4 0.6 0.8 1.0

0.2

0.4

0.6

0.8

l

0.93 0.94 0.95 0.96

0.1

0.2

0.3

0.4

0.5

0.6

0.7

4 Persuasion_SuppCode_20150221.nb

Page 25: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

In[47]:=

H*-------------For Proof of Lemma 4 -----------------*L

t2hat@l_D := t2 ê. NSolve@thetaBar@1 - t2, eE@1 - t2, lDD ã 1 ê 2, t2D@@1DD

H*To verify the condition, observe that the blue line is higher

than the purple line for all \theta_2 to the left of \hat8\theta<_2

H\lambdaL for any value of \lambda in @lUbar,lbarD in the slider *LManipulate@Show@

Plot@8

thetaBar@1 - t2, eE@1 - t2, lDD * HV1R@1, 1 - t2, lD - V1E@0, t2, lDL,H1 - thetaBar@1 - t2, eE@1 - t2, lDDL * V1E@1, t2, lD<,

8t2, 0, .5<, PlotStyle Ø 8Blue, Purple<D,

Plot@Evaluate@10 000 * Ht2 - t2hat@lDLD, 8t2, 0, .5<, PlotStyle Ø 8Green<D

D

, 8l, lLbar, lbar<D

Out[48]=

l

0.1 0.2 0.3 0.4 0.5

0.01

0.02

0.03

0.04

0.05

0.06

0.07

Persuasion_SuppCode_20150221.nb 5

Page 26: Online Appendix – NOT FOR PUBLICATION - static.cambridge.org... · agent consists of two functions e1 (x1),e2 (x1,e1,y1,x2)to[0,1] mapping histories to e↵ort. To allow for mixed

H*-------------For Proof of Proposition 3, Part 2 -----------------*LH*To verify the condition,observe that the function is <0 for for all \theta_2 \in @0,1D

and any value of \lambda \in @0,lbarD in the slider *L

Manipulate@Plot@8V1E@0, 1 - t2, lD ê Ht2 * lL - V1R@1, t2, lD ê Ht2 * lL<,8t2, 0, 1<, PlotRange Ø 8-1.5, 0<D, 8l, .001, lhat<D

l

0.2 0.4 0.6 0.8 1.0

-1.4

-1.2

-1.0

-0.8

-0.6

-0.4

-0.2

H*------------------------------------ENDPROOF SUPPLEMENTS----------------------------------*L

6 Persuasion_SuppCode_20150221.nb