game theoretical snapshots - hebrew university of jerusalem · game theoretical snapshots sergiu...

247
Game Theoretical Snapshots Sergiu Hart June 2015 SERGIU HART c 2015 – p. 1

Upload: others

Post on 19-Jan-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Game Theoretical Snapshots

Sergiu Hart

June 2015

SERGIU HART c© 2015 – p. 1

Game TheoreticalSnapshots

Sergiu HartCenter for the Study of Rationality

Dept of Mathematics Dept of EconomicsThe Hebrew University of Jerusalem

[email protected]://www.ma.huji.ac.il/hart

SERGIU HART c© 2015 – p. 2

Mathematical Snapshots (1939)

SERGIU HART c© 2015 – p. 3

Caleidoscop Matematic (1961)

SERGIU HART c© 2015 – p. 4

SERGIU HART c© 2015 – p. 5

Snapshot I

SERGIU HART c© 2015 – p. 5

Snapshot I

Game Dynamics

SERGIU HART c© 2015 – p. 5

Game Dynamics

SERGIU HART c© 2015 – p. 6

Game Dynamics

Next week (in the Summer School)

SERGIU HART c© 2015 – p. 6

SERGIU HART c© 2015 – p. 7

Snapshot II

SERGIU HART c© 2015 – p. 7

Snapshot II

Two(!) Good To Be True

SERGIU HART c© 2015 – p. 7

Two(!) Good ...

SERGIU HART c© 2015 – p. 8

Two(!) Good ...

Haven’t heard it yet ?

SERGIU HART c© 2015 – p. 8

Two(!) Good ...

Haven’t heard it yet ?

Two(!) bad ...

SERGIU HART c© 2015 – p. 8

SERGIU HART c© 2015 – p. 9

Snapshot III

SERGIU HART c© 2015 – p. 9

Snapshot III

Blotto, Lotto ...

SERGIU HART c© 2015 – p. 9

Snapshot III

Blotto, Lotto ...... and All-Pay

SERGIU HART c© 2015 – p. 9

Blotto, Lotto, and All-Pay

SERGIU HART c© 2015 – p. 10

Blotto, Lotto, and All-Pay

Sergiu Hart, Discrete Colonel Blotto andGeneral Lotto Games, International Journalof Game Theory 2008www.ma.huji.ac.il/hart/abs/blotto.html

Sergiu Hart, Allocation Games with Caps:From Captain Lotto to All-Pay Auctions,Center for Rationality 2014www.ma.huji.ac.il/hart/abs/lotto.html

Nadav Amir, Uniqueness of OptimalStrategies in Captain Lotto Games, Center forRationality 2015

SERGIU HART c© 2015 – p. 10

Colonel Blotto Games

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

Each urn is CAPTURED by the player who putmore marbles in it

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

Each urn is CAPTURED by the player who putmore marbles in it

The player who captures more urns WINS thegame

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

Each urn is CAPTURED by the player who putmore marbles in it

The player who captures more urns WINS thegame

(two-person zero-sum game:win = 1, lose = −1, tie = 0)

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

Each urn is CAPTURED by the player who putmore marbles in it

The player who captures more urns WINS thegame

Borel 1921SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K distinct urns

Each urn is CAPTURED by the player who putmore marbles in it

The player who captures more urns WINS thegame

Borel 1921. . .

SERGIU HART c© 2015 – p. 11

Colonel Blotto Games

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K indistinguishable urns

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K indistinguishable urns

One urn is selected at random (uniformly)

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

Player A has A aquamarine marbles

Player B has B blue marbles

The players distribute their marblesinto K indistinguishable urns

One urn is selected at random (uniformly)

The player who put more marbles in theselected urn WINS the game

SERGIU HART c© 2015 – p. 12

Colonel Lotto Games

SERGIU HART c© 2015 – p. 13

Colonel Lotto Games

The number of aquamarine marbles in theselected urn is a random variable X ≥ 0 withexpectation a = A/K

SERGIU HART c© 2015 – p. 13

Colonel Lotto Games

The number of aquamarine marbles in theselected urn is a random variable X ≥ 0 withexpectation a = A/K

The number of blue marbles in the selectedurn is a random variable Y ≥ 0 withexpectation b = B/K

SERGIU HART c© 2015 – p. 13

Colonel Lotto Games

The number of aquamarine marbles in theselected urn is a random variable X ≥ 0 withexpectation a = A/K

The number of blue marbles in the selectedurn is a random variable Y ≥ 0 withexpectation b = B/K

Payoff function:

H(X, Y ) = P[X > Y ] − P[X < Y ]

SERGIU HART c© 2015 – p. 13

Colonel Lotto Games

SERGIU HART c© 2015 – p. 14

General Lotto Games

SERGIU HART c© 2015 – p. 14

General Lotto Games

Player A chooses (the distribution of) arandom variable X ≥ 0 with expectation a

Player B chooses (the distribution of) arandom variable Y ≥ 0 with expectation b

SERGIU HART c© 2015 – p. 14

General Lotto Games

Player A chooses (the distribution of) arandom variable X ≥ 0 with expectation a

Player B chooses (the distribution of) arandom variable Y ≥ 0 with expectation b

Payoff function:

H(X, Y ) = P[X > Y ] − P[X < Y ]

SERGIU HART c© 2015 – p. 14

General Lotto Games: Solution

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

VALUE = 0

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

VALUE = 0

The unique OPTIMAL STRATEGY :X∗ = Y ∗ = UNIFORM(0, 2a)

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

VALUE = 0

The unique OPTIMAL STRATEGY :X∗ = Y ∗ = UNIFORM(0, 2a)

Bell & Cover 1980, Myerson 1993, Lizzeri 1999SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

VALUE = 0

The unique OPTIMAL STRATEGY :X∗ = Y ∗ = UNIFORM(0, 2a)

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

Theorem Let a = b > 0.

VALUE = 0

The unique OPTIMAL STRATEGY :X∗ = Y ∗ = UNIFORM(0, 2a)

Proof. Optimality of X∗:

P[Y > X∗] =

∫ 2a

0

P[Y > x]1

2adx

≤1

2aE[Y ] =

1

2aa =

1

2

SERGIU HART c© 2015 – p. 15

General Lotto Games: Solution

SERGIU HART c© 2015 – p. 16

General Lotto Games: Solution

Theorem Let a ≥ b > 0.

SERGIU HART c© 2015 – p. 16

General Lotto Games: Solution

Theorem Let a ≥ b > 0.

VALUE = a − ba = 1 − b

a

SERGIU HART c© 2015 – p. 16

General Lotto Games: Solution

Theorem Let a ≥ b > 0.

VALUE = a − ba = 1 − b

a

The unique OPTIMAL STRATEGY of A :

X∗ = UNIFORM(0, 2a)(

1 − ba

)

SERGIU HART c© 2015 – p. 16

General Lotto Games: Solution

Theorem Let a ≥ b > 0.

VALUE = a − ba = 1 − b

a

The unique OPTIMAL STRATEGY of A :

X∗ = UNIFORM(0, 2a)(

1 − ba

)

The unique OPTIMAL STRATEGY of B :

Y ∗ =(

1 − ba

)

10 + ba UNIFORM(0, 2a)

SERGIU HART c© 2015 – p. 16

General Lotto Games: Solution

Theorem Let a ≥ b > 0.

VALUE = a − ba = 1 − b

a

The unique OPTIMAL STRATEGY of A :

X∗ = UNIFORM(0, 2a)(

1 − ba

)

The unique OPTIMAL STRATEGY of B :

Y ∗ =(

1 − ba

)

10 + ba UNIFORM(0, 2a)

Sahuguet & Persico 2006, Hart 2008SERGIU HART c© 2015 – p. 16

Colonel Blotto Games: Solution

SERGIU HART c© 2015 – p. 17

Colonel Blotto Games: Solution

IMPLEMENT the optimal strategiesof the corresponding General Lotto game

by RANDOM PARTITIONS

SERGIU HART c© 2015 – p. 17

Colonel Blotto Games: Solution

IMPLEMENT the optimal strategiesof the corresponding General Lotto game

by RANDOM PARTITIONS

Hart 2008, Dziubinski 2013SERGIU HART c© 2015 – p. 17

All-Pay Auction

SERGIU HART c© 2015 – p. 18

All-Pay Auction

Object: worth vA to Player Aworth vB to Player B

SERGIU HART c© 2015 – p. 18

All-Pay Auction

Object: worth vA to Player Aworth vB to Player B

Player A bids XPlayer B bids Y

SERGIU HART c© 2015 – p. 18

All-Pay Auction

Object: worth vA to Player Aworth vB to Player B

Player A bids XPlayer B bids Y

Highest bid wins the object

SERGIU HART c© 2015 – p. 18

All-Pay Auction

Object: worth vA to Player Aworth vB to Player B

Player A bids XPlayer B bids Y

Highest bid wins the object

BOTH players pay their bids

SERGIU HART c© 2015 – p. 18

All-Pay: The Expenditure Game

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game⇒ value (and optimal strategies) already solved

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game⇒ value (and optimal strategies) already solved⇒ substitute in [E1]

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game⇒ value (and optimal strategies) already solved⇒ substitute in [E1]⇒ [E1] = simple game on a rectangle

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game⇒ value (and optimal strategies) already solved⇒ substitute in [E1]⇒ [E1] = simple game on a rectangle

→ find pure Nash equilibria of [E1]

SERGIU HART c© 2015 – p. 19

All-Pay: The Expenditure Game

[E1] Each player decides on his EXPECTED BID("expenditure"): a, b

[E2] Given a and b, the players choose X and Yso as to maximize the probability of winning

[E2] = General Lotto game⇒ value (and optimal strategies) already solved⇒ substitute in [E1]⇒ [E1] = simple game on a rectangle

→ find pure Nash equilibria of [E1]

Hart 2014SERGIU HART c© 2015 – p. 19

Colonel Lotto Games

SERGIU HART c© 2015 – p. 20

Colonel Lotto Games

CAPS = upper bounds on X and Y

SERGIU HART c© 2015 – p. 20

Captain Lotto Games

CAPS = upper bounds on X and Y

CAPTAIN LOTTO game = General Lotto gamewith CAPS

SERGIU HART c© 2015 – p. 20

Captain Lotto Games

CAPS = upper bounds on X and Y

CAPTAIN LOTTO game = General Lotto gamewith CAPS

→ All-pay auctions with CAPS

SERGIU HART c© 2015 – p. 20

Captain Lotto Games

CAPS = upper bounds on X and Y

CAPTAIN LOTTO game = General Lotto gamewith CAPS

→ All-pay auctions with CAPS

Hart 2014, Amir 2015

Einav Hart, Avrahami, Kareev, and Todd 2015SERGIU HART c© 2015 – p. 20

SERGIU HART c© 2015 – p. 21

Snapshot IV

SERGIU HART c© 2015 – p. 21

Snapshot IV

Complexity ofCorrelated Equilibria

SERGIU HART c© 2015 – p. 21

Complexity of Correlated Equilibria

SERGIU HART c© 2015 – p. 22

Complexity of Correlated Equilibria

Sergiu Hart and Noam NisanThe Query Complexity of CorrelatedEquilibriaCenter for Rationality 2013www.ma.huji.ac.il/hart/abs/corr-com.html

SERGIU HART c© 2015 – p. 22

Correlated Equilibria

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

Each player has 2 actions

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

Each player has 2 actions

CORRELATED EQUILIBRIUM :

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

Each player has 2 actions

CORRELATED EQUILIBRIUM :

2n unknowns, ≥ 0

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

Each player has 2 actions

CORRELATED EQUILIBRIUM :

2n unknowns, ≥ 0

2n + 1 linear inequalities

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

n-person games

Each player has 2 actions

CORRELATED EQUILIBRIUM :

2n unknowns, ≥ 0

2n + 1 linear inequalities

⇒ There is an algorithm for computingCORRELATED EQUILIBRIA withCOMPLEXITY = POLY(2n) = EXP(n)

SERGIU HART c© 2015 – p. 23

Correlated Equilibria

SERGIU HART c© 2015 – p. 24

Correlated Equilibria

BUT: Regret-Matching and other "no-regret"dynamics yield ǫ-CORRELATED EQUILIBRIA withhigh probability in O(log(n)/ǫ2) steps

SERGIU HART c© 2015 – p. 24

Correlated Equilibria

BUT: Regret-Matching and other "no-regret"dynamics yield ǫ-CORRELATED EQUILIBRIA withhigh probability in O(log(n)/ǫ2) steps

QUERY COMPLEXITY (QC) := maximalnumber of payoff queries (out of n · 2n)

SERGIU HART c© 2015 – p. 24

Correlated Equilibria

BUT: Regret-Matching and other "no-regret"dynamics yield ǫ-CORRELATED EQUILIBRIA withhigh probability in O(log(n)/ǫ2) steps

QUERY COMPLEXITY (QC) := maximalnumber of payoff queries (out of n · 2n)

⇒ There are randomized algorithms forcomputing ǫ-CORRELATED EQUILIBRIA withQC = POLY(n)

SERGIU HART c© 2015 – p. 24

Correlated Equilibria

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?perhaps not so much ...

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?perhaps not so much ...

There are CORRELATED EQUILIBRIA withsupport of size 2n + 1

basic solutions of Linear Programming

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?perhaps not so much ...

There are CORRELATED EQUILIBRIA withsupport of size 2n + 1

basic solutions of Linear Programming

There are ǫ-CORRELATED EQUILIBRIA withsupport of size O(log n/ǫ2)

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?perhaps not so much ...

There are CORRELATED EQUILIBRIA withsupport of size 2n + 1

basic solutions of Linear Programming

There are ǫ-CORRELATED EQUILIBRIA withsupport of size O(log n/ǫ2)

use Lipton and Young 1994

SERGIU HART c© 2015 – p. 25

Correlated Equilibria

Surprise ?perhaps not so much ...

There are CORRELATED EQUILIBRIA withsupport of size 2n + 1

basic solutions of Linear Programming

There are ǫ-CORRELATED EQUILIBRIA withsupport of size O(log n/ǫ2)

use Lipton and Young 1994no-regret dynamics

SERGIU HART c© 2015 – p. 25

Correlated Equilibria (recall)

SERGIU HART c© 2015 – p. 26

Correlated Equilibria (recall)

There are randomized algorithms forcomputing ǫ-CORRELATED EQUILIBRIA withQC = POLY(n)(Regret-Matching, ...)

SERGIU HART c© 2015 – p. 26

Correlated Equilibria

There are randomized algorithms forcomputing ǫ-CORRELATED EQUILIBRIA withQC = POLY(n)(Regret-Matching, ...)

Exact CORRELATED EQUILIBRIA ?

SERGIU HART c© 2015 – p. 26

Correlated Equilibria

There are randomized algorithms forcomputing ǫ-CORRELATED EQUILIBRIA withQC = POLY(n)(Regret-Matching, ...)

Exact CORRELATED EQUILIBRIA ?

Deterministic algorithms ?

SERGIU HART c© 2015 – p. 26

Query Complexity of CE

SERGIU HART c© 2015 – p. 27

Query Complexity of CE

Algorithm

Randomized Deterministic

ε-CE

exact CE

SERGIU HART c© 2015 – p. 27

Query Complexity of CE

Algorithm

Randomized Deterministic

ε-CEPOLY(n)

[1]

exact CE

[1] = Regret-Matching, No Regret

SERGIU HART c© 2015 – p. 27

Query Complexity of CE

Algorithm

Randomized Deterministic

ε-CEPOLY(n)

[1]

exact CEEXP(n)

[2]

[1] = Regret-Matching, No Regret[2] = Babichenko and Barman 2013

SERGIU HART c© 2015 – p. 27

Query Complexity of CE

Algorithm

Randomized Deterministic

ε-CEPOLY(n)

[1]

EXP(n)

[3]

exact CEEXP(n)

[3]

EXP(n)

[2]

[1] = Regret-Matching, No Regret[2] = Babichenko and Barman 2013[3] = this paper

SERGIU HART c© 2015 – p. 27

Complexity of CE

SERGIU HART c© 2015 – p. 28

Complexity of CE

CE is a special linear programming:dual decomposes into n problems

SERGIU HART c© 2015 – p. 28

Complexity of CE

CE is a special linear programming:dual decomposes into n problems"UNCOUPLED"

SERGIU HART c© 2015 – p. 28

Complexity of CE

CE is a special linear programming:dual decomposes into n problems"UNCOUPLED"

Question : Why does this help only forapproximate CE and randomized algorithms ?

SERGIU HART c© 2015 – p. 28

Complexity of CE

CE is a special linear programming:dual decomposes into n problems"UNCOUPLED"

Question : Why does this help only forapproximate CE and randomized algorithms ?

Question : Complexity of Nash Equilibria ?

SERGIU HART c© 2015 – p. 28

SERGIU HART c© 2015 – p. 29

Snapshot V

SERGIU HART c© 2015 – p. 29

Snapshot V

Smooth Calibration

SERGIU HART c© 2015 – p. 29

Snapshot V

Smooth Calibrationand Leaky Forecasts

SERGIU HART c© 2015 – p. 29

Smooth Calibration and Leaky Forecasts

SERGIU HART c© 2015 – p. 30

Smooth Calibration and Leaky Forecasts

Dean Foster and Sergiu HartSmooth Calibration, Leaky Forecasts, andFinite Recall2012 (in preparation)www.ma.huji.ac.il/hart/abs/calib-eq.html

SERGIU HART c© 2015 – p. 30

Calibration

SERGIU HART c© 2015 – p. 31

Calibration

Forecaster says: "The chance of raintomorrow is p"

SERGIU HART c© 2015 – p. 31

Calibration

Forecaster says: "The chance of raintomorrow is p"

Forecaster is CALIBRATED iffor every p: the proportion of rainy daysamong those days when the forecast was pequals p (or is close to p in the long run)

SERGIU HART c© 2015 – p. 31

Calibration

Forecaster says: "The chance of raintomorrow is p"

Forecaster is CALIBRATED iffor every p: the proportion of rainy daysamong those days when the forecast was pequals p (or is close to p in the long run)

Dawid 1982

SERGIU HART c© 2015 – p. 31

Calibration

SERGIU HART c© 2015 – p. 32

Calibration

CALIBRATION can be guaranteed , no matterwhat the weather is, provided that:

SERGIU HART c© 2015 – p. 32

Calibration

CALIBRATION can be guaranteed , no matterwhat the weather is, provided that:

Every day, the rain/no-rain decision isdetermined without knowing the forecast

SERGIU HART c© 2015 – p. 32

Calibration

CALIBRATION can be guaranteed , no matterwhat the weather is, provided that:

Every day, the rain/no-rain decision isdetermined without knowing the forecastForecaster uses mixed forecasting(e.g.: with probability 1/2 forecast = 30%

with probability 1/2 forecast = 60%)

SERGIU HART c© 2015 – p. 32

Calibration

CALIBRATION can be guaranteed , no matterwhat the weather is, provided that:

Every day, the rain/no-rain decision isdetermined without knowing the forecastForecaster uses mixed forecasting(e.g.: with probability 1/2 forecast = 30%

with probability 1/2 forecast = 60%)

Foster and Vohra 1998

SERGIU HART c© 2015 – p. 32

No Calibration

SERGIU HART c© 2015 – p. 33

No Calibration

CALIBRATION cannot be guaranteed when:

SERGIU HART c© 2015 – p. 33

No Calibration

CALIBRATION cannot be guaranteed when:Forecast is known before the rain/no-raindecision is made("LEAKY FORECASTS")

SERGIU HART c© 2015 – p. 33

No Calibration

CALIBRATION cannot be guaranteed when:Forecast is known before the rain/no-raindecision is made("LEAKY FORECASTS")Forecaster uses a deterministicforecasting procedure

SERGIU HART c© 2015 – p. 33

No Calibration

CALIBRATION cannot be guaranteed when:Forecast is known before the rain/no-raindecision is made("LEAKY FORECASTS")Forecaster uses a deterministicforecasting procedure

Oakes 1985

SERGIU HART c© 2015 – p. 33

Smooth Calibration

SERGIU HART c© 2015 – p. 34

Smooth Calibration

SMOOTH CALIBRATION: combine togetherthe days when the forecast was close to p

SERGIU HART c© 2015 – p. 34

Smooth Calibration

SMOOTH CALIBRATION: combine togetherthe days when the forecast was close to p(smooth out the calibration score)

SERGIU HART c© 2015 – p. 34

Smooth Calibration

SMOOTH CALIBRATION: combine togetherthe days when the forecast was close to p(smooth out the calibration score)

Main Result:

There exists a deterministic procedurethat is SMOOTHLY CALIBRATED.

SERGIU HART c© 2015 – p. 34

Smooth Calibration

SMOOTH CALIBRATION: combine togetherthe days when the forecast was close to p(smooth out the calibration score)

Main Result:

There exists a deterministic procedurethat is SMOOTHLY CALIBRATED.

Deterministic ⇒ result holds also whenthe forecasts are leaked

SERGIU HART c© 2015 – p. 34

Calibration

SERGIU HART c© 2015 – p. 35

Calibration

Set of ACTIONS: A ⊂ Rm (finite set)

Set of FORECASTS: C = ∆(A)

Example: A = {0, 1}, C = [0, 1]

SERGIU HART c© 2015 – p. 35

Calibration

Set of ACTIONS: A ⊂ Rm (finite set)

Set of FORECASTS: C = ∆(A)

Example: A = {0, 1}, C = [0, 1]

CALIBRATION SCORE at time T for asequence (at, ct)t=1,2,... in A × C:

SERGIU HART c© 2015 – p. 35

Calibration

Set of ACTIONS: A ⊂ Rm (finite set)

Set of FORECASTS: C = ∆(A)

Example: A = {0, 1}, C = [0, 1]

CALIBRATION SCORE at time T for asequence (at, ct)t=1,2,... in A × C:

KT =1

T

T∑

t=1

||at − ct||

where

at :=

∑Ts=1

1ct=csas

∑Ts=1

1ct=cs

SERGIU HART c© 2015 – p. 35

Smooth Calibration

SERGIU HART c© 2015 – p. 36

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

SERGIU HART c© 2015 – p. 36

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

SERGIU HART c© 2015 – p. 36

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

Example: Λ(x, c) = [δ − ||x − c||]+/δ

SERGIU HART c© 2015 – p. 36

Indicator Function

SERGIU HART c© 2015 – p. 37

Indicator Function

cx

1

1x=c

SERGIU HART c© 2015 – p. 37

Indicator and Λ Functions

cx

1

1x=c

c − δ c + δ

Λ(x, c)

SERGIU HART c© 2015 – p. 37

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

SERGIU HART c© 2015 – p. 38

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

Λ-CALIBRATION SCORE at time T :

SERGIU HART c© 2015 – p. 38

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

Λ-CALIBRATION SCORE at time T :

T =1

T

T∑

t=1

||aΛ

t − cΛ

t ||

SERGIU HART c© 2015 – p. 38

Smooth Calibration

A "smoothing function" is a Lipschitzfunction Λ : C × C → [0, 1] with Λ(c, c) = 1for every c.

Λ(x, c) = "weight" of x relative to c

Λ-CALIBRATION SCORE at time T :

T =1

T

T∑

t=1

||aΛ

t − cΛ

t ||

t =

∑Ts=1

Λ(cs, ct) as∑T

s=1Λ(cs, ct)

, cΛ

t =

∑Ts=1

Λ(cs, ct) cs∑T

s=1Λ(cs, ct)

SERGIU HART c© 2015 – p. 38

The Calibration Game

SERGIU HART c© 2015 – p. 39

The Calibration Game

At each period t = 1, 2, ... :

SERGIU HART c© 2015 – p. 39

The Calibration Game

At each period t = 1, 2, ... :Player C ("forecaster") chooses ct ∈ C

Player A ("action") chooses at ∈ A

SERGIU HART c© 2015 – p. 39

The Calibration Game

At each period t = 1, 2, ... :Player C ("forecaster") chooses ct ∈ C

Player A ("action") chooses at ∈ Aat and ct chosen simultaneously :REGULAR setup

SERGIU HART c© 2015 – p. 39

The Calibration Game

At each period t = 1, 2, ... :Player C ("forecaster") chooses ct ∈ C

Player A ("action") chooses at ∈ Aat and ct chosen simultaneously :REGULAR setupat chosen after ct is disclosed:LEAKY setup

SERGIU HART c© 2015 – p. 39

The Calibration Game

At each period t = 1, 2, ... :Player C ("forecaster") chooses ct ∈ C

Player A ("action") chooses at ∈ Aat and ct chosen simultaneously :REGULAR setupat chosen after ct is disclosed:LEAKY setup

Full monitoring, perfect recall

SERGIU HART c© 2015 – p. 39

Smooth Calibration

SERGIU HART c© 2015 – p. 40

Smooth Calibration

A strategy of Player C is

SERGIU HART c© 2015 – p. 40

Smooth Calibration

A strategy of Player C is

ε-SMOOTHLY CALIBRATED

SERGIU HART c© 2015 – p. 40

Smooth Calibration

A strategy of Player C is

ε-SMOOTHLY CALIBRATED

if there is T0 such that KΛT ≤ ε holds for:

every T ≥ T0,

every strategy of Player A, and

every smoothing function Λ with Lipschitzconstant ≤ 1/ε

SERGIU HART c© 2015 – p. 40

Smooth Calibration: Result

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Result

For every ε > 0

there exists a procedurethat is ε-SMOOTHLY CALIBRATED.

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Result

For every ε > 0

there exists a procedurethat is ε-SMOOTHLY CALIBRATED.

Moreover:• it is deterministic ,

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Result

For every ε > 0

there exists a procedurethat is ε-SMOOTHLY CALIBRATED.

Moreover:• it is deterministic ,• it has finite recall

(= finite window, stationary),

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Result

For every ε > 0

there exists a procedurethat is ε-SMOOTHLY CALIBRATED.

Moreover:• it is deterministic ,• it has finite recall

(= finite window, stationary),• it uses a fixed finite grid , and

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Result

For every ε > 0

there exists a procedurethat is ε-SMOOTHLY CALIBRATED.

Moreover:• it is deterministic ,• it has finite recall

(= finite window, stationary),• it uses a fixed finite grid , and

• forecasts may be leaked .

SERGIU HART c© 2015 – p. 41

Smooth Calibration: Implications

SERGIU HART c© 2015 – p. 42

Smooth Calibration: Implications

For forecasting :

SERGIU HART c© 2015 – p. 42

Smooth Calibration: Implications

For forecasting :nothing good ... (easy to pass the test)

SERGIU HART c© 2015 – p. 42

Smooth Calibration: Implications

For forecasting :nothing good ... (easy to pass the test)

For game dynamics :

SERGIU HART c© 2015 – p. 42

Smooth Calibration: Implications

For forecasting :nothing good ... (easy to pass the test)

For game dynamics :Uncoupled, finite recall, dynamics thatconverge to the set of CORRELATEDEQUILIBRIA

SERGIU HART c© 2015 – p. 42

Smooth Calibration: Implications

For forecasting :nothing good ... (easy to pass the test)

For game dynamics :Uncoupled, finite recall, dynamics thatconverge to the set of CORRELATEDEQUILIBRIA

Uncoupled, finite recall, dynamics that areclose most of the time to NASH EQUILIBRIA

SERGIU HART c© 2015 – p. 42

Previous Work

SERGIU HART c© 2015 – p. 43

Previous Work

Weak Calibration (deterministic):

SERGIU HART c© 2015 – p. 43

Previous Work

Weak Calibration (deterministic):Kakade and Foster 2004 / 2008Foster and Kakade 2006

SERGIU HART c© 2015 – p. 43

Previous Work

Weak Calibration (deterministic):Kakade and Foster 2004 / 2008Foster and Kakade 2006

Online Regression Problem:

SERGIU HART c© 2015 – p. 43

Previous Work

Weak Calibration (deterministic):Kakade and Foster 2004 / 2008Foster and Kakade 2006

Online Regression Problem:Foster 1991, 1999Vovk 2001Azoury and Warmuth 2001Cesa-Bianchi and Lugosi 2006

SERGIU HART c© 2015 – p. 43

SERGIU HART c© 2015 – p. 44

Snapshot VI

SERGIU HART c© 2015 – p. 44

Snapshot VI

Evidence Games:

SERGIU HART c© 2015 – p. 44

Snapshot VI

Evidence Games:Truth and Commitment

SERGIU HART c© 2015 – p. 44

Evidence Games

SERGIU HART c© 2015 – p. 45

Evidence Games

Sergiu Hart, Ilan Kremer, and Motty PerryEvidence Games: Truth and CommitmentCenter for Rationality 2015www.ma.huji.ac.il/hart/abs/st-ne.html

SERGIU HART c© 2015 – p. 45

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

Q: "Are you guilty and deserve punishment?"

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

Q: "Are you guilty and deserve punishment?"A: "Of course not."

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

Q: "Are you guilty and deserve punishment?"A: "Of course not."

How can one obtain reliable information?

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

Q: "Are you guilty and deserve punishment?"A: "Of course not."

How can one obtain reliable information?

How can one determine the "right" reward, orpunishment?

SERGIU HART c© 2015 – p. 46

Q: "Do you deserve a pay raise?"A: "Of course."

Q: "Are you guilty and deserve punishment?"A: "Of course not."

How can one obtain reliable information?

How can one determine the "right" reward, orpunishment?

How can one "separate" and avoid"unraveling" (Akerlof 70)?

SERGIU HART c© 2015 – p. 46

Evidence Games: General Setup

SERGIU HART c© 2015 – p. 47

Evidence Games: General Setup

AGENT who is informed

SERGIU HART c© 2015 – p. 47

Evidence Games: General Setup

AGENT who is informed

PRINCIPAL who takes decision but isuninformed

SERGIU HART c© 2015 – p. 47

Evidence Games: General Setup

AGENT who is informed

PRINCIPAL who takes decision but isuninformed

Agent TRANSMITS information to Principal(costlessly)

SERGIU HART c© 2015 – p. 47

Two Setups

SERGIU HART c© 2015 – p. 48

Two Setups

SETUP 1: Principal decides after receivingAgent’s message

SERGIU HART c© 2015 – p. 48

Two Setups

SETUP 1: Principal decides after receivingAgent’s message

SETUP 2: Principal chooses a policy beforeAgent’s message

SERGIU HART c© 2015 – p. 48

Two Setups

SETUP 1: Principal decides after receivingAgent’s message

SETUP 2: Principal chooses a policy beforeAgent’s message

policy : a function that assigns a decisionof Principal to each message of Agent

SERGIU HART c© 2015 – p. 48

Two Setups

SETUP 1: Principal decides after receivingAgent’s message

SETUP 2: Principal chooses a policy beforeAgent’s message

policy : a function that assigns a decisionof Principal to each message of Agent(Agent knows the policy when sending hismessage)

SERGIU HART c© 2015 – p. 48

Two Setups

SETUP 1: Principal decides after receivingAgent’s message

SETUP 2: Principal chooses a policy beforeAgent’s message

policy : a function that assigns a decisionof Principal to each message of Agent(Agent knows the policy when sending hismessage)Principal is committed to the policy

SERGIU HART c© 2015 – p. 48

Two Setups

GAME: Principal decides after receivingAgent’s message

MECHANISM: Principal chooses a policybefore Agent’s message

policy : a function that assigns a decisionof Principal to each message of Agent(Agent knows the policy when sending hismessage)Principal is committed to the policy

SERGIU HART c© 2015 – p. 48

Main Result

SERGIU HART c© 2015 – p. 49

Main Result

In EVIDENCE GAMES

SERGIU HART c© 2015 – p. 49

Main Result

In EVIDENCE GAMES

the GAME EQUILIBRIUM outcome(obtained without commitment )

SERGIU HART c© 2015 – p. 49

Main Result

In EVIDENCE GAMES

the GAME EQUILIBRIUM outcome(obtained without commitment )

and the OPTIMAL MECHANISM outcome(obtained with commitment )

SERGIU HART c© 2015 – p. 49

Main Result

In EVIDENCE GAMES

the GAME EQUILIBRIUM outcome(obtained without commitment )

and the OPTIMAL MECHANISM outcome(obtained with commitment )

COINCIDE

SERGIU HART c© 2015 – p. 49

Main Result: Equivalence

In EVIDENCE GAMES

the GAME EQUILIBRIUM outcome(obtained without commitment )

and the OPTIMAL MECHANISM outcome(obtained with commitment )

COINCIDE

SERGIU HART c© 2015 – p. 49

Evidence Games

SERGIU HART c© 2015 – p. 50

Evidence Games

AGENT has a value , known to him but not tothe PRINCIPAL

SERGIU HART c© 2015 – p. 50

Evidence Games

AGENT has a value , known to him but not tothe PRINCIPAL

PRINCIPAL decides on the reward

SERGIU HART c© 2015 – p. 50

Evidence Games

AGENT has a value , known to him but not tothe PRINCIPAL

PRINCIPAL decides on the reward

PRINCIPAL wants the reward to be as closeas possible to the value

SERGIU HART c© 2015 – p. 50

Evidence Games

AGENT has a value , known to him but not tothe PRINCIPAL

PRINCIPAL decides on the reward

PRINCIPAL wants the reward to be as closeas possible to the value

AGENT wants the reward to be as high aspossible (regardless of type)

SERGIU HART c© 2015 – p. 50

Evidence Games

AGENT has a value , known to him but not tothe PRINCIPAL

PRINCIPAL decides on the reward

PRINCIPAL wants the reward to be as closeas possible to the value

AGENT wants the reward to be as high aspossible (regardless of type)

↑Differs from signalling, screening,

cheap-talk, ...

SERGIU HART c© 2015 – p. 50

Evidence Games: Truth

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth

Agent reveals:

"the truth, nothing but the truth"

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth

Agent reveals:

"the truth, nothing but the truth"

NOT necessarily "the whole truth"

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth

Agent reveals:

"the truth, nothing but the truth"

all the evidence that the agent reveals mustbe true (it is verifiable)

NOT necessarily "the whole truth"

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth

Agent reveals:

"the truth, nothing but the truth"

all the evidence that the agent reveals mustbe true (it is verifiable)

NOT necessarily "the whole truth"

the agent does not have to reveal all theevidence that he has

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth

Agent reveals:

"the truth, nothing but the truth"

all the evidence that the agent reveals mustbe true (it is verifiable)

NOT necessarily "the whole truth"

the agent does not have to reveal all theevidence that he has

⇒ Agent can pretend to be a type thathas less information (less evidence)

SERGIU HART c© 2015 – p. 51

Evidence Games: Truth-Leaning

SERGIU HART c© 2015 – p. 52

Evidence Games: Truth-Leaning

Revealing the whole truth gets a slight(= infinitesimal) boost in payoff andprobability

SERGIU HART c© 2015 – p. 52

Evidence Games: Truth-Leaning

Revealing the whole truth gets a slight(= infinitesimal) boost in payoff andprobability

(T1) Revealing the whole truth is preferablewhen the reward is the same

SERGIU HART c© 2015 – p. 52

Evidence Games: Truth-Leaning

Revealing the whole truth gets a slight(= infinitesimal) boost in payoff andprobability

(T1) Revealing the whole truth is preferablewhen the reward is the same(lexicographic preference)

SERGIU HART c© 2015 – p. 52

Evidence Games: Truth-Leaning

Revealing the whole truth gets a slight(= infinitesimal) boost in payoff andprobability

(T1) Revealing the whole truth is preferablewhen the reward is the same(lexicographic preference)

(T2) The whole truth is revealed with infinitesimalpositive probability

SERGIU HART c© 2015 – p. 52

Evidence Games: Truth-Leaning

Revealing the whole truth gets a slight(= infinitesimal) boost in payoff andprobability

(T1) Revealing the whole truth is preferablewhen the reward is the same(lexicographic preference)

(T2) The whole truth is revealed with infinitesimalpositive probability(by mistake, or because the agent may benon-strategic, or ... [UK])

SERGIU HART c© 2015 – p. 52

Main Result: Equivalence

SERGIU HART c© 2015 – p. 53

Main Result: Equivalence

In EVIDENCE GAMES

the GAME EQUILIBRIUM outcome(obtained without commitment )

and the OPTIMAL MECHANISM outcome(obtained with commitment )

COINCIDE

SERGIU HART c© 2015 – p. 53

Previous Work

SERGIU HART c© 2015 – p. 54

Previous Work

UnravelingGrossman and O. Hart 1980Grossman 1981Milgrom 1981

SERGIU HART c© 2015 – p. 54

Previous Work

UnravelingGrossman and O. Hart 1980Grossman 1981Milgrom 1981

Voluntary disclosureDye 1985Shin 2003, 2006, ...

SERGIU HART c© 2015 – p. 54

Previous Work

UnravelingGrossman and O. Hart 1980Grossman 1981Milgrom 1981

Voluntary disclosureDye 1985Shin 2003, 2006, ...

MechanismGreen–Laffont 1986

SERGIU HART c© 2015 – p. 54

Previous Work

UnravelingGrossman and O. Hart 1980Grossman 1981Milgrom 1981

Voluntary disclosureDye 1985Shin 2003, 2006, ...

MechanismGreen–Laffont 1986

Persuasion gamesGlazer and Rubinstein 2004, 2006

SERGIU HART c© 2015 – p. 54

Evidence Games: Summary

SERGIU HART c© 2015 – p. 55

Evidence Games: Summary

EVIDENCE GAMES model very commonsetups

SERGIU HART c© 2015 – p. 55

Evidence Games: Summary

EVIDENCE GAMES model very commonsetups

In EVIDENCE GAMES there is equivalencebetween EQUILIBRIUM (without commitment)and OPTIMAL MECHANISM (with commitment)

SERGIU HART c© 2015 – p. 55

Evidence Games: Summary

EVIDENCE GAMES model very commonsetups

In EVIDENCE GAMES there is equivalencebetween EQUILIBRIUM (without commitment)and OPTIMAL MECHANISM (with commitment)

⇒ EQUILIBRIUM is constrained efficient(in the canonical case)

SERGIU HART c© 2015 – p. 55

Evidence Games: Summary

EVIDENCE GAMES model very commonsetups

In EVIDENCE GAMES there is equivalencebetween EQUILIBRIUM (without commitment)and OPTIMAL MECHANISM (with commitment)

⇒ EQUILIBRIUM is constrained efficient(in the canonical case)

The conditions of EVIDENCE GAMES areindispensable for this equivalence

SERGIU HART c© 2015 – p. 55

SERGIU HART c© 2015 – p. 56

c© Mick Stevens SERGIU HART c© 2015 – p. 56

Thanks !!

c© Mick Stevens SERGIU HART c© 2015 – p. 56