finite state mean field games - princeton university€¦ · discrete state space. this thesis...

221
Finite State Mean Field Games Peiqi Wang A Dissertation Presented to the Faculty of Princeton University in Candidacy for the Degree of Doctor of Philosophy Recommended for Acceptance by the Department of Operations Research and Financial Engineering Adviser: Professor Ren´ e Carmona April 2019

Upload: others

Post on 15-Jun-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Finite State Mean Field Games

Peiqi Wang

A Dissertation

Presented to the Faculty

of Princeton University

in Candidacy for the Degree

of Doctor of Philosophy

Recommended for Acceptance

by the Department of

Operations Research and Financial Engineering

Adviser: Professor Rene Carmona

April 2019

Page 2: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

c© Copyright by Peiqi Wang, 2019.

All rights reserved.

Page 3: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Abstract

Mean field game is a powerful framework for studying the strategic interactions within

a large population of rational agents. Although existing research has predominantly

relied on diffusion models to depict agents’ states, numerous applications, such as

epidemic control and botnet defense, can best be modeled by systems of particles in

discrete state space. This thesis tackles finite state mean field games. In the first part

of the thesis, we develop a probabilistic approach for finite state mean field games.

Based on the weak formulation of optimal control, the approach accommodates the

interactions through the players’ strategies and flexible information structures.

The second part of the thesis is devoted to finite state mean field games involving

a player possessing dominating influence. Two different mechanisms are explored. We

first study a form of Stackelberg games, in which the dominating player, referred to

as principal, moves first and chooses its strategy which impacts the dynamics and ob-

jective functions of every remaining player, referred to as agent. Having observed the

principal’s strategy, the agents reach a Nash equilibrium. We seek optimal strategies

of the principal, whose objective function depends on the statistical distribution of

the agents’ states in equilibrium. Using the weak formulation of finite state mean field

games developed previouly in the thesis, we transform the principal’s optimization

problem into a McKean-Vlasov control problem, and provide a semi-explicit solution

under the assumptions of linear transition rate, quadratic cost and risk-neutral utility.

In the second model, we assume that all players move simultaneously and we study

Nash equilibria formed jointly by major and minor players. We introduce finite player

games and derive mean field game formulation in the limit of infinitely many minor

players. In this limit, we characterize the best responses of major and minor players

via viscosity solutions of HJB equations, and we prove existence of Nash equilibria

under reasonable assumptions. We also derive approximate Nash equilibria for the

finite player game from the solution of the mean field game.

iii

Page 4: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Acknowledgements

I am profoundly indebted to my thesis advisor Rene Carmona. I am deeply grateful

that Rene believed in my potential and accepted me as his student. During the three

years I worked with him, I have come to know Rene as a brilliant scholar with deep

and broad knowledge, acute sense of intuition and infinite enthusiasm for inquisition

and creation. Rene taught me a great deal of mathematics, probability and statistics,

but the most invaluable qualities Rene has instilled upon me are perhaps the ability

to identify a meaningful problem and the tenacity of cracking the problem even after

a series of frustrations and failures. His passion, optimism and diligence towards

research are extremely contagious, which was a great source of inspiration during the

course of my thesis research, and will continue to inspire me to be a better researcher.

I would like to express my sincere gratitude to a group of great professors and

researchers both inside and outside Princeton for the advice and support I received

from them: Patrick Cheridito, Miklos Racz, Beatrice Acciaio, Julio Daniel Backhoff

Veraguas, Gaoyue Guo, Ronnie Sircar, Mykhaylo Shkolnikov and Ludovic Tangpi. I

would like to thank my thesis committee, Miklos, Mykhaylo and Ludovic for providing

invaluable advices and comments regarding the material and the presentation of this

thesis. It was a great pleasure working with Patrick during my first and second

years at Princeton. He helped me transition from an undergraduate learner to an

independent researcher and instructor, and I am grateful for his valuable time offered

to me, for the inspiring discussion on research and for the friendly encouragement

and support. My heartfelt appreciation goes to Beatrice, Julio and Gaoyue for their

enthusiastic discussions with me and sharing enlightening and novel ideas regarding

the theory of optimal transport. I would also like to thank them for their warm

hospitality during my visit to LSE and Oxford.

I will miss the time I spent with my officemates and fellow PhD students, Mark

Cerenzia, Xinyuan Fang, Christy Vaughn Graves, Juan Sagredo, Zongjun Tan and

iv

Page 5: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Xiuneng Zhu. I shared the same office with Juan and Xinyuan for two years and the

fact that we all work on seemingly unrelated topics did not prevent us from learning

from each other. As more senior members of graduate students in the department,

they were so generous in sharing their perspective and experience. Christy, Mark,

Xiuneng and Zongjun, on the other hand, are my academic sisters and brothers,

having the same passion for mathematics, probability and game theory as I do. I

thank them for the stimulating discussion, camaraderie and caring they provided.

I wish to thank the amazing group of staffs in the department, Michael Bino,

Connie Brown, Lauren Coleman, Kimberly Lupinacci, Melissa Thomas, Tara Zigler,

who worked so hard to take good care of me and provided generous help for research

and teaching duties. My special thanks go to Tabitha Mischler for her warm and

smiling friendship.

I would not have completed the thesis if it were not for the unconditional love

and support from my wonderful parents. They always believe in me, want the best

of me and stand ready to advice and comfort in time of need. So, thank you, Mom

and Dad, for being the most loving and supportive family one could every wish for.

v

Page 6: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

To my parents, Hailan Liu and Shunxing Wang

vi

Page 7: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Contents

Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii

Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv

1 Introduction 1

1.1 Past Work on Mean Field Games . . . . . . . . . . . . . . . . . . . . 2

1.2 Mean Field Games in Finite State Space . . . . . . . . . . . . . . . . 7

1.3 Organization of the Thesis . . . . . . . . . . . . . . . . . . . . . . . . 12

1.4 Notations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

2 Preliminaries on Finite State Mean Field Games 15

2.1 Controlled Continuous-Time Markov Chain . . . . . . . . . . . . . . . 15

2.1.1 Definition via Infinitesimal Generator . . . . . . . . . . . . . . 18

2.1.2 A Semimartingale Representation . . . . . . . . . . . . . . . . 21

2.2 Optimal Control of Continuous-Time Markov Chains . . . . . . . . . 27

2.2.1 Analytical Approach via the Dynamic Programming Principle 27

2.2.2 Probabilistic Approach by BSDE . . . . . . . . . . . . . . . . 30

2.3 Analytical Approach to Finite State Mean Field Games . . . . . . . . 36

2.3.1 Model Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.3.2 Existence and Uniqueness of Nash Equilibria . . . . . . . . . . 39

2.3.3 Game With Finitely Many Players . . . . . . . . . . . . . . . 44

2.3.4 Approximate Nash Equilibria . . . . . . . . . . . . . . . . . . 49

vii

Page 8: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

2.4 Appendix: BSDE Driven by Single Continuous-Time Markov Chain . 51

3 Probabilistic Analysis of Finite State Mean Field Games 54

3.1 Weak Formulation of Finite State Mean Field Games . . . . . . . . . 58

3.1.1 Mean Field Games with Interaction through States and Controls 58

3.1.2 Representative Player’s Optimization Problem . . . . . . . . . 62

3.2 Existence of Nash Equilibria . . . . . . . . . . . . . . . . . . . . . . . 66

3.2.1 Topology for the Mean Field . . . . . . . . . . . . . . . . . . . 67

3.2.2 Mapping for the Fixed Point Argument . . . . . . . . . . . . . 76

3.2.3 Proving the Existence of Nash Equilibria . . . . . . . . . . . . 79

3.3 Uniqueness of the Nash Equilibrium . . . . . . . . . . . . . . . . . . . 89

3.4 Approximate Nash Equilibria for Games with Finite Many Players . . 92

3.5 Appendix: BSDE Driven by Multiple Independent Continuous-Time

Markov Chains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100

4 Finite State Principal Agent Problem 108

4.1 Model Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112

4.1.1 Agent’s State Process . . . . . . . . . . . . . . . . . . . . . . . 112

4.1.2 Principal’s Contract and Agent’s Objective Function . . . . . 114

4.1.3 Principal’s Objective Function . . . . . . . . . . . . . . . . . . 114

4.1.4 Agents’ Mean Field Nash Equilibria . . . . . . . . . . . . . . . 115

4.1.5 Principal’s Optimal Contracting Problem . . . . . . . . . . . . 115

4.2 A McKean-Vlasov Control Problem . . . . . . . . . . . . . . . . . . . 116

4.2.1 Representation of Nash Equilibria Based on BSDEs . . . . . . 116

4.2.2 Principal’s Optimal Contracting Problem . . . . . . . . . . . . 120

4.3 Solving the Linear-Quadratic Model . . . . . . . . . . . . . . . . . . . 121

4.3.1 Model Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122

4.3.2 Reduction to Optimal Control on Flows of Probability Measures 123

viii

Page 9: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.3.3 Construction of the Optimal Contract . . . . . . . . . . . . . 128

4.4 Application to Epidemic Containment . . . . . . . . . . . . . . . . . 130

4.5 Appendix: System of ODEs for Epidemic Containment . . . . . . . . 137

4.5.1 Authority’s optimal planning . . . . . . . . . . . . . . . . . . 137

4.5.2 Mean field equilibrium in the absence of the authority . . . . . 139

5 Finite State Mean Field Game with Major and Minor Players 141

5.1 Game Model with Finitely Many Players . . . . . . . . . . . . . . . . 145

5.2 Mean Field Game Formulation . . . . . . . . . . . . . . . . . . . . . . 150

5.3 Convergence of Large Finite Player Games . . . . . . . . . . . . . . . 153

5.3.1 Major Player’s Problem with Finitely Many Minor Players . . 154

5.3.2 Deviating Minor Player’s Problem . . . . . . . . . . . . . . . . 161

5.4 Optimization Problem for Individual Players . . . . . . . . . . . . . . 163

5.5 Existence of Nash Equilibria . . . . . . . . . . . . . . . . . . . . . . . 169

5.6 The Master Equation for Nash Equilibria . . . . . . . . . . . . . . . . 173

5.7 Propagation of Chaos and Approximate Nash Equilibria . . . . . . . 175

5.7.1 Back to the (N + 1)-Player Game . . . . . . . . . . . . . . . . 175

5.7.2 Propagation of Chaos . . . . . . . . . . . . . . . . . . . . . . . 179

5.7.3 Approximate Nash Equilibria . . . . . . . . . . . . . . . . . . 186

5.8 Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186

5.8.1 Proof of Lemma 5.4.3 . . . . . . . . . . . . . . . . . . . . . . . 186

5.8.2 Proof of Theorem 5.4.4 . . . . . . . . . . . . . . . . . . . . . . 189

5.8.3 Proof of Lemma 5.5.8 . . . . . . . . . . . . . . . . . . . . . . . 194

5.8.4 Proof of Propositions 5.7.3 & 5.7.4 . . . . . . . . . . . . . . . 198

6 Conclusion 203

Bibliography 206

ix

Page 10: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 1

Introduction

The effort to depict and analyze a large system of particles dates back at least to

the era of ancient Greek philosophy, when Leucippus and Democritus, among other

theorists, invented the notion of atom. They posited that the origin of everything is

rooted in the interaction of indivisible bodies, as these atoms with simple intrinsic

properties like size and shape collide with one another and interlock in an infinite

void. Accordingly, properties of macroscopic bodies perceivable by man is the ag-

gregation of these same atomic interactions. Over the course of many centuries that

follows, philosophers and physicists would see these infinitesimal particles with their

own eyes (invention of microscope), uncover the fundamental rules of interactions be-

tween two individual particles (classical and quantum mechanics) and gain empirical

insights into the macroscopic effect of countless interacting particles (thermodynam-

ics and universal gravity). However, how to reconcile the microscopic behavior of an

individual particle with the macroscopic phenomenon of a large ensemble remained

obscure.

It was not until the advent of statistical mechanics and mean field theory that the

gap could finally be bridged. The groundbreaking idea is to consider the limit scenario

of infinitely many identical particles and study the statistical distribution of their

1

Page 11: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

states. The aggregation of the effects exerted on any given particle by all the others

can be approximated by a single averaged effect as a function of the distribution. In

turn, the evolution of the distribution over time can be deduced from the dynamics

governing an individual particle’s state. The potential of this innovative paradigm

quickly emerged beyond the realm of physics, as people soon realized the possibility

to deduce the macroscopic behavior of any large system of particles, provided that the

microscopic behaviors of individual particle is known. Indeed, by the term particle,

we might refer to substantial particles such as atoms or molecules, but we might also

refer to certain particles of more abstract forms: a bird in a migrating flock, a trader

in the stock market, a user of a social network or even a neuron in the human brain.

However, unlike substantial particles among which the interactions obey fundamental

principles of classical or quantum mechanics, there is no single canonical method to

depict the dynamics for these abstract particles.

Fortunately, game theory in modern mathematics and economics provides the

missing piece of the puzzle by endowing these imaginary particles with rationality.

Each particle is assumed to optimize an objective, which depends on its own action

and state as well as the actions and states of all the other particles. Consequently,

the dynamics of the particles are endogenously determined by each particle’s strategic

choice on the one hand and a certain type of equilibrium formed among the particles

on the other hand. With the rules of mechanics in physics replaced by the rules of

equilibrium in game theory, classical statistical mechanics readily evolved into the

theory of Mean Field Games as we know today.

1.1 Past Work on Mean Field Games

The theory of Mean Field Games was initiated by Lasry and Lions [2007], in an

attempt to analyze the seemingly intractable behavior of a large population of rational

2

Page 12: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

agents with mutual impact in the context of stochastic differential games. At about

the same time, the same notion of mean field game was independently introduced

by Huang et al. [2006] in the community of electrical engineering under the name

of Nash Certainty Equivalence. Since then, the research on mean field game have

attracted interests from multiple disciplines and rapidly advanced on both theoretical

and applicational fronts.

Tremendous efforts have been made on building the theoretical foundation of mean

field games. Among them, a voluminous body of the literature is concerned with the

fundamental question of existence and uniqueness of the Nash equilibrium. Named

after the Nobel laureate John Forbes Nash Jr. who invented it, Nash equilibrium

is perhaps one of the most fundamental notion of equilibrium in non-cooperative

games. It describes a situation where no player can benefit by deviating from its

current strategy while the others keep theirs unchanged. In the trailblazing work

Nash [1950], the authors showed that any game with a finite number of players in

which each player can choose from a finite set of pure strategies has at least one

mixed-strategy Nash equilibrium.

Starting from the pioneering work of Lasry and Lions [2007] and Huang et al.

[2006], these questions are first investigated via analytical methods, where the Nash

equilibrium is characterized by a system of partial differential equations (PDEs) com-

posed of a Hamilton-Jacobi-Bellman (HJB) equation coupled with a Fokker-Planck

equation for the dynamics of the mean field. Weak solutions to this system of PDEs

were studied in Lasry and Lions [2006b] with finite-horizon setup and in Lasry and

Lions [2006a] with stationary setup, whereas the ergodic setup was considered in

Cardaliaguet et al. [2012] as the limit of long-time average of mean field game. Later

Gomes et al. [2015], Gomes et al. [2016] and Gomes and Pimentel [2015] investigated

the existence and regularity of classical solutions under a variety of assumptions on

the nonlinearity. Inspired by Lions [2007], Porretta [2014] introduced the optimal

3

Page 13: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

planning problem by modifying the boundary condition of the mean field game PDEs

and Achdou et al. [2012] studied the related numerical problem through the penaliza-

tion method. An explicitly solvable instance, namely the linear-quadratic mean field

game, was examined in Bensoussan et al. [2016b].

The probabilistic framework of mean field games was initially developed in Car-

mona and Delarue [2013], where the authors used the Pontryagin maximum principal

to treat the player’s optimization problem. The Nash equilibrium can then be de-

picted by a system of McKean-Vlasov forward-backward stochastic differential equa-

tions (FBSDEs), for which topological arguments on spaces of measures were used

to show the existence of solutions. A more general form of McKean-Vlasov FBSDE

was closely studied in Carmona et al. [2015], as it can also be applied to the optimal

control problem of McKean-Vlasov type. The connection between mean field games

and McKean-Vlasov control problems was further explored in Carmona et al. [2013a].

A different probabilistic approach based on the weak formulation of optimal control

was proposed in Carmona and Lacker [2015] and was probed in a more general setup

in Lacker [2015].

Recently there has been renewed interest in the master equation of the mean field

game. First introduced in Lions [2007] as a PDE involving the space of probability

measures, the master equation emerges as the limit of the coupled system of HJB

equations which describes the Nash equilibrium of a symmetric N -player stochastic

differential game. In Cardaliaguet et al. [2015], the authors revealed the connection

between the master equation and the mean field game PDEs, and the existence of

a classical solution to the master equation was established. Relying on the classical

solution of the master equation, the authors also showed the convergence of the Nash

equilibrium of the N -player game to the Nash equilibrium of the mean field game as N

tends to infinity. The well-posedness of the master equation was also investigated in

Chassagneux et al. [2014]. From the probabilistic prospective, Carmona and Delarue

4

Page 14: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

[2014] showed that the master equation is the decoupling field of the McKean-Vlasov

FBSDE which describes the Nash equilibrium of the mean field game. This interpre-

tation was also exploited in Chassagneux et al. [2017] to develop a numerical scheme

for the master equation. In Bensoussan et al. [2015], the master equation was inter-

preted as the value function of McKean-Vlasov control problems, and recently Pham

and Wei [2017] studied the underlying dynamic programming principle which leads

to the master equation.

Meanwhile, inspired by practical applications, contributions have also been made

on a plethora of meaningful extensions. Carmona et al. [2016] studied the existence

and uniqueness of Nash equilibrium in the presence of common noise by introducing

the notion of weak Nash equilibrium of mean field games. Under the assumptions

of linear dynamics, quadratic running cost and convex terminal cost, Ahuja [2016]

showed a unique Nash equilibrium exists in the strong sense. An explicit solution

to a mean field game with common noise was obtained in Carmona et al. [2013b].

The effect of a dominant player’s presence in the mean field game was first studied

in Huang [2010] in the case of linear quadratic model via the PDE approach. Later,

Carmona and Zhu [2016] performed probabilistic analysis on mean field games with

major and minor players. The authors highlighted the fact that even in the limit of

infinite number of minor players, the major player’s influence does not average out and

behaves as a common noise. They further illustrated this by establishing the result of

conditional propagation of chaos for the minor players. In Carmona and Wang [2017],

a different scheme of fixed point based on strategies was proposed to solve mean field

games with major and minor players. They applied the scheme to the linear quadratic

model to derive semi-explicit solutions for both open-loop and Markov controls and

illustrated the conditional propagation of chaos through a flocking model. Meanwhile,

inspired by applications in bank run models, Carmona et al. [2017] examined mean

field interaction in the context of an optimal stopping game. The authors showed

5

Page 15: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

equilibria exist for pure stopping strategies under the assumption of complementarity

on cost functionals, while equilibria exist for randomized stopping time under a more

general setting. A different mean field game of stopping was formulated in Nutz [2016]

where the author derived the explicit solution for equilibrium strategy.

While mean field games have been mostly formulated in continuous state spaces,

some works considered the cases involving state processes in discrete space. Gomes

et al. [2013] applied the analytical framework to study finite-state mean field games

and established existence and uniqueness of the Nash equilibrium. Doncel et al.

[2016] looked into a similar model with a finite set of actions. Later Cecchin and

Fischer [2017] gave a probabilistic formulation based on stochastic differential equa-

tions (SDEs) driven by Poisson random measure and investigated existence of Nash

equilibria with relaxed open-loop and feedback controls. Most recently, the master

equation for finite state mean field game was investigated in Bayraktar and Cohen

[2017], where the authors showed the regularity of the solution and established con-

vergence of the value function of the N -player game towards the solution of the master

equation.

All of the theoretical efforts mentioned above have been greatly inspired and

invigorated by the ever-expanding applications of the mean field game theory in

biology, engineering, economics and finance, among other fields. Early illustrations of

mean field Nash equilibrium include the famous pedagogical models ‘Towels on Beach‘

and ‘When do the meeting start¿ in Lions [2007]. In the engineering community, mean

field game proved to be a versatile framework in modeling power grid management

Meriaux et al. [2013], Kizilkale and Malhame [2013], Ma et al. [2012], cellular networks

Aziz and Caines [2017], crowd congestion control Lachapelle and Wolfram [2011] and

botnet defense Kolokoltsov and Bensoussan [2016]. Among the many applications of

mean field game in economic theory, Lasry et al. [2008] studied the growth theory,

Chan and Sircar [2015] examined price formation by Bertrand and Cournot equilibria,

6

Page 16: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Shen and Turinici [2012] models the trading volume of a given asset by traders’

heterogeneous beliefs and costly estimations, and Carmona et al. [2013b] looked into

the systemic risk of banking system. Achdou et al. [2014] provided a number of

examples of PDEs that naturally arise in macroeconomics theory and Lachapelle

et al. [2010] tackled the numerical methods of mean field equilibria in economics.

Meanwhile in mathematical biology, Nourian et al. [2011] used mean field game theory

in providing an alternative approach to Cucker and Smale Model of flocking. Recently

Carmona and Graves [2018] explained the mechanism of jet lag by a mean field game

model on synchronization of circadian rhythm among neurons.

Finally, let us mention that interested readers can refer to Lions [2007] and Car-

mona and Delarue [2017] for an exhaustive presentation of the mean field game theory

from the analytical perspective and probabilistic perspective respectively.

1.2 Mean Field Games in Finite State Space

Although existing applications of mean field games have been predominantly focused

on diffusion-based model, there is a plethora of real life examples of interacting particle

systems in which the states of the particles belong to a discrete space. Epidemic

models are perhaps the most representative examples of this kind. From the modeling

perspective, the mechanism of an epidemic is extremely generic and versatile to fit

in a variety of scenarios such as cyber security in Kolokoltsov and Bensoussan [2016]

and opinion formation in Stella et al. [2013]. In its most simplest form, each particle

is either infected or not infected, and the dynamics of states are modeled as a Markov

chain, of which the transition rate is assumed to be a function of the fraction of

infected population. In a more realistic version of the model, a graph can be used

to depict the location of each particle. Accordingly, the transition rate is assumed

7

Page 17: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

to include the effect of only the neighboring particles, which allows to study how

infection spreads in the dimension of space.

Since we have just introduced the idea of a particle system on a graph, we would

like to mention the model of congestion studied in Gueant [2015]. The author con-

sidered a continuum of identical players distributed on a directed graph consisting

of N nodes. Naturally, the state of each player at any given moment is the node

where the player is located. Each player’s movement on the graph is modeled as a

continuous-time Markov chain. Each player can strategically choose the transition

rate from one node to another wherever the passage is allowed according to the edge

of the graph, and it pays a cost to make such choice. Each player then tries to max-

imizes its earning which consisted of a instantaneous and a terminal reward derived

from the location of the player. In order to model the effect of the congestion, the

reward is assumed to be a decreasing function of the fraction of players located at a

given node. Under some structural assumptions on the reward functions, the author

shows a unique equilibrium of spatial distribution exists for this congestion model.

Another interesting example we would like to curate here is the model of informa-

tion percolation proposed in Duffie et al. [2009]. Let us imagine a large population of

agents who wish to uncover some information of common concern. Each agent is en-

dowed with a signal regarding the information, which is characterized by its precision

Nt, assumed to be an integer. For a certain period of time, agents randomly meet in

pairs to share their signal. When two agents meet, the precisions of both agents’ sig-

nals after the meeting becomes the sum of the precisions before the meeting. During

the process, each agent chooses a search effort and faces a tradeoff between reaching

the maximal precision at the end of the game and reducing the cost incurred by its

effort. The key part of the model consists of the matching mechanism of the agents.

The authors assumed that an agent exerting a search effort α1 has an intensity of

α1α2q(α2) of being matched to some agent from the set of agents currently using

8

Page 18: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

effort level α2 at time t, where q(α2) is the proportion of the agents using effort α2.

This can be interpreted as the complementarity of search and matching technology,

which means that the more effort that an agent makes to be found, the more effective

are the efforts of his counterpart to find him.

Although the model was not explicitly formulated as a mean field game with

discrete state space in Duffie et al. [2009], we believe that it can be cast in such a

framework. Indeed, we may consider the precision of the signal Nt as the agent’s

state, which belongs to the discrete space of positive integers. When we assume that

the agent’s search effort is a function of time and the current precision of its signal

α(t, n), it can be shown that Nt evolves as a continuous-time Markov chain. Indeed,

according to the search and matching mechanism described above, an agent with

current precision n has an intensity α(t, n)α(t,m)µ(t,m) of being matched to another

agent with current precisionm, given that µ(t,m) is the current fraction of agents with

precision m, and α is the search effort adopted by all the other agents. This means

that the transition rate from state m to state m + n equals α(t, n)α(t,m)µ(t,m). It

is worth noticing that the agent feels the presence of the other agents through the

effort-weighted measure µα(n) := α(t,m)µ(t,m), instead of the distribution of the

states µ itself. Fortunately, this is the only singularity of the model compared to

other conventional mean field game models. Once the effort-weighted measure µα

is regarded as the mean field and fixed, the optimization problem of the agent is

straightforward to solve. The Nash equilibrium can be obtained by matching µα with

the effort-weighted measure resulting from the optimal search effort.

Inspired by the myriad of existing and potential applications, this thesis explores

several theoretical aspects of mean field games in finite state space. In the first

part of the thesis, we propose a probabilistic framework to analyze finite state mean

field game. Compared with the existing analytical approach in Gomes et al. [2013],

this new approach allows us not only to incorporate the interaction through agents’

9

Page 19: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

actions into the mean field game, but also to obtain equilibria among a broader class

of strategies. In the second part of the thesis, we study the mean field game in

which a player with dominating influence on the rest of the players is present. As the

terminology we shall use repeatedly in the thesis, we call this dominating player the

major player, while we refer to all the remaining players as the minor players. Our

quest is motivated by a list of fundamental questions prevailing in the theory of mean

field game:

(1) Formulation of games with finitely many players. What are the different

types of equilibria for a game involving a major player? What are the information

structure and admissible strategies for each player? Accordingly what is the proper

way to model the dynamics of the states and the interactions among the players?

(2) Passage to the mean field limit. When the number of minor players goes to

infinity, does the empirical distribution of the minor players’ states converge and in

what sense does it converge? What about the convergence of the state processes of

the players? How does the dominating player influence the dynamics of the entire

system in the limit? Can we formulate a mean field game based on the limit?

(3) Equilibrium. Does the equilibrium exist and is it unique? Does the finite nature

of the state space facilitate the numerical computation of the equilibria? What is the

master equation characterizing the equilibrium?

(4) Approximation. How well does the equilibrium of the mean field limit approx-

imate the equilibrium of the game with finitely many players?

When we investigate these problems in the context of the dynamic games in finite

state space, we quickly realize that the technical tools developed for stochastic differ-

ential mean field games can no longer be applied here. Let us mention some of the

concerns specific to the games in finite state space. The first subtlety lies in the choice

of proper information structure for the players. Recall that in the case of stochastic

differential games, we might want to consider open-loop strategies (each player can

10

Page 20: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

observe the past history of noise processes of all the other players), Markov strategies

(each player can observe the current states of all the other players), or distributed

strategies (each player can only observe its own noise process or state). Despite dif-

ferent information structures to be considered, the choice does not interfere with how

we describe the state processes of the agents. Indeed, the admissible control corre-

sponding to each of these information structures can be plugged in the controlled

SDE and produce a well-defined state process. Unfortunately, it is not the case for

the games in finite state space. As we shall see in Chapter 2, the most tractable way

to describe the dynamics in a finite state space is to rely on continuous-time Markov

chains and specify the transition rates between the states. This means that much

of the time Markovian strategy will be the only viable choice granting a well-defined

state process. Moreover, if we choose the Markovian information structure, we still

need to decide what state variables should be made available to the players, which

boils down to identifying the so-called sufficient statistics of the states of every players

in the game.

The second difficulty lies in how we can model the independence of the state

processes of different players. The assumption of independence regarding the noise

processes of different players is essential in any mean field game model, since it guar-

antees the the convergence of the empirical distribution as the number of players tends

to infinity. In diffusion-based mean field games, this requirement can be intuitively

fulfilled by assigning independent Wiener processes to the dynamics of the players.

However in the case of continuous-time Markov chains, it is more delicate to translate

independence into the transition rates between the states.

Due to the delicacies mentioned above, it is not immediately clear what the correct

formulation of mean field game with major and minor player should be. Therefore

we begin by formulating the game with finite number of players, and then study how

the states of each player evolve in the limit of infinity many players in the game.

11

Page 21: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

More specifically, we need to show the limiting state processes for both the major

and minor player are well-defined. Notice that this is very different from the more

commonly used top-down procedure to study mean field games, in which we first

formulate the mean field game based on the intuitions of the propagation of chaos

and then reverse-engineer the formulation of the game with finite many players.

1.3 Organization of the Thesis

We start with some preliminary results on finite state mean field games in Chapter 2.

Since the optimization problem is the centerpiece in analyzing the strategic behaviors

in any type of game, in Section 2.2 we review the optimal control of continuous-

time Markov chains based on the dynamic programming principal, and then intro-

duce a new probabilistic approach based on backward stochastic differential equation

(BSDE). When implemented in finite state mean field games, the first approach nat-

urally leads to the analytical formulation developed in Gomes et al. [2013]. In Section

2.3, we will give a detailed review of some important results of this approach, includ-

ing existence and uniqueness of the Nash equilibrium for the mean field game and

the convergence of the N -player game to the mean field game. In particular, the way

how we formulate the N -player game will be generalized to the game with major and

minor players in Chapter 5.

In Chapter 3, we present the weak formulation of finite state mean field games.

The probabilistic approach to the optimal control problem introduced in Chapter 2

serves as the backbone of this formulation, and its flexibility allows us to incorporate

the interaction through the strategies of players into the game, which is otherwise

intractable using the analytical approach. We give the definition of Nash equilibria

in the weak formulation in Section 3.1 and study the existence and uniqueness of

the equilibrium in Section 3.2 and Section 3.3. Then in Section 3.4, we formulate

12

Page 22: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the finite-player counterpart of the game. We show that when being assigned the

distributed strategies derived from the Nash equilibrium of the mean field game, the

players reach an approximative Nash equilibrium in the game with N players.

In the next two chapters, we study the scenario where a strong influential player

is present in the mean field game. We envisage two types of equilibria. In Chapter 4,

we study a type of Stackelberg equilibrium formulated as the so-called principal agent

problem. In term of the terminology, in order to be consistent with the existing litera-

tures dealing with similar models, we refer to the player with dominating influence as

the principal, while we call the remaining players as the agent. After introducing the

model in Section 4.1, we show in Section 4.2 that the principal’s optimal contracting

problem is equivalent to an optimal control problem of McKean-Vlasov SDE. This

results relies on the weak formulation of finite state mean field games developed in

Chapter 3. In Section 4.3, we treat the special case of the linear-quadratic model

where we further simplify the optimal contracting problem into the deterministic

control on the distribution of agents’ states. We illustrate the linear-quadratic model

with an example of epidemic containment in Section 4.4.

In Chapter 5, we study an entirely different game model with major and minor

players. Compared to the principal agent model studied in Chapter 4, we endow the

major player with not only an objective function, but also a state process and we aim

to study the Nash equilibrium formed by the major player and the minor players. We

start by formulating the finite player game in Section 5.1, in which we pay special

attention to the information structure of each player and the dynamics of the major

and minor player’s state process. Then in Section 5.3 we rigorously derive the limit

of the state processes of both major and minor players. Combined with a fixed-point

scheme for mean field games with major and minor player based on the strategies of

the players which we describe in Section 5.2, we formulate the mean field game. In

Section 5.4, we deal with the optimization problem of the major and minor player and

13

Page 23: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

derive their optimal response, respectively. Section 5.5 is devoted to the existence of

Nash equilibria and we derive the master equation in Section 5.6. In Section 5.7 we

investigate how the solution of the mean field game can provide an approximative

Nash equilibrium for the finite player game.

We conclude the thesis in Chapter 6 by discussing a few unsolved problems emerg-

ing during our investigation of finite state mean field games, which, we hope, will be

helpful in pointing out new directions for future research on this subject.

The thesis is built on material from four papers by the author and his thesis

adivsor, some of which are under review or revision at the time this thesis is written.

Chapter 2 and Chapter 3 interpolate material from Carmona and Wang [2018b].

Meanwhile, Chapter 4 is based on Carmona and Wang [2018a]. Finally, Chapter 5 is

based on Carmona and Wang [2016], and uses ideas from Carmona and Wang [2017].

1.4 Notations

We introduce some notations that will be used throughout the rest of the thesis. For

a real matrix M , we denote its transpose by M∗, and we denote by M+ the Moore-

Penrose pseudo inverse of M . The multiplication operator is denoted by ·, and is

always understood as matrix multiplications if the operands are vectors or matrices

of proper dimensions. For a random variable γ in a probability space (Ω,F ,P), we

denote by P#γ := P γ−1 the push-forward measure of P by γ. We denote by 1(·)

the indicator function. In some of the proofs, we use C to denote a positive constant

that might change its value from line to line.

Finally, considering that the players or agents in the game models we are about

to present are sometimes abstract objects, and in order to avoid the confusion of

the genders when referring to the players or agents, we decide to make the players

genderless and use the pronouns it/its throughout this thesis.

14

Page 24: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 2

Preliminaries on Finite State Mean

Field Games

In this chapter, we start with some basic facts about continuous-time Markov chain

and present two different approaches to the related stochastic control problem: one

based on dynamic programming principal and the other based on a probabilistic

representation by backward stochastic differential equation (BSDE). The analytical

approach naturally leads to the existing framework of finite state mean field game

proposed by Gomes et al. [2013], for which we shall give a detailed account at the

end of this chapter. The probabilistic approach, on the other hand, will provide us

with the necessary tools to develop the weak formulation of the finite state mean field

game in Chapter 3.

2.1 Controlled Continuous-Time Markov Chain

In preparation for finite state mean field games, in this section, we introduce the

controlled continuous-time Markov chain. This is a stochastic process taking values

in a discrete space, of which the dynamics, controlled by external actions, can be

15

Page 25: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

characterized in different fashions depending on the admissible set of controls, as well

as the techniques used to tackle the optimal control problem.

Throughout the thesis, we denote by E the state space of the process and we

assume that E is finite. The elements in E are denoted by 1, 2, . . . ,m, however we

sometimes identify these states with the m basis vectors e1, e2, . . . , em of Rm. We fix

a finite time horizon T . Let us first recall the definition of continuous-time Markov

chains.

Definition 2.1.1. A stochastic process (Xt)t∈[0,T ] in a probability space (Ω,F ,P) is

called a continuous-time Markov chain, if for any set of time 0 ≤ t0 < t1 < · · · <

tn ≤ T and any set of states i0, i1, . . . , in ∈ E, we have:

P[Xtn = in|Xtn−1 = in−1, . . . , Xt0 = i0] = P[Xtn = in|Xtn−1 = in−1].

For any continuous-time Markov chain X, we can associate it with a probability

transition matrix P (s, t). For s, t ∈ [0, T ] we define the component of P (s, t) on i−th

row and j−th column by Pij(s, t) := P[Xt = j|Xs = i]. Combined with the Markov

property described in Definition 2.1.1, the probability transition matrix allows us to

compute the marginal law of the stochastic process X. However, a better way to

describe the law of a continuous-time Markov chain is to look at the derivative of the

probability transition matrix, which is called the transition rate matrix. For t ∈ [0, T ],

the element of the transition rate matrix Q(t) is defined as:

Qii(t) := limh→0

1

h(P[Xt+h = i|Xt = i]− 1) = lim

h→0

1

h(Pii(t, t+ h)− 1), (2.1)

Qij(t) := limh→0

1

hP[Xt+h = j|Xt = i] = lim

h→0

1

hPij(t, t+ h). (2.2)

16

Page 26: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

It can be shown that the transition rate matrix Q(t) of any continuous-time Markov

chain satisfies several properties. We summarize these properties in the following

definition, and we call any matrix satisfying these properties the Q-matrix.

Definition 2.1.2. We say a real-valued matrix Q = [Qij]i,j∈E is a Q-matrix if Qij ≥ 0

for i 6= j and ∑j 6=i,j∈E

Qij = −Qii, for all i ∈ E.

Conversely, given a measurable mapping [0, T ] 3 t → Q(t) such that Q(t) is a

Q-matrix for all t ∈ [0, T ], we can construct the corresponding probability transition

matrix P (s, t) satisfying (2.1) and (2.2). Moreover, we can construct a probability

space and a stochastic process X such that X is a continuous-time Markov chain with

P (s, t) as its probability transition matrix.

Theorem 2.1.3. Let µ be a probability measure on E and [0, T ] 3 t → Q(t) be a

measurable mapping such that Q(t) is a Q-matrix for all t ∈ [0, T ] and the components

Qij(t) are bounded uniformly in i, j ∈ E and t ∈ [0, T ]. Let p be a probability

distribution on the state space E. Then there exists a probability space (Ω,F ,P) and

a continuous-time Markov chain X such that the distribution of X0 is p and we have

for all t ∈ [0, T ]:

Qii(t) = limh→0

1

h(P[Xt+h = i|Xt = i]− 1),

Qij(t) = limh→0

1

hP[Xt+h = j|Xt = i].

Proof. From Section 8.9.2 in Iosifescu [2014], we see that there exists a mapping

[0, T ]2 3 (s, t)→ P (s, t) solving the forward and backward Kolmogorov equations:

P (s, t) =Im +

∫ t

s

Q(u) · P (u, t)du,

P (s, t) =Im +

∫ t

s

P (s, u) ·Q(u)du,

17

Page 27: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where Im is the identity matrix of size m. In addition, for all (s, t) ∈ [0, T ]2, all the

components of P (s, t) are positive and∑

j∈E Pij(s, t) = 1 for all i ∈ E.

Now for each k ∈ N, and finite sequence of distinct times t1, . . . , tk ∈ [0, T ], we

define νt1,...,tk the probability measure on Ek as follows:

νt1,...,tk(F1 × · · · × Fk)

:=∑i0∈E

∑i1∈Fσ(1)

· · ·∑

ik∈Fσ(k)

pi0Pi0i1(0, tσ(1))Pi1i2(tσ(1), tσ(2)) . . . Pik−1ik(tσ(k−1), tσ(k)).(2.3)

Here σ is the permutation which puts t1, . . . , tk in increasing order, i.e. tσ(1) < · · · <

tσ(k). It is plain to check that the family of measures νt1,...,tk is consistent. We

can therefore apply Kolmogorov extension theorem, which allow us to construct a

probability space (Ω,F ,P) and a stochastic process X such that νt1,...,tk(F1 × · · · ×

Fk) = P[Xt1 ∈ F1, . . . , Xtk ∈ Fk] for all k ∈ N, t1, . . . , tk ∈ [0, T ] and F1, . . . , Fk ⊂ E.

Now from the definition of the measure νt1,...,tk , we easily check that X is a continuous-

time Markov chain with probability transition matrix P (·, ·) (see Definition 2.1.1), and

the distribution of X0 is p. Finally, from the Kolmogorov equation we verify that Q

is the transition rate function of X.

From Theorem 2.1.3, we see that in order to describe the continuous-time Markov

chain, we only need to specify its transition rate function and its initial distribution.

This provides a convenient way to model controlled continuous-time Markov chain,

since we only need to specify how the transition rate function can be controlled.

2.1.1 Definition via Infinitesimal Generator

Recall the case of controlled diffusion process, in which the dynamics of the process

are completely determined by a drift function b : (t, x, α) → b(t, x, α), a volatility

function σ : (t, x, α)→ σ(t, x, α), and finally, a control process t→ αt which satisfies

a certain information structure. This is a pathwise characterization of the dynamics,

18

Page 28: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

as we consider the strong solution to the following stochastic differential equation

(SDE):

dYt = b(t, Yt, αt)dt+ σ(t, Yt, αt)dWt.

However, another interesting perspective is to regard b and σ as the local mean and

standard deviation of the increment of the process, during an infinitesimally small

interval of time. This amounts to defining the infinitesimal generator G of a Markov

process which characterizes how the probability distribution of the process - starting

from a certain point in space-time - will evolve onward. Due to the Markovian nature

of the process, we need to restrict the admissible control to be Markovian control,

which is of the form αt = φ(t, Yt). Formally, we define G as an operator on the space

of functions:

GF (t, y) := limh→0

1

hE[F (t+ h, Yt+h)− F (t, Yt)|Yt = y].

Under suitable condition of regularity on F and by Ito’s formula we obtain:

GF (t, y) = ∂tF (t, y) + b(t, y, φ(t, y))∂yF (t, y) +1

2σ2(t, y, φ(t, y))∂2

yyF (t, y).

When it comes to the controlled continuous-time Markov chain, it is still possible to

define the dynamics of the process in a pathwise fashion by using the jump process.

However, it is more intuitive to directly specify the infinitesimal generator of the

process, which grants an interpretation in terms of the jump rate and the probability

of transitions between the states. Let A be a compact subset of Rl in which we pick

the control. We introduce a function q:

[0, T ]× E2 × A→ R

(t, i, j, α)→ q(t, i, j, α),

19

Page 29: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

and for later use, we denote by Q(t, α) the matrix [q(t, i, j, α)]1≤i,j≤m. Throughout

this chapter, we make the following assumptions on q:

Assumption 2.1.4. For all (t, α) ∈ [0, T ]× A, the matrix Q(t, α) is a Q-matrix.

Assumption 2.1.5. There exists C1, C2 > 0 such that for all (t, i, j, α) ∈ [0, T ] ×

E2 × A such that i 6= j, we have 0 < C1 < q(t, i, j, α) < C2.

We consider a Markovian control α of the form αt = φ(t,Xt), where φ : [0, T ] ×

E → A is a measurable mapping. φ is called the feedback function associated with the

Markovian control α. As we have seen in Theorem 2.1.3, we can define the dynamics

of the process X by specifying the transition probability in an infinitesimal interval

of time h, or the transition rate function:

P[Xt+h = j|Xt = i, Xu, 0 ≤ u < t] =q(t, i, j, φ(t, i))h+ o(h), if j 6= i (2.4)

P[Xt+h = i|Xt = i, Xu, 0 ≤ u < t] =1−∑j 6=i

q(t, i, j, φ(t, i))h+ o(h) (2.5)

This is equivalent to defining the following infinitesimal generator Gφ, acting on the

space of mappings F : [0, T ]× E → R which are C1 in t:

GφF (t, i) := ∂tF (t, i) +∑j 6=i

(F (t, j)− F (t, i))q(t, i, j, φ(t, i)). (2.6)

A more intuitive way to interpret the dynamics of X described by Gφ is to con-

sider the hazard rate of the jumps. More specifically, conditioned on the past path

of X and given that Xt = i, the event that X jumps from i to j has a hazard

rate q(t, i, j, α(t, i)). If we assume that q and φ does not depend on t, i.e. the dy-

namics are time-homogeneous, then conditioned on the past path of X and given

that Xt = i, the time until the next jump follows an exponential distribution of

parameter∑

j 6=i q(i, j, α(i)) and the probability of jumping from i to j is given by

[q(i, j, α(i))/∑

j 6=i q(t, i, j, α(i))].

20

Page 30: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

2.1.2 A Semimartingale Representation

We now present an alternative formalism of finite state continuous-time Markov chain

based on semimartingales, which was first introduced in Elliott et al. [1995] and later

developed in Cohen and Elliott [2008] and Cohen and Elliott [2010]. We shall see

that by an argument of change of measure, such a representation can accommodate

not only the controlled continuous-time Markov chains, but also a larger class of

controlled point processes.

Consider X = (Xt)t∈[0,T ] a continuous-time Markov chain with m states. We

identify these states with the unit vector ei in Rm and therefore we denote E :=

e1, . . . , em. We assume that the sample path t→ Xt is cadlag, i.e., right continuous

and admits left limit. In addition, we assume that X is continuous at T .

We now construct a canonical probability space for X. Let Ω be the space of

cadlag functions from [0, T ] to E which are continuous at T and let X be the canonical

process. We denote by Ft := σ(Xs, s ≤ t) the natural filtration generated by X

and F := FT . Now let us fix a probability measure p on E. On (Ω, (Ft)t∈[0,T ], F)

we consider the probability measure P under which X is a continuous-time Markov

chain of which the law of X0 is p and the transition rate between any two different

states is 1. This means that for i, j ∈ 1, . . . ,m, i 6= j and ∆t > 0, we have

P[Xt+∆t = ej|Ft] = P[Xt+∆t = ej|Xt] and P[Xt+∆t = ej|Xt = ei] = ∆t + o(∆t). By

Appendix B in Elliott et al. [1995], the process X has the following representation:

Xt = X0 +

∫(0,t]

Q0 ·Xt−dt+Mt. (2.7)

Here and in the rest of this chapter, Q0 is the m by m square matrix with diagonal

elements all equal to −(m− 1) and off-diagonal elements all equal to 1. M is a Rm-

valued P-martingale. The multiplication · is understood as matrix multiplication.

21

Page 31: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Indeed, Q0 is the transition rate matrix of X under the probability measure P. We

shall refer to the probability measure P as the reference measure of the sample space.

Remark 2.1.6. The representation originally proposed in Elliott et al. [1995] is:

Xt = X0 +

∫(0,t]

Q0 ·Xtdt+Mt.

However, since we know that almost surely Xt is only discontinuous on a countable

set, we can replace Xt by Xt− in the integral. The reason for this slight change of

representation is to make the integrand a predictable process, which will be suitable

for the change of measure argument in the following.

Let us recall the definitions of predictable quadratic (co)variation 〈·, ·〉 and

quadratic (co)variation [·, ·]. See Chapter II.6 in Protter [2005].

Definition 2.1.7. Let (Ω, F ,P) be a probability space supporting a filtration F :=

(Ft)t≥0 and a square integrable F-martingale X. Then by Doob-Meyer decomposition

theorem, there exists a unique increasing predictable process (At)t≥0 such that A0 = 0

and the process t → (X2t − At) is a martingale. The process (At)t≥0 is called the

predictable quadratic variation of X, which we denote by 〈X,X〉t := At. In addition,

if X, Y are two square integrable F-martingales, the predictable quadratic covaration

of X and Y , denoted by (〈X, Y 〉t)t≥0 is defined by:

〈X, Y 〉t :=1

4

(〈X + Y,X + Y 〉t − 〈X − Y,X − Y 〉t

)(2.8)

Definition 2.1.8. Let X, Y be semimartingales. The quadratic variation of X, de-

noted by ([X,X]t)t≥0, is defined by:

[X,X]t := X2t − 2

∫ t

0

Xs−dXs (2.9)

22

Page 32: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

The quadratic covariation of X and Y , denoted by ([X, Y ]t)t≥0, is defined by:

[X, Y ]t :=1

4

([X + Y,X + Y ]t − [X − Y,X − Y ]t

)(2.10)

The predictable quadratic variation of the martingale M, which appears in the

canonical decomposition of the continuous-time Markov chain in equation (2.7), is

given in the following lemma. The proof can be found in Cohen and Elliott [2008]:

Lemma 2.1.9. The predictable quadratic variation of the martingale M under P is:

〈M,M〉t =

∫ t

0

ψsds, (2.11)

where the process (ψt)t∈[0,T ] is given by:

ψt := diag(Q0 ·Xt−)−Q0 · diag(Xt−)− diag(Xt−) ·Q0. (2.12)

Let us define the matrix ψi := diag(Q0ei)−Q0diag(ei)−diag(ei)Q0. Clearly we have

ψt =∑m

i=1 1(Xt− = ei)ψi. Since each ψi is a semi-definite positive matrix, so is ψt.

We define the corresponding (stochastic) seminorm ‖ · ‖Xt− on Rm by:

‖Z‖2Xt− := Z∗ · ψt · Z. (2.13)

The seminorm ‖ · ‖Xt− can be rewritten in a more explicit way. For i ∈ 1, . . . ,m,

let us define the seminorm ‖ · ‖ei on Rm by :

‖Z‖2ei

:= Z∗ · ψi · Z =∑j 6=i

|Zj − Zi|2. (2.14)

Then it is easy to verify that ‖Z‖Xt− =∑m

i=1 1(Xt− = i)‖Z‖ei .

23

Page 33: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

For later use, we also denote by ψ+t the Moore-Penrose generalized inverse of the

matrix ψt. Since ψt is symmetric, we have (ψ+t )∗ = ψ+

t . On the other hand, it is

straightforward to verify that for all t ≤ T and ω ∈ Ω, the image of the matrix ψt

(i.e. the linear space spanned by the columns of ψt) is the space q ∈ Rm,∑qi = 0.

Therefore for all q ∈ Rm with∑qi = 0, we have ψt · ψ+

t · q = q. This holds in

particular for any row vector from any Q-matrix, or any vectors of the form ej − ei.

Let us also recall the definition of the Doleans-Dade exponential.

Definition 2.1.10. For a semimartingale L with L0 = 0, the Doleans-Dade expo-

nential, denoted as E(L), is the (unique) semimartingale Z such that for all t ≥ 0,

we have Zt = 1 +∫ t

0Zs−dXs.

To define the controlled probability measure for X, we shall use the following ver-

sion of Girsanov Theorem regarding the change of probability measure. See Theorem

III.41 in Protter [2005] or Lemma 4.3 in Sokol and Hansen [2015].

Theorem 2.1.11. Let T > 0 and L be a martingale defined on [0, T ] with ∆L ≥

−1. Assume that E(L), which is the Doleans-Dade exponential of L, is a uniformly

integrable martingale and let Q be the probability measure having Radon-Nikodym

derivative E(L)T with respect to P. If [M, L] is integrable under P, then M−〈M, L〉

is a martingale under Q, where the predictable quadratic covariation process 〈M, L〉

is computed under the measure P.

We now describe how the control process α affects the dynamics of X through the

probability measure of its sample path. Let q : [0, T ] × E2 × A → R be a transition

rate function that satisfies Assumption 2.1.4 and Assumption 2.1.5 and denote by

Q(t, α) the matrix q(t, i, j, α). Compared to the previous characterization of controlled

continuous-time Markov chains based on the infinitesimal generator where we only

allow Markovian controls, here we allow a larger class of admissible controls. Let us

24

Page 34: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

define the set A to be the collection of F -predictable processes (αt)0≤t≤T such that

αt ∈ A for 0 ≤ t ≤ T . For α ∈ A, we define the scalar martingale L(α) under P:

L(α)t :=

∫ t

0

X∗s− · (Q(s, αs)−Q0) · ψ+s · dMs. (2.15)

Clearly, the jumps of the martingale L(α) are given by:

∆L(α)t = X∗t− · (Q(t, αt)−Q0) · ψ+

t ·∆Xt. (2.16)

It can be easily verified that ψ+t · (ej−Xt−) = m−1

mej−

∑i 6=j

1mei when Xt− = ei 6= ej.

Therefore when Xt− = ei 6= ej = Xt, we have:

∆L(α)t = X∗t− · (Q(t, αt)−Q0) · ψ+

t · (ej −Xt−) = e∗i · (Q(t, αt)−Q0) ·

m− 1

mej −

∑k 6=j

1

mek

=

m− 1

m(q(t, i, j, αt)− q0

i,j)−1

m

∑k 6=j

(q(t, i, k, αt)− q0i,k)

= q(t, i, j, αt)− q0i,j = q(t, i, j, αt)− 1,

where the last equality is due to the fact that∑

k(q(t, i, k, αt)− q0i,k) = 0. Therefore

we have ∆L(α)t ≥ −1. According to Theorem III.45 in Protter [2005] and the remark

that follows, in order to show that E(L(α)) is uniformly integrable, it suffices to show

E[exp(〈L(α), L(α)〉T )] <∞. This is straightforward since we have:

〈L(α), L(α)〉T =

∫ T

0X∗s− · (Q(s, αs)−Q0) · ψ+

s · d〈M,M〉s · (X∗s− · (Q(s, αs)−Q0) · ψ+s )∗ds

=

∫ T

0X∗s− · (Q(s, αs)−Q0) · ψ+

s · (Q∗(s, αs)−Q0) ·Xs−ds,

and the integrand is bounded by some constant according to Assumption 2.1.5. Now

using the definition of L(α) in equation (2.15), we obtain:

〈M, L(α)〉t =

∫ t

0d〈M,M〉s · (ψ+

s )∗ · (Q∗(s, αs)−Q0) ·Xs− =

∫ t

0ψs · ψ+

s · (Q∗(s, αs)−Q0) ·Xs−ds

25

Page 35: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

=

∫ t

0(Q∗(s, αs)−Q0) ·Xs−ds.

In the last equality, we use the fact that (Q∗(s, αs)−Q0) ·Xs is the difference between

two row vectors coming from Q-matrices, and therefore is invariant by ψs · ψ+s . Let

us define the measure Q(α) by:

dQ(α)

dP:= E(L(α))T . (2.17)

By Theorem 2.1.11, we know that the process M(α), defined as:

M(α)t :=Mt −

∫ t

0

(Q∗(s, αs)−Q0) ·Xs−ds (2.18)

is a Q(α)-martingale. Therefore the canonical decomposition of X can be rewritten

as:

Xt = X0 +

∫(0,t]

Q∗(s, αs) ·Xs−ds+M(α)t . (2.19)

This means that under the measure Q(α), the stochastic intensity rate of X is given

by Q(t, αt). In addition, since Q(α) and P coincide on F0, the law of X0 under Q(α)

is p. In particular, when α is a Markovian control, i.e. of the form αt = φ(t,Xt−)

for some measurable feedback function φ, X is a continuous-time Markov chain with

intensity rate q(t, i, j, φ(t, i)) and initial distribution p under the measure Q(α).

Remark 2.1.12. In classical literatures of optimal control, admissible controls are of-

ten classified into the categories of open-loop controls and closed-loop controls. Open-

loop controls are refers to controls adapted to the underlying filtration, which is often

generated by the noise process. Closed-loop controls, on the other hand, are controls

that are adapted to the filtration generated by the history of the state process. In our

setup, however, we see that the underlying filtration is indeed the one generated by

the past path of state process and therefore this difference vanishes.

26

Page 36: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

2.2 Optimal Control of Continuous-Time Markov

Chains

2.2.1 Analytical Approach via the Dynamic Programming

Principle

We first recall the analytical approach to optimal control of continuous-time Markov

chain. Let A be a compact subset of Rl. Let f : [0, T ]×E×A→ R and g : E → R be

respectively the running cost function and the terminal cost function. In the rest of

Section 2.2, we let Assumption 2.1.4 and Assumption 2.1.5 hold. In addition, we make

the following assumption on the regularity of the cost functions and the transition

rate:

Assumption 2.2.1. For all i, j ∈ E, the mappings (t, α)→ q(t, i, j, α) and (t, α)→

f(t, i, α) are continuous.

Let A0 be the collection of Markovian controls, i.e.

A0 = t→ φ(t,Xt)|φ : [0, T ]× E → A is measurable.. (2.20)

We use the notation α ↔ φ to denote an element in A0, emphasizing the feedback

function φ. When we pick a control α↔ φ ∈ A0, the total expected cost is given by:

J(α, t, i) := E[∫ T

t

f(s,Xs, φ(s,Xs))ds+ g(XT )|Xt = i

], (2.21)

where the dynamics of X are determined by the infinitesimal generator Gφ:

GφF (t, i) := ∂tF (t, i) +∑j 6=i

(F (t, j)− F (t, i))q(t, i, j, φ(t, i)). (2.22)

27

Page 37: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We consider the optimization problem:

V (t, i) := infα↔φ∈A0

J(α, t, i). (2.23)

For i = 1, . . . ,M , we define the reduced Hamiltonian Hi by:

[0, T ]× Rm × A→ R

(t, z, α)→ Hi(t, z, α) := f(t, i, α) +∑j 6=i

(zj − zi)q(t, i, j, α).(2.24)

Let us denote the infimum of the Hamiltonian by Hi(t, z) := infα∈AHi(t, z, α). We

have the following result on the minimizer of the Hamiltonian.

Lemma 2.2.2. For all i ∈ E, there exists a measurable mapping ai : [0, T ]×Rm → A

such that Hi(t, z, ai(t, z)) = infα∈AHi(t, z, α) for all (i, t, z) ∈ E × [0, T ]× A.

Proof. This is a direct consequence of the measurable selection theorem (see Theorem

18.19 in Aliprantis and Border [2006]).

Recall the definition of the seminorm ‖ · ‖ei defined in (2.14). By the regularity of

q and f , we can show the following lemma:

Lemma 2.2.3. For all i ∈ E, the mapping (t, z) → Hi(t, z) is continuous. In

addition, there exists a constant C > 0 such that for all t ∈]0, T ] and z, z′ ∈ Rm, we

have:

|Hi(t, z)− Hi(t, z′)| ≤ C‖z − z′‖ei . (2.25)

Proof. By Berge’s maximum theorem, the continuity of Hi and the compactness of

A implies the continuity of Hi. Now let z, z′ ∈ Rm. Since the transition rate function

q is bounded uniformly in α, we have:

infα∈A

Hi(t, z, α)−Hi(t, z′, α) ≤ Hi(t, z, α)−Hi(t, z

′, α)

28

Page 38: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

=∑j 6=i

[(zj − z′j)− (zi − z′i)]q(t, i, j, α) ≤ C‖z − z′‖ei .

Since the above is true for all α, taking supremum of the left-hand side, we obtain

Hi(t, z) − Hi(t, z′) ≤ C‖z − z′‖ei . Exchanging the roles of z and z′, we obtain the

desired inequality.

The Markovian setting allows us to apply the Dynamic Programming Principle

to the optimal control problem (2.23), which leads to the following Hamilton-Jacobi

equation:

∂tv(t, i) = − Hi(t, v(t, ·)), i ∈ E

v(T, i) = g(i),

(2.26)

where the unknown is a mapping v : [0, T ] × E → R. Since E is finite, v(t, ·) can

be identified with an element in Rm, and the above equation is nothing but a cou-

pled system of m ordinary differential equations. Using the Lipschitz regularity of

Hi we just proved, the existence and uniqueness of a local solution can be obtained

by applying the Cauchy-Lipschitz theorem. In order to extend the existence of the

solution to the entire interval [0, T ], it suffices to find a uniform bound for the solu-

tion. This can be done by linking the solution to the total expected cost associated

with the Markovian control with feedback function φ(t, i) := ai(t, v(t, ·)), where ai

is the measurable minimizer of the reduced Hamiltonian Hi as mentioned in Lemma

2.2.2. It is plain to check that the feedback function φ is measurable, therefore it is

an admissible Markov control. Since the control belongs to a compact set and the

running cost is continuous on the control, the total expected cost can be bounded

by a constant. This shows existence of the solution on the entire interval [0, T ] for

any T > 0. The uniqueness of the solution on [0, T ] follows immediately from the

uniqueness of the local solution.

Once we show the HJB equation (2.26) admits a unique solution v, by applying

the Dynkin formula, it is then straightforward to verify that v is the value function

29

Page 39: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

and φ(t, i) := ai(t, v(t, ·)) is an optimal control. This leads to the following result of

the verification argument.

Theorem 2.2.4. Equation (2.26) admits a unique solution, i.e. a mapping v : [0, T ]×

E → R that is C1 in t and satisfies (2.26), and the solution v(t, i) equals the value

function of the optimal control problem V (t, i) defined in 2.23. In addition, given any

measurable minimizer ai of the reduced Hamiltonian Hi as defined in Lemma 2.2.2,

the Markovian control with the feedback function φ defined by φ(t, i) := ai(t, v(t, ·)) is

optimal.

2.2.2 Probabilistic Approach by BSDE

We now introduce the weak formulation of stochastic optimal control of continuous-

time Markov chains which is based on the semimartingale representation. Our ap-

proach is based on the theory of BSDE driven by Markov chain, developed in Cohen

and Elliott [2008] and Cohen and Elliott [2010]. For the sake of completeness, we

collect a few useful results in the appendix of this chapter.

We reuse the notations of transition rate function and cost functions as in Section

2.1.2 and Section 2.2.1. Let A be a compact subset of Rm. Let f : [0, T ]×E×A→ R

and g : E → R be, respectively, the running cost function and the terminal cost

function. As is mentioned in Section 2.1.2, we shall consider all the closed-loop

controls as the admissible controls in the weak formulation. In other words, the set

of admissible controls A is the set of all F-predictable processes taking values in A.

When the control is α ∈ A, the total expected cost is defined as:

J(α) := EQ(α)

[∫ T

0

f(t,Xt, αt)dt+ g(XT )

](2.27)

30

Page 40: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where Q(α) is the probability measure defined in (2.17). We aim to solve the following

optimization problem:

V := infα∈A

J(α). (2.28)

We now show that this optimization problem can be characterized by a BSDE driven

by the continuous-time Markov chain X. To begin with, we define the Hamiltonian

H : [0, T ]× E × Rm × A→ R by

H(t, x, z, α) := f(t, x, α) + x∗ · (Q(t, α)−Q0) · z (2.29)

Recall the definition of reduced Hamiltonian Hi in (2.24) and the fact that x takes

values in the set e1, . . . , em. It is straightforward to verify that H(t, x, z, α) =∑i∈E 1(x = ei)[Hi(t, z, α)−

∑j 6=i(zj − zi)].

In the following, we let Assumption 2.1.4, 2.1.5 and 2.2.1 hold. We have shown in

Lemma 2.2.2 that there exists a measurable mapping ai such that Hi(t, z, ai(t, z)) =

infα∈AHi(t, z, α). We then define the mapping H and a by:

H(t, x, z) :=m∑i=1

1(x = ei)[Hi(t, z)−∑j 6=i

(zj − zi)], (2.30)

a(t, x, z) :=m∑i=1

1(x = ei)ai(t, z). (2.31)

It is clear that for all t ∈ [0, T ], x ∈ E and z ∈ Rm, a(t, x, z) is a minimizer of

the mapping α → H(t, x, z, α), and the minimum is given by H(t, x, z). Combining

Lemma 2.2.3 and the definition of the seminorm ‖ · ‖Xt− (see equation (2.14)), we

obtain:

Lemma 2.2.5. There exists a constant C > 0 such that for all (ω, t, α) ∈ Ω×]0, T ]×

A, and z, z′ ∈ Rm, we have:

|H(t,Xt−, z, α)−H(t,Xt−, z′, α)| ≤ C‖z − z′‖Xt− , (2.32)

31

Page 41: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

|H(t,Xt−, z)− H(t,Xt−, z′)| ≤ C‖z − z′‖Xt− . (2.33)

Let us fix an admissible control α ∈ A and consider the following BSDE:

Yt = g(XT , pT ) +

∫ T

t

H(s,Xs−, Zs, αs)ds−∫ T

t

Z∗s · dMs. (2.34)

Lemma 2.2.6. Given α ∈ A, BSDE (2.34) admits a unique solution (Y, Z) and we

have J(α) = EP[Y0].

Proof. By Lemma 2.2.5, the driver function H of the BSDE (2.34) is Lipschitz in z

with regard to the stochastic semi norm ‖ · ‖Xt− , therefore by Lemma 2.4.1, it admits

a unique solution (Y, Z). We have:

Y0 = g(XT , pT ) +

∫ T

0

H(t,Xt−, Zt, αt)dt−∫ T

t

Z∗s · dMs

= g(XT , pT ) +

∫ T

0

f(t,Xt−, αt)dt−∫ T

0

Z∗t · (dMt − (Q∗(t, αt)−Q0) ·Xt−dt)

= g(XT , pT ) +

∫ T

0

f(t,Xt−, αt)dt−∫ T

0

Z∗t · dM(α)t .

Since M(α) is a martingale under the measure Q(α), we take the expectation under

Q(α) and obtain that EQ(α)[Y0] = J(α). Since Y0 is F0-measurable, and P coincides

with Q(α) on F0, we have EP[Y0] = J(α).

Now let us consider the following BSDE:

Yt = g(XT ) +

∫ T

t

H(s,Xs−, Zs)ds−∫ T

t

Z∗s · dMs. (2.35)

We show that the above BSDE characterizes the optimality of the control problem

(2.28):

Proposition 2.2.7. BSDE (2.35) admits a unique solution (Y, Z). In addition, the

value function of the optimal control problem (2.28) is given by V = E[Y0] and the

32

Page 42: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

control α defined by:

αt = a(t,Xt−, Zt) (2.36)

is an optimal control.

Proof. The existence and uniqueness of solution to (2.35) is easily verified by using

the Lipschitz property of H stated in Lemma 2.2.5. Let (Y, Z) be the solution and

define the process α by αt := a(t,Xt−, Zt). We first show that α is an admissible

control. Recall the definition of a in (2.31). We have:

a(t,Xt−, Zt) =m∑i=1

1(Xt− = ei)ai(t, Zt) = X∗t− ·

(m∑i=1

ai(t, Zt)ei

)

Since ai is measurable for each i ∈ 1, . . . ,m, we see that a is a measurable mapping

from [0, T ]×Rm×Rm to A. Since both the processes t→ Xt− and Z are predictable,

we conclude that α is a predictable process and therefore an admissible control.

Now let us fix an arbitrary control α ∈ A. Let us denote by (Y (α), Z(α)) the

solution to BSDE (2.34) corresponding to the control α. Moreover, (Y, Z) is also the

solution to BSDE (2.34) corresponding to the control α. We denote ∆Y := Y (α)− Y

and ∆Z := Z(α) − Z. Computing the difference of the two BSDEs, we notice that

∆Y and ∆Z are the solutions to the following BSDE:

∆Yt =

∫ T

t

[H(s,Xs−, Zαs , αs)−H(s,Xs−, Zs, αs)]ds−

∫ T

t

∆Z∗s · dMs.

We can further decompose the driver of the above BSDE as:

H(s,Xs−, Z(α)s , αs)−H(s,Xs−, Zs, αs)

= H(s,Xs−, Z(α)s , αs)−H(s,Xs−, Zs, αs) +H(s,Xs−, Zs, αs)−H(s,Xs−, Zs, αs)

= [H(s,Xs−, Zs, αs)−H(s,Xs−, Zs, αs)] +X∗s− · (Q(s, αs)−Q0) ·∆Z.

33

Page 43: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Define the processes ψt := H(t,Xt−, Zt, αt)−H(t,Xt−, Zt, αt) and γt := (Q∗(t, αt)−

Q0) ·Xt−. Therefore (∆Y,∆Z) is the solution to the linear BSDE of the form (2.69)

with ψ and γ defined previously and β = 0. Clearly ψ and γ are both predictable.

Since αt minimizes the Hamiltonian, ψ is nonnegative. The boundedness of γ follows

from the boundedness of the transition rate function q, which is a consequence of the

boundedness of A and the Lipschitz property of q as stated in Assumption 2.2.1. It

remains to check that 1 + γ∗t · ψ+t · (ej −Xt−) ≥ 0.

Indeed, when Xt− = ej, it is trivial that the above inequality holds. We now

assume that Xt− = ei 6= ej. We have ψ+t · (ej −Xt−) = m−1

mej −

∑i 6=j

1mei. Therefore

when Xt− = ei 6= ej, we have:

γ∗t · ψ+t · (ej −Xt−)

=X∗t− · (Q(t, αt)−Q0) · ψ+t · (ej −Xt−) = e∗i · (Q(t, αt)−Q0) ·

m− 1

mej −

∑k 6=j

1

mek

=m− 1

m(q(t, i, j, αt)− q0

i,j)−1

m

∑k 6=j

(q(t, i, k, αt)− q0i,k) = q(t, i, j, αt)− q0

i,j ,

where the last equality is due to the fact that∑

k(q(t, i, k, αt)− q0i,k) = 0. Therefore

we have:

1 + γ∗t · ψ+t · (ej −Xt−) = 1 + q(t, i, j, αt)− q0

i,j = q(t, i, j, αt) ≥ 0

By Lemma 2.4.3, we conclude that ∆Y is nonnegative and in particular E[Y(α)

0 ] ≥

E[Y0]. Since α is an arbitrary admissible control, in light of Lemma 2.2.6, this means

that E[Y0] ≤ infα∈A J(α) = V . Finally, we notice that E[Y0] is the total expected cost

when the control is α. We conclude that α is an optimal control and E[Y0] = V .

Remark 2.2.8. In both the analytical and probabilistic approach we presented above,

we chose to work under weak assumptions of regularity on the transition rate function

and the cost function (see Assumption 2.1.4, 2.1.5 and 2.2.1). As a result, we have to

34

Page 44: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

rely on an argument of measurable selection, in order to obtain the regularity for the

minimizer of the Hamiltonian. Indeed, we can construct an optimal control from any

minimizer of the Hamiltonian. However, if we assume in addition that the minimizer

of the Hamiltonian is unique, then the optimal control is unique.

Proposition 2.2.9. Assume in addition that for all (i, t, z) ∈ E × [0, T ] × Rm, the

reduced Hamiltonian A 3 α → Hi(t, z, α) admits a unique minimizer ai(t, z). Then

for any optimal control α′ ∈ A, we have α′t = αt := a(t,Xt−, Zt), dt⊗ dP-a.e., where

(Y, Z) is the solution to BSDE (2.35) and a is defined in (2.31).

Proof. We consider the solution (Y ′, Z ′) to the following BSDE:

Y ′t =

∫ T

t

H(s,Xs−, Z′s, α

′s)ds−

∫ T

t

(Z ′s)∗ · dMs. (2.37)

Since α′ is optimal, we have EP[Y ′0 ] = J(α′, p, ν) = V (p, ν) = EP[Y0]. Now taking the

difference of the BSDE (2.35) and (2.37), we obtain:

Y0 − Y ′0 =

∫ T

0

[H(t,Xt−, Zt, αt)−H(t,Xt−, Z

′t, α′t)]dt−

∫ T

0(Zt − Z ′t)∗ · dMt

=

∫ T

0

[X∗t− · (Q(t, αt)−Q0) · Zt −X∗t− · (Q(t, α′t)−Q0) · Z ′t

]dt

+

∫ T

0

[f(t,Xt−, αt)− f(t,Xt−, α

′t)]dt−

∫ T

0(Zt − Z ′t)∗ · dMt

=

∫ T

0

[f(t,Xt−, αt)− f(t,Xt−, α

′t) +X∗t− · (Q(t, αt)−Q(t, α′t)) · Zt

]dt

−∫ T

0(Zt − Z ′t)∗ ·

[dMt − (Q∗(t, α′t)−Q0) ·Xt−dt

]=

∫ T

0

[H(t,Xt−, Zt, αt)−H(t,Xt−, Zt, α

′t)]dt−

∫ T

0(Zt − Z ′t)∗ · dM

(α′)t .

Taking the expectation under the measure Q(α′) and using the fact that Q(α′) coincides

with P in F0, we have:

0 = EP[Y0 − Y ′0 ] = EQ(α′)[Y0 − Y ′0 ] = EQ(α′)

[∫ T

0

[H(t,Xt−, Zt, αt)−H(t,Xt−, Zt, α

′t)]dt

]≤ 0,

35

Page 45: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where the last inequality is due to the fact that αt minimizes the Hamiltonian. In

fact, we have αt = α′t, dt⊗dQ(α′)-a.e. If we assume otherwise, the last inequality would

be strict, since the minimizer of the Hamiltonian is unique. Since P is equivalent to

Q(α′), we have αt = α′t, dt⊗ dP-a.e.

2.3 Analytical Approach to Finite State Mean

Field Games

In this section, we shall revisit the analytical formulation of finite state mean field

games. We shall see that Nash equilibria of finite state mean field games can be

characterized by a system of two strongly coupled ordinary differential equations.

The first equation - the Hamilton-Jacobi equation - translates a representative player’s

optimal response facing the other players in the game, while the second equation -

the Kolmogorov equation - depicts the evolution of the distribution of players’ states.

We will provide a version of Lasry-Lions monotonicity condition for finite state mean

field games, which guarantees the uniqueness of the Nash equilibrium. We shall also

formulate the N -player game and explained how the solution of the mean field game

consists of an approximate Nash equilibrium for the N -player game. For the results

and their proofs in this section, we will loosely follow the presentation in Gomes et al.

[2013]. We also refer the readers to Chapter 7, Volume I of Carmona and Delarue

[2017] for a detailed account of analytical formulation to finite state mean field games

and numerical examples.

2.3.1 Model Setup

We consider a continuous-time game of duration T with a large population of play-

ers, whose states can be in one of m possible states. As mentioned previously, the

dynamics according to which each player switches their states are determined by a

36

Page 46: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

transition rate matrix. To simplify the model, we assume that each player is allowed

to control directly the transition rate. Using the notations of Section 2.2.1, we let A

be a compact subset of (R+)m. For a control α = (α1, . . . , αm) ∈ A, we define the

transition rate q:

q(t, i, j, α) :=αj, for j 6= i,

q(t, i, i, α) :=−∑j 6=i

αj.(2.38)

We assume that each player can only observe its own state and chooses a Marko-

vian control α↔ φ ∈ A0 of the form αt = φ(t,Xt), where φ is the feedback function.

Each player then incurs a running cost and a terminal cost which depend on its state,

its control, and the statistical distribution of all the players’ states. Since there are

only m possible states, the distribution of all the players’ states at any given time t

can be described by an element in the m-dimensional simplex, which we denote by

S:

S := p ∈ Rm,∑

pi = 1, pi ≥ 0. (2.39)

We consider the mean field limit where the number of players tends to infinity

and each player has the same cost functions and uses the Markov control with the

same feedback function φ. Under such a limit, a version of law of large number

applies and the empirical distribution of all the players’ states converges to the law

of any single player’s state, which can be represented by a deterministic mapping

µ : [0, T ] 3 t→ µ(t) ∈ S. Assuming that all the players use the control α↔ φ ∈ A0,

the flow of distribution µ evolves according to the Kolmogorov equation:

dµi(t)

dt=∑j∈E

µj(t) · q(t, j, i, φ(t, j)), µ(0) = p. (2.40)

Here p is the distribution of players’ states at t = 0. Now we single out one player

which we call the representative player. When it picks a control α ↔ φ ∈ A0, and

37

Page 47: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the distribution of all the players’ states at time t is given by µ(t) for t ≤ T , the

representative player incurs a total expected cost which it tries to minimize:

J(α, µ) := E[∫ T

0

f(t,Xt, αt, µ(t)) + g(XT , µ(T ))

]. (2.41)

Our goal is to identify Nash equilibria of the mean field game. In such an equi-

librium which is characterized by a control α ∈ A0 and a flow of distribution µ of

players’ states, the representative player cannot do better by deviating from the con-

trol α, provided that all the other players maintain their controls α. Notice that in

the limit of infinitely many players, the effect on the distribution of players’ states

resulting from the unilateral deviation of the representative player is indeed negligi-

ble. Therefore, when we derive the optimal response of the representative player, we

can treat the distribution of players’ states as fixed. We give the formal definition of

Nash equilibrium below:

Definition 2.3.1. Given p ∈ S. We say a tuple (α ↔ φ, µ) is a Nash equilibrium

of the finite state mean field game with initial distribution p if we have:

α = arg infα↔φ∈A0

E[∫ T

0

f(t,Xt, αt, µ(t)) + g(XT , µ(T ))

], (2.42)

dµi(t)

dt=∑j∈E

µj(t) · q(t, j, i, φ(t, j)), µ(0) = p. (2.43)

Quite naturally, the search of Nash equilibrium can be carried out in these two

steps:

Step 1 (Representative player’s optimization problem)

Given a flow of measures µ, solve the optimization problem:

infα↔φ∈A0

E[∫ T

0

f(t,Xt, αt, µ(t)) + g(XT , µ(T ))

], (2.44)

38

Page 48: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where Xt has the transition rate q(t, i, j, φ(t, i)). We denote by α(µ) ↔ φ(µ) the

optimal control of the representative player facing the mean field µ.

Step 2 (Consistency of the flow of measures)

Find the measure flow µ, such that for all t ∈ [0, T ], the distribution of Xt, which is

the state of the player using the optimal control α(µ), coincides with µ(t). This can

be translated into the Kolmogorov equation:

dµi(t)

dt=∑j∈E

µj(t) · φ(µ)i (t, j), µ(0) = p. (2.45)

where φ(µ)i (t, j) is the i-th component of φ(µ)(t, j).

2.3.2 Existence and Uniqueness of Nash Equilibria

We now implement the above scheme to identify the Nash equilibrium. We make the

following assumption on the cost functions f and g.

Assumption 2.3.2. (i) f is Lipschitz in µ, uniformly in t, α and i.

(ii) For i ∈ E, α → f(t, i, α, µ) does not depend on the i-th component of α and

is differentiable with respect to α. In addition, the gradient of f with regard to α is

Lipschitz in µ, uniformly in t, i and α.

(iii) f is strongly convex in α, in the sense that there exists a constant γ > 0 such

that for all (t, i, µ) ∈ [0, T ]× E × S and all α, α′ ∈ A we have:

f(t, i, α′, µ)− f(t, i, α, µ)−∆αf(t, i, α, µ) · (α′ − α) ≥ γ‖α− α′‖2. (2.46)

(iv) g is Lipschitz in µ, uniformly in i.

Representative player’s optimization problem

We now fix a flow of measures µ : [0, T ]→ S and solve the optimization problem

(2.44). Using the notation introduced in Section 2.2.1, the reduced Hamiltonian Hi

39

Page 49: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

becomes:

Hi(t, z, α, µ) := f(t, i, α, µ) +∑j 6=i

(zj − zi)αj.

Since A is convex and compact and f is uniformly convex in α, we can prove that

the mapping [α1, . . . , αi−1, αi+1, . . . , αm] → Hi(t, z, α, µ) admits a unique minimizer

a(t, i, z, µ), of which the components are denoted by aj(t, i, z, µ) for j 6= i. We denote

the minimum of the Hamiltonian by Hi(t, z, µ). In addition, we define ai(t, i, z, µ) :=

−∑

j 6=i aj(t, i, z, µ). Using Assumption 2.3.2, in particular the uniform convexity of

f on α and the Lipschitz property of ∆αf , we can show the following result:

Lemma 2.3.3. aj(t, i, z, µ) is Lipschitz in z and µ, uniformly in t, i, j. Hi is locally

Lipschitz in z and µ.

From Theorem 2.2.4, we see that the optimal control of the representative player

is characterized by the following Hamilton-Jacobi equation:

∂tv(t, i) = − Hi(t, v(t, ·), µ(t)), i ∈ E,

v(T, i) = g(i, µ(T )).

(2.47)

More specifically, the ODE (2.47) admits a unique solution v(µ) and the optimal

control is given by φ(µ)j (t, i) = aj(t, i, v

(µ)(t, ·), µ(t)).

Consistency of the flow of measures

The second step of the search for Nash equilibrium can be formulated as a fixed-

point problem. Let us consider the following Kolmogorov equation:

dpi(t)

dt=∑j∈E

pj(t) · ai(t, j, v(µ)(t, ·), p(t)), i ∈ 1, . . . ,m, (2.48)

with the initial condition p(0) = p. By the Lipschitz property and boundedness of

the mapping a, the ODE (2.48) admits a unique solution which we denote by p(µ).

We then consider the mapping Φ : µ → p(µ). It is clear that in order to show the

40

Page 50: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

existence of a Nash equilibrium, we need to show that the mapping Φ admits a fixed

point.

Let us denote C := C([0, T ],S) the space of continuous mapping from [0, T ] to the

simplex S endowed with the uniform norm ‖ · ‖∞. This will be the space where the

flow of measures lives and clearly it is a Banach space. Indeed, the Lipschitz property

and boundedness of the mapping a implies that any solution p(µ) of (2.48) is Lipschitz

in t with Lipschitz constant bounded uniformly in µ. Therefore by the Arzela-Ascoli

theorem, we can construct a convex and compact subset of C which is stable by the

mapping Φ. In addition, the Lipschitz continuity of the mappings ai and Hi leads to

the continuity of the mapping Φ with regard to the norm ‖ · ‖∞. Finally, we apply

Schauder’s fixed point theorem to the mapping C. To sum it up, we have just proved:

Theorem 2.3.4. Under Assumption 2.3.2, the following system of forward-backward

ODEs admits a solution:

∂tv(t, i) = Hi(t, v(t, ·), p(t)), v(T, i) = g(i, p(T )), i ∈ 1, . . . ,m (2.49)

dpi(t)

dt=

m∑j=1

pj(t) · ai(t, j, v(t, ·), p(t)), p(0) = p, i ∈ 1, . . . ,m. (2.50)

The finite state mean field game admits a Nash equilibrium in the sense of Definition

2.3.1, and at the equilibrium, the distribution of the states of the players is given

by (p(t))t∈[0,T ] and the strategies of the players are given by the feedback function

φj(t, i) := ai(t, j, v(t, ·), p(t)).

We now investigate the uniqueness of the Nash equilibrium. In reminiscence of

the monotonicity condition proposed in Lasry and Lions [2006a], Lasry and Lions

[2006b] and Lasry and Lions [2007], we work under the following assumption:

Assumption 2.3.5. (i) The cost functional f takes the following form:

f(t,X, α, p, ν) = f0(t,X, p) + f1(t,X, α). (2.51)

41

Page 51: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(ii) For all t ∈ [0, T ], i ∈ 1, . . . ,m and p, p′ ∈ S, we have:

m∑i=1

(g(ei, p)− g(ei, p′))(pi − p′i) ≥ 0, (2.52)

m∑i=1

(f0(t, ei, p)− f0(t, ei, p′))(pi − p′i) ≥ 0, (2.53)

and the inequality is strict if p 6= p′.

Theorem 2.3.6. Under Assumption 2.3.5, the system of forward-backward ODEs

(2.49)-(2.50) has at most one solution.

Proof. Let (p(1), v(1)) and (p(2), v(2)) be two solutions to the system (2.49)-(2.50). We

denote by α(1)j (t, i) := aj(t, v

(1)(t, ·), p(1)(t)) and α(2)j (t, i) := aj(t, v

(2)(t, ·), p(2)(t)) the

optimal control corresponding to the two solutions respectively. Observe that:

d

dt

[m∑i=1

(p(1)i (t)− p(2)

i (t))(v(1)(t, i)− v(2)(t, i))

]

=m∑i=1

(p(1)i (t)− p(2)

i (t))(−Hi(t, v(1)(t, ·), p(1)(t)) + Hi(t, v

(2)(t, ·), p(2)(t)))

+m∑i=1

m∑j=1

(p(1)j (t)α

(1)i (t, j)− p(2)

j (t)α(2)i (t, j))

(v(1)(t, i)− v(2)(t, i))

=m∑i=1

(p(1)i (t)− p(2)

i (t))

f(t, i, α(2)(t, i), p(2)(t)) +m∑j=1

v(2)(t, j)α(2)j (t, i)

m∑i=1

(p(1)i (t)− p(2)

i (t))

f(t, i, α(1)(t, i), p(1)(t)) +m∑j=1

v(1)(t, j)α(1)j (t, i)

+

m∑i=1

m∑j=1

(p(1)j (t)α

(1)i (t, j)− p(2)

j (t)α(2)i (t, j))

(v(1)(t, i)− v(2)(t, i)).

Interchanging the index i and j in the summation and rearranging the terms, we

obtain:

d

dt

[m∑i=1

(p(1)i (t)− p(2)i (t))(v(1)(t, i)− v(2)(t, i))

]

42

Page 52: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

=

m∑i=1

p(2)i (t)

f(t, i, α(1)(t, i), p(1)(t))− f(t, i, α(2)(t, i), p(2)(t)) +

m∑j=1

(α(1)j (t, i)− α(2)

j (t, i))v(1)(t, j)

m∑i=1

p(1)i (t)

f(t, i, α(1)(t, i), p(1)(t))− f(t, i, α(2)(t, i), p(2)(t)) +

m∑j=1

(α(1)j (t, i)− α(2)

j (t, i))v(2)(t, j)

=

m∑i=1

p(2)i (t)

[Hi(t, v

(1)(t, ·), α(1)(t, i), p(1)(t))−Hi(t, v(1)(t, ·), α(2)(t, i), p(2)(t))

]−

m∑i=1

p(1)i (t)

[Hi(t, v

(2)(t, ·), α(1)(t, i), p(1)(t))−Hi(t, v(2)(t, ·), α(2)(t, i), p(2)(t))

].

We have assumed that f is of the form f(t,X, α, p) = f0(t,X, p) + f1(t,X, α), when

we minimize the hamiltonian Hi(t, z, α, p) over α, the minimizer does not depend on

p. Therefore, since α(1)(t, i) minimizes α→ Hi(t, v(1)(t, ·), α, p(1)(t)), it also minimizes

α→ Hi(t, v(1)(t, ·), α, p(2)(t)), which implies that

Hi(t, v(1)(t, ·), α(2)(t, i), p(2)(t)) ≥ Hi(t, v

(1)(t, ·), α(1)(t, i), p(2)(t)).

Interchanging the indices, we also have:

Hi(t, v(2)(t, ·), α(1)(t, i), p(1)(t)) ≥ Hi(t, v

(2)(t, ·), α(2)(t, i), p(1)(t)).

Injecting these two inequalities into the computation of the derivative, we have:

d

dt

[m∑i=1

(p(1)i (t)− p(2)

i (t))(v(1)(t, i)− v(2)(t, i))

]

≤m∑i=1

p(2)i (t)

[Hi(t, v

(1)(t, ·), α(1)(t, i), p(1)(t))−Hi(t, v(1)(t, ·), α(1)(t, i), p(2)(t))

]−

m∑i=1

p(1)i (t)

[Hi(t, v

(2)(t, ·), α(2)(t, i), p(1)(t))−Hi(t, v(2)(t, ·), α(2)(t, i), p(2)(t))

]=−

m∑i=1

(p(1)i (t)− p(2)

i (t))(f0(t, ei, p(1)(t))− f0(t, ei, p

(2)(t)).

43

Page 53: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Using the initial and terminal conditions p(1)(0)−p(2)(0) = 0, v(1)(T, i) = g(ei, p(1)(T ))

and v(2)(T, i) = g(ei, p(2)(T )), we integrate the inequality and obtain:

0 ≥∫ T

0

m∑i=1

(p(2)i (t)− p(1)

i (t))(f0(t, ei, p(2)(t))− f0(t, ei, p

(1)(t))dt

+

m∑i=1

(p(2)i (T )− p(1)

i (T ))(g(ei, p(2)(T )− g(ei, p

(1)(T )).

Now we assume by contradiction that there exists j ∈ E and t ∈ [0, T ] such that

p(1)j (t) 6= p

(2)j (t). Then by the continuity of p(1) and p(2), there exists ε > 0 such

that p(1)j (s) 6= p

(2)j (s) for s ∈ [t − ε, t + ε] ∩ [0, T ]. By the monotonicity condition

stated in Assumption 2.3.5, we deduce that the integrand in the above inequality is

positive. In addition, we have (p(2)i (s)− p(1)

i (s))(f0(s, ei, p(2)(s))− f0(s, ei, p

(1)(s)) > 0

for s ∈ [t − ε, t + ε] ∩ [0, T ]. This implies that the integral on the right hand side

is strictly positive, which leads to a contradiction. It follows that p(1) = p(2) and

subsequently v(1) = v(2).

2.3.3 Game With Finitely Many Players

We now present the model for the game with N players in the finite state space E.

Let us denote by Xn,Nt the state of player n at time t. We assume that each player n

can observe its own state Xn,Nt as well as the empirical distribution of the states of

all players in the game µNt . Clearly, µNt is given by the following vector:

µNt :=1

N

[ N∑n=1

1(Xn,Nt = 1),

N∑n=1

1(Xn,Nt = 2), . . . ,

N∑n=1

1(Xn,Nt = m)

], (2.54)

which belongs to SN , a finite subset of the m-dimensional simplex S:

SN :=

[n1

N, . . . ,

nmN

] ∣∣∣∣ ni ∈ N,m∑i=1

ni = N

. (2.55)

44

Page 54: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

As in the mean field game, we assume that agents can directly control the transition

rates of their states. Accordingly, the control space is A = (R+)m and the admissible

strategy of player n at time t is a function of t, Xn,Nt and µNt taking values in Rm.

We define the set of admissible strategies for player n as:

An,N := t→ φn(t,Xn,Nt , µNt )|φn : [0, T ]× E × S → A. (2.56)

Here φn is the feedback function of player n and we denote by φnj (t, i, p) the j-th

component of φn(t, i, p). We now describe the joint dynamics for the states of all the

players in the game. We assume that the process t→ (X1t , . . . , X

Nt ) is a continuous-

time Markov chain and conditioned on the current states of the players in the game,

the changes of state are independent for different players. Accordingly we define the

transition rate to be:

P[X1t+∆t = j1, . . . , XN

t+∆t = jN |X1t = i1, . . . , XN

t = iN ]

:=

N∏n=1

1(in 6= jn)[φnjn(t, in, µNt )∆t+ o(∆t)]

× 1(in = jn)[1−∑j 6=in

φnj (t, in, µNt )∆t+ o(∆t)].

(2.57)

where we recall the definition of the empirical measure µNt in (2.54). Notice that

by expanding the above product and retaining only the term of the same order as

∆t, we obtain the transition rate matrix for the the continuous-time Markov chain

(X1, . . . , XN), which is a sparse matrix of dimension mN . Moreover, due to the

independence assumption, at most one player changes its state at any given time. We

then define the total expected cost of player n:

E[∫ T

0

f(t,Xn,Nt , φn(t,Xn,N

t , µNt ), µNt ) + g(Xn,NT , µNT )

]. (2.58)

Since our objective is to explore the connection with the mean field game, we

shall consider only symmetric Nash equilibria, in which the players use Markovian

45

Page 55: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

control with the same feedback function φ. Without loss of generality, we single out

player n = 1 as the deviating player who uses a different strategy φ and we assume

that all the other players use the strategy φ. We now show that (X1,Nt , µNt ) is a

continuous-time Markov chain. Let us denote (FNt )t∈[0,T ] the natural filtration of

(X1,N , . . . , XN,N). For any j ∈ E, π ∈ SN and t ≤ T and h > 0 such that t+ h ≤ T ,

by the Markovian property, we have:

P[X1,Nt+h = j, µNt+h = π|FNt ] = P[X1,N

t+∆t = j, µNt+∆t = π|X1,Nt , . . . , XN,N

t ].

Now assuming that at time t, we have Xn,Nt = in for n = 1, . . . , N and we denote

p ∈ SN the empirical measure corresponds to the states (i1, . . . , iN). Clearly we have:

P[X1,Nt+h = j, µNt+h = π| ∩Nn=1 X

n,Nt = in]

=P[X1,Nt+h = j, µNt+h = π| ∩Nn=2 X

n,Nt = in ∩ X1,N

t = i1 ∩ µNt = p].

Since all the players except player 1 use the same strategy, the above probability will

remain the same if we permute the states among players 2 to N , provided that the

empirical distribution of all players remains to be p. In fact, all these permutations

constitute the event X1t = i1 ∩ µNt = p. Therefore we have:

P[X1,Nt+h = j, µNt+h = π| ∩Nn=1 X

n,Nt = in] = P[X1,N

t+∆t = j, µNt+∆t = π|X1,Nt = i1, µNt = p],

and subsequently:

P[X1,Nt+h = j, µNt+h = π|FNt ] = P[X1,N

t+∆t = j, µNt+∆t = π|X1,Nt , µNt ].

Now that we have showed the Markovian property for the process (X1,Nt , µNt ), we

determine its dynamics by computing the infinitesimal generator. To this end, let

us fix a function F : [0, T ] × E × SN → R. Let i ∈ E and p ∈ SN . Denote

nj := N · pj − 1(i = j) the number of players in state j for i ∈ E, without counting

46

Page 56: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

player 1. Using the same argument of symmetry as above, we have:

E[F (t+ h,X1,Nt+h , µ

Nt+h)|X1

t = i, µNt = p]

=E[F (t+ h,X1,Nt+h , µ

Nt+h)|X1

t = i,X2:(n1+1),Nt = 1, . . . , X

(n1+···+nm−1+2):(n1+···+nm+1),Nt = m].

Using the transition rate defined in equation (2.57), we obtain the infinitesimal

generator Gφ,φ evaluated based on the following limit:

Gφ,φF (t, i, p) = limh→0

1

h

[E[F (t+ h,X1,N

t+h , µNt+h)|X1,N

t = i, µNt = p]− F (t, i, p)]

=d

dtF (t, i, p) +

∑j,j 6=i

[F (t, j, p+1

N(ej − ei))− F (t, i, p)]φj(t, i, p)

+∑

(j,k),j 6=k

[F (t, i, p+1

N(ej − ek))− F (t, i, p)](N · pk − 1(k = i))φj(t, k, p).

Here the ei’s are the basis vector of Rm. On the right hand side, the first sum-

mation corresponds to the change of states of player 1, while the second summation

corresponds to the change of states of the remaining players. Having specified the

dynamics of the deviating player and the empirical distribution of players’ states, we

define the total expected cost of the deviating players:

JNφ,φ(t, i, p) := E[∫ T

tf(s,X1,N

s , φ(s,X1,Ns , µNs ), µNs ) + g(X1,N

T , µNT )

∣∣∣∣X1,Nt = i, µNt = p

],

(2.59)

where f and g are respectively the running cost and terminal cost. To sum it up,

given that all the other players use the Markov strategy φ, the deviating player solves

the following control problem:

V Nφ (t, i, p) := inf

α↔φ∈A1,NJNφ,φ(t, i, p). (2.60)

Notice that the Markov process (X1,N , µN) takes values in a finite set 1, . . . ,m ×

SN and therefore it is a continuous-time finite state Markov chain. The analytical

approach to optimal control problem presented in Section 2.2.1 applies. Using the

47

Page 57: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

infinitesimal generator of the process (X1,N , µN) we just derived, it is straightforward

to write down the HJB equation, which is a system of coupled ODEs:

d

dtV Nφ (t, i, p) =− inf

α∈A

∑j,j 6=i

[V Nφ (t, j, p+

1

N(ej − ei))− V N

φ (t, i, p)]αj + f(t, i, α, p)

∑(j,k),j 6=k

[V Nφ (t, i, p+

1

N(ej − ek))− V N

φ (t, i, p)](N · pk − 1(k = i))φj(t, k, p),

(2.61)

V Nφ (T, i, p) = g(i, p). (2.62)

To identify the symmetric Nash equilibria, we search for a feedback function φ such

that when all the remaining players use Markovian controls given by feedback function

φ, the optimal strategy of the deviating player is also φ. Therefore if the feedback

function φ leads to a symmetric Nash equilibrium, we must have:

φj(t, i, p) = aj(t, i, VNφ

(t, ·, p+1

N(e· − ei)), p), (2.63)

where by abuse of notation, we denote by Vφ(t, ·, p + 1N

(e· − ei)) the m-dimensional

vector of which the j-th component is Vφ(t, j, p + 1N

(ej − ei)). This leads to the

following system of non-linear coupled ODEs:

d

dtvN (t, i, p) =−

∑j,j 6=i

[vN (t, j, p+1

N(ej − ei))− vN (t, i, p)]φj(t, i, p)− f(t, i, α, p) (2.64)

−∑

(j,k),j 6=k

[vN (t, i, p+1

N(ej − ek))− vN (t, i, p)](N · pk − 1(k = i))φj(t, k, p),

vN (T, i, p) = g(i, p), (2.65)

φj(t, i, p) = aj(t, i, vN (t, ·, p+

1

N(e· − ei)), p). (2.66)

To show the well-posedness of the system (2.64)-(2.66), we verify that the right hand

side of equation (2.64) is Lipschitz in vN , and any local solution of the system (2.64)-

48

Page 58: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(2.66) is bounded by a uniform constant. Finally, we obtain the existence of the Nash

equilibrium by a verification argument.

Theorem 2.3.7. There exists a symmetric Nash equilibrium for the N-player game.

Let vN be the unique solution to the system (2.64)-(2.66). Then for each n ∈

1, . . . , N, the total expected cost of the player n in equilibrium is given by vN(0, i,p)

if the player is in state i and the empirical distribution of players’s states is p at

time 0, i.e. Xn,N0 = i and 1

N

∑Nn=1 1(Xn,N

0 = j) = pj . Moreover, the optimal strategy

of the players in equilibrium is given by the feedback function:

φj(t, i, p) = aj(t, i, vN (t, ·, p+

1

N(e· − ei)), p).

Remark 2.3.8. In full analogy with N-player stochastic differential games studied

in Chassagneux et al. [2014], when we let N tend to infinity in the system of ODEs

(2.64)-(2.66), formally we should obtain the master equation of the finite state mean

field game. The solution to the master equation gives the expected total cost of the

Nash equilibrium at any given time, state and the distribution of the populations,

and provides a way to compute the Nash equilibrium without searching for a fixed

point. In the case of finite state mean field game, the master equation is a first-

order partial differential equation, thus is much easier to handle numerically than the

master equation for diffusion-based mean field game, where the space of distribution

is of infinite dimension. When studying finite state mean field games with major

and minor players in Chapter 5, we shall derive the master equation using the same

technique.

2.3.4 Approximate Nash Equilibria

To conclude our presentation on the analytical approach to finite state mean field

games, we state without proof the result on the convergence of the Nash equilibrium

49

Page 59: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

of the N -player game to the Nash equilibrium of the mean field game. We refer the

interested readers to Section 4 in Gomes et al. [2013] for detailed proofs.

Let us fix p ∈ S. At time t = 0, we assign states to the N players randomly

according to the distribution p and these random assignments are independent. We

denote µN0 the resulting empirical distribution at time 0. These N players then

enter the game described in Section 2.3.3. Clearly, a symmetric Nash equilibrium is

formed when each agent adopts the Markovian strategy specified in Theorem 2.3.7.

Using the same argument of symmetry when we show the Markov property of the

deviating player’s state and the empirical distribution, we can show that the flow of

the empirical distribution (µNt )t∈[0,T ] in a symmetric Nash equilibrium is a continuous-

time Markov chain. Now let us denote by (v, p) the solution to the system (2.49) -

(2.50) which characterizes the Nash equilibria of the mean field game. We also recall

that vN is the solution to the system (2.64) - (2.66). We state the following result of

convergence:

Theorem 2.3.9. There exists two positive constants T ∗ and C, independent of N ,

such that if T < T ∗ and TC < 1, then for all t ∈ [0, T ] we have:

E[‖µNt − p(t)‖22

]+ E

[‖vN (t, ·, µNt )− v(t, ·)‖22

]≤ C

(1− TC)N, (2.67)

where ‖ · ‖2 is the Euclidean norm on Rm.

Remark 2.3.10. Theorem 2.3.9 only provides the result of approximate Nash equi-

libria for games with duration smaller than T ∗. A closer look at the proof of Theorem

2.3.9 suggests that the upper bound of the duration T ∗ is a decreasing function of the

Lipschitz constant of the terminal cost function g (with regard to the argument of the

measure p). It is not clear whether this upper bound is optimal, or whether the result

of approximate Nash equilibria hold for any duration of the game.

50

Page 60: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

2.4 Appendix: BSDE Driven by Single Continuous-

Time Markov Chain

We recall several useful results on BSDEs driven by continuous-time Markov chain,

which were developed in Cohen and Elliott [2008] and Cohen and Elliott [2010]. Recall

thatM is the P-martingale in the canonical decomposition of the Markov chain X in

(2.7). We consider the following BSDE with unknown (Y, Z), where Y is an adapted

and cadlag process in R, and Z is an adapted and left continuous process in Rm:

Yt = ξ +

∫ T

t

F (w, s, Ys, Zs)ds−∫ T

t

Z∗sdMs. (2.68)

Here ξ is a FT -measurable P-square integrable random variable and F is the driver

function such that the process t→ F (w, t, y, z) is predictable for all y, z.

Recall the stochastic seminorm ‖ · ‖Xt− defined in (2.13). We have the following

result on existence and uniqueness of the solution to the BSDE. See Theorem 1.1 in

Cohen and Elliott [2010].

Lemma 2.4.1. Assume that there exists C > 0 such that dP⊗dt-a.e., for all y1, y2 ∈

R and z1, z2 ∈ Rm we have:

|F (w, t, y1, z1)− F (w, t, y2, z2)| ≤ C(|y1 − y2|+ ‖z1 − z2‖Xt−).

Then the BSDE (2.68) admits a solution (Y, Z) satisfying

E[∫ T

0

|Yt|2dt]< +∞, E

[∫ T

0

‖Zt‖2Xt−dt

]< +∞.

In addition, the solution is unique in the sense that if (Y (1), Z(1)) and (Y (2), Z(2))

are two solutions, then Y (1) and Y (2) are indistinguishable and we have E[∫ T

0‖Z(1)

t −

Z(2)t ‖2

Xt−dt] = 0.

51

Page 61: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We also have the following stability property of the solution, which can be proved

by mimicking the argument used in the proof of Theorem 2.1 in Hu and Peng [1997].

Lemma 2.4.2. For n ≥ 0, let (Y (n), Z(n)) be the solution to the BSDE with driver

F (n) and terminal condition ξ(n). Assume that for all n, F (n) satisfies the Lipschitz

condition in Lemma 2.4.1 and the following conditions hold:

(i) limn→+∞ E[|ξ(n) − ξ(0)|2] = 0.

(ii) For all t ≤ T , we have:

limn→+∞

E

[(∫ T

t

(F (n)(w, s, Y (0)s , Z(0)

s )− F (0)(w, s, Y (0)s , Z(0)

s ))ds

)2]

= 0.

(iii) There exists C > 0 such that for all t ≤ T and n ≥ 0, we have:

E

[(∫ T

t

(F (n)(w, s, Y (0)s , Z(0)

s )− F (0)(w, s, Y (0)s , Z(0)

s ))ds

)2]≤ C

Then we have:

limn→+∞

E[∫ T

t

‖Z(n)s − Z(0)

s ‖2Xs−ds

]+ E[|Y (n)

t − Y (0)t |2] = 0.

Finally, we state a result of comparison for linear BSDEs. See Theorem 3.16 in

Cohen and Elliott [2010].

Lemma 2.4.3. Let γ be a bounded predictable process in Rm and β be a bounded

predictable process in R. Let φ be a non-negative predictable process in R such that

E[∫ T

0‖φt‖2dt] < +∞ and ξ a non-negative square-integrable FT measurable random

variable in R. Let (Y, Z) be the solution of the linear BSDE:

Yt = ξ +

∫ T

t

(φu + βuYu + γ∗u · Zu)du−∫ T

t

Z∗u · dMu. (2.69)

52

Page 62: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Assume that for all t ∈ (0, T ] and j such that e∗j ·Q0 ·Xt− > 0, we have:

1 + γ∗t · ψ+t · (ej −Xt−) ≥ 0, (2.70)

where ψ+t is the Moore-Penrose inverse of the matrix ψt defined in (2.12). Then Y

is nonnegative.

53

Page 63: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 3

Probabilistic Analysis of Finite

State Mean Field Games

In this chapter, we construct a probabilistic framework for continuous-time mean

field games in finite state space. We draw our inspiration from the weak formulation

of stochastic differential mean field games developed in Carmona and Lacker [2015].

Evoking the notion of weak solutions of stochastic differential equations in which

the probability measure and the underlying Wiener process both become part of the

solution, the key idea is to view the optimization problem of a deviating player as

controlling the probability law of the state process, as opposed to controlling the state

process in a pathwise fashion. This perspective seems to be tailor-made to continuous-

time Markov chains. Indeed, as we have seen in Chapter 2, the most convenient way to

describe the dynamics of a continuous-time Markov chain is to specify the transition

rates between the states, which is essentially to specify the probability distribution

of the time until the next change of state, and the probability of each new possible

state after the change.

Similar to the approach taken in Carmona and Lacker [2015], our starting point

is the semimartingale representation of controlled continuous-time Markov chain in-

54

Page 64: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

troduced in Section 2.1.2 and the weak formulation of optimal control presented

in Section 2.2.2 of Chapter 2. As a reminder, we constructed a probability space

of sample paths under which the canonical process X has the representation Xt =

X0 +∫

]0,t]Q0 ·Xs−ds+Mt where Q0 is the transition rate matrix of a continuous-time

Markov chain with constant transition rate 1. By applying Girsanov theorem, we

showed how to construct an equivalent probability measure under which the process

X has the desired controlled transition rates. We then identify the expected total

cost under the controlled probability measure with the solution to a BSDE based

on Markov chain, and the optimality of the control problem can be obtained by the

comparison principle for linear BSDEs.

As we have already seen in Section 2.3 of Chapter 2, the paradigm of mean field

games can be summarized as a two-step procedure. In the first step, we fix the mean

field, i.e. the distribution of all players’ states, and try to derive the optimal strategy

of a representative player in the game. Fixing the mean field makes perfect sense in the

limit of infinitely many players, because the unilateral deviation of any single player

has negligible impact on the mean field. In the second step, we solve the fixed point

problem of Nash equilibrium by matching the mean field generated by the optimal

strategy to the one we fix in the first step. Different from the analytical approach

presented in Section 2.3, in the probabilistic approach we are about to present in this

chapter, we achieve the first step through the weak formulation of optimal control.

We then recast the search for the Nash equilibrium into a fixed point problem on the

space of probability measures of sample of players’ state processes

The advantage of the weak formulation over the traditional analytical approach are

three-fold. First of all, the flexibility of the probabilistic approach allows us to consider

mean field games where the interactions are not only realized through the states, but

also through the strategies of the players. Mean field games with interactions through

55

Page 65: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the strategies of the players are known to be notoriously intractable via PDE/ODE

methods due to the difficulty in deriving the equation obeyed by the flow of probability

measure of the optimal control. Under the probabilistic framework however, the mean

field of state and the mean field of control can be treated in similar manners, although

the treatment of the mean field of control is more involved in terms of topological

arguments. Secondly, using the weak formulation we are able to show Nash equilibria

exist among all closed-loop strategies, including the strategies depending on the past

history of player’s states, whereas the analytical approach presented in Chapter 2 can

only accommodate Markovian controls.

Lastly, the weak formulation we develop for finite state mean field games will

play an indispensable role when we tackle the finite state mean field principal agent

problems in Chapter 4. This is a special model of Stackelberg game in which the

principal fixes a contract first and a large population of competitive agents form a

Nash equilibrium according to the contract proposed by the principal. Mathemati-

cally speaking, the contract is a control chosen by the principal which enters into each

agent’s dynamics and cost functions. One meaningful direction in probing mean field

principal agent problems is to understand how the principal can choose the optimal

contract in order to minimize its own cost function which depends on the distribution

of the agents’ states. To the best of our knowledge, this type of problems was first

investigated in Elie et al. [2016] where agent’s dynamics are given by a controlled

diffusion process. The main idea is to formulate the optimal contract problem as a

McKean-Vlasov optimal control problem, in which the state process to be controlled

is the McKean-Vlasov BSDE characterizing the Nash equilibrium in the weak formu-

lation of the mean field game. We shall see in Chapter 4 that with the help of the

weak formulation developed in this chapter, mean field principal agent problems in

finite state space can be tackled in an analogous way. Moreover, thanks to the finite

nature of the state space, the optimal contract can be computed with straightforward

56

Page 66: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

numerical schemes, which could lead to potential applications in epidemic control and

cyber security.

We would also like to mention some literature related to the weak probabilistic

approach we present in this chapter. In Cecchin and Fischer [2017] the authors pro-

posed a probabilistic framework for finite state mean field games where the state

dynamics of each player are represented by stochastic differential equations driven by

Poisson random measures. By using Ky Fan’s fixed point theorem, the authors ob-

tained existence and uniqueness of Nash equilibrium in relaxed open-loop controls as

well as relaxed feedback controls. Then under a stronger assumption that guarantees

uniqueness of optimal non-relaxed feedback control, the authors deduced existence

of Nash equilibria in non-relaxed feedback form. In Doncel et al. [2016], continuous-

time mean field games with finite state space and finite action space were studied.

The authors proved existence of Nash equilibria among relaxed feedback controls. In

Benazzoli et al. [2017] the authors investigated mean field games where each player’s

state follows a jump-diffusion process and the player controls the sizes of the jumps.

Their approach is based on weak formulation of stochastic controls and martingale

problems. Existence of Nash equilibria among relaxed controls and Markov controls

was established.

The rest of this chapter is organized as follows. In Section 3.1, we introduce the

weak formulation of finite state mean field games, which is based on the semimartin-

gale representation of continuous-time Markov chains and an argument of change of

measure. We state the assumptions of the model and give a precise definition of

Nash equilibria in the weak formulation. We also analyze the representative player’s

optimal control problem when facing a fixed mean field of state and control, as a

straightforward application of weak formulation of optimal control introduced in Sec-

tion 2.2 of Chapter 2. Section 3.2 and Section 3.3 are devoted to the existence and

57

Page 67: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the uniqueness of Nash equilibria. Finally in Section 3.4, we formulate the game with

finitely many players and show the Nash equilibrium of the mean field game is an

approximative Nash equilibrium of the game with finite number of players.

3.1 Weak Formulation of Finite State Mean Field

Games

3.1.1 Mean Field Games with Interaction through States and

Controls

We consider a large population of players whose states belong to a finite state space

E constituted of m elements. These states are identified with the m basis vectors of

Rm, which are denoted by e1, . . . , em. We denote by S the m-dimensional simplex

S := p ∈ Rm,∑pi = 1, pi ≥ 0 equipped with the euclidean norm ‖ · ‖, which we

identify with the space of probability distribution on E.

Let Ω be the space of cadlag functions from [0, T ] to E that are continuous on T ,

and X be the canonical process. Later we will use X to represent the state process

of a representative player. We denote by Ft := σ(Xs, s ≤ t) the natural filtration

generated by X and F := FT . Now let us fix p ∈ S a probability measure on E.

On (Ω, (Ft)t∈[0,T ], F), we consider the probability measure P under which X is the

continuous-time Markov chain such that the law of X0 is p and the transition rate

between any two different states is 1. We recall that under P the process X has the

following representation:

Xt = X0 +

∫]0,t]

Q0 ·Xs−ds+Mt, (3.1)

where M is a P-martingale and Q0 is a m by m matrix with diagonal elements all

equal to −(m− 1) and off-diagonal elements all equal to 1.

58

Page 68: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Let A be a compact subset of Rl in which the players can choose their controls.

Denote by P(A) the space of probability measures on A. We introduce a function q:

[0, T ]× 1, . . . ,m2 ×A× S × P(A)→ q(t, i, j, α, p, ν), (3.2)

and we denote by Q(t, α, p, ν) the matrix [q(t, i, j, α, p, ν)]1≤i,j≤m. Throughout this

chapter, we make the following assumption on q:

Assumption 3.1.1. For all (t, α, p, v) ∈ [0, T ]×A×S×P(A), the matrix Q(t, α, p, ν)

is a Q-matrix. (See Definition 2.1.2)

We now describe how the player’s control and the distribution of all the players’

states affect the probability law of X in our model. Let us define the player’s strategy

set A to be the collection of F -predictable processes (αt)t∈[0,T ] such that αt ∈ A for

0 < t ≤ T . Given (pt)t∈[0,T ], a flow of measures on 1, . . . ,m, and (νt)t∈[0,T ], a flow

of measures on A, we define the scalar martingale L(α,p,ν) under P by:

L(α,p,ν)t :=

∫ t

0X∗s− · (Q(s, αs, ps, νs)−Q0) · ψ+

s · dMs, (3.3)

where we recall that ψt is the density of the predictable quadratic variation of the

martingale M under P as defined in equation (2.12), and ψ+t is the Moore-Penrose

inverse of ψt. We then define the measure Q(α,p,ν) through the Doleans-Dade expo-

nential of L(α,p,ν):

dQ(α,p,ν)

dP:= E(L(α,p,ν))T . (3.4)

Following the same arguments developed in Section 2.1.2, Chapter 2, we know that

the process M(α,p,ν), defined as:

M(α,p,ν)t :=Mt −

∫ t

0(Q∗(s, αs, ps, νs)−Q0) ·Xs−ds, (3.5)

59

Page 69: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

is a Q(α,p,ν)-martingale. Therefore the canonical decomposition of X can be rewritten

as:

Xt = X0 +

∫]0,t]

Q∗(s, αs, ps, νs) ·Xs−dt+M(α,p,ν)t . (3.6)

This means that under the measure Q(α,p,ν), the stochastic intensity rate of X is

given by Q(t, αt, pt, νt). In particular, when α is a Markovian control, i.e. of the

form αt = φ(t,Xt−) for some measurable feedback function φ, X is a continuous-time

Markov chain with intensity rate q(t, i, j, φ(t, i), pt, νt) under the measure Q(α,p,ν).

Moreover, we have Q(α,p,ν) = P on F0, and therefore the distribution of X0 under

Q(α,p,ν) is p.

Let f : [0, T ] × E × A × S × P(A) → R and g : E × S → R be respectively the

running cost function and the terminal cost function. When the player picks a control

α ∈ A and the mean field is (p, ν), its cost is:

J(α, p, ν) := EQ(α,p,ν)

[∫ T

0f(t,Xt, αt, pt, νt)dt+ g(XT , pT )

](3.7)

Each player aims at minimizing J(α, p, ν), that is, it solves the optimization problem:

V (p, ν) := infα∈A

EQ(α,p,ν)

[∫ T

0f(t,Xt, αt, pt, νt)dt+ g(XT , pT )

](3.8)

In the classical setup of mean field games, a single player’s strategy α does not

alter the mean field (p, ν). Therefore when each player solves its own optimization

problem, it considers (p, ν) as given. A Nash equilibrium is then reached when the

law of Xt under the player’s controlled probability measure, along with the law of its

control at time t under the same probability measure, coincides with (pt, νt) for all

t ≤ T . Accordingly, we have the following definition of a Nash equilibrium for the

weak formulation of finite state mean field games.

60

Page 70: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Definition 3.1.2. Let p∗ : [0, T ]→ S, ν∗ : [0, T ]→ P(A) be two measurable functions

and α∗ ∈ A. We say that the tuple (α∗, p∗, ν∗) is a Nash equilibrium for the weak

formulation of finite state mean field game if:

(i) α∗ minimizes the cost when the mean field is given by (p∗, ν∗):

α∗ ∈ arg infα∈A

EQ(α,p∗,ν∗)[∫ T

0f(t,Xt, αt, p

∗t , ν∗t )dt+ g(XT , p

∗T )

]. (3.9)

(ii) (α∗, p∗, ν∗) satisfies the consistency conditions:

p∗t = Q(α∗,p∗,ν∗)[Xt = ei]1≤i≤m, ∀t ≤ T, (3.10)

ν∗t = Q(α∗,p∗,ν∗)#α∗t

, ∀t ≤ T. (3.11)

Throughout the rest of this chapter, we make the following assumptions on the

transition rate q and the cost functionals f and g.

Assumption 3.1.3. There exists C1, C2 > 0 such that for all t, i, j, α, p, ν ∈ [0, T ]×

E2 × A× S × P(A) with i 6= j, we have 0 < C1 < q(t, i, j, α, p, ν) < C2.

Assumption 3.1.4. There exists a constant C > 0 such that for all (t, i, j) ∈ [0, T ]×

E2, α, α′ ∈ A, p, p′ ∈ S and ν, ν ′ ∈ P(A), we have:

|q(t, i, j, α, p, ν)− q(t, i, j, α′, p′, ν ′)| ≤C(‖α− α′‖+ ‖p− p′‖+W1(ν, ν ′)), (3.12)

|f(t, ei, α, p, ν)− f(t, ei, α′, p′, ν ′)| ≤C(‖α− α′‖+ ‖p− p′‖+W1(ν, ν ′)), (3.13)

|g(ei, p)− g(ei, p′)| ≤C‖p− p′‖. (3.14)

Remark 3.1.5. Assumption 3.1.3 is analogous to the non-degeneracy condition in

the diffusion-based mean field game model. It guarantees that the probability measure

Q(α,p,ν) is equivalent to the reference measure P.

Remark 3.1.6. In certain application of continuous-time Markov chain models, it is

possible that the jumps from some states to others are forbidden. That is we consider

61

Page 71: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

a transition rate function q such that q(t, i, j, α, p, ν) ≡ 0 for some tuples (i, j). this

is indeed the case in the botnet defense model proposed by Kolokoltsov and Bensous-

san [2016] as well as in an extended version of model including an attacker studied

in Carmona and Wang [2016]. If this were the case, we need to define a different

reference probability P: we set the transition rate to 1 for the jumps, except those

jumps that are forbidden, for which we set the transition rate to 0. Fortunately, this

is the only modification we need to make in order to accommodate for this kind of

special case. The arguments presented in the following can be trivially extended to be

compatible with this modified reference probability.

3.1.2 Representative Player’s Optimization Problem

We now apply the weak formulation introduced in Chapter 2 to tackle a representative

player’s optimization problem. As before, we define the Hamiltonian H : [0, T ]×E×

Rm × A× S × P(A)→ R by:

H(t, x, z, α, p, ν) := f(t, x, α, p, ν) + x∗ · (Q(t, α, p, ν)−Q0) · z, (3.15)

as well as the reduced Hamiltonian:

Hi(t, z, α, p, ν) := H(t, ei, z, α, p, ν). (3.16)

We make the following assumption on the minimizer of the reduced Hamiltonian.

Assumption 3.1.7. (i) For any t ∈ [0, T ], i ∈ 1, . . . ,m, z ∈ Rm, p ∈ S and

ν ∈ P(A), the mapping α → Hi(t, z, α, p, ν) admits a unique minimizer which does

not depend on the mean field of control ν. We denote the minimizer by ai(t, z, p).

62

Page 72: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(ii) ai is measurable in [0, T ] × Rm × S. There exist constants C1 > 0 and C2 ≥ 0

such that for all i ∈ 1, . . . ,m, z, z′ ∈ Rm, p, p′ ∈ S:

‖ai(t, z, p)− ai(t, z′, p′)‖ ≤ C1‖z − z′‖ei + (C1 + C2‖z‖ei)‖p− p′‖. (3.17)

We denote by Hi the minimum of the reduced Hamiltonian: Hi(t, Z, p, ν) :=

Hi(t, Z, αi(t, Z, p), p, ν). Now we define the mapping H and a by:

H(t,X,Z, p, ν) :=

m∑i=1

1(X = ei) · Hi(t, Z, p, ν), (3.18)

a(t,X,Z, p) :=

m∑i=1

1(X = ei) · ai(t, Z, p). (3.19)

Remark 3.1.8. For the sake of convenience and generality, we choose to make the as-

sumption directly on the uniqueness and the regularity of the minimizer of the Hamil-

tonian. Alternatively, we could also impose structural constraints on the transition

rate function q and require strong convexity for the running cost function f . For

example, we can make the following assumption:

Assumption 3.1.9. (i) A is a convex and compact subset of Rl

(ii) The transition rate function q takes the form q(t, i, j, α, p, ν) = q0(t, i, j, p, ν) +

q1(t, i, j, p) · α, where q0 : [0, T ]× E2 × S × P(A)→ R and q1 : [0, T ]× E2 × S → Rl

are two continuous mapping.

(iii) The running cost function f is of the form f(t,X, α, p, ν) = f0(t,X, α, p) +

f1(t,X, p, ν), where for each i ∈ 1, . . . ,m, the mapping f0(·, ei, ·, ·) (resp.

f1(·, ei, ·, ·)) is continuous in [0, T ]× A× P (resp. [0, T ]× S × P(A)).

(iv) For all (t, i, p) ∈ [0, T ]×E×S, the mapping α→ f0(t, ei, α, p) is once continuous

differentiable. Moreover there exists a constant C > 0 such that:

‖∇αf0(t, ei, α, p)−∇αf0(t, ei, α, p′)‖ ≤ C‖p− p′‖ (3.20)

63

Page 73: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(v) f0 is γ−strongly convex in α, i.e., for all (t, i, p) ∈ [0, T ]× E × S and α, α′ ∈ A,

we have:

f0(t, ei, α, p)− f0(t, ei, α′, p)− (α− α′) · ∇αf0(t, ei, α, p) ≥ γ‖α′ − α‖2 (3.21)

Under Assumption 3.1.7, it is clear from the definition of a in equation (3.19)

that a(t, x, z, p) is the unique minimizer of the mapping α → H(t, x, z, α, p, ν), and

the minimum is given by H(t, x, z, p, ν). In addition, from Assumption 3.1.3 and

Assumption 3.1.4, we can show the following result on the Lipschitz property of H:

Lemma 3.1.10. There exists a constant C > 0 such that for all (ω, t) ∈ Ω× (0, T ],

p, p′ ∈ S, ν, ν ′ ∈ P(A) and z, z′ ∈ Rm, we have:

|H(t,Xt−, z, p, ν)− H(t,Xt−, z′, p′, ν ′)|

≤ C‖z − z′‖Xt− + C(1 + ‖z‖Xt−)(‖p− p′‖+W1(ν, ν ′)),(3.22)

|a(t,Xt−, z, p)− a(t,Xt−, z′, p′)| ≤ C‖z − z′‖Xt− + C(1 + ‖z‖Xt−)‖p− p′‖. (3.23)

Proof. Inequality (3.23) is an easy consequence of Assumption 3.1.7 and the definition

of the stochastic seminorm ‖ · ‖Xt− . We now deal with the regularity of H. By

Berge’s maximum theorem, the continuity of Hi and the compactness of A imply the

continuity of Hi. Let z, z′ ∈ Rm, p, p′ ∈ S and ν, ν ′ ∈ P(A). For any α ∈ A, we have:

Hi(t, z, p, ν)−Hi(t, z′, α, p′, ν ′) ≤ Hi(t, z, α, p, ν)−Hi(t, z

′, α, p′, ν ′)

= f(t, ei, α, p, ν)− f(t, ei, α, p′, ν ′) +

∑j 6=i

[(zj − zi)− (z′j − z′i)]q(t, i, j, α, p′, ν ′)

+∑j 6=i

(zj − zi)[q0(t, i, j, p, ν)− q0(t, i, j, p′, ν ′)] + (zj − zi)[q1(t, i, j, p)− q1(t, i, j, p′)] · α

≤ C‖z − z′‖ei + C(1 + ‖z‖ei)(‖p− p′‖+W1(ν, ν ′)),

64

Page 74: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where we have used the Lipschitz property of f and q, and the boundedness of A

and q. Since the above is true for all α ∈ A, taking supremum of the left-hand side,

we obtain:

Hi(t, z, p, ν)− Hi(t, z′, p′, ν ′) ≤ C‖z − z′‖ei + C(1 + ‖z‖ei)(‖p− p′‖+W1(ν, ν ′)).

Exchanging the roles of z and z′, we obtain:

|Hi(t, z, p, ν)− Hi(t, z′, p′, ν ′)| ≤ C‖z − z′‖ei + C(1 + ‖z‖ei)(‖p− p′‖+W1(ν, ν ′)),

and (3.22) follows immediately from the definition of the seminorm ‖ · ‖Xt− .

Let us fix measurable functions p : [0, T ]→ S and ν : [0, T ]→ P(A). We consider

the following BSDE:

Yt = g(XT , pT ) +

∫ T

tH(s,Xs−, Zs, ps, νs)ds−

∫ T

tZ∗s · dMs. (3.24)

Piggybacking on the argument developed in Section 2.2.2 of Chapter 2, in partic-

ular Proposition 2.2.7 and 2.2.9, we show that the above BSDE characterizes the

optimality of the control problem (3.8), and the optimal control is unique.

Proposition 3.1.11. For any measurable function p from [0, T ] to S and any measur-

able function ν from [0, T ] to P(A), the BSDE (3.24) admits a unique solution (Y, Z).

The value function of the optimal control problem (3.8) is given by V (p, ν) = EP[Y0]

and the control α(p,ν) defined by:

α(p,ν)t = a(t,Xt−, Zt, pt). (3.25)

is optimal. In addition, if α′ ∈ A is an optimal control, we have α′t = α(p,ν)t , dt⊗ dP-

a.e.

65

Page 75: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

3.2 Existence of Nash Equilibria

Theorem 3.2.1. Under Assumption 3.1.3, 3.1.4 and 3.1.7, there exists a Nash equi-

librium (α∗, p∗, ν∗) for the weak formulation of the finite-state mean field game in the

sense of Definition 3.1.2.

The rest of this section is devoted to the proof of Theorem 3.2.1. As in the case of

diffusion-based mean field games, we shall rely on an argument of fixed point to show

the existence of Nash equilibria. We start from a measurable function p : [0, T ]→ S

and a measurable function ν : [0, T ] → P(A) where we recall that S is the m-

dimensional simplex isomorphic to the space of probability measures on E, while

P(A) is the space of probability measures on A. We then solve the BSDE (3.24) which

allows us to obtain the solution (Y (p,ν), Z(p,ν)) as well as the optimal control α(p,ν)

given by (3.25). Finally, we compute the probability measure Q(p,ν) := Q(α(p,ν),p,ν)

as defined in (3.4) and consider the push-forward measures of Q(p,ν) by Xt as well

as α(p,ν)t . Clearly, a Nash equilibrium exists if the mapping (p, ν)→ (Q(p,ν)

#Xt, Q(p,ν)

#α(p,ν)t

)

admits a fixed-point.

In practice however, the implementation of the fixed-point argument mentioned

above is prone to several difficulties. The foremost challenge lies in the fact that the

space of the mean fields of states and controls, or more specifically the space of mea-

surable mappings from [0, T ] to S or to P(A), where p or ν belong respectively, does

not have a convenient topological description to grant the compactness. This makes

it difficult to apply Schauder’s theorem or similar versions of fixed point theorems.

Due to the reasons mentioned above, we shall resort to different descriptions of

the mean fields for the states and the controls. For the mean field of the states,

since we have assumed from the very beginning that X is a cadlag process, we will

directly deal with the probability measures on D, which is the space of all cadlag

functions from [0, T ] to E = e1, . . . , em endowed with the Skorokhod topology. The

66

Page 76: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

space of probability measures on D and its topological properties have been studied

thoroughly (see Jacod and Shiryaev [1987] for a detailed account). In particular a

simple criterion for compactness is available.

On the other hand, the construction of the space for mean field on control is more

involved. Here, we adopt the same technique as used in Carmona and Lacker [2015]

based on the stable topology. Indeed, a measurable function from [0, T ] to P(A) can

be viewed as a random variable defined on the space ([0, T ],B([0, T ]),L) taking values

in P(A), where B([0, T ]) is the collection of Borel subsets of [0, T ], L is the uniform

distribution on [0, T ] and P(A) is endowed with the Wasserstein-1 distance. To obtain

compactness, the idea is to use randomization. We consider the space of probability

measures on [0, T ] × P(A), denoted by P([0, T ] × P(A)). Then for each measurable

function ν from [0, T ] to P(A), we consider the measure η on [0, T ]×P(A) given by

η(dt, dm) := L(dt)× δνt(dm) where δ is the Dirac measure. We may endow the space

P([0, T ] × P(A)) with the so-called stable topology proposed by Jacod and Memin

[1981], for which convenient results on compactness are readily available.

In the following, we detail the steps that lead to the existence of Nash equilibria.

We start by specifying the topology in the space of mean fields of the states as well as

the controls. We then properly define the mapping compatible with the definition of

the Nash equilibrium. We show the continuity and construct a compact subset that

is stable by the mapping. Once these ingredients are in place, we apply Schauder’s

fixed point theorem to conclude.

3.2.1 Topology for the Mean Field

We start by defining the topological space for the mean field of the states. To this

end, we first endow the space of states E := e1, . . . , em with the discrete metric

dE(x, y) := 1(x 6= y). It is well known that (E, dE) is a Polish space. Now let’s define

67

Page 77: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the Skorokhod space D:

D := x : [0, T ]→ E, x is cadlag and left continuous on T. (3.26)

We endow D with the J1 metric defined as follow:

dD(x, y) := infλ∈Λ

max supt∈[0,T ]

|λ(t)− t|, supt∈[0,T ]

|y(λ(t))− x(t)|, (3.27)

where Λ is the set of all strictly increasing, continuous bijections from [0, T ] to itself.

It can be proved that dD is a metric on D and the metric space (D, dD) is a Polish

space. Let us denote by P the collection of probability measures on (D, dD) endowed

with the weak topology. Recall that the reference measure P defined in the very

beginning is an element of P . Consider P0 a subset of P defined as:

P0 := Q :dQdP

= L, with EP[L2] ≤ C0, (3.28)

where C0 is a constant which we will choose later (see the proof of Proposition 3.2.10).

We have the following result:

Proposition 3.2.2. P0 is convex and relatively compact in P.

Proof. The convexity of P0 is trivial. Let us show that P0 is relatively compact. We

proceed in three steps.

Step 1 For K ∈ N and δ > 0, we define Dδ,K the collection of paths in D which

meets the following criteria: (a) the path has no more than K discontinuities, (b) the

first jump time if there is any happens on or after δ, (c) the last jump happens on or

before T − δ, and (d) the time between jumps are greater or equal than δ. We now

show that Dδ,K is compact in D. Since D is a Polish space it is enough to show the

sequential compactness. Let us fix a sequence xn in Dδ,K . For each xn, we use the

following notation: kn is the number of its jumps, δ ≤ t1n < t2n < · · · < tknn ≤ T − δ

are the moments of its jumps. ∆t1n := t1n and ∆tin := tin− ti−1n for i = 2, . . . , kn are the

68

Page 78: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

time elapsed between consecutive jumps and x0n, x

1n, . . . , t

knn are the value taken by xn

in each interval defined by the jumps. Then we can represent xn using the vector yn

of dimension 2(K + 1):

yn = [kn,∆t1n,∆t

2n, . . . ,∆t

knn , 0, . . . , 0, x

0n, x

1n, . . . , x

knn , 0, . . . , 0].

In the above representation, the first coordinate of yn is the number of jumps. Coor-

dinate 2 to K + 1 are the time elapsed between jumps defined above, and if there are

fewer than K jumps, we complete the vector by 0. Coordinates K + 2 to 2(K + 1)

are the values taken by the path x and completed with 0. Clearly there is a bijection

from xn to yn by this representation. By the definition of the set Dδ,K , we have

∆tin ∈ [δ, T ] for i ≤ kn and∑kn

i=1 ∆tin ≤ T − δ, whereas the rest of the coordinates

of yn belong to a finite set. This implies that yn lives in a compact set and therefore

we can extract a converging subsequence which we still denote by yn. Again, since kn

and the last K+ 1 components can only take finitely many values by their definition,

there exists N0 such that for n ≥ N0, we have kn = k and xin = xi for all i ≤ k. In

addition we have ∆tin converges to ∆ti for all i ≤ k, where ∆ti ≥ δ for all i ≤ k and∑ki=1 ∆ti ≤ T − ε. We consider the path represented by the vector y:

y = [k,∆t1,∆t2, . . . ,∆tk, 0, . . . , 0, x0, x1, . . . , xk, 0, . . . , 0]

Clearly x belongs to the set Dδ,K and it is straightforward to verify that xn converges

to x in the J1 metric, where xn is the path represented by the vector yn. This implies

that Dδ,K is compact.

Step 2 Now we show that for any ε > 0, there exists δ > 0 and K ∈ N such

that P(Dδ.K) ≥ 1 − ε. Recall that P is the reference measure and under P the

canonical process X is a continuous-time Markov chain with transition rate matrix

Q0. Therefore the time of the first jump, as well as the time between consecutive

jumps thereafter, which we denote by ∆t1,∆2, . . . are i.i.d. exponential random

69

Page 79: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

variables of parameter (m− 1) under the measure P. We have:

P(Dδ,K)

=P[∆t1 > T ] +K∑k=1

P

[∆t1 ≥ δ ∩ · · · ∩ ∆tk ≥ δ ∩

k+1∑i=1

∆ti > T ∩ k∑i=1

∆ti ≤ T − δ

].

For each k = 1, . . . , K, we have:

P

[∆t1 ≥ δ ∩ · · · ∩ ∆tk ≥ δ ∩

k+1∑i=1

∆ti > T ∩ k∑i=1

∆ti ≤ T − δ

]

≥P [∆t1 ≥ δ ∩ · · · ∩ ∆tk ≥ δ] + P

[k+1∑i=1

∆ti > T ∩ k∑i=1

∆ti ≤ T − δ

]− 1

=(P[∆t1 ≥ δ])k + P

[k+1∑i=1

∆ti > T ∩ k∑i=1

∆ti ≤ T − δ

]− 1

=(exp(−k(m− 1)δ)− 1) + exp(−(m− 1)T )(m− 1)k(T − δ)k

k!.

It follows that:

P(Dδ,K) ≥K∑k=1

(exp(−k(m− 1)δ)− 1) + exp(−(m− 1)T )

K∑k=0

(m− 1)k(T − δ)k

k!

≥K∑k=1

(exp(−k(m− 1)δ)− 1) + exp(−(m− 1)T )K∑k=0

(m− 1)kT k

k!− (1− exp(−(m− 1)δ)).

We can first pick K large enough such that (exp(−(m−1)T )∑K

k=0(m−1)kTk

k!) is greater

than 1− ε/2 and then pick δ small enough to make the rest of the terms greater than

−ε/2, which eventually makes P(Dδ,K) greater than 1− ε.

Step 3 Finally, we show that P0 is tight. For any ε > 0, by Step 2 we can pick δ > 0

and K ∈ N such that P(D \ Dδ,K) ≤ (ε/C0)2. For all Q ∈ P0, we have dQ/dP = L

and (EP[L2])1/2 ≤ C0 and by Cauchy-Schwartz inequality we obtain:

Q(D \Dδ,K) = EP[L · 1(x ∈ D \Dδ,K)] ≤ (EP[L2])1/2P(D \Dδ,K)1/2 ≤ ε.

70

Page 80: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

This implies the tightness of P0. Finally by Prokhorov’s Theorem we conclude that

P0 is relatively compact.

We now need to link the convergence of measures on path space to the convergence

in S, i.e. measures on state space. We define the function π:

π : [0, T ]× P(E)→S

(t, µ)→[µ#Xt(e1), µ#Xt(e2), . . . , µ#Xt(em)].

We have the following result:

Lemma 3.2.3. Let µn converge to µ in P. Then there exists a subset D(µ) of [0, T [

at most countable such that for all t 6∈ D(µ):

π(t, µn)→ π(t, µ), n→ +∞. (3.29)

Proof. Denote the set D(µ) := 0 ≤ t ≤ T, µ(Xt−Xt− 6= 0) > 0. By Lemma 3.12 in

Jacod and Shiryaev [1987], the set D(µ) is at most countable. In addition, we have

T 6∈ D(µ) since all the paths in D are left-continuous on T . In light of Proposition

3.14 in Jacod and Shiryaev [1987], we have µn#Xt converges to µ#Xt weakly for all

t 6∈ D(µ). To conclude, we use the fact that µn#Xt for all t ∈ [0, T ] and n are counting

measures on the discrete set E.

We now turn to the mean field of control. Let (P(A),W1) be the space of probabil-

ity measures on the compact set A ⊂ Rl endowed with the weak topology and metri-

cized by the Wasserstein-1 distance. (P(A),W1) is therefore a Polish space. Since A

is compact, it is easy to show that P(A) is tight and therefore by Prokhorov’s theo-

rem (P(A),W1) is compact. We endow P(A) with the Borel σ−algebra denoted by

B(P(A)). For the set [0, T ], we endow it with the Borel σ−algebra B([0, T ]) and the

71

Page 81: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

uniform distribution on [0, T ]. Finally, we construct the product space [0, T ]×P(A)

endowed with the σ-algebra B([0, T ])⊗ B(P(A)). The space of probability measures

on [0, T ]×P(A) can be viewed as a randomized version of the space of mean field of

control. We introduce the stable topology on this space:

Definition 3.2.4. Let us denote by R the space of probability measures on ([0, T ]×

P(A),B([0, T ]) ⊗ B(P(A))). We call the stable topology of R the coarsest topology

such that the mapping η →∫g(t,m)η(dt, dm) is continuous for all bounded and

measurable mappings g defined on [0, T ]×P(A) such that m→ g(t,m) is continuous

for all t ∈ [0, T ].

We collect a few useful results on the space R endowed with the stable topology.

Proposition 3.2.5. The topological space R is metrizable and Polish.

Proof. Notice that both [0, T ] and P(A) are Polish with the topology we endow with

them. This implies that the σ-algebra B([0, T ]) ⊗ B(P(A))) is separable. It follows

from Proposition 2.10 in Jacod and Memin [1981] that R is metrizable.

We now show that R is compact. Notice that for an element η in R, its first

marginal is a probability measure on [0, T ] and its second marginal is a probability

measure on P(A). It is trivial to see that both the spaces of probability measures on

[0, T ] and on P(A) are tight and therefore relatively compact by Prokhorov’s theorem.

We then apply Theorem 2.8 in Jacod and Memin [1981] and obtain the compactness

of R.

Having showed that R is compact and metrizable, we see that R is separable.

Compactness also leads to completeness. Therefore R is Polish space. Finally, we

notice that R is also sequential compact since R is metrizable.

The following result provides a more convenient way to characterize the conver-

gence in the stable topology.

72

Page 82: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Lemma 3.2.6. Denote by H the collection of mappings f of the form f(t, ν) =

1B(t) · g(ν) where B is a Borel subset of [0, T ] and g : P(A) → R is a bounded

Lipschitz function (with regard to the Wasserstein-1 distance on P(A)). Then the

stable topology introduced in Definition 3.2.4 is the coarsest topology which makes the

mapping η →∫

[0,T ]×P(A)f(t, ν)η(dt, dν) continuous for all f ∈ H.

Proof. Let H0 be the collection of mappings f of the form f(t, ν) = 1B(t) · g(ν)

where B is a Borel subset of [0, T ] and g : P(A) → R is a bounded and uniformly

continuous function. Then clearly we have H ⊂ H0. By Proposition 2.4 in Jacod and

Memin [1981], the stable topology is the coarsest topology under which the mapping

η →∫

[0,T ]×P(A)f(t, ν)η(dt, dν) is continuous for all f ∈ H0. Therefore, we only

need to show that if ηn is a sequence of elements in R such that∫f(t, ν)ηn(dt, dν)→∫

f(t, ν)η0(dt, dν) for all f ∈ H, then we have∫f(t, ν)ηn(dt, dν)→

∫f(t, ν)η0(dt, dν)

for all f ∈ H0 as well.

Now let us fix f ∈ H0 with f(t, ν) = 1B(t) · g(ν), Note that P(A) is a compact

metric space and g is a bounded, uniformly continuous and real-valued function. A

famous result from Georganopoulos [1967] (see also Miculescu [2000]) shows that g

can be approximated uniformly by bounded Lipschitz continuous function. That is,

for all ε > 0, we can find gε ∈ H such that supν∈P(A) |gε(ν) − g(ν)| ≤ ε/3. By

our assumption we have∫

1B(t)gε(ν)ηn(dt, dν) →∫

1B(t)gε(ν)η0(dt, dν). Therefore

there exists N0 such that |∫

1B(t)gε(ν)ηn(dt, dν) −∫

1B(t)gε(ν)η0(dt, dν)| ≤ ε/3 for

all n ≥ N0. Combining these facts we have, for n ≥ N0:

∣∣∣∣∫ 1B(t)g(ν)ηn(dt, dν)−∫

1B(t)g(ν)η0(dt, dν)

∣∣∣∣≤∣∣∣∣∫ 1B(t)gε(ν)ηn(dt, dν)−

∫1B(t)gε(ν)η0(dt, dν)

∣∣∣∣+

∫1B(t) |gε(ν)− g(ν)| ηn(dt, dν)

+

∫1B(t) |gε(ν)− g(ν)| η0(dt, dν)

≤ε/3 + ε/3 + ε/3 = ε,

73

Page 83: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

which shows that∫f(t, ν)ηn(dt, dν)→

∫f(t, ν)η0(dt, dν).

Now consider the following subset of R:

R0 := η ∈ R, the marginal distribution of η on [0, T ] is L. (3.30)

where L is the uniformly distributed measure on [0, T ]. We have the following result:

Lemma 3.2.7. R0 is a convex and compact subset of R.

Proof. We apply Theorem 2.8 in Jacod and Memin [1981]. In particular, we verify

without difficulty that η[0,T ], η ∈ R0 = L is compact and ηP(A), η ∈ R0 is a

subset of P(P(A)), which is relatively compact as well.

For any η ∈ R0, since the distribution on its first marginal is L, by the dis-

integration theorem, we can write η(dt, dm) = L(dt) × ηt(dm) where the mapping

[0, T ] 3 t → ηt(·) ∈ P(P(A)) is a measurable mapping and the decomposition is

unique up to almost everywhere equality. On the other hand, for any measurable

function ν : [0, T ]→ P(A), we may construct an element Ψ(ν) in R0 by:

Ψ(ν)(dt, dm) := L(dt)× δνt(dm). (3.31)

Since we have changed the way we represent the mean field of control, we need

to modify accordingly the definition of transition rate matrix as well as the cost

functionals in order to make them compatible with the randomization procedure.

For any function F : P(A) → R possibly containing other arguments, we denote

F : P(P(A)) → R by F (m) :=∫ν∈P(A)

F (ν)m(dν), which we call the randomized

version of F . Obviously we have F (δν) = F (ν). In this way, without any ambiguity,

we define q the randomized version of the rate function q, as well as its matrix

representation Q. We also define f the randomized version of cost functional f .

74

Page 84: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Since the terminal cost g does not depend on the mean field of control, we do not

need to consider its randomized version.

Recall from Assumption 3.1.7 that the minimizer ai of the reduced Hamiltonian

is only a function of t, z and p. Consequently, for H, H i, H and H i, which are

respectively the randomized version of respectively H, Hi, H and Hi, we still have:

H(t, x, z, p,m) = infα∈A

H(t, x, z, α, p,m),

a(t, x, z, p) = arg infα∈A

H(t, x, z, α, p,m).

In addition, we have the following result on the Lipschitz property of H and a:

Lemma 3.2.8. There exists a constant C > 0 such that for all (ω, t) ∈ Ω ×

(0, T ],p, p′ ∈ S, α, α′ ∈ A, Z,Z ′ ∈ Rm and m,m′ ∈ P(P(A)), we have:

|q(t, i, j, α, p,m)− q(t, i, j, α′, p′,m′)| ≤ C(‖α− α′‖+ ‖p− p′‖+ W1(m,m′)), (3.32)

|H(t,Xt−, z, p,m)− H(t,Xt−, z′, p′,m′)|

≤ C‖z − z′‖Xt− + C(1 + ‖z‖Xt−)(‖p− p′‖+ W1(m,m′)).(3.33)

Proof. For the Lipschitz property of H, we have:

|H(t,Xt−, z, p,m)− H(t,Xt−, z′, p′,m′)|

≤|H(t,Xt−, z, p,m)− H(t,Xt−, z, p,m′)|+ |H(t,Xt−, z, p,m

′)− H(t,Xt−, z′, p′,m′)|

∣∣∣∣∣∫ν∈P(A)

H(t,Xt−, z, p, ν)(m(dν)−m′(dν))

∣∣∣∣∣+

∫ν∈P(A)

|H(t,Xt−, z, p, ν)− H(t,Xt−, z′, p′, ν)|m′(dν)

∣∣∣∣∣∫ν∈P(A)

H(t,Xt−, z, p, ν)(m(dν)−m′(dν))

∣∣∣∣∣+ C(1 + ‖z‖Xt−)‖p− p′‖+ C‖z − z′‖Xt− ,

75

Page 85: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where the last inequality is due to Lemma 3.1.10. Since the space P(A) is compact

and the mapping ν → H(t,Xt−, z, p, ν) is Lipschitz, with Lipschitz constant C(1 +

‖z‖Xt−), the Kantorovich-Rubinstein duality theorem implies that:

∣∣∣∣∣∫ν∈P(A)

H(t,Xt−, z, p, ν)(m(dν)−m′(dν))

∣∣∣∣∣ ≤ C(1 + ‖z‖Xt−)W1(m,m′),

where W1(m,m′) is the Wasserstein-1 distance on the space of probability measure

on P(A) defined as followed:

W1(m,m′) := infπ∈P(P(A)×P(A))

∫P(A)×P(A)

W1(ν, ν ′)π(dν, dν ′). (3.34)

Combined with the estimation above, we obtain the desired inequality for H. The

Lipschitz property for q can be proved in the same way.

3.2.2 Mapping for the Fixed Point Argument

We now define the mapping of which the fixed points characterize Nash equilibria of

the mean field game in the weak formulation. For any (µ, η) ∈ P ×R0, where η has

the disintegration η(dt, dm) = L(dt)×ηt(dm), we consider the solution (Y (µ,η), Z(µ,η))

to the BSDE:

Yt = g(XT , pT ) +

∫ T

tH(s,Xs−, Zs, π(s, µ), ηs)ds−

∫ T

tZ∗s · dMs. (3.35)

Denote by α(µ,η) the predictable process t → a(t,Xt−, Z(µ,η)t , π(t, µ)), which is the

optimal control of the player faced with the mean field (µ, η) ∈ P × R0. Next, we

consider the scalar martingale L(µ,η) defined by:

L(µ,η)t :=

∫ t

0X∗s− · (Q(s, α(µ,η)

s , π(s, µ), ηs)−Q0) · ψ+s · dMs. (3.36)

76

Page 86: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Define the probability measure P(µ,η) by:

dP(µ,η)

dP:= E(L(µ,η))T , (3.37)

where E(L(µ,η)) is the Doleans-Dade exponential of the martingale L(µ,η). Finally we

define the mappings Φµ, Φη and Φ respectively by:

Φµ : P ×R0 → P

(µ, η)→ P(µ,η),(3.38)

Φη : P ×R0 → R0

(µ, η)→ L(dt)× δP(µ,η)

#α(µ,η)t

(dν),(3.39)

Φ : P ×R0 → P ×R0

(µ, η)→(Φµ(µ, η),Φη(µ, η)

).

(3.40)

Remark 3.2.9. Before delving into the properties of Φ, we need to show that

the mapping Φ is well-defined. More specifically, we need to show that given

(µ, η) ∈ P × R0, the output of the mapping (P(µ,η) and L(dt) × δP(µ,η)

#α(µ,η)t

(dν)) does

not depend on which solution to the BSDE (3.35) we use to construct successively

α(µ,η), L(µ,η)t and E(L(µ,η)). To this end, let us consider (Y, Z) and (Y ′, Z ′) two

solutions to BSDE (3.35), α and α′ the corresponding optimal controls, L and

L′ the corresponding martingales defined in (3.36), and Q and Q′ the resulting

probability measures defined in (3.37). By uniqueness of the solution to (3.35), we

have E[∫ T

0‖Z ′t − Zt‖2

Xt−dt]

= 0. Using the Lipschitz continuity of a and q, it is

straightforward to show E[∫ T

0‖α′t − αt‖2dt

]= 0 and eventually Q = Q′.

Proposition 3.2.10. Denote P0 the closure of the set P0 defined in (3.28). Then

the set P0 ×R0 is stable for the mapping Φ.

Proof. It suffices to show that for all (µ, η) ∈ P × R0, we have Φµ(µ, η) ∈ P0. By

the definition of P0 in (3.28), this boils down to showing that there exists a constant

77

Page 87: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

C0 > 0 such that for all (µ, η), we have:

EP[(E(L(µ,η))T )2] ≤ C0.

Let us denote Wt := E(L(µ,η))t. By Ito’s lemma we have:

d(W 2t ) = 2Wt−dWt + d[W,W ]t,

since dL(µ,η)t = X∗t− · (Q(t, α

(µ,η)t , π(t, µ), ηt)−Q0) · ψ+

t · dMt and dWt = Wt−dL(µ,η)t .

Denoting It := ψ+t · (Q∗(t, α

(µ,η)t , π(t, µ), ηt)−Q0) ·Xt−, we have:

d(W 2t ) = 2W 2

t−dL(µ,η)t +W 2

t−I∗t · d[M,M]t · It.

We know that the optional quadratic variation of M can be decomposed as:

[M,M]t = Gt + 〈M,M〉t = Gt +

∫ t

0ψsds,

where G is a martingale. Therefore we have:

d(W 2t ) = 2W 2

t−dL(µ,η)t +W 2

t−I∗t · dGt · It +W 2

t−I∗t · ψt · Itdt.

Let Tn be a sequence of stopping times converging to +∞ which localize both the

local martingales∫ t

0W 2s−dL

(µ,η)s and

∫ t0W 2s−I

∗s · dGs · Is. Then integrating the above

SDE between 0 and T ∧ Tn and taking the expectation under P we obtain:

EP[W 2T∧Tn ] = 1 + EP

[∫ T∧Tn

0W 2t−I∗t · ψt · Itdt

]= 1 + EP

[∫ T∧Tn

0W 2t−I∗t · ψt · Itdt

]≤1 +

∫ T

0EP [W 2

t∧TnI∗t∧Tn · ψt∧Tn · It∧Tn

]dt ≤ 1 + C0

∫ T

0EP[W 2

t∧Tn ].

Here we have used Tonelli’s theorem as well as the fact that I∗s · ψs · Is is bounded

by a constant C0 independent of µ, η and n, which is a consequence of the bound-

edness of the transition rate function q. Now applying Gronwall’s lemma we obtain

78

Page 88: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

EP[W 2T∧Tn ] ≤ C0, where the constant C0 does not depend on n, µ or η. Notice that

W 2T∧Tn converges to W 2

t almost surely. We then apply Fatou’s lemma and obtain

EP[W 2T ] ≤ C0.

3.2.3 Proving the Existence of Nash Equilibria

The last missing piece in applying Schauder’s fixed point theorem is to show the

continuity of the mapping Φ in P × R0, which is the product topology space of P

and R0. To this end, we show the continuity of the mapping Φµ and Φη, respectively.

Notice that both P and R0 are metrizable, so we only need to show the sequential

continuity.

Let us fix a sequence (µ(n), η(n))n≥1 converging to (µ(0), η(0)) in P × R0, with

the decomposition η(n)(dt, dν) = L(dt) × η(n)t (dν). To simplify the notation

we denote Y (µ(n),η(n)), Z(µ(n),η(n)), α(µ(n),η(n)), L(µ(n),η(n)),P(µ(n),η(n)), respectively, by

Y (n), Z(n), α(n), L(n),Q(n). We also denote by E(n) the expectation evaluated under

the measure Q(n) and p(n)t = π(t, µ(n)), whereas E still denotes the expectation under

the reference measure P.

Let us start with the proof of continuity of Φµ, or equivalently the convergence of

Q(n) to Q(0). We shall divide the proof into several intermediary results.

Lemma 3.2.11. Without loss of generality, we may assume that there exists a con-

stant C such that the mapping ‖Z(0)t ‖Xt− ≤ C for all (ω, t) ∈ Ω× [0, T ].

Proof. We consider the following ODE with unknown Vt = [V1(t), . . . , Vm(t)] ∈ Rm:

0 =dVi(t)

dt+ H i(t, V (t), p

(0)t , η

(0)t ) +

∑j 6=i

[Vj(t)− Vi(t)],

Vi(T ) = g(ei, p(0)T ), i = 1, . . . ,m.

(3.41)

Denote ζ : [0, T ] × Rm 3 (t, v) → [ζ1(t, v), . . . , ζm(t, v)] ∈ Rm where ζi(t, v) :=

H i(t, v, p(0)t , η

(0)t )+

∑j 6=i[vj−vi]. By Lemma 3.2.8, we see that t→ ζ(t, v) is measurable

79

Page 89: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

for all v ∈ Rm and v → ζ(t, v) is Lipschitz in v uniformly in t. By Theorem 1 and

Theorem 2 in Filippov [2013], the ODE (3.41) admits a unique solution on the interval

[0, T ], which is absolutely continuous. Now we define Yt =∑m

i=1 1(Xt = ei)Vi(t) and

Zt = Vt. By continuity of V , we have ∆Yt := Yt − Yt− = V ∗t · (Xt −Xt−) = Z∗t · dXt.

Applying Ito’s formula to Y , we obtain:

Yt = YT −∫ T

t

m∑i=1

1(Xt = ei)dVidt

(s)ds−∑t<s≤T

∆Ys

= g(XT , p(0)T ) +

∫ T

t

m∑i=1

1(Xs = ei)H i(s, V (s), p(0)s , η(0)

s )

+

∫ T

t

m∑i=1

1(Xs = ei)∑j 6=i

[Vj(s)− Vi(s)]−∫ T

tZ∗s · dXs

= g(XT , p(0)T ) +

∫ T

tH i(s,Xs, Zs, p

(0)s , η(0)

s )ds−∫ T

tZ∗s · dMs,

where in the last equality we used the fact that dXs = Q0Xs−ds+dMs and Vt = Zt.

Therefore (Y, Z) and (Y (0), Z(0)) solve the same BSDE. As we have discussed in

Remark 3.2.9, we may assume that Z(0) = Z. Therefore Z(0)t = V (t). It follows from

the continuity of t→ V (t) that t→ ‖Z(0)t ‖Xt− is bounded for all ω ∈ Ω.

Now we show Z(n) converges to Z(0).

Proposition 3.2.12. We have:

limn→+∞

E[∫ T

0‖Z(n)

t − Z(0)t ‖2Xt−dt

]= 0. (3.42)

Proof. By Lemma 2.4.2, we need to check

In(t) := E

[(∫ T

tH(s,Xs−, Z

(0)s , p(n)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s )ds

)2]

converges to 0 for all t ≤ T and In(t) is bounded by C uniformly in t and n. We also

need to check Jn := E[|g(XT , p(n)T ) − g(XT , p

(0)T )|2] converges to 0. By the Lipschitz

80

Page 90: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

continuity of the cost functional g and Lemma 3.2.3, we have:

Jn ≤ C‖p(n)T − p

(0)T ‖

2 = C‖π(µ(n), T )− π(µ(0), T )‖2 → 0, n→ +∞.

To check the uniform boundedness for In(t), we recall from Lemma 3.1.10 that

|H(t,Xt−, Z(0)t , p

(n)t , ηnt )− H(t,Xt−, Z

(0)t , p

(0)t , η

(0)t )|

≤ C(1 + ‖Z(0)t ‖Xt−)(‖p(n)

t − p(0)t ‖+ W1(η

(n)t , η

(0)t )),

where W1 is the Wasserstein distance on the space P(P(A)). Clearly ‖p(n)t − p

(0)t ‖

can be bounded by a constant since p(n)t is in the simplex S. On the other hand, we

have:

W1(η(n)t , η

(0)t ) ≤

∫(ν1,ν2)∈P(A)2

W1(ν1, ν2)η(n)t (dν1)η

(0)t (dν2).

Since A is compact, W1(ν1, ν2) for (ν1, ν2) ∈ P(A)2 is bounded, which implies that

W1(η(n)t , η

(0)t ) is also bounded by a constant uniformly in n and t. This implies:

In(t) ≤ CE[∫ T

t(1 + ‖Z(0)

s ‖Xs−)ds

]≤ C + C

(E[∫ T

0‖Z(0)

s ‖2Xs−ds])1/2

< +∞,

which means that In(t) is uniformly bounded in n and t. To show that In(t) converges

to 0, we write:

In(t) ≤ 2E

[(∫ T

t(H(s,Xs−, Z

(0)s , p(n)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(n)s ))ds

)2]

+ 2E

[(∫ T

t(H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s ))ds

)2]

≤ 2CE[∫ T

t(1 + ‖Z(0)

s ‖Xs−)2‖p(n)s − p(0)

s ‖2ds]

+ 2E

[(∫ T

t(H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s ))ds

)2].

81

Page 91: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

By Lemma 3.29, we have (1 + ‖Z(0)s ‖Xs−)2‖p(n)

s − p(0)s ‖2 → 0, ds × P-almost surely.

On the other hand, we have:

(1 + ‖Z(0)s ‖Xs−)2‖p(n)

s − p(0)s ‖2 ≤ C(1 + ‖Z(0)

s ‖Xs−)2,

where the right hand side is ds× P-integrable. Therefore by dominated convergence

theorem, we obtain:

limn→+∞

E[∫ T

t(1 + ‖Z(0)

s ‖Xs−)2‖p(n)s − p(0)

s ‖2ds]

= 0.

It remains to show that

Kn := E

[(∫ T

t(H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s ))dt

)2].

converges to 0. For fixed w ∈ Ω and s ≤ T , we have:

∫ T

t(H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s ))ds

=

∫ T

t

∫ν∈P(A)

H(s,Xs−, Z(0)s , p(0)

s , ν)(η(n)s − η(0)

s )(dν)ds

=

∫[0,T ]×P(A)

κ(s, ν)η(n)(ds, dν)−∫

[0,T ]×P(A)κ(s, ν)η(0)(ds, dν),

where we have defined the mapping κ(s, ν) := 1(t ≤ s ≤ T )H(s,Xs−, Z(0)s , p

(0)s , ν).

Clearly κ is continuous in ν for all s. On the other hand, by inequality (3.22) in

Lemma 3.1.10, for all t ≤ s ≤ T and ν ∈ P(A) we have:

|H(s,Xs−, Z(0)s , p(0)

s , ν)|

≤ |H(s,Xs−, 0, 0, 0)|+ C‖Z(0)s ‖Xs− + C(1 + ‖Z(0)

s ‖Xs−)(‖p(0)s ‖+W1(ν, 0)).

Therefore by Lemma 3.2.11 and the boundedness of P(A), we conclude that the

mapping (s, ν) → κ(s, ν) is bounded. By the definition of stable topology, the fact

82

Page 92: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

that η(n) → η(0) implies:

limn→+∞

∫ T

t[H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s )]ds = 0

for all ω ∈ Ω. In addition, we have:

(∫ T

t(H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s ))ds

)2

≤(T − t)∫ T

t|H(s,Xs−, Z

(0)s , p(0)

s , η(n)s )− H(s,Xs−, Z

(0)s , p(0)

s , η(0)s )|2ds

≤C∫ T

t(1 + ‖Z(0)

s ‖Xs−)2(W1(η(n)s , η(0)

s ))2ds ≤ C∫ T

t(1 + ‖Z(0)

s ‖Xs−)2ds,

and∫ Tt

(1 + ‖Z(0)s ‖Xs−)2ds is integrable. Applying once again the dominated conver-

gence theorem, we conclude that Kn converges to 0. This completes the proof.

We will also need a result on a more convenient representation of the Doleans-Dade

exponential of L(n).

Lemma 3.2.13. Denote by W (n) the Doleans-Dade exponential of L(n). Then the Ito

differential of log(W (n)) satisfies:

d[log(W(n)t )] = X∗t− · (Q(t, α

(n)t , p

(n)t , η

(n)t )−Q0 + U(t, α

(n)t , p

(n)t , η

(n)t ) ·Q0) ·Xt−dt

+X∗t− · U(t, α(n)t , p

(n)t , η

(n)t ) · dMt,

where U(t, α(n)t , p

(n)t , η

(n)t ) is the matrix with log(q(t, i, j, α

(n)t , p

(n)t , η

(n)t )) as off-

diagonal elements and zeros on the diagonal.

Proof. Since W (n) is the Doleans-Dade exponential of L(n), W (n) satisfies the SDE

dW(n)t = W

(n)t− dL

(n)t . Applying Ito’s formula and noticing that the continuous mar-

tingale part of Ln is zero, we have:

d log(W(n)t ) = dL

(n)t −∆L

(n)t + log(1 + ∆L

(n)t ).

83

Page 93: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Then using dL(n)t = X∗t− · (Q(t, α

(n)t , p

(n)t , η

(n)t )−Q0) ·ψ+

t · dMt and noticing that the

jumps of Ln are driven by the jumps of M, and hence X, we obtain:

d log(W(n)t ) =−X∗t− · (Q(t, α

(n)t , p

(n)t , η

(n)t )−Q0) · ψ+ ·Q0 ·Xt−dt+ log(1 + ∆L

(n)t )

=X∗t− · (Q(t, α(n)t , p

(n)t , η

(n)t )−Q0) ·Xt−dt+ log(1 + ∆L

(n)t ),

where we have used the fact that for all q-matrices A, we have X∗t− ·A ·ψ+ ·Q0 ·Xt− =

−X∗t− · A ·Xt−. Piggybacking on the derivation following equation (2.16) in Section

2.1.2, for Xt− = ei and Xt = ej we have:

log(1 + ∆L(n)t ) = log(q(t, i, j, α

(n)t , p

(n)t , η

(n)t )).

Using matrix notation and recalling the definition of U in the statement of Lemma

3.2.13, we may write:

log(1 + ∆L(n)t ) = X∗t− · U(t, α

(n)t , p

(n)t , η

(n)t ) ·∆Xt.

Using again the equality ∆Xt = dXt = Q0 · Xt−dt + dMt, we arrive at the desired

representation of the differential of log(Wt).

We now show that the mapping Φµ is sequentially continuous.

Proposition 3.2.14. Q(n) converges to Q(0) in P.

Proof. For two probability measures Q and Q′ in P , we define the total variation

distance dTV between Q and Q′:

dTV (Q,Q′) := sup|Q(A)−Q′(A)|, A ∈ B(D), (3.43)

where B(D) is the σ-algebra generated by the Borel sets of D. It is well-known that

convergence of total variation distance implies the weak convergence of the measure

84

Page 94: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

and hence the convergence in the topology space P . Therefore our aim is to show

that dTV (Q(n),Q(0))→ 0 as n→ +∞.

By Pinsker’s inequality, we have:

d2TV (Q(0),Q(n)) ≤ 1

2E(0)

[log

(dQ(0)

dQ(n)

)].

Since dQndP = E(Ln)T , we have:

d2TV (Q(0),Q(n)) ≤ E(0)[log(E(L(0))T )− log(E(L(n))T )].

Using Lemma 3.2.13, we have:

E(0)[log(E(L(0))T )− log(E(L(n))T )]

=E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(0)t )−Q(t, α

(n)t , p

(n)t , η

(n)t )) ·Xt−dt

]+ E(0)

[∫ T

0X∗t− · (U(t, α

(0)t , p

(0)t , η

(0)t )− U(t, α

(n)t , p

(n)t , η

(n)t )) ·Q0 ·Xt−dt

]+ E(0)

[∫ T

0X∗t− · (U(t, α

(0)t , p

(0)t , η

(0)t )− U(t, α

(n)t , p

(n)t , η

(n)t )) · dMt

].

By Assumption 3.1.3, the process t→∫ t

0X∗s−·(O(s, α

(0)s , p

(0)s , ν

(n)s )−O(s, α

(n)s , p

(n)s , ν

(n)s ))·

dMs is a martingale, and therefore has zero expectation. We now deal with the con-

vergence of the term E(0)[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(0)t )−Q(t, α

(n)t , p

(n)t , η

(n)t )) ·Xt−dt],

while the term E(0)[∫ T

0X∗t− · (U(t, α

(0)t , p

(0)t , η

(0)t ) − U(t, α

(n)t , p

(n)t , η

(n)t )) · Q0 · Xt−dt]

can be dealt with in the exact the same way. Using the Lipschitz property in Lemma

3.2.8, we obtain:

E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(0)t )−Q(t, α

(n)t , p

(n)t , η

(n)t )) ·Xt−dt

]≤E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(n)t )−Q(t, α

(n)t , p

(n)t , η

(n)t )) ·Xt−dt

]+ E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(0)t )−Q(t, α

(0)t , p

(0)t , η

(n)t )) ·Xt−dt

]

85

Page 95: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

≤ E(0)

[∫ T

0C(‖α(n)

t − α(0)t ‖+ ‖p(n)

t − p(0)t ‖)dt

]+ E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(n)t )−Q(t, α

(0)t , p

(0)t , η

(0)t )) ·Xt−dt

]≤E(0)

[∫ T

0C‖Z(n)

t − Z(0)t ‖Xt−dt

]+ E(0)

[∫ T

0C(1 + ‖Z(0)

t ‖Xt−)‖p(n)t − p

(0)t ‖dt

]+ E(0)

[∫ T

0X∗t− · (Q(t, α

(0)t , p

(0)t , η

(n)t )−Q(t, α

(0)t , p

(0)t , η

(0)t )) ·Xt−dt

].

We deal with these terms separately. For the first expectation, by Cauchy-Schwartz

inequality, we have:

(E(0)

[∫ T

0C‖Z(n)

t − Z(0)t ‖Xt−dt

])2

=

(EP[W

(0)T

∫ T

0C‖Z(n)

t − Z(0)t ‖Xt−dt

])2

≤EP[(W(0)T )2]EP

[(∫ T

0C‖Z(n)

t − Z(0)t ‖Xt−dt

)2]≤ CEP[(W

(0)T )2]EP

[∫ T

0‖Z(n)

t − Z(0)t ‖2Xt−dt

].

This converges to 0 by Proposition 3.2.12. For the second expectation, we notice

from Lemma 3.2.11 that ‖Z(0)t ‖Xt− is bounded by a constant for all (ω, t) ∈ Ω× [0, T ].

Therefore we have:

E(0)

[∫ T

0

C(1 + ‖Z(0)t ‖Xt−)‖p(n)

t − p(0)t ‖dt

]≤ C

∫ T

0

‖p(n)t − p

(0)t ‖dt,

where the right-hand side converges to 0 by dominated convergence theorem. Finally

for the third expectation, we rewrite the integrand as:

∫ T

0

X∗t− · (Q(t, α(0)t , p

(0)t , η

(n)t )−Q(t, α

(0)t , p

(0)t , η

(0)t )) ·Xt−dt

=

∫[0,T ]×P(A)

X∗t− ·Q(t, α(0)t , p

(0)t , ν) ·Xt−(η(n)(dt, dν)− η(0)(dt, dν)).

This converges to 0, since η(n) converges to η(0) in stable topology and the mapping

ν → Q(t, α(0)t , p

(0)t , ν) is continuous for all t. Notice also that the integrand is bounded

by a constant, since q is bounded according to Assumption 3.1.3. Then by dominated

86

Page 96: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

converges theorem the thrid expectation converges to 0 as well. This completes the

proof.

To show the continuity of Φη, we need the following lemma:

Lemma 3.2.15. Let (ν(n))n∈N be a sequence of measurable functions from [0, T ] to

P(A) such that∫ T

0W1(ν

(n)t , ν

(0)t ) → 0. Then L(dt) × δ

ν(n)t

(dν) converges to L(dt) ×

δν(0)t

(dν) in R0 in the sense of the stable topology.

Proof. Denote λ(n)(dt, dν) := L(dt)× δν(n)t

(dν) for n ≥ 0. Let f : [0, T ]× P(A)→ R

be a mapping of the form f(t, ν) = 1(t ∈ B) · g(ν) where B is measurable subset of

[0, T ] and g a bounded Lipschitz function on P(A). We then have:

∣∣∣∣∣∫

[0,T ]×P(A)f(t, ν)λ(n)(dt, dν)−

∫[0,T ]×P(A)

f(t, ν)λ(0)(dt, dν)

∣∣∣∣∣≤ 1

T

∫t∈B|g(ν

(n)t )− g(ν

(0)t )|dt ≤ C

∫ T

0W1(ν

(n)t , ν

(0)t )dt.

By Lemma 3.2.6, we conclude that λ(n) converges to λ(0) in the stable topology.

Proposition 3.2.16. L(·)×δQ(n)

#α(n)t

(·) converges to L(·)×δQ(0)

#α(0)t

(·) in R0 in the sense

of the stable topology.

Proof. By Lemma 3.2.15, we only need to show∫ T

0W1(Q(n)

#α(n)t

,Q(0)

#α(0)t

)dt converges to

0. Notice that:

∫ T

0W1(Q(n)

#α(n)t

,Q(0)

#α(0)t

)dt ≤∫ T

0W1(Q(n)

#α(n)t

,Q(0)

#α(n)t

)dt+

∫ T

0W1(Q(0)

#α(n)t

,Q(0)

#α(0)t

)dt.

By the definition of total variation distance (see equation 3.43), we clearly have:

dTV (Q(n)

#α(n)t

,Q(0)

#α(n)t

) ≤ dTV (Q(n),Q(0)),

which converges to 0 according to the proof of Proposition 3.2.14. By Theorem 6.16

in Villani [2008], since A is bounded and Qn#αnt∈ P(A), there exists a constant C

87

Page 97: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

such that:

W1(Q(n)

#α(n)t

,Q(0)

#α(n)t

) ≤ C · dTV (Q(n)

#α(n)t

,Q(0)

#α(n)t

).

This shows that W1(Q(n)

#α(n)t

,Q(0)

#α(n)t

) converges to 0. In addition, it is also bounded

since A is bounded. The dominated convergence theorem then implies that:

limn→+∞

∫ T

0W1(Q(n)

#α(n)t

,Q(0)

#α(n)t

)dt = 0.

Now for the other term, we have:

∫ T

0W1(Q(0)

#α(n)t

,Q(0)

#α(0)t

)dt ≤∫ T

0E(0)[‖α(n)

t − α(0)t ‖]dt

=EP[W

(0)T

∫ T

0‖α(n)

t − α(0)t ‖dt

]≤ (EP[(W

(0)T )2])1/2

(EP[T

∫ T

0‖α(n)

t − α(0)t ‖2dt

])1/2

.

Using the Lipschitz property of a and Proposition 3.2.12, we verify easily that

EP[T∫ T

0‖α(n)

t − α(0)t ‖2dt

]converges to 0.

We are now ready to show the existence of Nash equilibria.

Proof. (of Theorem 3.2.1) Consider the product space Γ := P × R endowed with

the product topology of the weak topology on P and the stable topology on R. By

Proposition 3.2.5, Γ is a metrizable Polish space. By Proposition 3.2.2 and Lemma

3.2.7, Γ0 := P0 × R0 is a compact convex subset of Γ. By Proposition 3.2.10, Γ0 is

stable for the mapping Φ defined in (3.40). In addition, we see from Proposition 3.2.14

and Proposition 3.2.16 that Φ is continuous on Γ0. Therefore, applying Schauder’s

fixed point theorem, we conclude that Φ admits a fixed point (µ∗, η∗) ∈ P0 ×R0.

Now let us define p∗t := π(t, µ∗) ∈ S and α∗t := a(t,Xt−, Z∗t , π(t, µ∗)) where

(Y ∗, Z∗) is the solution to the BSDE (3.35) with µ = µ∗ and η = η∗. We then

define P∗ := Pµ∗,η∗ and ν∗t := P∗#α∗t . Since (µ∗, η∗) is the fixed point of the mapping Φ,

we have η∗t = δν∗t and p∗ = P∗. It follows that p∗t = π(t,P∗) = [P∗(Xt = ei)]1≤i≤m−1.

By Proposition 3.1.11, we see that α∗ is the solution to the optimal control problem

88

Page 98: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

3.8 when the mean field of state is p∗ and the mean field of control is ν∗. this implies

that (α∗, p∗, ν∗) is a Nash equilibrium.

3.3 Uniqueness of the Nash Equilibrium

Assumption 3.3.1. The transition rate function q does not depend on the mean field

of state p and the mean field of control ν. The cost functional f is of the form:

f(t,X, α, p, ν) = f0(t,X, p) + f1(t,X, α) + f2(t, p, ν). (3.44)

For all t ∈ [0, T ], i ∈ 1, . . . ,m, z ∈ Rm, p ∈ S and ν ∈ P(A), the mapping

α→ Hi(t, z, α, p, ν) admits a unique minimizer denoted by ai(t, z). In addition, there

exists a constant C > 0 such that for all i ∈ 1, . . . ,m, z, z′ ∈ Rm:

‖ai(t, z)− ai(t, z′)‖ ≤ C‖z − z′‖ei . (3.45)

Here is a version of the Lasry-Lions monotonicity condition:

Assumption 3.3.2. For all p, p′ ∈ S and t ∈ [0, T ], we have:

m∑i=1

(g(ei, p)− g(ei, p′))(pi − p′i) ≥ 0, (3.46)

m∑i=1

(f0(t, ei, p)− f0(t, ei, p′))(pi − p′i) ≥ 0. (3.47)

Theorem 3.3.3. Under Assumption 3.1.3, 3.1.4, 3.3.1 and 3.3.2, there exists at most

one Nash equilibrium for the weak formulation of the finite-state mean field game.

Proof. Let (α(1), p(1), ν(1)) and (α(2), p(2), ν(2)) be two Nash equilibria of the mean field

game. For i = 1, 2, we denote by (Y (i), Z(i)) the solution to the BSDE (3.24) with

p = p(i), ν = ν(i), which is written as:

Y(i)

0 = g(XT , p(i)T ) +

∫ T

0H(t,Xt−, Z

(i)t , p

(i)t , ν

(i)t )dt−

∫ T

0(Z

(i)t )∗ · dMt.

89

Page 99: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Let us denote α(i)t := a(t, Z

(i)t ) the optimal control, Q(i) := Q(α(i),p(i),ν(i)) the con-

trolled probability measure defined in (3.4), under which M(i) := M(α(i),p(i),ν(i)) is

a martingale. In addition, we use the abbreviation f(i)t := f(t,Xt−, α

(i)t , p

(i)t , ν

(i)t ),

g(i) = g(XT , p(i)T ) and Q

(i)t := Q(t, α

(i)t ). Taking the difference of the BSDEs we

obtain:

Y(1)

0 − Y (2)0

=g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q0) · Z(1)

t −X∗t− · (Q(2)t −Q0) · Z(2)

t )dt

+

∫ T

0(Z

(1)t − Z

(2)t )∗ · dMt

=g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(1)

t )dt+

∫ T

0(Z

(1)t − Z

(2)t )∗ · dM(2)

t

=g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(2)

t )dt+

∫ T

0(Z

(1)t − Z

(2)t )∗ · dM(1)

t .

We then take the expectation under Q(1) and Q(2) respectively. Notice that EP[Y(1)

0 −

Y(2)

0 ] = EQ(1)[Y

(1)0 − Y (2)

0 ] = EQ(2)[Y

(1)0 − Y (2)

0 ], we obtain the following equality:

EQ(1)

[g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(2)

t )dt

]=EQ(2)

[g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(1)

t )dt

].

(3.48)

Now we take a closer look at the term f(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) ·Z(2)

t . We have:

f1t +X∗t− ·Q

(1)t · Z2

t = f(t,Xt−, α(1)t , p

(1)t , ν

(1)t ) +X∗t− ·Q(t, α

(1)t ) · Z(2)

t

= H(t,Xt−, α(1)t , Z

(2)t , p

(1)t , ν

(1)t )

≥ H(t,Xt−, α(2)t , Z

(2)t , p

(1)t , ν

(1)t )

= H(t,Xt−, α(2)t , Z

(2)t , p

(2)t , ν

(2)t ) + (f1(t,Xt−, p

(1)t )− f1(t,Xt−, p

(2)t ))

+ (f2(t, p(1)t , ν

(1)t )− f2(t, p

(2)t , ν

(2)t )).

90

Page 100: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Here we use the inequality:

H(t,Xt−, α(1)t , Z

(2)t , p

(1)t , ν

(1)t ) ≥ H(t,Xt−, α

(2)t , Z

(2)t , p

(1)t , ν

(1)t ),

which is due to the fact that α(2)t minimizes the Hamiltonian α→ H(t,Xt−, α, Z

(2)t , p

(1)t , ν

(1)t )

and Assumption 3.3.1 that the minimizer does not depend on the mean field terms.

It follows that:

f(1)t −f

(2)t +X∗t−(Q

(1)t −Q

(2)t )Z

(2)t ≥ (f1(t,Xt−, p

(1)t )−f1(t,Xt−, p

(2)t ))+(f2(t, p

(1)t , ν

(1)t )−f2(t, p

(2)t , ν

(2)t )).

We can interchange the indices and obtain:

f(2)t −f

(1)t +X∗t−·(Q

(2)t −Q

(1)t )·Z(1)

t ≥ (f1(t,Xt−, p(2)t )−f1(t,Xt−, p

(1)t ))+(f2(t, p

(2)t , ν

(2)t )−f2(t, p

(1)t , ν

(1)t )).

Injecting these inequalities into equation (3.48) we have:

0 = EQ(1)

[g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(2)

t )dt

]− EQ(2)

[g(1) − g(2) +

∫ T

0(f

(1)t − f

(2)t +X∗t− · (Q

(1)t −Q

(2)t ) · Z(1)

t )dt

]≥ EQ(1)

[g(1) − g(2) +

∫ T

0(f1(t,Xt−, p

(1)t )− f1(t,Xt−, p

(2)t ) + f2(t, p

(1)t , ν

(1)t )− f2(t, p

(2)t , ν

(2)t ))dt

]− EQ(2)

[g(1) − g(2) +

∫ T

0(f1(t,Xt−, p

(1)t )− f1(t,Xt−, p

(2)t ) + f2(t, p

(1)t , ν

(1)t )− f2(t, p

(2)t , ν

(2)t ))dt

]= EQ(1)

[g(1) − g(2) +

∫ T

0(f1(t,Xt−, p

(1)t )− f1(t,Xt−, p

(2)t ))dt

]− EQ(2)

[g(1) − g(2) +

∫ T

0(f1(t,Xt−, p

(1)t )− f1(t,Xt−, p

(2)t ))dt

],

where the last equality is due to the fact that [f2(t, p(2)t , ν

(2)t ) − f2(t, p

(1)t , ν

(1)t )] is

deterministic.

From Proposition 3.1.11, since α(i) is the optimal control with regard to the mean

field p(i) and ν(i), we have α(i)t = α

(i)t , dt ⊗ dP-a.e. This implies that Q(i)[Xt− =

ek] = Q(α(i),p(i),ν(i))[Xt− = ek] for all i = 1, 2 and k = 1, . . . ,m. Since (α(i), p(i), ν(i))

91

Page 101: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

is a Nash equilibrium, we have Q(α(i),p(i),ν(i))[Xt− = ek] = [p(i)t ]k. Therefore we obtain

Q(i)[Xt− = ek] = [p(i)t ]k for all i = 1, 2 and k = 1, . . . ,m. Now using Assumption

3.3.2, we have:

0 ≥m∑i=1

(g(ei, p(1)T )− g(ei, p

(2)T ))([p

(1)T ]i − [p

(2)T ]i)

+

∫ T

0

m∑i=1

(f0(t, ei, p(1)t )− f0(t, ei, p

(2)t ))([p

(1)t ]i − [p

(2)t ]i)dt ≥ 0.

(3.49)

Assume that there exists a measurable subset N of [0, T ]×Ω with strictly positve dt⊗

dQ(1) measure, such that α(1)t 6= α

(2)t on N . By Assumption 3.3.1, the mapping α →

H(t,Xt−, α, Z(2)t , p

(1)t , ν

(1)t ) admits a unique minimizer and therefore for all (t, w) ∈ N ,

we have:

H(t,Xt−, α(1)t , Z

(2)t , p

(1)t , ν

(1)t ) > H(t,Xt−, α

(2)t , Z

(2)t , p

(1)t , ν

(1)t )

Piggybacking on the argument laid out above, we see that the first inequality is strict

in (3.49) which leads to a contradiction. Therefore we have α(1)t = α

(2)t , dt⊗ dQ1-a.e.,

and dt⊗ dP-a.e., since P is equivalent to Q(1). It follows that α(1)t = α

(2)t , dt⊗ dP-a.e.

Finally, using the same type of argument as in the proof of Proposition 3.2.14, we

obtain Q(1) = Q(2) which finally leads to (p(1), µ(1)) = (p(2), µ(2)). This completes the

proof of the uniqueness.

3.4 Approximate Nash Equilibria for Games with

Finite Many Players

In this section we show that the solution of the mean field game is an approximate

Nash equilibrium for the game of finite number of players. We first set the stage for

the weak formulation of the game with N players in finite state. Recall that Ω is

the space of cadlag mappings from [0, T ] to E = e1, . . . , em which are continuous

on T , t → Xt is the canonical process and F is the natural filtration generated by

92

Page 102: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

X. Let us fix p ∈ S a distribution on state space E. Let P be the probability

on (Ω,FT ) under which X is a continuous-time Markov chain with transition rate

matrix Q0 and initial distribution p. Let ΩN be the product space of N copies

of Ω, and PN be the product probability measure of N identical copies of P. For

n = 1, . . . , N , define the process Xnt (w) := wnt of which the natural filtration is

denoted by (Fn,Nt )t∈[0,T ]. We also denote by (FNt )t∈[0,T ] the natural filtration generated

by the process (X1, X2, . . . , XN). DenoteMnt := Xn

t −X0t −∫ t

0Q0 ·Xn

s−ds. It is clear

that under PN , X1, . . . , XN are N independent continuous-time Markov chains with

initial distribution p and Q0 as the transition rate matrix, and M1, . . . ,MN are

independent FN -martingales.

Throughout this section, we adopt the following assumption:

Assumption 3.4.1. The transition rate function q does not depend on either the

mean field of state or the mean field of control.

We assume that each player can observe the entire past history of every player’s

state. We denote by AN the collection of FN -predictable processes taking values in

A. Each player n chooses a strategy αn ∈ AN . We consider the martingale L(α1,...,αN )

by:

L(α1,...,αN )t :=

∫ t

0

N∑n=1

(Xns−)∗ · (Q(s, αns )−Q0) · (ψns )+ · dMn

s . (3.50)

We define the probability measure Q(α1,...,αN ) by:

dQ(α1,...,αN )

dPN= E(α1,...,αN )

T , (3.51)

where we denote by E (α1,...,αN ) the Doleans-Dade exponential of L(α1,...,αN ). Finally

we introduce the empirical distribution of the states:

pNt :=1

N

[N∑n=1

1(Xnt = e1),

N∑n=1

1(Xnt = e2), . . . ,

N∑n=1

1(Xnt = em)

]∈ S, (3.52)

93

Page 103: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

as well as the empirical distribution of the controls:

ν(α1t , . . . , α

Nt ) :=

1

N

N∑n=1

δαnt ∈ P(A), (3.53)

where δa(·) is the Dirac measure on a. The total expected cost of player n in the

game with N players, denoted by Jn,N(α1, . . . , αN), is defined as:

Jn,N (α1, . . . , αN )

:= EQ(α1,...,αN )

[∫ T

0f(t,Xn

t , αnt , p

Nt , ν(α1

t , . . . , αNt ))dt+ g(Xn

T , pNT )

].

(3.54)

Now let us consider an equilibrium of the mean field game (α∗, p∗, ν∗) as is in Def-

inition 3.1.2. Recall that α∗ is a predictable process with respect to the natural

filtration generated by the canonical process X. For each n = 1, . . . , N , we may

define the control αn of player n by:

αn(w1, . . . , wN) := α∗(wn). (3.55)

Clearly, αn is Fn,N -predictable. In other words, it only depends on the observation

of player n’s own path. Accordingly, the strategy profile αN := (α1, . . . , αN) is a

distributed strategy profile, which means that every player’s strategy is only based

on the observation of its own path.

In the following, we will show that α is an approximate Nash equilibrium. To

this end, we first give a result on the propagation of chaos, which compares players

n’s total expected cost in the mean field game with its total expected cost in the

finite player game. To simplify the notation, we use the abbreviation (β, α−n,N) :=

(α1, . . . , αn−1, β, αn+1, . . . , αN), QN := QαN , EN := EQN and EN := E αN . We start

from the following lemmas:

94

Page 104: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Lemma 3.4.2. There exists a sequence (δN)N≥0 such that δN → 0, N → +∞ and

for all N ≥ 1, n ≤ N and t ≤ T we have:

maxEN [W2

1 (ν(βt, α−1,Nt ), ν∗t )], EN [‖pNt − p∗t ‖2]

≤ δN . (3.56)

Proof. By the definition of QN = QαN as well as the fact that (α∗, p∗, ν∗) is an

equilibrium of the mean field game, we deduce that under the measure QN , the states

of players X1t , . . . , X

Nt are independent and have the same distribution characterized

by p∗t and the controls of players α1t , . . . , α

Nt are independent and have the same

distribution ν∗t . Therefore, for i ∈ 1, . . . ,m, we have:

EN( 1

N

N∑n=1

1(Xnt = ei)− QN [X1

t = ei]

)2 =

1

N(QN [X1

t = ei]− (QN [X1t = ei])

2) ≤ 1

4N,

which leads to:

EN [‖pNt − p∗t ‖2] =

M∑i=1

EN( 1

N

N∑n=1

1(Xnt = ei)− QN [X1

t = ei]

)2 ≤ M

4N.

On the other hand, ν(βt, α−1,Nt ) and ν∗t are in P(A) with A being a compact subset

of Rd. We have:

EN [W21 (ν(βt, α

−1,Nt ), ν∗t )]

≤ CEN [W1(ν(βt, α−1,Nt ), ν∗t )] ≤ CEN [W1(ν(βt, α

−1,Nt ), ν(αN )) +W1(ν(αN ), ν∗t )]

≤ C(EN [1

N‖βt − α1,N

t ‖] + EN [W1(ν(αN ), ν∗t )]) ≤ C(1

N+ EN [W1(ν(αN ), ν∗t )]),

where C is a constant only depending on supa∈A ‖a‖ which changes its value from

line to line. Now applying Theorem 1 in Fournier and Guillin [2015], we have:

EN [W1(ν(αN ), ν∗t )]) ≤ supa∈A‖a‖·[1(d ≤ 2)(N−1/2 log(1+N)+N−2/3)+1(d > 2)(N−1/d+N−1/2)].

Combining with the estimates previously shown, we obtain the desired result.

95

Page 105: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Lemma 3.4.3. There exists a constant C which only depends on the bound of the

transition rate q, such that for all N > 0 and β ∈ AN we have:

EN(E(β,α−1,N )

T

ENT

)2 ≤ C. (3.57)

Proof. Let us denote Wt := E (β,α−1,N )t /ENt . By Ito’s formula we have:

dWt =d

(E(β,α−1,N )t

ENt

)=dE(β,α−1,N )

t −∆E(β,α−1,N )t

ENt−−E(β,α−1,N )t− (dENt −∆ENt )

(ENt−)2+ ∆Wt

=Wt−

(dE(β,α−1,N )

t −∆E(β,α−1,N )t

E(β,α−1,N )t−

− dENt −∆ENtENt−

)+ ∆Wt

Recall that:

dENtENt−

=N∑n=1

(Xnt−)∗ · (Q(t, αnt )−Q0) · (ψnt )+ · dMn

t ,

dE(β,α−1,N )t

E(β,α−1,N )t−

= (X1t−)∗ · (Q(t, βt)−Q0) · (ψ1

t )+ · dM1

t +N∑n=2

(Xnt−)∗ · (Q(t, αnt )−Q0) · (ψnt )+ · dMn

t ,

and dMnt = ∆Mn

t − Q0 ·Xnt−dt. Noticing that for n 6= 1, the jumps of Mn

t do not

result in the jumps of Wt, we obtain:

∆Wt = ∆

(E(β,α−1,N )t

E αt

)=E(β,α−1,N )t−E αt−

·(

1 + (X1t−)∗ · (Q(t, βt)−Q0) · (ψ1

t )+ ·∆M1

t

1 + (X1t−)∗ · (Q(t, α1

t )−Q0) · (ψ1t )

+ ·∆M1t

− 1

)= Wt−

(X1t−)∗ · (Q(t, βt)−Q(t, α1

t )) · (ψ1t )

+ ·∆M1t

1 + (X1t−)∗ · (Q(t, α1

t )−Q0) · (ψ1t )

+ ·∆M1t

.

Piggybacking on the computation in equation (2.16), we see that when X1t− = ei 6=

ej = Xt, we have ∆M1t = ∆X1

t = ej − ei and:

∆Wt = Wt−q(t, i, j, βt)− q(t, i, j, α1

t )

q(t, i, j, α1t )

.

Let us define Ξβt to be an m by m matrix where the diagonal elements are 0 and the

element on the i-th row and the j-th column isq(t,i,j,βt)−q(t,i,j,α1

t )

q(t,i,j,α1t )

. Then it is clear that

96

Page 106: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

∆Wt = e∗i · Ξβt · (ej − ei). It follows that:

∆Wt = Wt− · (X1t−)∗ · Ξβt ·∆M1

t .

Injecting the above equation into the Ito decomposition of Wt, we obtain:

dWt = Wt−[(Q(t, α1t )−Q(t, βt)) · (ψ1

t )+ ·Q0 ·X1

t−dt+ (X1t−)∗ · Ξβt ·∆M1

t ]

= Wt−[(Q(t, α1t )−Q(t, βt)) · (ψ1

t )+ ·Q0 ·X1

t−dt+ (X1t−)∗ · Ξβt · (dM1

t +Q∗(t, α1t ) ·X1

t−dt)].

In the second equality, we use the fact that under the measure QN , the state process

X1t has the canonical decomposition dX1

t = dM1t + Q∗(t, α1

t ) · X1t dt where M1 is

a QN -martingale. We also use the equality ∆M1t = ∆X1

t = dX1t . In addition, by

replacing X1t with ei for i = 1, . . . ,M , it is plain to check the following equality:

(Q(t, α1t )−Q(t, βt)) · (ψ1

t )+ ·Q0 ·X1

t− + (X1t−)∗ · Ξβt ·Q∗(t, α1

t ) ·X1t− = 0.

This leads to the following representation of Wt:

dWt = Wt− · (X1t−)∗ · Ξβt · dM1

t ,

which is a local martingale under the measure QN . At this stage, the rest of the

proof is exactly the same as the proof of Proposition 3.2.10. In particular, we make

use of Assumption 3.1.3, that is the transition rate q being bounded uniformly with

regard to the controls.

We are now ready to show the propagation of chaos.

Proposition 3.4.4. There exists a sequence (εN)N≥0 such that εN → 0, N → +∞

and for all N ≥ 0, n ≤ N and β ∈ AN :

∣∣∣∣Jn,N (β, α−n,N )− EQ(β,α−n,N )

[∫ T

0f(t,Xn

t , βt, p∗t , ν∗t )dt+ g(Xn

T , p∗T )

]∣∣∣∣ ≤ εN . (3.58)

97

Page 107: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Proof. Due to symmetry, we only need to show the claim for n = 1. Let N > 0 and

β ∈ AN . Using successively Cauchy-Schwartz inequality, Assumption 3.1.4, Lemma

3.4.2 and Lemma 3.4.3, we have:

∣∣∣∣∣Jn,N (β, α−1,N )− EQ(β,α−1,N )

[∫ T

0

f(t,X1t , βt, p

∗t , ν∗t )dt+ g(X1

T , p∗T )

]∣∣∣∣∣≤ EQ(β,α−1,N )

[∫ T

0

|f(t,X1t , βt, p

∗t , ν∗t )− f(t,X1

t , βt, pNt , ν(βt, α

−1,Nt ))|dt+ |g(X1

T , p∗T )− g(X1

T , pNT )|

]

= EN[E(β,α

−1,N )T

ENT

∫ T

0

|f(t,X1t , βt, p

∗t , ν∗t )− f(t,X1

t , βt, pNt , ν(βt, α

−1,Nt ))|dt+ |g(X1

T , p∗T )− g(X1

T , pNT )|

]

≤ EN(E(β,α−1,N )

T

ENT

)21/2

× EN(∫ T

0

|f(t,X1t , βt, p

∗t , ν∗t )− f(t,X1

t , βt, pNt , ν(βt, α

−1,Nt ))|dt+ |g(X1

T , p∗T )− g(X1

T , pNT )|

)21/2

≤ CEN(E(β,α−1,N )

T

ENT

)21/2

×

[∫ T

0

(EN [‖pNt − p∗t ‖2] + EN [W21 (ν(βt, α

−1,Nt ), ν∗t )])dt+ EN [‖pNT − p∗T ‖2]

]1/2≤ C

√δN ,

where δN is as appeared in Lemma 3.4.2, and C is a constant only depending on T ,

the Lipschitz constant of f and g and the constant appearing in Lemma 3.4.3. this

gives us the desired inequality.

As a direct consequence of the above result on the propagation of chaos, we show

that the Nash equilibrium of the mean field game consists of an approximate Nash

equilibrium for the game with finite many players.

Theorem 3.4.5. There exists a sequence εN converging to 0 such that for all N > 0,

β ∈ AN and n ≤ N , we have:

Jn,N (β, α−n,N ) ≤ Jn,N (α) + εN .

98

Page 108: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Proof. Recall that the strategy profile is αN = (α1, . . . , αN) is defined as:

αn(w1, w2, . . . , wN ) := α∗(wn),

where α∗ is the strategy of the mean field game equilibrium, together with p∗ as

the mean field of states and ν∗ as the mean field of control. For a strategy profile

(α1, . . . , αN) we denote Kn,N(α1, . . . , αN) by:

Kn,N (α1, . . . , αN ) := EQ(α1,...,αN )

[∫ T

0f(t,Xn

t , αnt , p∗t , ν∗t )dt+ g(Xn

T , p∗T )

]

Now taking n = 1, we observe that K1,N(αN) = EPN [Y(αN )

0 ], where Y(αN )

0 is the

solution (at time t = 0) of the following BSDE:

Yt = g(X1T , p

∗T ) +

∫ T

tH(s,X1

s−, Z1s , α

1s, p∗s, ν∗s )ds−

∫ T

t(Z1

s )∗ · dM1s. (3.59)

By the optimality of the equilibrium, we know that for all t ∈ [0, T ], α1t minimizes

the mapping α → H(t,X1t−, Z

1t , α, p

∗t , ν∗t ). Clearly, the solution of the above BSDE

(3.59) is also the unique solution to the following BSDE:

Yt = g(X1T , p

∗T ) +

∫ T

t(H(s,X1

s−, Z1s , α

1s, p∗s, ν∗s ) +

N∑n=2

(Xnt )∗ · (Q(s, αns )−Q0) · Zns ds

−∫ T

t

N∑n=1

∫ T

t(Zns )∗ · dMn

s ,

(3.60)

with Znt = 0 for n = 2, . . . , N . Indeed, the existence and uniqueness of the BSDE

(3.60) can be checked easily by applying Theorem 3.5.2. On the other hand, by

following exactly the same argument as in the proof of Lemma 2.2.6, we can show

that K1,N(β, α−1,N) = EPN [Y(β,α−1,N )

0 ], where Y(β,α−1,N )

0 is the solution (at time t = 0)

99

Page 109: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

of the following BSDE:

Yt = g(X1T , p

∗T ) +

∫ T

t(H(s,X1

s−, Z1s , βs, p

∗s, ν∗s ) +

N∑n=2

(Xnt )∗ · (Q(s, αns )−Q0) · Zns ds

−∫ T

t

N∑n=1

∫ T

t(Zns )∗ · dMn

s .

(3.61)

Notice that H(s,X1s−, Z

1s , α, p

∗s, ν∗s ) = f(s,X1

s−, α, p∗s, ν∗s )+(X1

s−)∗ ·(Q(s, α)−Q0) ·Z1s ,

and H(s,X1s−, Z

1s , α

1s, p∗s, ν∗s ) ≥ H(s,X1

s−, Z1s , βs, p

∗s, ν∗s ). Applying the comparison

principle as stated in Theorem 3.5.3 to the BSDEs (3.60) and (3.61), we conclude

that K1,N(β, α−1,N) ≤ K1,N(αN) for all β ∈ AN . Now thanks to symmetry, we have

Kn,N(β, α−n,N) ≤ K1,N(αN) for all β ∈ AN and n = 1, . . . , N . The desired results

immediately follows by applying Proposition 3.4.4.

3.5 Appendix: BSDE Driven by Multiple Inde-

pendent Continuous-Time Markov Chains

Let us consider a probability space (Ω,F ,P) supporting N independent continuous-

time Markov chains X1t , . . . , X

Nt . For each n = 1, . . . , N , Xn has mn states, which

are represented by the basis vectors of the space Rmn . We assume that under P, the

transition rate matrix of Xn is Q0,n, which is an mn by mn matrix where all the

diagonal elements equal −(mn − 1) and all the off-diagonal elements equal 1. We

denote by (Ft)t∈[0,T ] the natural filtration generated by (X1t , . . . , X

Nt ). It is clear that

for each n, we can decompose the Markov chain Xn as Xnt = Xn

0 +∫ t

0Q0,n ·Xn

s−ds+

dMnt , where Mn is an F -martingale. In addition, due to the independence of the

Markov chains, for all n1 6= n2 and t ≤ T , P-almost surely we have ∆Xn1t = 0 or

∆Xn2t = 0. In other words, any two Markov chains cannot jump simultaneously.

Let us consider the process X t := X1t ⊗ X2

t ⊗ · · · ⊗ XNt where ⊗ stands for the

Kronecker product. Indeed, X is a Markov chain encoding the joint states of the the

100

Page 110: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

N independent Markov chains, and X only takes values among the unit vectors of

the space Rm1×···×mN . We have the following result on the decomposition of X.

Lemma 3.5.1. X is a continuous-time Markov chain with transition rate matrix Q0

given by:

Q0 :=

N∑n=1

Im1 ⊗ · · · ⊗ Imn−1 ⊗Q0,n ⊗ Imn+1 ⊗ · · · ⊗ ImN . (3.62)

In addition it has the canonical decomposition:

dXt = Q0 ·Xt−dt+ dMt, (3.63)

where M is a F-martingale and satisfies:

dMt =N∑n=1

(X1t− ⊗ · · · ⊗Xn−1

t− ⊗ Imn ⊗Xn+1t− ⊗ · · · ⊗XN

t−) · dMnt . (3.64)

Proof. We show the claim for N = 2. Applying Ito’s formula to X1t ⊗X2

t and noticing

that X1t and X2

t have no simultaneous jumps, we obtain:

d(X1t ⊗X2

t ) = dX1t ⊗X2

t− +X1t− ⊗ dX2

t

= (Q0,1 ·X1t−)⊗X2

t−dt+ dM1t ⊗X2

t− +X1t− ⊗ (Q0,2 ·X2

t−)dt+X1t− ⊗ dM2

t .

Using the properties of the Kronecker product, we have:

(Q0,1 ·X1t−)⊗X2

t− = (Q0,1 ·X1t−)⊗ (Im2 ·X2

t−) = (Q0,1 ⊗ Im2) · (X1t− ⊗X2

t−)

dM1t ⊗X2

t− = (Im1 · dM1t )⊗ (X2

t− · 1)

= (Im1 ⊗X2t−) · (dM1

t ⊗ 1) = (Im1 ⊗X2t−) · dM1

t

X1t− ⊗ (Q0,2 ·X2

t−) = (Im1 ·X1t−)⊗ (Q0,2 ·X2

t−) = (Im1 ⊗Q0,2) · (X1t− ⊗X2

t−)

X1t− ⊗ dM2

t = (X1t− · 1)⊗ (Im2 · dM2

t )

= (X1t− ⊗ Im2) · (1⊗ dM2

t ) = (X1t− ⊗ Im2) · dM2

t .

101

Page 111: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Plugging the above equalities into the Ito decomposition yields the desired result for

N = 2. The case N > 2 can be treated by applying a simple argument of induction,

which we will not detail here.

As in the case of a single Markov chain, we define the stochastic matrix ψnt :=

diag(Q0,n · Xnt−) − Q0,n · diag(Xn

t−) − diag(Xnt−) · Q0,n for n = 1, . . . , N as well as

ψt := diag(Q0 ·X t−)−Q0 ·diag(X t−)−diag(X t−) ·Q0. For n = 1, . . . , N , we define

the stochastic seminorm ‖ · ‖Xnt−

by ‖Z‖2Xnt−

:= Z∗ · ψnt · Z where Z ∈ Rmn . We then

define the stochastic seminorm ‖·‖Xt− by ‖Z‖2Xt−

:= Z∗·ψt·Z whereZ ∈ Rm1×···×mN .

Our objective is to show the existence and uniqueness of the following BSDE:

Yt = ξ +

∫ T

tF (w, s, Ys, Z

1s , . . . , Z

ns )ds−

N∑n=1

∫ T

t(Zns )∗ · dMn

s . (3.65)

Here ξ is a FT -measurable P-square integrable random variable and F : Ω ×

[0, T ] × R × Rm1 × · · · × RmN → R is the driver function such that the process

t → F (w, t, y, z1, . . . , zN) is predictable for all y, z1, . . . , zN ∈ R × Rm1 × · · · × RmN .

The unknowns of the equation are a cadlag process Y taking values in R and

predictable processes Z1, . . . , ZN taking valus in Rm1 , . . . ,RmN respectively.

Theorem 3.5.2. Assume that there exists a constant C > 0 such that dt × P-a.s.,

we have:

|F (w, t, y, z1, . . . , zN )− F (w, t, y, z1, . . . , zN )|

≤ C

(|y − y|+

N∑n=1

‖zn − zn‖Xnt−

) (3.66)

Then the BSDE (3.65) admits a solution (Y, Z1, . . . , ZN) satisfying:

E[∫ T

0|Yt|2dt

]< +∞, E

[N∑n=1

∫ T

0‖Znt ‖2Xn

t−dt

]< +∞

Moreover, the solution is unique in the sense that if (Y (1), Z(1)) and (Y (2), Z(2)) are

two solutions, then Y (1) and Y (2) are indistinguishable and we have E[∫ T

0‖Z(1)

t −

Z(2)t ‖2

Xt−dt] = 0.

102

Page 112: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Proof. For simplicity of the presentation, we give the proof for N = 2, which can

be trivially generalized to any N > 2. Our first step is to show that the following

equality holds for all Z ∈ Rm1×m2 :

‖Z‖2Xt− = ‖(Im1 ⊗ (X2t−)∗) ·Z‖2X1

t−+ ‖((X1

t−)∗ ⊗ Im2) ·Z‖2X2t−

(3.67)

By the definition of the semi-norm ‖ · ‖X1t−

, we have:

‖(Im1 ⊗ (X2t−)∗) ·Z‖2X1

t−= Z∗ · (Im1 ⊗X2

t−) · ψ1t · (Im1 ⊗ (X2

t−)∗) ·Z

= Z∗ · (Im1 ⊗X2t−) · (ψ1

t ⊗ 1) · (Im1 ⊗ (X2t−)∗) ·Z = Z∗ · (ψ1

t ⊗X2t−) · (Im1 ⊗ (X2

t−)∗) ·Z

= Z∗ · [ψ1t ⊗ (X2

t− · (X2t−)∗)] ·Z = Z∗ · (ψ1

t ⊗ diag(X2t−)) ·Z.

Similarly we have ‖((X1t−)∗ ⊗ Im2) ·Z‖2

X2t−

= Z∗ · (diag(X1t−)⊗ ψ2

t ) ·Z. Now by the

definition of ψt, we have:

ψt = diag(Q0 ·Xt−)−Q0 · diag(Xt−)− diag(Xt−) ·Q0

= diag((Im1 ⊗Q0,2 +Q0,1 ⊗ Im2) · (X1t− ⊗X2

t−))− (Im1 ⊗Q0,2 +Q0,1 ⊗ Im2) · diag(X1t− ⊗X2

t−)

− diag(X1t− ⊗X2

t−) · (Im1 ⊗Q0,2 +Q0,1 ⊗ Im2)

= diag(X1t− ⊗ (Q0,2 ·X2

t−)) + diag((Q0,1 ·X1t−)⊗X2

t−)

− diag(X1t−)⊗ (Q0,2 · diag(X2

t−))− (Q0,1 · diag(X1t−))⊗ diag(X2

t−)

− diag(X1t−)⊗ (diag(X2

t−) ·Q0,2)− (diag(X1t−) ·Q0,1)⊗ diag(X2

t−)

= ψ1t ⊗ diag(X2

t−) + diag(X1t−)⊗ ψ2

t ,

where we have used the fact that for any two vectors X1, X2 we have diag(X1⊗X2) =

diag(X1)⊗diag(X2). This immediately leads to the equality (3.67). Now we consider

the BSDE driven by the continuous-time Markov chain X with terminal condition ξ

and the driver function F defined by:

F (w, t, Y,Z) := F (w, t, Y, (Im1 ⊗ (X2t−)∗) ·Z, ((X1

t−)∗ ⊗ Im2) ·Z).

103

Page 113: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

By equality (3.67) and the assumption on the regularity of F , we have:

|F (w, t, Y1,Z1)− F (w, t, Y2,Z2)|

≤ C(|Y1 − Y2|+ ‖(Im1 ⊗ (X2t−)∗) · (Z1 −Z2)‖X1

t−+ ‖((X1

t−)∗ ⊗ Im2) · (Z1 −Z2)‖X2t−

)

≤ C(|Y1 − Y2|+√

2‖Z1 −Z2‖Xt−).

Applying Lemma 2.4.1 we obtain the existence of the solution to the BSDE:

Yt = ξ +

∫ T

tF (s, Ys,Zs)ds+

∫ T

tZ∗s · dMs

Now we set Z1t := (Im1 ⊗ (X2

s−)∗) · Zs and Z2t := ((X1

s−)∗ ⊗ Im2) · Zs. From the

definition of the driver F and M in equation (3.64), we see that:

Yt = ξ +

∫ T

tF (w, s, Ys, (Im1 ⊗ (X2

s−)∗) ·Zs, ((X1s−)∗ ⊗ Im2) ·Zs)ds

+

∫ T

tZ∗s · [(Im1 ⊗X2

t−) · dM1t + (X1

t− ⊗ Im2) · dM2t ]

= ξ +

∫ T

tF (w, s, Ys, Z

1s , Z

2s )ds+

∫ T

t(Z1

s )∗ · dM1t +

∫ T

t(Z2

s )∗ · dM2t

This shows that (Y, Z1, Z2) is a solution to BSDE (3.65).

We also state a comparison principle for linear BSDEs driven by multiple inde-

pendent Markov chains.

Theorem 3.5.3. For each n ∈ 1, . . . , N, let γn be a bounded predictable process in

Rmn such that∑mn

i=1[γnt ]i = 0 for all t ∈ [0, T ], and β a bounded predictable process in

R. Let φ be a non-negative predictable process in R such that E[∫ T

0‖φt‖2dt] < +∞

and ξ a non-negative square-integrable FT measurable random variable in R. Let

(Y, Z) be the solution of the linear BSDE:

Yt = ξ +

∫ T

t(φs + βsYs +

N∑n=1

(γns )∗ · Zns )ds−N∑n=1

∫ T

t(Zns )∗ · dMn

s . (3.68)

104

Page 114: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Assume that for all n = 1, . . . , N , t ∈]0, T ] and j such that (enj )∗ ·Q0,n ·Xnt− > 0, we

have 1 + (γnt )∗ · (ψnt )+ · (enj −Xnt−) ≥ 0 where (ψnt )+ is the Moore-Penrose inverse of

the matrix ψnt . Then Y is nonnegative.

Proof. As before we treat the case for N = 2, for which the argument can be trivially

generalized to any N > 2. Since γn and β are bounded processes and∑mn

i=1[γnt ]i = 0

for all t ≤ T and n ≤ N , we easily verify that the Lipschitz condition (3.66) as stated

in Theorem 3.5.2 is satisfied and therefore the BSDE (3.68) admits a unique solution.

Now consider the following BSDE driven by M:

Yt = ξ +

∫ T

t(φs + βsYs + γ∗s ·Zs)ds−

2∑n=1

∫ T

tZ∗s · dMs, (3.69)

where γt := (γ1t ⊗X2

t−) + (X1t− ⊗ γ2

t ). It is easy to verify the BSDE (3.69) admits a

unique solution (Y, Z) and following the same argument as in the proof of Theorem

3.5.2, we verify that (Yt, Z1t , Z

2t ) := (Yt, (Im1⊗ (X2

s−)∗) ·Zs, ((X1s−)∗⊗ Im2) ·Zs) solves

the BSDE (3.68), which is also its unique solution. Therefore we only need to show

that the solution Y to BSDE (3.68) is nonnegative. To this ends, we need to apply

the comparison principal for the case of a single Markov chain, as is stated in Lemma

2.4.3. Note that X1 and X2 do not jump simultaneously and X t = X1t ⊗X2

t . For the

jump of X resulting from the jump of X1, we need to show that for k = 1, . . . ,m1:

1 + γ∗t ·ψ+t · (e1

k ⊗X2t− −X1

t− ⊗X2t−) ≥ 0. (3.70)

Let us assume that X1t− = e1

i , X2t− = e2

j . If k = i, the above equality is trivial. In

the following, we consider the case k 6= i. Then by the assumption of the theorem,

we have:

1 + (γ1t )∗ · (ψ1

t )+ · (e1

k − e1i ) ≥ 0. (3.71)

105

Page 115: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

It can be easily verified that:

(diag(e1i )⊗ ψ2

t + ψ1t ⊗ diag(e2

j )) ·

(m1 +m2 − 2)e1k ⊗ e2

j −∑k0 6=k

e1k0 ⊗ e

2j −

∑j0 6=j

e1i ⊗ e2

j0

= e1

k ⊗ e2j − e1

i ⊗ e2j ,

so that we have:

ψ+t · (e1

k ⊗X2t− −X1

t− ⊗X2t−)

=1

m1 +m2 − 1

(m1 +m2 − 2)e1k ⊗ e2

j −∑k0 6=k

e1k0 ⊗ e

2j −

∑j0 6=j

e1i ⊗ e2

j0

.It follows that:

γ∗t ·ψ+t · (e1

k ⊗X2t− −X1

t− ⊗X2t−)

=1

m1 +m2 − 1(γ1t ⊗ e2

j + e1i ⊗ γ2

t )∗ ·

(m1 +m2 − 2)e1k ⊗ e2

j −∑k0 6=k

e1k0 ⊗ e

2j −

∑j0 6=j

e1i ⊗ e2

j0

=

1

m1 +m2 − 1

(m1 +m2 − 2)(e1k)∗ · γ1

t −∑k0 6=k

(e1k0)∗ · γ1

t − (e2j )∗ · γ2

t −∑j0 6=j

(e2j0)∗ · γ2

t

=

1

m1 +m2 − 1

(m1 +m2 − 1)(e1k)∗ · γ1

t −∑k0

(e1k0)∗ · γ1

t − (e2j )∗ · γ2

t −∑j0

(e2j0)∗ · γ2

t

= (e1

k)∗ · γ1

t ,

where in the last equality we used the assumption that∑mn

i=1[γnt ]i = 0 for n = 1, 2.

Now noticing that (e1k)∗ · γ1

t = (γ1t )∗ · (ψ1

t )+ · (e1

k − e1i ), we obtain:

1 + γ∗t ·ψ+t · (e1

k ⊗X2t− −X1

t− ⊗X2t−) = 1 + (γ1

t )∗ · (ψ1t )

+ · (e1k − e1

i ).

106

Page 116: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Combining this with the inequality (3.71), we obtain the inequality (3.70). Proceed-

ing in a similar way we can also show that for k = 1, . . . ,m2:

1 + γ∗t ·ψ+t · (X1

t− ⊗ e2k −X1

t− ⊗X2t−) ≥ 0.

Applying Lemma 2.4.3 to the BSDE (3.69), we obtain the desired result.

107

Page 117: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 4

Finite State Principal Agent

Problem

In many real life situations, not only are we interested in understanding whether and

how a system of many interacting particles form equilibria, but we also attempt to

control or manage such equilibria so that the macroscopic behavior of these particles

reflects certain preferences. How should the government encourage citizens to vacci-

nate during the flu season to battle the flu outbreak? How should tax be collected

to influence people’s decision on consumption, saving and investment? How can an

employer compensate its employees in order to boost their productivity? Present

in all these scenarios are two parties: the principal - who devices a contract ac-

cording to which incentives are given to, or penalties are imposed on - the agents.

Agents are assumed to be rational, in the sense that they optimally choose their

action to maximize their utility function, which translates the tradeoff between the

rewards/penalties they received and the effort they put in. The principal’s problem

is therefore to conceive a contract that maximizes the principal’s own utility, which is

often a function of the population’s states and the transactions with the agents. To

make the model even more realistic, we can consider a situation where the principal

108

Page 118: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

can only partially observe the agent’s action. We can also model the agent’s choice

of accepting or declining the contract by introducing a reservation utility.

Historically, principal agent problems were first studied under the framework of

contract theory in economics and most contributions dealt with models involving a

principal and a single agent. In the seminal work Holmstrom and Milgrom [1991],

the authors considered a discrete-time model in which the agent’s effort influences

the drift of the state process. By assuming both the principal and the agent have a

constant absolute risk aversion (CARA) utility function, the authors showed that the

optimal contract is linear. This model was further extended in Schattler and Sung

[1993], Sung [1995], Muller [1998], and Hellwig and Schmidt [2002]. The breakthrough

in understanding the continuous-time principal agent problem can be attributed to

Sannikov [2008] and Sannikov [2012], where the author exploited the dynamic nature

of the agent’s value function stemming from the dynamic programming principal.

This remarkable observation allows the principal’s optimal contracting problem to

be formulated as a tractable optimal control problem. This approach was further

investigated and generalized in Cvitanic et al. [2015] and Cvitanic et al. [2016] with

the help of the second-order BSDE.

It is however a prevailing situation in real life applications that the principal

is faced with multiple, if not a large population of agents. Needless to say, the

interactions or competition among the agents are likely to play a key role in shaping

their behaviors. This adds an extra layer of complexity since the principal not only

needs to figure out the optimal response of the agents, but also needs to deduce the

equilibrium resulting from the interaction among the agents. Equilibria formed by a

large population of agents with complex dynamics used to be notoriously intractable,

not to mention how to solve an optimization problem on the equilibria which are

parameterized by principal’s contracts. To circumvent these difficulties, Elie et al.

[2016] borrowed the idea from the theory of mean field games and considered the

109

Page 119: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

limit scenario of infinitely many agents. Relying on the weak formulation of stochastic

differential mean field games developed in Carmona and Lacker [2015], the authors

obtained a succinct representation of the equilibrium strategies of individual agents as

well as the evolution of the equilibrium distribution across the population. Based on

this convenient description of the equilibria, the author managed to obtain a tractable

formulation of the optimal contracting problem, which can be solved by the technique

of McKean-Vlasov optimal control.

Harnessing the probabilistic approach of finite state mean field games developed in

Chapter 3, we believe that the approach adopted in Elie et al. [2016] can be extended

to principal agent problem in finite state space. In this chapter, we consider a game

involving one principal and a population of infinitely many agents whose states belong

to a finite space. Similar to the mean field game model discussed in Chapter 2 and

Chapter 3, we use a continuous-time Markov chain to model the evolution of the

agents’ states, and we assume that the transition rate between different states is a

function of the agent’s control and the statistical distribution of all the agents’ states.

At the beginning of the game, the principal chooses a continuous reward rate and a

terminal reward to be paid to each agent, which is assumed to be a function of the

past history of the agent’s state. Each agent then maximizes their objective function

by choosing their optimal effort, accounting for the principal’s reward and the states

of all the other agents. Finally, the principal records a utility which depends on the

total reward offered to the agents as well as the states of the population of the agents.

Assuming that a Nash equilibrium is formed among the population of the agents, we

would like to investigate the principal’s optimal choice of rewards to achieve the

maximal payoff.

Relying on the weak formulation developed in Chapter 3, the Nash equilibrium can

be readily described by a McKean-Vlasov backward stochastic differential equation

(BSDE). This BSDE is parameterized by the principal’s continuous payment stream,

110

Page 120: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

as well as the principal’s terminal payment as the terminal condition. Using a similar

technique in Cvitanic et al. [2016] and Elie et al. [2016], we show that controlling the

terminal condition of the BSDE is indeed equivalent to controlling the initial condition

and the martingale term of the corresponding SDE. This allows us to transform the

rather intractable formulation of the principal’s optimal contracting problem into an

optimal control problem of McKean-Vlasov SDE.

We then focus on a special case where the agents are risk-neutral in the terminal

reward, their state transition rates are linear and their instantaneous cost is quadratic

in the control. We show that the optimal contracting problem can be further simplified

to a deterministic control problem on the flow of distribution of the agents’ states,

and the agents use Markovian strategies when the principal announces the optimal

contract. By applying the Pontryagin maximum principle, the optimal strategies of

the agents and the resulting dynamics of the states under the optimal contract can

be obtained by solving a system of forward-backward ODEs.

The rest of this chapter is organized as follow. We introduce the model in Section

4.1, where we will revisit some key elements of the weak formulation of finite state

mean field games and formulate the principal’s optimal contracting problem. Our

first main result is stated in Section 4.2, where we show the equivalence between the

optimal contracting problem and a McKean-Vlasov control problem. In Section 4.3,

we deal with the linear-quadratic model, where the search for the optimal contract can

be reduced to a deterministic control problem on the flow of distribution of agents’

states. We illustrate the numerical aspect of this linear-quadratic model in Section

4.4 through an example of epidemic containment.

111

Page 121: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.1 Model Setup

4.1.1 Agent’s State Process

Given a finite time horizon T , we assume that at any time t ∈ [0, T ] each agent finds

itself in one of m different states. We denote by E := e1, e2, . . . , em the space of

these possibles states, where the ei’s are the basis vectors in Rm. As previously, we

denote by S the m-dimensional simplex S := p ∈ Rm,∑pi = 1, pi ≥ 0, which we

identify with the space of probability measures on E.

Let Ω be the space of cadlag mappings from [0, T ] to E. Let X be the canonical

process, where Xt represents a representative agent’s state at time t. We denote by

Ft := σ(Xs, s ≤ t) the natural filtration generated by X and F := FT . Now

let us fix an initial distribution for the agents’ states p ∈ S. On (Ω, (Ft)t∈[0,T ], F)

we consider the probability measure P under which the law of X0 is p and the

canonical process (Xt)t∈[0,T ] is a continuous-time Markov chain where the transition

rate between any two different states is 1. Recall that the process X has the following

representation:

Xt = X0 +

∫ t

0Q0 ·Xs−ds+Mt, (4.1)

where M is a Rm-valued P-martingale, and Q0 is the square matrix with diagonal

elements all equal to −(m− 1) and off-diagonal elements all equal to 1.

Let A be a compact subset of Rl in which the players can choose their controls.

Denote by A the collection of F -predictable processes (αt)t∈[0,T ] taking values in A.

We introduce the transition rate function q:

[0, T ]× 1, . . . ,m2 ×A× S →R

(t, i, j, α, p)→q(t, i, j, α, p),

and we denote by Q(t, α, p) the matrix [q(t, i, j, α, p)]1≤i,j≤m. In the rest of the

chapter, we make the following assumption on the transition rate:

112

Page 122: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Assumption 4.1.1. (i) For all (t, α, p) ∈ [0, T ] × A × S, the matrix Q(t, α, p) is a

Q-matrix.

(ii) There exists C1, C2 > 0 such that for all t, i, j, α, p, ν ∈ [0, T ]× E2 × A× S with

i 6= j, we have 0 < C1 < q(t, i, j, α, p) < C2.

(iii) There exists a constant C > 0 such that for all (t, i, j) ∈ [0, T ]× E2, α, α′ ∈ A,

p, p′ ∈ S, we have:

|q(t, i, j, α, p)− q(t, i, j, α′, p′)| ≤ C(‖α− α′‖+ ‖p− p′‖). (4.2)

Now we fix (p(t))0≤t≤T a flow of distributions in E such that p(0) = p, with p(t)

representing the probability distribution of the agents’ states at time t. Following the

presentation in Section 2.1.2 of Chapter 2, we construct a probability measure Q(α,p)

which is absolutely continuous with regard to P such that the process X admits the

following representation:

Xt = X0 +

∫ t

0Q∗(s, αs, p(s)) ·Xs−ds+M(α,p)

t , (4.3)

whereM(α,p)t is a martingale under the measure Q(α,p). The Radon-Nikodym deriva-

tive of Q(α,p) with regard to P is given by dQ(α,p)

dP = E(L(α,p))T , where E(L(α,p)) is the

Doleans-Dade exponential (see Definition 2.1.10) of the martingale L(α,p) defined as

follows:

dL(α,p)t = X∗t− · (Q(t, αt, p(t))−Q0) · ψ+

t · dMt, L(α,p)0 = 0.

The decomposition of X in equation (4.3) means that under the measure Q(α,p),

the stochastic intensity rate of jumps for the process X is given by the matrix

Q(t, αt, p(t)). In addition, since Q(α,p) and the reference measure P coincide on F0,

the distributions of X0 under Q(α,p) and under P are both p.

113

Page 123: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.1.2 Principal’s Contract and Agent’s Objective Function

The principal fixes a contract consisting of a payment stream (rt)t∈[0,T ] which is

a F−predictable process and a terminal payment ξ which is a FT− measurable

P−square integrable random variable. Let c : [0, T ]×E×A×S → R be the running

cost function of the agent. Let u : R → R (resp. U : R → R) be the agent’s utility

function with regard to continuous payments (resp. terminal payment). When an

agent enters into the contract (r, ξ) with the principal and the distribution of all the

agents’ states at time t is p(t) for t ≤ T , the agent’s total expected cost, Jr,ξ(α, p), is

defined as:

Jr,ξ(α, p) := EQ(α,p)

[∫ T

0[c(t,Xt, αt, p(t))− u(rt)]dt− U(ξ)

]. (4.4)

4.1.3 Principal’s Objective Function

The principal’s objective function depends on the distribution of the agents’ states

and the payments it makes to the agents. Let c0 : [0, T ]×S → R be the running cost

function and C0 : S → R the terminal cost function resulting from the distribution

of the agents’ states. Given that all the agents choose α as their controls and the

distribution of the agents’ states is given by (p(t))0≤t≤T and the principal offers a

contract (r, ξ), the principal records a total expected cost, Jα,p0 (r, ξ), defined as:

Jα,p0 (r, ξ) := EQ(α,p)

[∫ T

0[c0(t, p(t)) + rt]dt+ C0(p(T )) + ξ

]. (4.5)

114

Page 124: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.1.4 Agents’ Mean Field Nash Equilibria

We assume that the population of infinitely many agents reach a symmetric Nash

equilibrium for a given contract (r, ξ) proposed by the principal. We recall the fol-

lowing definition of the Nash equilibrium introduced in Chapter 3 and adapted to the

present situation. See Definition 3.1.2 in Section 3.1.1.

Definition 4.1.2. Let (r, ξ) be a contract, p : [0, T ] → S be a measurable mapping

such that p(0) = p, and α ∈ A be an admissible control for the agents. We say that

the tuple (α, p) is a Nash equilibrium of the agents for the contract (r, ξ), denoted by

(α, p) ∈ N (r, ξ), if:

(i) α minimizes the cost when the agent is committed to the contract (r, ξ) and the

distribution of all the agents is given by the flow p:

α = arg infα∈A

EQ(α,p)

[∫ T

0[c(t,Xt, αt, p(t))− u(rt)]dt− U(ξ)

]. (4.6)

(ii) (α, p) satisfies the consistency conditions:

p(t) = EQ(α,p)[Xt], ∀t ∈ [0, T ]. (4.7)

Note that equation (4.7) is equivalent to pi(t) = Q(α,p)[Xt = ei], ∀t ∈ [0, T ], i ∈

1, . . . ,m.

4.1.5 Principal’s Optimal Contracting Problem

We now turn to the principal’s optimal choice of the contract, which consists of

minimizing the objective function computed based on the Nash equilibria formed by

the agents. Of course we need to address the existence of Nash equilibria. However,

in the following formulation of the optimal contracting problem, we shall avoid the

problem of existence by only considering the contracts (r, ξ) that will result in at least

one Nash equilibrium. We call such contracts admissible contracts, and we denote by

115

Page 125: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

C the collection of all admissible contracts. In addition, among all the possible Nash

equilibria, we would like to disregard the equilibria in which the agent’s expected

total cost is above a given threshold κ. The motivation for imposing this additional

constraint is to model the take-it-or-leave-it behavior of the agents in the contract

theory: if the agent’s expected total cost exceeds a certain amount, it will turn down

the contract. Summarizing the constraints mentioned above, we propose the following

optimization problem for the principal:

V (κ) := inf(r,ξ)∈C

inf(α,p)∈N (r,ξ)

Jr,ξ(α,p)≤κ

EQ(α,p)

[∫ T

0[c0(t, p(t)) + rt]dt+ C0(p(T )) + ξ

], (4.8)

where we adopt the convention that the infimum over an empty set equals +∞.

4.2 A McKean-Vlasov Control Problem

The original formulation of the principal’s optimal contract is intuitive but far from

tractable. In this section we shall provide an equivalent formulation of the principal’s

optimal contracting problem which turns out to be an optimal control problem of

McKean-Vlasov type. To this end, we rely on the probabilistic characterization of

the agents’ Nash equilibria developed in Chapter 3.

4.2.1 Representation of Nash Equilibria Based on BSDEs

We define H : [0, T ] × E × Rm × A × S × R → R the Hamiltonian function for the

agent:

H(t, x, z, α, p,R) := c(t, x, α, p)−R+ x∗ · (Q(t, α, p)−Q0) · z, (4.9)

as well as the reduced Hamiltonian Hi(t, z, α, p, R) := H(t, ei, z, α, p, R).

We make the following assumption on the minimizer of the Hamiltonian:

116

Page 126: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Assumption 4.2.1. For all t ∈ [0, T ], i ∈ 1, . . . ,m, z ∈ Rm and p ∈ S, the

mapping α → Hi(t, z, α, p, R) admits a unique minimizer. We denote the minimizer

by ai(t, z, p). In addition, for all i ∈ 1, . . . ,m, ai is a measurable mapping on

[0, T ] × Rm × S, and there exists a constant C > 0 such that for all i ∈ 1, . . . ,m,

z, z′ ∈ Rm and p ∈ S:

‖ai(t, z, p)− ai(t, z′, p)‖ ≤ C‖z − z′‖ei (4.10)

The above assumption holds, for example, when the cost function c is strongly

convex in α and the transition rate function is linear in α (see Remark 3.1.8 in Chapter

3).

We denote by Hi the minimum of the reduced Hamiltonian:

Hi(t, z, p, R) := Hi(t, z, ai(t, z, p), p, R). (4.11)

Now we define the mapping H and a by:

H(t, x, z, p, R) :=m∑i=1

Hi(t, z, p, R)1(x = ei), (4.12)

a(t, x, z, p) :=m∑i=1

ai(t, z, p)1(x = ei). (4.13)

Under Assumption 4.2.1, it is clear that a(t, x, z, p) is the unique minimizer of the

mapping α → H(t, x, z, α, p, R), and the minimum is given by H(t, x, z, p, R). In

addition, from Assumption 4.1.1 and Assumption 4.2.1, we can show the following

result on the Lipschitz property of H and a.

Lemma 4.2.2. There exists a constant C > 0 such that for all (ω, t) ∈ Ω × (0, T ],

R ∈ R, p ∈ S and z, z′ ∈ Rm, we have:

|H(t,Xt−, z, p, R)− H(t,Xt−, z′, p, R)| ≤ C‖z − z′‖Xt− (4.14)

117

Page 127: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

|a(t,Xt−, z, p)− a(t,Xt−, z′, p)| ≤ C‖z − z′‖Xt− . (4.15)

Now we consider the following McKean-Vlasov BSDE:

Yt = − U(ξ) +

∫ T

tH(s,Xs−, Zs, p(s), u(rs))ds−

∫ T

tZ∗s · dMs, (4.16)

Et = 1 +

∫ t

0Es−X∗s− · (Q(s, αs, p(s))−Q0) · ψ+

s · dMs, (4.17)

αt = a(t,Xt−, Zt, p(t)), (4.18)

p(t) = EQ[Xt],dQdP

= ET . (4.19)

Definition 4.2.3. We say that a tuple (Y, Z, α, p,Q) is a solution to the McKean-

Vlasov BSDE (4.16)-(4.19) if Y is cadlag and adapted such that EP[∫ T

0Y 2t dt] < +∞,

Z is left-continuous and adapted such that EP[∫ T

0‖Zt‖2

Xt−dt] < +∞, α ∈ A, p :

[0, T ] → S is a measurable mapping, Q is a probability measure on Ω and equations

(4.16)-(4.19) are satisfied P-a.s. for all t ≤ T .

The following result links the solution of the McKean-Vlasov BSDE (4.16)-(4.19)

to the Nash equilibrium of the agents.

Theorem 4.2.4. Let Assumption 4.1.1 and Assumption 4.2.1 hold. Let (r, ξ) be a

contract. Assuming that the BSDE (4.16)-(4.19) admits a solution (Y, Z, α, p,Q),

then (α, p) is a Nash equilibrium. Conversely if (α, p) is a Nash equilibrium, then the

BSDE (4.16)-(4.19) admits a solution (Y, Z, α, p,Q) such that α = α, dP ⊗ dt-a.e.

and p(t) = p(t), dt-a.e..

Proof. The first part of the theorem is proved in Chapter 3 (see the end of Section

3.2.3). We now show the second part of the claim. Let (α, p) be a Nash equilibrium

and set Q := Q(α,p). By item (ii) in Definition 4.1.2, we see that (4.17) and (4.19) are

satisfied. Therefore it remains to show (4.16) and (4.18). Let us consider the BSDE:

Yt = −U(ξ) +

∫ T

tH(s,Xs−, Zs, αs, p(s), u(rs))ds−

∫ T

tZ∗s · dMs. (4.20)

118

Page 128: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

By the regularity of the cost function and transition rate, it can be shown that the

above BSDE admits a unique solution which we denote by (Y 0, Z0). In addition, we

have EP[Y 00 ] = Jr,ξ(α, p). On the other hand, we consider the following BSDE:

Yt = −U(ξ) +

∫ T

tH(s,Xs−, Zs, p(s), u(rs))ds−

∫ T

tZ∗s · dMs. (4.21)

From the regularity of H, we can show that BSDE (4.21) also admits a unique

solution which we denote by (Y 1, Z1). In addition, the expectation of Y 10 equals

the value function of the agent’s optimization problem facing a population of agents

with distribution p, i.e. EP[Y 10 ] = infα∈A J

r,ξ(α, p). By the definition of the Nash

equilibrium, we have:

α ∈ arg infα∈A

Jr,ξ(α, p),

which implies that EP[Y 10 ] = EP[Y 0

0 ]. Note that Y 10 and Y 0

0 are F0-measurable and P

coincides with Q on F0. Therefore we have EQ[Y 10 − Y 0

0 ] = 0. Now we set α′t :=

a(t,Xt−, Z1t , p(t)) which minimizes the mapping α → H(s,Xt−, Z

1t , α, p(t), u(rt)).

Since (Y 1, Z1) solves the BSDE (4.21), we have:

Y 1t = −U(ξ) +

∫ T

tH(s,Xs−, Z

1s , α

′t, p(s), u(rs))ds−

∫ T

t(Z1

s )∗ · dMs. (4.22)

Taking the difference of the BSDEs (4.22) and (4.20), we obtain:

Y 00 − Y 1

0 =

∫ T

0

[H(t,Xt−, Z

0t , αt, p(t), u(rt))−H(t,Xt−, Z

1t , α

′t, p(t), u(rt))

]dt−

∫ T

0

(Z0t − Z1

t )∗ · dMt

=

∫ T

0

[X∗t− · (Q(t, αt, p(t))−Q0) · Z0

t −X∗t− · (Q(t, α′t, p(t))−Q0) · Z1t

]dt

+

∫ T

0

[c(t,Xt−, αt, p(t))− c(t,Xt−, α

′t, p(t))

]dt−

∫ T

0

(Z0t − Z1

t )∗ · dMt

=

∫ T

0

[c(t,Xt−, αt, p(t))− c(t,Xt−, α

′t, p(t)) +X∗t− · (Q(t, αt, p(t))−Q(t, α′t, p(t))) · Z1

t

]dt

−∫ T

0

(Z0t − Z1

t )∗ · dM(α,p)t .

119

Page 129: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Using the optimality of αt and taking the expectation under the measure Q(α,p), we

obtain that:

0 = EQ(α,p)[H(t,Xt−, Z

1t , αt, p(t), u(rt))−H(t,Xt−, Z

1t , α

′t, p(t), u(rt))] ≥ 0.

Since the Hamiltonian admits a unique minimizer, we deduce from an argument of

contradiction that α′t = αt, dQ(α,p)⊗ dt-a.e., which implies that α′t = αt, dP⊗ dt-a.e..

Comparing BSDE (4.20) and BSDE (4.22) and using the uniqueness of the solu-

tion, we obtain E[∫ T

0‖Z1

t −Z0t ‖2

Xt−dt] = 0. Now by the regularity of a in Assumption

4.2.1, we have:

E[∫ T

0‖αt − a(t,Xt−, Z

0t , p(t))‖2dt

]= E

[∫ T

0‖α′t − a(t,Xt−, Z

0t , p(t))‖2dt

]= E

[∫ T

0‖a(t,Xt−, Z

1t , p(t))− a(t,Xt−, Z

0t , p(t))‖2dt

]≤ CE[

∫ T

0‖Z1

t − Z0t ‖2Xt−dt] = 0.

Therefore we have αt = a(t,Xt−, Z0t , p(t)), dP ⊗ dt-a.e., which immediately implies

(4.16) and (4.18). This completes the proof.

4.2.2 Principal’s Optimal Contracting Problem

We now turn to the principal’s optimal choice of the contract. Recall that we have

defined C as the collection of all contracts that result in at least one Nash equilibrium.

In addition, in order to model the fact that agents are not accepting contracts that

will incur a cost above a certain threshold, we further restrict our optimization to the

collection of Nash equilibria in which the agent’s expected total cost is below a given

threshold κ. Therefore we propose to solve the following optimization problem for

the principal:

V (κ) := inf(r,ξ)∈C

inf(α,p)∈N (r,ξ)

Jr,ξ(α,p)≤κ

EQ(α,p)

[∫ T

0[c0(t, p(t)) + rt]dt+ C0(p(T )) + ξ

], (4.23)

where we adopt the convention that the infimum over an empty set equals +∞.

120

Page 130: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Formulated in this way, the problem seems rather intractable. However, thanks to

the probabilistic characterization of the agents’ equilibria stated in Theorem 4.2.4, it

is possible to transform it into a McKean-Vlasov control problem. Let us denote by

H2X the collection of F−adapted and left-continuous processes Z such that Zt ∈ Rm

for t ∈ [0, T ] and E[∫ T

0‖Zt‖Xt−dt] < +∞. We also denote by R the collection of

F−predictable process taking values in R. Given Z ∈ H2X , r ∈ R and Y0 a F0-

measurable random variable, we consider the following SDE of McKean-Vlasov type:

Yt = Y0 −∫ t

0H(s,Xs−, Zs, p(s), u(rs))ds+

∫ t

0Z∗s · dMs, (4.24)

Et = 1 +

∫ t

0Es−X∗s− · (Q(s, αs, p(s))−Q0) · ψ+

s · dMs, (4.25)

αt = a(t,Xt−, Zt, p(t)), (4.26)

p(t) = EQ[Xt],dQdP

= ET . (4.27)

These are exactly the same equations as in the McKean-Vlasov BSDE (4.16)-(4.19),

except that we write the dynamic of Y in the forward manner. We denote the solution

by (Y (Z,r,Y0), Z(Z,r,Y0), α(Z,r,Y0), p(Z,r,Y0),P(Z,r,Y0)) and the expectation under P(Z,r,Y0) by

E(Z,r,Y0). We consider the following optimal control problem:

V (κ) := infEP[Y0]≤κ

infZ∈H2

Xr∈R

E(Z,r,Y0)

[∫ T

0[c0(t, p(Z,r,Y0)(t)) + rt]dt+ C0(p(Z,r,Y0)(T )) + U−1(−Y (Z,r,Y0)

T )

].

(4.28)

As a direct consequence of Theorem 4.2.4, we have the following result:

Theorem 4.2.5. Let Assumption 4.1.1 and Assumption 4.2.1 hold. Then V (κ) =

V (κ).

4.3 Solving the Linear-Quadratic Model

In this section, we focus on a particular setup of the principal agent problem where

the transition rate has a linear structure and the cost function is quadratic in the

121

Page 131: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

control. When the agent is risk-neutral in terms of their utility function with regard

to the terminal reward, we show that the principal’s optimal contracting problem can

be further reduced to a deterministic control problem on the probability distribution

of agents’ states.

4.3.1 Model Setup

We set the initial distribution of the agents to be p ∈ S. Each agent is allowed

to pick a control α which belongs to the set A = [α, α] ⊂ R. We assume that the

transition rate is a linear function of the control and we define:

q(t, i, j, α, p) := qi,j(t, p) + λi,j(α− α), for i 6= j, (4.29)

q(t, i, i, α, p) := −∑j 6=i

q(t, i, j, α, p), (4.30)

where λi,j ∈ R+,∑

j 6=i λi,j > 0 for all i ∈ 1, . . . ,m and qi,j : [0, T ] × S → R+ are

continuous mappings. We define the agent’s cost function (not including the utility

derived from the payment stream):

c(t, ei, α, p) := c1(t, ei, p) +γi2α2, (4.31)

with γi > 0 for all i ∈ 1, . . . ,m. Finally we define the agent’s utility function of

terminal reward to be U(ξ) = ξ and the utility function of continuous reward u to be

a concave and increasing function.

Under the setup outlined above, it is straightforward to compute the optimizer of

the Hamiltonian, which takes the form:

a(ei, Z) = b

− 1

γi

∑j 6=i

λi,j(Zj − Zi)

, (4.32)

for i ∈ 1, . . . ,m, where we have defined b(z) := minmaxz, α, α.

122

Page 132: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.3.2 Reduction to Optimal Control on Flows of Probability

Measures

We now proceed to reduce the principal’s optimal contracting problem to a determin-

istic control problem on flows of probability measures corresponding to a continuous-

time Markov chain. From the SDE (4.24) and the definition of H, we have:

Y(Z,r,Y0)T = Y0 −

∫ T

0

[c(t,Xt, a(Xt−, Zt), p(Z,r,Y0)(t))− u(rt)]dt+

∫ T

0

Z∗t · dM(Z,r,Y0)t ,

whereM(Z,r,Y0) is a martingale under the measure P(Z,r,Y0). Now using U−1(−Y (Z,r,Y0)T ) =

−Y (Z,r,Y0)T and injecting the SDE into the objective function of the principal defined

in (4.23), we may rewrite the principal’s optimal contracting problem into:

V (κ) = infEP[Y0]≤κ

infZ∈H2

Xr∈R

E(Z,r,Y0)

[ ∫ T

0

[c0(t, p(Z,r,Y0)(t)) + c(t,Xt, a(Xt−, Zt), p

(Z,r,Y0)(t)) + rt − u(rt)]dt

+ C0(p(Z,r,Y0)(T ))− Y0]

=− κ+ infZ∈H2

Xr∈R

E(Z,r,Y0)

[ ∫ T

0

[c0(t, p(Z,r,Y0)(t)) + c(t,Xt, a(Xt−, Zt), p

(Z,r,Y0)(t)) + rt − u(rt)]dt

+ C0(p(Z,r,Y0)(T ))

],

where we have used the equality EP[Y0] = E(Z,r,Y0)[Y0]. Notice that both the transition

rate matrix Q and the optimizer a do not depend on the reward r or the agent’s total

expected cost Y0, and thus we can drop the dependency of p(Z,r,Y0), P(Z,r,Y0) and

α(Z,r,Y0) on r and Y0. We can also isolate the optimal choice of r from the principal’s

optimization problem. Indeed, let r be the minimizer of the mapping r → r − u(r).

It is immediately clear that for the principal it is optimal to choose rt = r for all

t ∈ [0, T ]. It follows that:

V (κ) = infZ∈H2

X

E(Z)

[∫ T

0

(c0(t, p(Z)(t)) +

m∑i=1

1(Xt = ei)[c1(t, ei, p(Z)(t)) +

γi2a2(ei, Zt)]

)dt+ C0(p(Z)(T ))

]

− κ+ T (r − u(r)).

(4.33)

123

Page 133: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We now focus on:

W := infZ∈H2

X

E(Z)

[∫ T

0

(c0(t, p(Z)(t)) +

m∑i=1

1(Xt = ei)[c1(t, ei, p(Z)(t)) +

γi2a2(ei, Zt)]

)dt+ C0(p(Z)(T ))

],

(4.34)

where p(Z)(t) = P(Z)[Xt] and under P(Z), X has the following decomposition with

M(Z) being a P(Z)-martingale:

Xt = X0 +

∫ t

0Q∗(s, a(Xs−, Zs), p

(Z)(s)) ·Xs−ds+M(Z)t .

The key observation is that the control Z affects the value function only through

the optimal control α(Z)t := a(t,Xt−, Zt) and the mapping H2

X 3 Z → α(Z) ∈ A is

surjective. For any Z ∈ H2X , it is clear from the definition of a that α(Z) ∈ A. On the

other hand, given an arbitrary α ∈ A, we set the j-th component of the process Z to

be:

Zjt :=∑i 6=j

1(Xt− = i)γiαt

(m− 1)∑

k 6=i λi,k. (4.35)

By the boundedness of α we have Z ∈ H2X and it is easy to verify that a(t,Xt−, Zt) =

αt. Therefore we can transform the optimization problem W into W = infα∈A I(α),

where we define:

I(α) := E(α)

[∫ T

0

(c0(t, p(α)(t)) +

m∑i=1

1(Xt = ei)[c1(t, ei, p(α)(t)) +

γi2α2t ]

)dt+ C0(p(α)(T ))

].

(4.36)

Here with a mild abuse of notation, we denote by E(α) the expectation under P(α),

and P(α) is the probability measure defined by the following McKean-Vlasov SDE:

dP(α)

dP=E(α)

T , p(α)(t) = E(α)[Xt],

E(α)t =1 +

∫ t

0Es−X∗s− · (Q(s, αs, p

(α)(s))−Q0) · ψ+s · dMs.

(4.37)

124

Page 134: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Recall that under P(α), X has the decomposition:

Xt = X0 +

∫ t

0Q∗(s, αs, p

(α)(s)) ·Xs−ds+M(α)t ,

where M(α) is a P(α)-martingale. Taking expectation under P(α), we have:

p(α)(t) = p(α)(0) +

∫ t

0E(α)[Q∗(s, αs, p

(α)(s)) ·Xs−]ds

= p +

∫ t

0

m∑j=1

p(α)j (s)E(α)[Q∗(s, αs, p

(α)(s)) · ej |Xs− = ej ] ds.

Expanding further the coordinates p(α)i (t) of p(α)(t) and using the linear structure of

Q, we have:

p(α)i (t) = pi +

∫ t

0

m∑j=1

p(α)j (s)E(α)[e∗j ·Q(s, αs, p

(α)(s)) · ei|Xs− = ej ] ds

= pi +

∫ t

0

m∑j=1

p(α)j (s)E(α)[q(s, j, i, αs, p

(α)(s))|Xs− = ej ] ds

= pi +

∫ t

0

∑j 6=i

p(α)j (s)E(α)[q(s, j, i, αs, p

(α)(s))|Xs− = ej ]ds

−∫ t

0

∑j 6=i

p(α)i (s)E(α)[q(s, i, j, αs, p

(α)(s))|Xs− = ei]ds

= pi +

∫ t

0

∑j 6=i

p(α)j (s)

[λj,i(E(α)[αs|Xs− = ej ]− α) + qj,i(s, p

(α)(s))]ds

−∫ t

0

∑j 6=i

p(α)i (s)

[λi,j(E(α)[αs|Xs− = ei]− α) + qi,j(s, p

(α)(s))]ds

We see that the dynamic of p(α) is completely driven by the deterministic processes

t → E(α)[αt|Xt− = ei] for i ∈ E. We shall denote αit := E(α)[αt|Xt− = ei] and

αt := [α1t , . . . , α

mt ] in the following. On the other hand, by Jensen’s inequality, for all

α ∈ A we have:

I(α) =

∫ T

0

[c0(t, p(α)(t)) +

m∑i=1

(c1(t, ei, p

(α)(t)) +γi2E(α)[α2

t |Xt− = ei])p(α)i (t)

]dt+ C0(p(α)(T ))

≥∫ T

0

[c0(t, p(α)(t)) +

m∑i=1

(c1(t, ei, p

(α)(t)) +γi2

(E(α)[αt|Xt− = ei])2)p(α)i (t)

]dt+ C0(p(α)(T ))

125

Page 135: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

This leads to the deterministic control problem:

W := infα∈A

I(α), (4.38)

where we define A as the collection of all measurable mappings from [0, T ] to Am,

and π(α) as the solution to the following system of coupled ODEs:

dπ(α)i (t)

dt=∑j 6=i

π(α)j (t)

[λj,i(α

jt − α) + qj,i(t, π

(α)(t))]

−∑j 6=i

π(α)i (t)

[λi,j(α

it − α) + qi,j(t, π

(α)(t))], i ∈ 1, . . . ,m,

π(α)(0) =p,

(4.39)

and finally the objective function I is given by:

I(α) :=

∫ T

0

[c0(t, π(α)(t)) +

m∑i=1

[c1(t, ei, π(α)(t)) +

γi2

(αit)2]π

(α)i (t)

]dt+C0(π(α)(T )). (4.40)

The following results show that W and W are two equivalent optimization problems.

Proposition 4.3.1. We have W = W . If α is a solution to the optimization problem

W , then the predictable process α defined by αt =∑m

i=1 1(Xt− = ei)αit is an optimal

control of W . Moreover, under the probability measure at optimum, the agent’s state

evolves as a continuous-time Markov chain.

Proof. From the derivation above, we see that W ≥ W . We now show that W ≥ W .

Given any α ∈ A, we set:

αt :=

m∑i=1

1(Xt− = ei)αit. (4.41)

Clearly we have α ∈ A. By standard results on ordinary differential equation,

equation (4.39) admits a unique solution (π(α)(t))t∈[0,T ]. Now let us consider the

126

Page 136: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

probability measure P defined by:

dPdP

=ET ,

dEt =Et−X∗t− · (Q(t, αt, π(α)(t))−Q0) · ψ+

t · dMt, E0 = 1.

(4.42)

It is easy to see that under P, X has the decomposition:

Xt = X0 +

∫ t

0Q∗(s, αs, π

(α)(s)) ·Xs−ds+ Mt = X0 +

∫ t

0Q∗(s) ·Xs−ds+ Mt,

where Mt is a martingale and Q(t) is the q-matrix with components Qij(t) =

q(t, i, j, αit, π(α)(t)). This implies that X is a continuous-time Markov chain under

P and it is then straightforward to write the Kolmogorov equation for the law of X

under P. Comparing this Kolmogorov equation with the ODE (4.39), we conclude by

the uniqueness of the solution that E[Xt] = π(α)(t), where E stands for the expecta-

tion under P. Now in light of equation (4.42), we conclude that P is the solution to

the McKean-Vlasov SDE defined in (4.37) corresponding to the control α. It follows

that P = P(α) and π(α)(t) = p(α)(t). We now compute I(α):

I(α) = E(α)

[∫ T

0

[c0(t, p(α)(t)) +

m∑i=1

1(Xt = ei)[c1(t, ei, p(α)(t)) +

γi2α2t ]

]dt+ C0(p(α)(T ))

]

=

∫ T

0

[c0(t, p(α)(t)) +

m∑i=1

c1(t, ei, p(α)(t))P(α)[Xt = ei] +

m∑i=1

γi2

(αit)2P(α)[Xt = ei]

]dt

+ C0(p(α)(T ))

=

∫ T

0

[c0(t, π(α)(t)) +

m∑i=1

c1(t, ei, π(α)(t))π

(α)i (t) +

m∑i=1

γi2

(αit)2π

(α)i (t)

]dt+ C0(π(α)(T ))

= I(α).

From this, we deduce that I(α) = I(α) ≥ W and thus W ≥ W . Therefore we have

W = W . Let α ∈ A be the optimizer of W and set α ∈ A as in (4.41). Then from

the computations above, we see that W = I(α) = I(α). This immediately implies

I(α) = W and α is the optimal control of W .

127

Page 137: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

4.3.3 Construction of the Optimal Contract

We continue to investigate the deterministic control problem W as defined in equa-

tions (4.38)-(4.40). Once we have identified the optimal strategy of the agents at

the equilibrium from the control problem W , we can then provide a semi-explicit

construction for the optimal contract.

Lemma 4.3.2. An optimal control exists for the deterministic control problem W .

Proof. It is straightforward to verify that: (1) The space of controls is convex and

compact. (2) The right-hand side of the ODE (4.39) is C1 and is linear in α. (3) The

running cost is C1 and convex in α for all (t, π) ∈ [0, T ]× S and the terminal cost is

C1. This allows us to apply Theorem I.11.1 in Fleming and Soner [2006] and obtain

the existence of the optimal control.

Having verified that W admits an optimal solution, we now apply the necessary

part of the Pontryagin maximum principle (see for example Theorem I.6.3 in Fleming

and Soner [2006]), and derive a system of ODEs that characterizes the optimal control

and the corresponding flow of probability measures. The Hamiltonian of the control

problem W is a mapping from [0, T ]× S × Rm × Am to R defined by:

H(t, π, y, α) :=m∑i=1

∑j 6=i

yi [πj (λj,i(αj − α) + qj,i(t, π))− πi (λi,j(αi − α) + qi,j(t, π))]

+ c0(t, π) +m∑i=1

πi

[c1(t, ei, π) +

γi2

(αi)2].

It is straightforward to obtain:

∂αiH(t, π, y, α) = πi

∑k 6=i

(yk − yi)λi,k + γiαi

,

128

Page 138: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

and the minimizer of α → H(t, π, y, α) is a vector [a1(y), . . . , am(y)]∗ where ai(y) is

given by:

ai(y) := b(−∑k 6=i

λi,k(yk − yi)/γi),

and b(z) := minmaxz, α, α. By the necessary condition of Pontryagin’s maxi-

mum principle, if (π(t))t∈[0,T ] is the flow of measures associated with the optimal con-

trol, and (y(t))t∈[0,T ] is the adjoint process associated with π, then (π(t), y(t))t∈[0,T ] is

the solution to the following system of forward-backward ODEs:

dπ(t)

dt= ∂yH(t, π(t), y(t), a(y(t))), π(0) = p, (4.43)

dy(t)

dt= − ∂πH(t, π(t), y(t), a(y(t))), y(T ) = ∇C0(π(T )). (4.44)

Summarizing the above discussion, as well as the arguments which allows us to

reduce the optimal contracting problem to the deterministic control problem we just

solve, we provide a semi-explicit construction of the optimal contract and the optimal

strategy of the agents at the equilibrium.

Theorem 4.3.3. Let (π, y) be the solution to the system (4.43)-(4.44). We define

the process α ∈ A and Z ∈ H2X by:

αt :=m∑i=1

1(Xt− = ei)ai(y(t)), (4.45)

Zt :=[Z1t , . . . , Z

mt ]∗, Zit :=

∑j 6=i

1(Xt− = ej)γj aj(y(t))

(m− 1)∑

k 6=j λj,k. (4.46)

Let Y0 ∈ Rm such that (p)∗ · Y0 ≤ κ and r be the minimizer of the mapping

r → r − u(r). We define the random variable ξ by the following Stieltjes integral:

ξ := −X∗0 ·Y0 +

∫ T

0[c(t,Xt−, αt, π(t))− r+X∗t− ·Q(t, αt, π(t)) · Zt]dt−

∫ T

0Z∗t ·dXt−. (4.47)

129

Page 139: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Then (r, ξ) is an optimal contract. Moreover, under the optimal contract, every agent

adopts the Markovian strategy where they pick the control ai(t, y(t)) when in the state

i at time t, and the flow of distributions of agents’ states is given by (π(t))t∈[0,T ].

4.4 Application to Epidemic Containment

To illustrate the linear quadratic model studied above, we consider an example of

epidemic containment. We imagine a disease control authority aims at containing

the spread of a virus over a time period T within its jurisdiction, which consists of

two cities A and B. Each individual’s state is encoded by whether it is infected

(denoted by I) or healthy (denoted by H), and by its location (denoted by A or B).

Therefore the state space is E = AI,AH,BI,BH, and we use πAI , πAH , πBI , πBH

to denote the proportion of individuals in each of these 4 states. We assume that each

individual’s state evolves as a continuous-time Markov chain, and our modeling of the

transition rate accounts for the following set of mechanisms regarding the contraction

of the virus and the migration of the individuals between the two cities:

(1) Within each city, the rate of contracting the virus depends on the proportion

of infected individuals in the city, and accordingly we assume that the transition

rate from state AH to state AI is θ−A( πAIπAI+πAH

) and the transition rate from state

BH to state BI is θ−B( πBIπBI+πBH

). Here θ−A and θ−B are two increasing, positive and

differentiable mappings from [0, 1] to R+, and are related to the quality of health care

in city A and B, respectively.

(2) Likewise, the rate of recovery in each city is a function of the proportion of

healthy individuals in the city. We thus assume that the transition rate from state

AI to state AH is θ+A( πAH

πAI+πAH) and the transition rate from state BH to state BI

is θ+B( πBH

πBI+πBH). Similarly, θ+

A and θ+B are two increasing, positive and differentiable

130

Page 140: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

mappings from [0, 1] to R+, characterizing the quality of health care in city A and B

respectively.

(3) Each individual can choose an effort α which will give it the chance to move

to the other city. We assume that the efficacy of that effort depends on whether

the individual is healthy or infected. Accordingly, we set νIα as the transition rates

between the states AI and BI, and we set νHα as the transition rates between the

states AH and BH.

(4) To model the inflow of infection, we assume that each individual’s status of infec-

tion does not change when it moves between the cities. This means that we set the

transition rates between the state AI and BH, and the transition rates between the

state AH and BI to 0.

To summarize, we define the transition rate matrix Q to be:

Q(t, α, π) :=

· · · θ+

A( πAHπAI+πAH

) νIα 0

θ−A( πAIπAI+πAH

) · · · 0 νHα

νIα 0 · · · θ+B( πBH

πBI+πBH)

0 νHα θ−B( πBIπBI+πBH

) · · ·

. (4.48)

Here the transition rate matrix is written with the order of states AI,AH,BI,BH.

For simplicity of the notation, we also omit the diagonal elements. Indeed, since Q

is a transition rate matrix, each diagonal element equals the opposite of the sum of

off-diagonal elements in the same row.

We resume the description of our model in terms of the cost functions for the

individuals and the disease control authority. We assume that individuals living

in the same city incur the same cost, which depends on the proportion of infected

individuals in that city. On the other hand, the cost for exerting the effort to move

depends on the status of infection. Using the notations of the cost function for the

131

Page 141: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

linear quadratic model as in (4.31), we define:

c1(t, AI, π) = c1(t, AH, π) := φA

(πAI

πAI + πAH

), (4.49)

c1(t, BI, π) = c1(t, BH, π) := φB

(πBI

πBI + πBH

), (4.50)

γAI = γBI := γI , γAH = γBH := γH , (4.51)

where φA and φB are two increasing mappings on R. For the authority, we propose

to use the following running cost and terminal cost:

c0(t, π) = exp(σAπAI + σBπBI), (4.52)

C0(π) =σP · (πAI + πAH − π0A)2, (4.53)

where π0A is the population of city A at time 0. Intuitively speaking, the above cost

function translates the trade-off between the epidemic control and the population

planning. On the one hand, the authority attempts to minimize the infection rate of

both cities. On the other hand, as we shall see shortly in the numerical simulation,

individuals tends to move away from the city with higher infection rate and poorer

health care, which might result in the overpopulation of the other city. Therefore, the

authority also wishes to maintain the population of both cities at a steady level. The

coefficients σA, σB and σP reflect the importance the authority attributes to each of

these objectives.

It can be easily verified that the setup outlined above satisfies the assumptions of

the linear quadratic model studied in Section 4.3. Although the forward-backward

system of ODEs (4.43) - (4.44) which characterizes the optimal contract can be readily

derived, for the sake of completeness, we shall give the details of the equations to be

solved in the appendix of this chapter.

132

Page 142: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

For the numerical simulations, we consider a scenario in which city A has a higher

quality of health care than city B. Accordingly we set:

θ+A(q) := 0.4 · q, θ−A(q) := 0.1 · q, θ+

B(q) := 0.2 · q, θ−B(q) := 0.2 · q. (4.54)

This means that it is easier to recover and harder to get infected in city A than in

city B. In addition, we assume that individuals suffer a higher cost associated with

the epidemic in city B than in city A, and we set the cost function of individuals in

each city to be:

φA(q) := q, φB(q) := 2 · q. (4.55)

We set the maximal possible effort of individuals to α := 10 and the coefficients for

quadratic cost of efforts to γI := 2.0 and γH = 0.5. Finally, the parameters for the

cost of the authority in equation (4.52) and (4.53) are set to:

σA = σB := 1, σP := 0. (4.56)

Notice that the authority has attributed the same importance to the infection rates

of city A and city B while disregarding the problem of overpopulation.

In the following, we shall visualize the effect of the disease control authority’s

intervention by comparing the equilibrium computed from the principal agent problem

with the equilibrium from the mean field game of anarchy. By the term anarchy, we

refer to the situation where the states of individuals are still governed by the same

transition rate, but the individuals do not receive any rewards or penalties from the

authority. More specifically, the total expected cost of each individual is given by:

EQ(α,π)

[∫ T

0c(t,Xt, αt, π(t))dt

],

133

Page 143: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

with the instantaneous cost c is given by:

c(t,X, α, π) := 1(X ∈ AI,AH) · φA(

πAIπAI + πAH

)+ 1(X ∈ BI,BH) · φB

(πBI

πBI + πBH

)+ (1(X ∈ AI,BI)γI + 1(X ∈ AH,BH)γH) · α

2

2.

Applying the analytical approach of finite-state mean field games which we have

presented in Section 2.3 of Chapter 2, it is straightforward to derive the system of

forward-backward ODEs characterizing the Nash equilibrium (see equations (2.49)-

(2.50)). We display this system of ODEs in the appendix of this chapter (see Section

4.5.2).

The effect of the authority’s intervention is conspicuous in diminishing the infec-

tion rate among the entire population, as is shown in the upper panel of Figure 4.1.

However, when we visualize the respective infection rate of each city in the lower

panels of Figure 4.1, we do observe a surge of infection in city A as a result of the

authority’s intervention, although the epidemic eventually dies down. This is due

to the inflow of infected individuals from city B in the early stage of the epidemic.

Indeed, since city A provides better health care and maintains a lower rate of infec-

tion compared to city B, an infected individual has a better chance of recovery if it

moves to city A. The cost of moving prevents individuals from seeking better health

care in the scenario of anarchy, whereas individuals seem to receive subsidy to move

when the authority tries to intervene. This outcome of the authority’s intervention is

further corroborated when we visualize an individual’s optimal strategy in Figure 4.2.

We observe that individuals have a greater propensity to move when the authority

provides incentives.

134

Page 144: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

0 2 4 6 8 10Time

10

20

30

40

% In

fect

ed

Infection Rate - Entire PopulationInterventionAnarchy

0 2 4 6 8 10Time

0

10

20

30

% In

fect

ed

Infection Rate - City A

0 2 4 6 8 10Time

80

85

90

95

100Infection Rate - City B

Figure 4.1: Evolution of infection rate with and without authority’s intervention.

0 2 4 6 8 10Time

0

2

4

6

8

10

Effort of M

oving

Infected - City AAnarchyIntervention

0 2 4 6 8 10Time

0

2

4

6

8

10Healthy - City A

0 2 4 6 8 10Time

0

2

4

6

8

10

Effort of M

oving

Infected - City B

0 2 4 6 8 10Time

0

2

4

6

8

10Healthy - City B

Figure 4.2: Optimal effort of moving for an individual in different states.

135

Page 145: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Recall that in the above numerical computation, by setting σP = 0 in its termi-

nal cost function (4.53), the authority does not attempt to maintain the balance of

population between the two cities. We now investigate how the behavior of the indi-

viduals changes when the authority seeks to prevent the occurrence of overpopulation.

To this end, we redo the computations with σP = 1.5 and all the other parameters

unchanged.

0 2 4 6 8 10Time

20

30

40

50

60

70

80

90

100

% of T

otal Pop

ulation

Population of City A

0 2 4 6 8 10Time

5

10

15

20

25

30

35

40Total Infection Rate

Intervention: Gamma_P = 0Intervention: Gamma_P = 1.5Anarchy

Figure 4.3: Evolution of population in city A and the total infection rate. The solid

curve corresponds to intervention without population planning (σP = 0), the dashed-

dot curve corresponds to intervention with population planning (σP = 1.5), and the

dashed curve corresponds to the absence of intervention.

In Figure 4.3, we compare the evolution of the population in city A as well as the

total infection rate with and without the population planning. When the authority

does not try to control the flow of the population, the entire population ends up in

city A. However, when a terminal cost related to population planning is introduced,

we see a more balanced population distribution while the infection rate is still well

managed. This can be explained by Figure 4.4, from which we have a more detailed

perspective on the change of individual behavior when the authority implements the

population planning. We see that healthy individuals in city A are now encouraged

136

Page 146: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

to move to city B, in order to compensate the exodus caused by the epidemic in city

B. On the other hand, healthy individuals in city B are now incentivized to stay in

place.

0 2 4 6 8 10Time

0

2

4

6

8

10

Effort of Moving

Infected - City AIntervention: Gamma_P = 1.5Intervention: Gamma_P = 0

0 2 4 6 8 10Time

0

2

4

6

8

10Healthy - City A

0 2 4 6 8 10Time

0

2

4

6

8

10

Effort of Moving

Infected - City B

0 2 4 6 8 10Time

0

2

4

6

8

10Healthy - City B

Figure 4.4: Individual’s optimal effort of moving with and without population plan-

ning.

4.5 Appendix: System of ODEs for Epidemic Con-

tainment

4.5.1 Authority’s optimal planning

Using the transition rate (4.48) and the cost functions (4.49)-(4.53), the system of

ODEs (4.43)-(4.44) becomes:

y0(t) = − [y0(t)− y1(t)]π1(t)

(π0(t) + π1(t))2·[π1(t)(θ−A)′

(π0(t)

π0(t) + π1(t)

)+ π0(t)(θ+

A)′(

π1(t)

π0(t) + π1(t)

)]

137

Page 147: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

+ [y0(t)− y1(t)] · θ+A

(π1(t)

π0(t) + π1(t)

)− y2(t)νI a0(y(t))− σA exp(σAπ0(t) + σBπ2(t))

− 1

2γI((a0(y(t)))2 − φA

(π0(t)

π0(t) + π1(t)

)− φ′A

(π0(t)

π0(t) + π1(t)

)· π1(t)

π0(t) + π1(t),

y1(t) = − [y1(t)− y0(t)]π0(t)

(π0(t) + π1(t))2·[π1(t)(θ−A)′

(π0(t)

π0(t) + π1(t)

)+ π0(t)(θ+

A)′(

π1(t)

π0(t) + π1(t)

)]+ [y1(t)− y0(t)] · θ+

A

(π0(t)

π0(t) + π1(t)

)− y3(t)νH a1(y(t))

− 1

2γH((a1(y(t)))2 − φA

(π0(t)

π0(t) + π1(t)

)+ φ′A

(π0(t)

π0(t) + π1(t)

)· π0(t)

π0(t) + π1(t),

y2(t) = − [y2(t)− y3(t)]π3(t)

(π2(t) + π3(t))2·[π3(t)(θ−B)′

(π2(t)

π2(t) + π3(t)

)+ π2(t)(θ+

B)′(

π3(t)

π2(t) + π3(t)

)]+ [y2(t)− y3(t)] · θ+

B

(π3(t)

π2(t) + π3(t)

)− y0(t)νI a2(y(t))− σB exp(σAπ0(t) + σBπ2(t))

− 1

2γI((a2(y(t)))2 − φB

(π2(t)

π2(t) + π3(t)

)− φ′B

(π2(t)

π2(t) + π3(t)

)· π3(t)

π2(t) + π3(t),

y3(t) = − [y3(t)− y2(t)]π2(t)

(π2(t) + π3(t))2·[π3(t)(θ−B)′

(π2(t)

π2(t) + π3(t)

)+ π2(t)(θ+

B)′(

π3(t)

π2(t) + π3(t)

)]+ [y3(t)− y2(t)] · θ+

B

(π2(t)

π2(t) + π3(t)

)− y1(t)νH a3(y(t))

− 1

2γH((a3(y(t)))2 − φB

(π2(t)

π2(t) + π3(t)

)+ φ′B

(π2(t)

π2(t) + π3(t)

)· π2(t)

π2(t) + π3(t),

π0(t) = π1(t)θ−A

(π0(t)

π0(t) + π1(t)

)+ π2(t)νI a2(y)− π0(t)

[θ+A

(π1(t)

π0(t) + π1(t)

)+ νI a0(y)

],

π1(t) = π0(t)θ+A

(π1(t)

π0(t) + π1(t)

)+ π3(t)νH a3(y)− π1(t)

[θ−A

(π0(t)

π0(t) + π1(t)

)+ νH a1(y)

],

π2(t) = π3(t)θ−B

(π2(t)

π2(t) + π3(t)

)+ π0(t)νI a0(y)− π2(t)

[θ+B

(π3(t)

π2(t) + π3(t)

)+ νI a2(y)

],

π3(t) = π2(t)θ+B

(π3(t)

π2(t) + π3(t)

)+ π1(t)νH a1(y)− π3(t)

[θ−B

(π2(t)

π2(t) + π3(t)

)+ νH a3(y)

],

138

Page 148: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where the optimal control is defined by:

a0(y) := b

(νI(y0 − y2)

γI

), a1(y) := b

(νH(y1 − y3)

γH

),

a2(y) := b

(νI(y2 − y0)

γI

), a3(y) := b

(νH(y3 − y1)

γH

),

and the terminal conditions are π(0) = p and y(T ) = ∇C0(π(T )).

4.5.2 Mean field equilibrium in the absence of the authority

The system of ODEs characterizing the mean field game equilibrium consists of the

Hamilton-Jacobi equation and the Kolmogorov equation.

v0(t) = [v1(t)− v0(t)]θ+A

(π1(t)

π0(t) + π1(t)

)+ [v2(t)− v0(t)]νI a0(v(t)) +

1

2γI(a0(v(t)))2

+ φA

(π0(t)

π0(t) + π1(t)

),

v1(t) = [v0(t)− v1(t)]θ−A

(π0(t)

π0(t) + π1(t)

)+ [v3(t)− v1(t)]νH a1(v(t)) +

1

2γH(a1(v(t)))2

+ φA

(π0(t)

π0(t) + π1(t)

),

v2(t) = [v3(t)− v2(t)]θ+B

(π3(t)

π2(t) + π3(t)

)+ [v0(t)− v2(t)]νI a2(v(t)) +

1

2γI(a2(v(t)))2

+ φB

(π2(t)

π2(t) + π3(t)

),

v3(t) = [v2(t)− v3(t)]θ−B

(π2(t)

π2(t) + π3(t)

)+ [v1(t)− v3(t)]νH a3(v(t)) +

1

2γH(a3(v(t)))2

+ φB

(π2(t)

π2(t) + π3(t)

),

139

Page 149: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

π0(t) = π1(t)θ−A

(π0(t)

π0(t) + π1(t)

)+ π2(t)νI a2(v(t))− π0(t)

[θ+A

(π1(t)

π0(t) + π1(t)

)+ νI a0(v(t))

],

π1(t) = π0(t)θ+A

(π1(t)

π0(t) + π1(t)

)+ π3(t)νH a3(v(t))− π1(t)

[θ−A

(π0(t)

π0(t) + π1(t)

)+ νH a1(v(t))

],

π2(t) = π3(t)θ−B

(π2(t)

π2(t) + π3(t)

)+ π0(t)νI a0(v(t))− π2(t)

[θ+B

(π3(t)

π2(t) + π3(t)

)+ νI a2(v(t))

],

π3(t) = π2(t)θ+B

(π3(t)

π2(t) + π3(t)

)+ π1(t)νH a1(v(t))− π3(t)

[θ−B

(π2(t)

π2(t) + π3(t)

)+ νH a3(v(t))

],

where the optimal control is defined by:

a0(v) := b

(νI(v0 − v2)

γI

), a1(v) := b

(νH(v1 − v3)

γH

),

a2(v) := b

(νI(v2 − v0)

γI

), a3(v) := b

(νH(v3 − v1)

γH

),

and the terminal conditions are π(0) = p and v(T ) = 0.

140

Page 150: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 5

Finite State Mean Field Game

with Major and Minor Players

Mean field games with major and minor players were introduced to accommodate

the presence of one single player whose influence on the behavior of the remaining

population does not vanish in the asymptotic regime of large games. In this chapter

we develop the theory of these dynamic games when the states of the players belong

to a finite space. In Chapter 2 and Chapter 3, we have given a brief account of

the theory of finite state mean field games for a single homogeneous population of

players via both the analytical and the probabilistic method. The present chapter

is concerned with the extension of the analytical method to models with major and

minor players. We search for closed-loop Nash equilibria and for this reason, we

use the approach which was advocated in Carmona and Wang [2017], and called an

alternative approach in Chapter 13 of Carmona and Delarue [2017].

Our interest in mean field games with major and minor players when the state

space is finite was sparked by the four-state model by Kolokoltsov and Bensoussan

[2016] for the behavior of computer owners facing cyber attacks. Even though the

model was not introduced and treated as a game with major and minor players,

141

Page 151: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

clearly, it is of this type if the behaviors of the attacker and the targets are strategic.

Practical applications amenable to these models abound and a better theoretical

understanding of their structures should lead to sorely needed numerical procedures

to compute Nash equilibria.

Early forms of mean field games with major and minor players appeared in Huang

[2010] in an infinite-horizon setting, in Nguyen and Huang [2012a] for finite time

horizons, and Nourian and Caines [2013] offered a first generalization to non linear-

quadratic cases. In these models, the state of the major player does not enter the

dynamics of the states of the minor players: it only appears in their cost functionals.

This was remedied in Nguyen and Huang [2012b] for linear quadratic models. The

recent technical report Jaimungal and Nourian [2015] adds a major player to the

particular case (without idiosyncratic random shocks) of an extended mean field game

model of optimal execution introduced in Chapter 1 and solved in Chapter 4 of

Carmona and Delarue [2017].

The asymmetry between major and minor players was emphasized in Bensoussan

et al. [2016a] where the authors insist on the fact that the statistical distribution of

the state of a generic minor player should be derived endogenously. Like Nourian

et al. [2011], Bensoussan et al. [2016a] characterizes the limiting problem by a set

of stochastic partial differential equations. However, Bensoussan et al. [2016a] seems

to be solving a Stackelberg game, a particular type of equilibrium we treated in

Chapter 4. Recall that in a Stackelberg game, the major player first chooses an

action and the population of minor players then reach a Nash equilibrium according

to the major player’s strategy. A Stackelberg equilibrium is reached when the major

player optimizes its cost function which depends on the distribution of the minor

players’ states at the Nash equilibrium. In some sense, a Stackelberg game can be

formulated as an optimal control problem of Nash equilibria. This is in contrast with

142

Page 152: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Carmona and Zhu [2016] which also insists on the endogenous nature of the statistical

distribution of the state of a generic minor player, but which formulates the search

for a mean field equilibrium as the search for a Nash equilibrium in a two player game

over the time evolutions of states, some of which being of McKean-Vlasov type.

In this chapter, we cast the search for Nash equilibria as a search for fixed points

of the best response function constructed from the optimization problems of both

types of players. Typically, in a mean field game with major and minor players, the

dynamics of the state X0t and the cost functions of the major player depend upon

the statistical distribution µt of the state Xt of a generic minor player. Meanwhile,

the dynamics of the state Xt and the cost of a generic minor player depend upon

the values of the state X0t and the control α0

t of the major player as well as the

statistical distribution µt which captures the mean field interactions between the

minor players. In the rest of this chapter, we shall refer to this generic minor player

as the deviating minor player, since our objective is to study the optimal response

of any given minor player given the strategies of its peers and the major player.

As the first important result in this chapter, we prove that the processes (X0t , µt)

and (X0t , Xt, µt) are Markovian and we characterize their laws by their infinitesimal

generators. We start from the finite player version of the model and show convergence

when the number of minor players goes to infinity. We rely on standard results from

the theory of the convergence of Markov semigroups.

Note that the control of the major player implicitly influences µt through the

major player’s state, so the major player’s optimization problem should be treated

as an optimal control problem for McKean-Vlasov dynamics. On the other hand,

for the deviating minor player’s problem, we are just dealing with a classical Markov

decision problem in continuous time. This allows us to adapt to the finite state space

the approach introduced in Carmona and Wang [2017] and reviewed in Chapter 13 of

143

Page 153: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Carmona and Delarue [2017], to define and construct Nash equilibria. We emphasize

that these are Nash equilibria for the whole system major + minor players and not

only for the minor players. This is fully justified by our results on the propagation of

chaos and their applications to the proof that our mean field game equilibria provide

approximate Nash equilibria for games with finitely many players, including both

major and minor players.

The rest of this chapter is structured as follows. Games with finitely many minor

players and a major player are introduced in Section 5.1 where we explain the conven-

tions and notations we use to describe continuous-time controlled Markov processes

in finite state spaces. We also identify the major and minor players by specifying the

information structures available to them, the types of actions they can take, and the

costs they incur. The short and non-technical Section 5.2 describes the mean field

game strategy and emphasizes the steps needed in the search for Nash equilibria of

the system. This is in contrast with some earlier works where the formulation of the

problem lead to Stackelberg equilibria, in which only the minor players end up in an

approximate Nash equilibrium. To keep with the intuition that the mean field game

strategy is to implement a form of limit when the number of minor players grows to

infinity, Section 5.3 considers the convergence of the Markov processes describing the

states of the major + minor players system in this limit, and identifies the optimiza-

tion problems which the major player and the deviating minor player need to solve in

order to construct their best responses. This leads to the formalization of the search

for a mean field equilibrium as the search of fixed points for the best response map

constructed in this way. The optimization problems underpinning the definition of

the best response map are studied in Section 5.4. There, we use dynamic program-

ming to prove that the value functions of these optimization problems are viscosity

solutions of HJB type Partial Integro - Differential Equations (PIDEs). Section 5.5

proves existence of the best response map and of Nash equilibria under reasonable

144

Page 154: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

conditions. Next, Section 5.6 gives a verification theorem based on the existence of

a classical solution to the master equation. The longer Section 5.7 proves that the

solution of the mean field game problem provides approximate Nash equilibria for the

finite player games. This vindicates our formulation as the right formulation of the

problem if the goal is to find Nash equilibria for the system including both major and

minor players. The proof is technical but standard in the literature on mean field

games and it relies on the results of propagation of chaos. However, the latter are

usually derived for stochastic differential systems with mean field interactions, and

because we could not find the results we needed in the existing literature, we provide

proofs of the main steps of the derivations of these results in the context of controlled

Markov evolutions in finite state spaces. Finally, an appendix provides the proofs of

some of the technical results we used in the text.

5.1 Game Model with Finitely Many Players

We consider a stochastic game in continuous time, involving a major player indexed

by 0, and N minor players indexed from 1 to N . The states of all the players

X0t , X

1t , . . . , X

Nt are described by a continuous-time finite-state Markov process. Let

us denote 1, 2, . . . ,m0 the set of possible values taken by the state X0t of the major

player, and 1, 2, . . . ,m the set of possible values taken by the state Xnt of the minor

players. We introduce the empirical distribution of the states of the minor players at

time t:

µNt =

[1

N

N∑n=1

1(Xnt = 1),

1

N

N∑n=1

1(Xnt = 2), . . . ,

1

N

N∑n=1

1(Xnt = m− 1)

].

We denote by S the following set:

S := x ∈ Rm−1|xi ≥ 0,∑

xi ≤ 1.

145

Page 155: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

This is the convex hull of the (m − 1)-dimensional simplex, which can be identified

with the space of probability measures on the state space of the minor players. We

consider Markovian dynamics in continuous time, according to which the rates of

jumps of the state of a generic minor player depend upon the value of its control,

the empirical distribution of the states of all the minor players, as well as the major

player’s control and state. We denote by A0 (resp. A) a convex set in which the

major player (resp. all the minor players) can choose their controls. We introduce a

function q:

[0, T ]× 1, . . . ,m2 × A× 1, . . . ,m0 × A0 × S →R

(t, i, j, α, i0, α0, x)→q(t, i, j, α, i0, α0, x),

and we make the following assumption on q:

Assumption 5.1.1. For all (t, α, i0, α0, x) ∈ [0, T ] × A × 1, . . . ,m0 × A0 × S,

the matrix [q(t, i, j, α, i0, α0, x)]1≤i,j≤m is a Q-matrix.

Then we assume that at time t if the state of minor player n is i, the player will jump

to state j at a rate given by:

q(t, i, j, αnt , X0t , α

0t , µ

Nt ),

if X0t is the major player state, α0

t ∈ A0 is the major player control, αnt ∈ A is the n-th

minor player control and µNt ∈ S is the empirical distribution of the minor player’s

states. Our goal is to use these rates to completely specify the law of a continuous-

time process for every player’s state in the following way: if at time t the n-th minor

player is in state i and uses control αnt , if the major player is in state X0t and uses the

control α0t , and if the empirical distribution of the states of the population of minor

players is µNt , then the probability of player n remaining in the same state during the

146

Page 156: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

infinitesimal time interval [t, t+∆t) is [1+q(t, i, i, αnt , X0t , α

0t , µ

Nt )∆t+o(∆t)], whereas

the probability of this state changing to another state j during the same time interval

is given by [q(t, i, j, αnt , X0t , α

0t , µ

Nt )∆t+ o(∆t)].

Similarly, to describe the evolution of the state of the major player we introduce

a function q0:

[0, T ]× 1, . . . ,m02 × A0 × S →R

(t, i0, j0, α0, x)→q0(t, i0, j0, α0, x),

which satisfies the following assumption:

Assumption 5.1.2. For each (t, α0, x) ∈ [0, T ]×A0× S, [q0(t, i0, j0, α0, x)]1≤i0,j0≤m0

is a Q-matrix.

So if at time t the state of the major player is i0, its control is α0t ∈ A0, and the

empirical distribution of the states of the minor players is µNt , we assume the state

of the major player will jump to state j0 at rate q0(t, i0, j0, α0t , µ

Nt ).

We now define the control strategies which are admissible to the major and minor

players. In our model, we assume that the major player can only observe its own

state and the empirical distribution of the states of the minor players, whereas each

minor player can observe its own state, the state of the major player as well as the

empirical distribution of the states of all the minor players. Furthermore, we only

allow for Markovian strategies given by feedback functions. Therefore the major

player’s control should be of the form α0t = φ0(t,X0

t , µNt ) for some feedback function:

φ0 : [0, T ]× 1, · · · ,m0 × S → A0

(t, i0, x)→ φ0(t, i0, x),

147

Page 157: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

and the control of minor player n should be of the form αnt = φn(t,Xnt , X

0t , µ

Nt ) for

some feedback function:

φn : [0, T ]× 1, · · · ,m × 1, · · · ,m0 × S → A

(t, i, i0, x)→ φn(t, i, i0, x).

We denote the sets of admissible control strategies by A0 and An respectively. In line

with the notations used in Section 2.2.1 of Chapter 2, for each n ∈ 0, 1, . . . , N, we

denote an element of An by αn ↔ φn to emphasize the role of the feedback function.

When we study the convergence of N -player game to the mean field game and

identify the Nash equilibria for mean field game later in this chapter, we shall consider

the admissible strategies of which the feedback function is Lipschitz in the measure.

More specifically, we define the collection of Lipschitz strategies for the major and

minor players as follow:

L0 :=α0 ↔ φ0| φ0 is Lipschitz continuous in x, uniformly in t and i0., (5.1)

Ln :=αn ↔ φn| φn is Lipschitz continuous in x, uniformly in t, i and i0.. (5.2)

Let us define the joint dynamics of the states of all the players. We assume that

conditioned on the current state of the system, the changes of states are independent

for different players. This means that for all i0, i1, . . . , iN and j0, j1, . . . , jN , where

i0, j0 ∈ 1, 2, . . . ,m0 and in, jn ∈ 1, 2, . . . ,m for n = 1, 2, . . . , N , we have:

P[X0t+∆t = j0, X

1t+∆t = j1, . . . , X

Nt+∆t = jN |X0

t = i0, X1t = i1, . . . , X

Nt = iN ]

= [1(i0 = j0) + q0(t, i0, j0, φ0(t, i0, µNt ), µNt )∆t+ o(∆t)]

×N∏n=1

[1(in = jn) + qn(t, in, jn, φn(t, in, i0, µNt ), i0, φ0(t, i0, µ

Nt ), µNt )∆t+ o(∆t)].

(5.3)

Formally, this statement is equivalent to the definition of the Q-matrix, say Q(N),

of the continuous-time Markov chain (X0t , X

1t , X

2t , . . . X

Nt ). The state space of this

148

Page 158: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Markov chain is the Cartesian product of each player’s state space. Therefore Q(N)

is a square matrix of size m0 · mN . The non-diagonal entry of Q(N) can be found

by simply retaining the first order term in ∆t when expanding the above product

of probabilities. Because we assume that the transitions of states are independent

among the individual players, Q(N) is a sparse matrix.

Each individual player aims to minimize its expected cost in the game. We assume

that these costs are given by:

J0,N (α0,α1, . . . ,αN ) :=E[∫ T

0f0(t,X0

t , φ0(t,X0t , µ

Nt ), µNt )dt+ g0(X0

T , µNT )

],

Jn,N (α0,α1, . . . ,αN ) :=E[ ∫ T

0fn(t,Xn

t , φn(t,Xnt , X

0t , µ

Nt ), X0

t , φ0(t,X0t , µ

Nt ), µNt )dt

+ gn(XnT , X

0T , µ

NT )

],

where f0 : [0, T ]×1, . . . ,m0×A0×S → R, g0 : 1, . . . ,m0×S → R are respectively

the running cost and terminal cost of the major player, and fn : [0, T ]×1, . . . ,m×

A×1, . . . ,m0×A0×S → R, gn : 1, . . . ,m×1, . . . ,m0×S → R are respectively

the running cost and terminal cost of the minor player n.

In this chapter, we focus on the special case of symmetric games, for which all the

minor players share the same transition rate function and cost function, i.e.: qn := q,

fn := f , gn := g, Jn,N := JN , and we search for symmetric Nash equilibria. We say

that a couple of feedback functions (φ0, φ) form a symmetric Nash equilibrium if the

controls (α0, α1, · · · , αN) given by α0t = φ0(t,X0

t , µNt ) and αnt = φ(t,Xn

t , X0t , µ

Nt ) for

n = 1, · · · , N , form a Nash equilibrium in the sense that:

J0,N (α0, α1, . . . , αN ) ≤ J0,N (α0, α1, . . . , αN ),

JN (α0, α1, . . . , αn, . . . αN ) ≤ JN (α0, α1, . . . ,αn, . . . αN ),

for any choices of alternative admissible controls α0 and αn of the forms α0t =

φ0(t,X0t , µ

Nt ) and αnt = φ(t,Xn

t , X0t , µ

Nt ).

149

Page 159: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

In order to simplify the notations, we will systematically use the following no-

tations when there is no risk of possible confusion. When α0 ∈ A0 is given by a

feedback function φ0 and α ∈ A is given by a feedback fucntion φ, we denote by qφ00 ,

qφ0,φ, fφ00 and fφ0,φ the functions:

qφ00 (t, i0, j0, x) := q0(t, i0, j0, φ0(t, i0, x), x),

qφ0,φ(t, i, j, i0, x) := q(t, i, j, φ(t, i, i0, x), i0, φ0(t, i0, x), x),

fφ00 (t, i0, x) := f0(t, i0, φ0(t, i0, x), x),

fφ0,φ(t, i0, i, x) := f(t, i, φ(t, i, i0, x), i0, φ0(t, i0, x), x).

5.2 Mean Field Game Formulation

Solving for Nash equilibria when the number of players is finite is challenging. There

are many reasons why the problem becomes quickly intractable. Among them is the

fact that as the number of minor players increases, the dimension of the Q - matrix of

the system increases exponentially. The paradigm of mean field games consists in the

analysis of the limiting case where the number N of minor players tends to infinity. In

this asymptotic regime, one expects that simplifications due to averaging effects will

make it easier to find asymptotic solutions which could provide approximate equilibria

for finite player games when the number N of minor players is large enough. The

rationale for such a belief is based on the intuition provided by classical results on

the propagation of chaos for large particle systems with mean field interactions. We

will develop these results later in the paper.

The advantage of considering the limit case is two-fold. First, when N goes to

infinity, the empirical distribution of the minor players’ states converges to a random

measure µt which we expect to be the conditional distribution of any minor player’s

150

Page 160: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

state, i.e.:

µNt → µt :=(P[Xn

t = 1|X0t ],P[Xn

t = 2|X0t ], . . . ,P[Xn

t = m− 1|X0t ]).

As we shall see later on in the next section, when considered together with the major

player’s state and one of the minor player’s state, the resulting process is Markovian

and its infinitesimal generator has a tractable form. Also, when the number of minor

players goes to infinity, small perturbations of a single minor player’s strategy will

have negligible influence on the distribution of the minor player’s states. This gives

rise to a simple formulation of the typical minor player’s search for the best response

to the control choices of the major player. In the limit N → ∞, we understand a

Nash equilibrium as a situation in which neither the major player, nor a typical minor

player could be better off by changing their control strategies. In order to formulate

this limiting problem, we need to define the joint dynamics of the states of the major

player and a deviating minor player, making sure that the dynamics of the state of

the major player depend upon the statistical distribution of the states of the minor

players, and that the dynamics of the state of the deviating minor player depend

upon the values of the state and the control of the major player, their own state, and

the statistical distribution of the states of all the minor players.

As argued in Carmona and Wang [2017], and echoed in Chapter 13 of Carmona

and Delarue [2017], the best way to search for Nash equilibria in the mean field limit

of games with major and minor players is first to identify the best response map of

the major player and the deviating minor player by solving the optimization problems

for the strategies of (1) the major player in response to the field of the minor players,

and (2) the deviating minor player in response to the behavior of the major player

and the other minor players. Solving these optimization problems separately provides

a definition of the best response map for the system. One can then search for a fixed

151

Page 161: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

point for this best response map. So the search for Nash equilibria for the mean field

game with major and minor players can be summarized in the following two steps.

Step 1: Identifying the Best Response Map

Step 1.1: Major Player’s Problem

Fix an admissible strategy α ∈ A of the form αt = φ(t,Xt, X0t , µt) for the minor

players, and solve for the optimal control problem of the major player given that all

the minor players use the feedback function φ. We denote by φ∗0(φ) the feedback

function giving the optimal strategy of this optimization problem.

Notice that, in order to formulate properly this optimization problem, we need to

define Markovian dynamics for the couple (X0t , Xt) where Xt is interpreted as the state

of a deviating minor player, and the (random) measure µt has to be defined clearly.

This is done in the next section as the solution of the major player’s optimization

problem, the Markovian dynamics being obtained from the limit of games with N

minor players.

Step 1.2: Deviating Minor Player’s Problem

We single out the deviating minor player and we search for its best response to the

rest of the other players. So we fix an admissible strategy α0 ∈ A0 of the form

α0t = φ0(t,X0

t , µt) for the major player, and an admissible strategy α ∈ A of the

form αt = φ(t,Xt, X0t , µt) for the remaining minor players. We then assume that

the deviating minor player which we singled out responds to the other players by

choosing an admissible strategy α ∈ A of the form αt = φ(t, Xt, X0t , µt). Clearly, if

we want to find the best response of the deviating minor player to the behavior of

the major player and the field of the other minor players, we need to determine the

dynamics for the triple (X0t , Xt, µt), and define clearly what we mean by the (random)

measure µt. This is done in the next section as the solution of the deviating minor

player optimization problem, the Markovian dynamics being obtained from the limit

152

Page 162: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

of games with N minor players. We denote by φ∗(φ0, φ) the feedback function giving

the optimal strategy of this optimization problem.

Step 2: Search for a Fixed Point of the Best Response Map

A Nash equilibrium for the mean field game with major and minor players is a pair

of Markov control strategies (φ0, φ) which provides a fixed point of the best response

mappings, i.e. [φ0, φ] = [φ∗0(φ),φ∗(φ0, φ)].

Clearly, in order to carry out Step 1, we need to formulate properly the search for

these two best responses, and study the limit N →∞ of both optimization problems

of interest.

5.3 Convergence of Large Finite Player Games

Throughout the rest of this chapter, we make the following assumptions on the reg-

ularity of the transition rate and cost functions:

Assumption 5.3.1. There exists a constant L > 0 such that for all i, j ∈ 1, . . . ,m,

i0, j0 ∈ 1, . . . ,m0 and all t, t′ ∈ [0, T ], x, x′ ∈ S, α0, α′0 ∈ A0 and α, α′×A, we have:

|(f, f0, g, g0, q, q0)(i, j, i0, j0, t, x, α0, α)− (f, f0, g, g0, q, q0)(i, j, i0, j0, t′, x′, α′0, α

′)|

≤ L(|t− t′|+ ‖x− x′‖+ ‖α0 − α′0‖+ ‖α− α′‖).(5.4)

Assumption 5.3.2. There exists a constant C > 0 such that for all i, j ∈ 1, . . . ,m,

i0 ∈ 1, . . . ,m0 and all t ∈ [0, T ], x ∈ S, α0 ∈ A0 and α ∈ A, we have:

|q(t, i, j, α, i0, α0, x)| ≤ C. (5.5)

Finally, we add a boundary condition on the Markovian dynamics of the minor

players. Intuitively speaking, this assumption rules out the extinction: it says that

153

Page 163: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

a minor player can no longer change its state, when the percentage of minor players

who are in the same state falls below a certain threshold.

Assumption 5.3.3. There exists a constant ε > 0 such that for all t ∈ [0, T ], i, j ∈

1, . . . ,m− 1, i 6= j and α0 ∈ A0 and α ∈ A, we have:

xi < ε =⇒ q(t, i, j, α, i0, α0, x) = 0, (5.6)

1−m−1∑k=1

xk < ε =⇒ q(t,m, i, α, i0, α0, x) = 0. (5.7)

The purpose of this section is to identify the state dynamics which should be posited

in the formulation of the mean field game problem with major and minor players. In

order to do so, we formulate the search for the best response of each player by first

setting the game with finitely many minor players, and then letting the number of

minor players go to infinity to identify the dynamics over which the best response

should be computed in the limit.

5.3.1 Major Player’s Problem with Finitely Many Minor

Players

For any integer N (fixed for the moment), we consider a game with N minor players,

and we compute the best response of the major player when the minor players choose

control strategies αn = (αnt )0≤t≤T given by the same feedback function φ so that

αnt = φ(t,Xn,Nt , X0,N

t , µNt ) for n = 1, · · · , N . Here, Xn,Nt denotes the state of the

n-th minor player at time t, X0,Nt the state of the major player, and µNt the empirical

distribution of the states of the N minor players at time t. The latter is a probability

measure on the state space E = 1, · · · ,m, and we shall identify it with an element

in S:

µNt =1

N

( N∑n=1

1(Xn,Nt = 1),

N∑n=1

1(Xn,Nt = 2), . . . ,

N∑n=1

1(Xn,Nt = m− 1)

). (5.8)

154

Page 164: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

So for each i ∈ E, µNt (i) is the proportion of minor players whose state at time t

is equal to i. Consequently, for N fixed, µNt can be viewed as an element of the finite

space 0, 1/N, · · · , (N − 1)/N, 1m−1. We denote by SN the set of possible values of

µNt , in other words, we set:

SN := 1

N(n1, n2, . . . nm−1); ni ∈ N,

∑i

ni ≤ N. (5.9)

Given the choice of control strategies made by the minor players, we denote by

α0 = (α0t )0≤t≤T the control strategy of the major player, and we study the time

evolution of the state of the system given these choices of control strategies. Later

on, we shall find the optimal choice for the major player’s control α0 given by a

feedback function φ0 in response to the choice of the feedback function φ of the minor

players. While this optimization should be done over the dynamics of the whole state

(X0,Nt , X1,N

t , · · · , XN,Nt ), we notice that the process (X0,N

t , µNt )0≤t≤T is sufficient to

define the optimization problem of the major player, and that it is also a continuous-

time Markov process in the finite state space 1, . . . ,m0 × SN .

Our goal is to show that asN →∞, the Markov process (X0,Nt , µNt )0≤t≤T converges

in some sense to a Markov process (X0t , µt)0≤t≤T . This will allow us to formulate the

optimization problem of the major player in the mean field limit in terms of this

limiting Markov process.

For each integer N , we denote by Gφ0,φ0,N the infinitesimal generator of the Markov

process (X0,Nt , µNt )0≤t≤T . Since the process is not time-homogeneous, when we say

infinitesimal generator, we mean the infinitesimal generator of the space-time process

(t,X0,Nt , µNt )0≤t≤T . Except for the partial derivative with respect to time, this in-

finitesimal generator is given by the Q-matrix of the process, namely the instantaneous

rates of jump in the state space 1, . . . ,m0 × SN . So if F : [0, T ]× 1, 2, . . . ,m0 ×

155

Page 165: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

SN → R is C1 in time,

[Gφ0,φ0,N F ](t, i0, x) = ∂tF (t, i0, x) +∑j0 6=i0

[F (t, j0, x)− F (t, i0, x)]qφ00 (t, i0, j0, x)

+∑j 6=i

[F (t, i0, x+1

Neij)− F (t, i0, x)]Nxiq

φ0,φ(t, i, j, i0, x),(5.10)

where the first summation in the right hand side corresponds to jumps in the state of

the major player and the terms in the second summation account for the jumps of the

state of one minor player from i to j. Here we code the change in the empirical distri-

bution x of the states of the minor players caused by the jump from i ∈ 1, · · · ,m

to j ∈ 1, · · · ,m with j 6= i, of the state of a single minor player as (1/N)eij with

the notation eij := ej1(j 6= m)− ei1(i 6= m) where ei stands for the i-th vector in the

canonical basis of the space Rm−1. We have also used the notation xm = 1−∑m−1

i=1 xi

for the sake of simplicity.

Notice that the two summations appearing in (5.10) correspond to finite difference

operators which are bounded. So the domain of the operator Gφ0,φ0,N is nothing else

than the domain of the partial derivative with respect to time. Notice also that the

sequence of generators Gφ0,φ0,N converges, at least formally, toward a limit which can

easily be identified. Indeed, it is clear from the definition (5.10) that [Gφ0,φ0,N F ](t, i0, x)

still makes sense if x ∈ S, where S is the m − 1 dimensional simplex. Moreover,

if F : [0, T ] × 1, 2, . . . ,m0 × S → R is C1 in both variables t and x, we have

[Gφ0,φ0,N F ](t, i0, x)→ [Gφ0,φ0 F ](t, i0, x), where the operator Gφ0,φ0 is defined by:

[Gφ0,φ0 F ](t, i0, x) := ∂tF (t, i0, x) +∑j0 6=i0

[F (t, j0, x)− F (t, i0, x)]qφ00 (t, i0, j0, x)

+

m−1∑i,j=1

∂xjF (t, i0, x)xiqφ0,φ(t, i, j, i0, x)

+(1−m−1∑k=1

xk)

m−1∑j=1

∂xjF (t, i0, x)qφ0,φ(t,m, j, i0, x).

(5.11)

156

Page 166: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

So far, we have a sequence of time-inhomogeneous Markov processes (X0,Nt , µNt )N≥0

characterized by their infinitesimal generators Gφ0,φ0,N which converge to Gφ0,φ0 . We now

aim to show the existence of a limiting Markov process with infinitesimal generator

Gφ0,φ0 . The proof consists of first showing the existence of a Feller semigroup generated

by the limiting generator Gφ0,φ0 , and then applying an argument of convergence of

semigroups.

Remark 5.3.4. The standard results in the theory of semigroups are tailor-made

for time-homogeneous Markov processes. However, they can easily be adapted to the

case of time-inhomogeneous Markov processes by simply considering the space-time

expansion, specifically by augmenting the process (X0,Nt , µNt ) into (t,X0,N

t , µNt ) and

considering the uniform convergence on all bounded time intervals.

Let us introduce some additional notations which are useful for the functional

analysis of the infinitesimal generators and their corresponding semigroups. We set

EN = [0, T ]×1, . . . ,m0×SN and E∞ = [0, T ]×1, . . . ,m0×S for the state spaces,

and we denote by C(E∞) the Banach space for the norm ‖F‖∞ = supt,i0,x |F (t, i0, x)|,

of the real valued continuous functions defined on E∞. We also denote by C1(E∞)

the collection of functions in C(E∞) that are C1 in t and x for all i0 ∈ 1, . . . ,m0.

Note that the Markov process (t,X0,Nt , µNt ) lives in EN while the candidate limiting

process (t,X0,Nt , µNt ) lives in E∞. The difference is that µNt only takes values in SN ,

which is a finite subset of S. Thus if we want to show the convergence, we need to

reset all the processes on the same state space, and our first step should be to extend

the definition of (t,X0,Nt , µNt ) to a Markov process taking value in E∞. To do so, we

extend the definition of the generator Gφ0,φ0,N to accommodate functions F defined on

the whole E∞:

[Gφ0,φ0,N F ](t, i0, x) = ∂tF (t, i0, x) +∑j0 6=i0

(F (t, j0, x)− F (t, i0, x)

)qφ00 (t, i0, j0, x)

+∑j 6=i

(F (t, i0, x+

1

Neij)− F (t, i0, x)

)Nxi1(xi ≥

1

N)qφ0,φ(t, i, j, i0, x).

157

Page 167: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We claim that for N large enough, Gφ0,φ0,N generates a Markov process with a Feller

semigroup taking values in E∞. Indeed, when the initial distribution is a probability

measure on 1, . . . ,m0 × SN , the process has exactly the same law as (X0,Nt , µNt ).

To see why this is true, let us denote for all x ∈ S the set SxN := (x + 1NZm−1) ∩ S.

Then we can construct a Markov process starting from (i, x) and living in the space of

finite states 1, . . . ,m× SxN , which has the same transition rates as those appearing

in the definition of Gφ0,φ0,N . In particular, the indicator function 1(xi ≥ 1N

) forbids

the component x to exit the domain S. Assumption 5.3.3 implies that the transition

function is continuous on E when N ≥ 1/ε, where ε is the extinction threshold in

the assumption. So this process is a continuous-time Markov process with continuous

probability kernel in a compact space. By Proposition 4.4 in Swart and Winter [2013],

it is a Feller process. In the following, we will still denote this extended version of

the process as (X0,Nt , µNt ).

Proposition 5.3.5. There exists a Feller semigroup T = (Tt)t≥0 on the space C(E∞)

such that the closure of Gφ0,φ0 is the infinitesimal generator of T .

Proof. We use the perturbation argument. Observe that Gφ0,φ0 is the sum of two linear

operators H and K on C(E∞):

[HF ](t, i0, x) :=∂tF (t, i0, x) + v(t, i0, x) · ∇F (t, i0, x),

[KF ](t, i0, x) :=∑j0 6=i0

[F (t, j0, x)− F (t, i0, x)]qφ00 (t, i0, j0, x),

where we denote by ∇F (t, i0, x) the gradient of F with respect to x, and by v the

vector field:

vj(t, i0, x) :=m−1∑i=1

xiqφ0,φ(t, i, j, i0, x) +

(1−

m−1∑i=1

xi

)qφ0,φ(t,m, j, i0, x).

Being a finite difference operator, K is a bounded operator on C(E∞) so the proof

reduces to showing that H generates a Feller semigroup (see for example Theorem

158

Page 168: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

7.1, Chapter 1 in Ethier and Kurtz [2009]). To show that the closure of H generates

a strongly continuous semigroup on C(E∞) we use the characteristics of the vector

field v. For any (t, i0, x) ∈ E∞, let (Y(t,i0,x)u )u≥0 be the solution of the following ODE:

dY (t,i0,x)u = v(t+ u, i0, Y

(t,i0,x)u )du, Y

(t,i0,x)0 = x.

The existence and uniqueness of the solution to this ODE are guaranteed by the Lip-

schitz continuity of the vector field v, which in turn is a consequence of the Lipschitz

property of qφ0,φ. Notice that by Assumption 5.3.3, the process Y(t,i0,x)u is confined to

S. So we can define the linear operator Ts on C(E∞):

[TsF ](t, i0, x) := F (s+ t, i0, Y(t,i0,x)s ).

Uniqueness of the solution implies that (Ts)s≥0 is a semigroup. This semigroup is

also strongly continuous. Indeed, by the boundedness of v, for a fixed h > 0, there

exists a constant C0 such that |Y (t,i0,x)s − x| ≤ C0s for all s ≤ h and (t, i0, x) ∈ E∞.

Combining this estimation with the fact that F is uniformly continuous in (t, x) for all

F ∈ C(E∞), we obtain that ‖TsF−F‖ → 0, s→ 0. Finally the semigroup T is Feller,

since the solution of ODE, Y(t,i0,x)s , depends continuously on the initial condition as

a consequence of the Lipschitz property of the vector field v.

It is plain to check that H is the infinitesimal generator of the semigroup T , and

the domain of H is C1(E∞).

The following lemma is a simple adaptation of Theorem 6.1, Chapter 1 in Ethier

and Kurtz [2009] and is an important ingredient in the proof of the convergence. It

says that the convergence of the infinitesimal generators implies the convergence of

the corresponding semigroups.

Lemma 5.3.6. For N = 1, 2, . . . , let TN(t) and T (t) be strongly continuous

contraction semigroups on a Banach space L with infinitesimal generators GN and G,

159

Page 169: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

respectively. Let D be a core for G and assume that D ⊂ D(GN) for all N ≥ 1. If

limN→∞ GNF = GF for all F ∈ D, then for each F ∈ L, limN→∞ TN(t)F = T (t)F

for all t ≥ 0.

We are now ready to state and prove the main result of this section: the Markov

process (X0,Nt , µNt ) describing the dynamics of the state of the major player and the

empirical distribution of the states of the N minor players converges weakly to a

Markov process with infinitesimal generator Gφ0,φ0 , when the players choose Lipschitz

strategies.

Theorem 5.3.7. Assume that the major player chooses a control strategy α0 given

by a Lipschitz feedback function φ0 and that all the minor players choose control

strategies given by the same Lipschitz feedback function φ. Let i0 ∈ 1, . . . ,m0 and

for each integer N ≥ 1, let xN ∈ SN with limit x ∈ S. Then the sequence of processes

(X0,Nt , µNt ) with initial conditions X0,N

0 = i0, µNt = xN converges weakly to a Markov

process (X0t , µt) with initial condition X0

0 = i0, µ0 = x. The infinitesimal generator

for (X0t , µt) is given by:

[Gφ0,φ0 F ](t, i0, x) := ∂tF (t, i0, x) +∑j0 6=i0

[F (t, j0, x)− F (t, i0, x)]qφ00 (t, i0, j0, x)

+

m−1∑i,j=1

∂xjF (t, i0, x)xiqφ0,φ(t, i, j, i0, x)

+(1−m−1∑k=1

xk)m−1∑j=1

∂xjF (t, i0, x)qφ0,φ(t,m, j, i0, x).

(5.12)

Proof. Let us denote by (T Nt )t≥0 the semigroup associated with the time-homogeneous

Markov process (t,X0,Nt , µNt ) and the infinitesimal generator Gφ0,φ0,N . Recall that by the

procedure of extension we described above, the process (t,X0,Nt , µNt ) now lives in E∞

and the domain for Gφ0,φ0,N is C(E∞). In light of Theorem 2.5, Chapter 4 in Ethier

and Kurtz [2009] and Proposition 5.3.5 we just proved, it boils down to proving that

for any F ∈ E∞ and t ≥ 0, T Nt F converges to TtF , where (Tt)t≥0 is the strongly

continuous semigroup generated by the closure of G0φ0,φ

.

160

Page 170: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

To show the convergence, we apply Lemma 5.3.6. It is easy to see that C1(E∞)

is a core for Gφ0,φ0 and C1(E∞) is included in the domain of Gφ0,φ0,N . Therefore it only

remains to show that for all F ∈ C1(E∞), Gφ0,φ0,N F converges to Gφ0,φ0 F in the space

(C(E∞), ‖ · ‖). Using the notation xm := 1−∑m−1

i=1 xi, we have:

∣∣∣[Gφ0,φ0,N F ](t, i0, x)− [Gφ0,φ

0 F ](t, i0, x)∣∣∣

=∑j 6=i

∣∣∣∣N(F (t, i0, x+1

Neij)− F (t, i0, x))− (1(j 6= m)∂xjF (t, i0, x)− 1(i 6= m)∂xiF (t, i0, x))

∣∣∣∣xiqN (t, i, j, i0, x)

≤∑j 6=i

[∣∣∣∣∂xjF (t, i0, x+λi,jN

ei,j)− ∂xjF (t, i0, x)

∣∣∣∣+

∣∣∣∣∂xiF (t, i0, x+λi,jN

ei,j)− ∂xiF (t, i0, x)

∣∣∣∣]xiqN (t, i, j, i0, x),

where we applied the intermediate value theorem at the last inequality and λi,j ∈

[0, 1]. Note that λi,j also depends on t, x, i0, but we omit them for the sake of simplic-

ity. We notice that F ∈ C1(E∞) and E∞ is compact, and therefore ∂xiF is uniformly

continuous on E∞ for all i, which immediately implies that ‖Gφ0,φ0,N F − Gφ0,φ0 F‖ →

0, N → +∞. This completes the proof.

Major Player’s Optimization Problem in the Mean Field Limit

Given that all the minor players use control strategies based on the same feedback

function φ, the best response of the major player is to use the strategy α0 given by

the feedback function φ0 solving the optimal control problem:

infα0↔φ0∈A0

E[∫ T

0f0(t,X0

t , φ0(t,X0t , µt), µt)dt+ g0(X0

T , µT )

],

where (X0t , µt)0≤t≤T is the continuous-time Markov process with infinitesimal gener-

ator Gφ0,φ0 .

5.3.2 Deviating Minor Player’s Problem

We turn to the computation of the best response of the deviating minor player, which

we assume to be player 1. We assume that the major player chooses a strategy

α0 ∈ A0 of the form α0t = φ0(t,X0,N

t , µNt ) and that the minor players n ∈ 2, · · · , N

161

Page 171: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

all use strategies αn ∈ A of the form αnt = φ(t,Xn,Nt , X0,N

t , µNt ), and that the first

minor player uses strategy α ∈ A of the form αt = φ(t,X1,Nt , X0,N

t , µNt ). Clearly, by

symmetry, whatever we are about to say after we singled the first minor player out

as the deviating minor player, can be said if we single out any other minor player. As

before, for each fixed N , the process (X0,Nt , X1,N

t , µNt ) is a finite-state continuous-time

Markov process with state space 1, . . . ,m0 × 1, . . . ,m × SN whose infinitesimal

generator Gφ0,φ,φN is given, up to the time derivative, by the corresponding Q-matrix of

infinitesimal jump rates. In the present situation, its value on any real valued function

F defined on [0, T ]× 1, . . . ,m0 × 1, . . . ,m × SN such that t→ F (t, i0, i, x) is C1

for any i0, i and x, is given by the formula:

[Gφ0,φ,φN F ](t, i0, i, x) := ∂tF (t, i0, i, x) +∑

j0,j0 6=i0

[F (t, j0, i, x)− F (t, i0, i, x)]qφ00 (t, i0, j0, x)

+∑j,j 6=i

[F (t, i0, j, x+1

Neij)− F (t, i0, i, x)]qφ0,φ(t, i, j, i0, x)

+∑

(j,k),j 6=k

[F (t, i0, i, x+1

Nekj)− F (t, i0, i, x)](Nxk − 1(k = i))qφ0,φ(t, k, j, i0, x).

(5.13)

As before, the summations appearing above correspond to single jumps when 1) only

the state of the major player changes from state i0 to j0, 2) only the state of the

singled out first minor player changes from state i to j, and finally 3) the state of one

of the last N − 1 minor players jumps from state k to j.

Following the same treatment as in the major player’s problem, we have the

convergence result for the process (XNt , X

0,Nt , µNt ):

Theorem 5.3.8. Assume that for each integer N , the major player chooses a control

α0 given by a Lipschitz feedback function φ0, the first minor player chooses a control α

given by a Lipschitz feedback function φ, all the other minor players choose strategies

given by the same Lipschitz feedback function φ, and that these three feedback functions

do not depend upon N . Let i0 ∈ 1, . . . ,m0 and for each integer N ≥ 2, let xN ∈ SN

with limit x ∈ S. Then the sequence of processes (XNt , X

0,Nt , µNt )0≤t≤T with initial

162

Page 172: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

conditions XN0 = i, X0,N

0 = i0 and µN0 = xN converges weakly to a Markov process

(Xt, X0t , µt) with initial condition X0 = i, X0

0 = i0 and µ0 = x. Its infinitesimal

generator is given by:

[Gφ0,φ,φF ](t, i, i0, x) := ∂tF (t, i, i0, x) +∑

j0,j0 6=i0

[F (t, i, j0, x)− F (t, i, i0, x)]qφ00 (t, i0, j0, x)

+∑j,j 6=i

[F (t, j, i0, x)− F (t, i, i0, x)]qφ0,φ(t, i, j, i0, x) +m−1∑i,j=1

∂xjF (t, i, i0, x)xiqφ0,φ(t, i, j, i0, x)

+(1−m−1∑k=1

xk)

m−1∑j=1

∂xkF (t, i, i0, x)qφ0,φ(t,m, j, i0, x).

(5.14)

Deviating Minor Player’s Optimization Problem in the Mean Field Limit

Accordingly, in the mean field game limit, we define the search for the best response

of the deviating minor player (i.e. the minor player we singled out) to the strategies

adopted by the major player and the field of minor players as the following optimal

control problem. Assuming that the major player uses a feedback function φ0 and all

the other minor players use the feedback function φ, the best response of the deviating

minor player is given by the solution of:

infα↔φ∈A

E[∫ T

0

f(t,Xt, φ(t,Xt, X0t , µt), X

0t , φ0(t,X0

t , µt), µt)dt+ g(XT , X0T , µT )

],

where (Xt, X0t , µt)0≤t≤T is a Markov process with infinitesimal generator Gφ0,φ,φ. We

shall denote by φ = φ(φ0, φ) the optimal feedback function providing the solution of

this optimal control problem.

5.4 Optimization Problem for Individual Players

In this section, we use the dynamic programming principle to characterize the value

functions of the major and minor players’ optimization problems as viscosity solutions

of the corresponding Hamilton-Jacobi-Bellman (HJB for short) equations. We follow

163

Page 173: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the detailed arguments given in Chapter II of Fleming and Soner [2006]. For both the

major and deviating minor player, we show that the value function solves a weakly

coupled system of Partial Differential Equations (PDEs for short) in viscosity sense.

We also prove an important uniqueness result for these solutions. This uniqueness

result is important indeed because as the reader noticed, in defining the best response

map, we implicitly assumed that these optimization problems could be solved and that

their solutions were unique.

We first consider the value function of the major player’s optimization problem

assuming that all the minor players use the feedback function φ:

V φ0 (t, i0, x) := inf

α0↔φ0∈A0

E[∫ T

tfφ00 (s,X0

s , µs)ds+ g0(X0T , µT )|X0

t = i0, µt = x

]. (5.15)

Theorem 5.4.1. Assume that for all i0 ∈ 1, . . . ,m0, the mapping (t, x) →

V φ0 (t, i0, x) is continuous on [0, T ]× S. Then V φ

0 is a viscosity solution to the system

of m0 PDEs on [0, T ]× S:

0 = ∂tv0(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +

∑j0 6=i0

[v0(t, j0, x)− v0(t, i0, x)]q0(t, i0, j0, α0, x)

+(1−m−1∑k=1

xk)

m−1∑k=1

∂xkv0(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+

m−1∑i,j=1

∂xjv0(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

,

0 = g0(i0, x)− v0(T, i0, x).

(5.16)

The notion of viscosity solution in the above result is specified by the following

definition:

Definition 5.4.2. A real valued function v0 defined on [0, T ]×1, . . . ,m0× S such

that v0(·, i0, ·) is continuous on [0, T ] × 1, . . . ,m0 × S for all i0 ∈ 1, . . . ,m0 is

said to be a viscosity subsolution (resp. supersolution) if for any (t, i0, x) ∈ [0, T ] ×

1, . . . ,m0 × S and any C∞ function θ defined on [0, T ] × S such that the function

164

Page 174: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(v0(·, i0, ·)−θ) attains a maximum (resp. minimum) at (t, x) and v0(t, i0, x) = θ(t, x),

then the following inequalities hold:

0 ≤ (resp. ≥) ∂tθ(t, x) + infα0∈A0

f0(t, i0, α0, x) +

∑j0 6=i0

[v0(t, j0, x)− v0(t, i0, x)]q0(t, i0, j0, α0, x)

+(1−m−1∑k=1

xk)m−1∑k=1

∂xkθ(t, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+

m−1∑i,j=1

∂xjθ(t, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

, if t < T,

0 ≤ (resp. ≥) g0(i0, x)− v0(T, i0, x).

(5.17)

If v0 is both a viscosity subsolution and supersolution, we call it a viscosity solution.

Proof. Define C(S)m0 the collection of mappings (i0, x) → θ(i0, x) defined on

1, . . . ,m0 × S such that θ(i0, ·) is continuous on S for all i0. Define the dynamic

programming operator Tt,s on C(S)m0 by:

[Tt,sθ](i0, x) := infα0↔φ0

E[∫ s

tfφ00 (u,X0

u, µu)du+ θ(X0s , µs)|X0

t = i0, µt = x

]. (5.18)

where the Markov process (X0u, µu)0≤t≤T has infinitesimal generator Gφ0,φ0 . Then the

value function can be expressed as:

V φ0 (t, i0, x) = [Tt,T g0](i0, x),

and the dynamic programming principle says that:

V φ0 (t, i0, x) = [Tt,sV φ

0 (s, ·, ·)](i0, x), (t, s, i0, x) ∈ [0, T ]2 × 1, . . . ,m0 × S.

We will use the following lemma for which the proof is given in the appendix.

Lemma 5.4.3. Let Φ be a function on [0, T ]×1, . . . ,m0× S and i0 ∈ 1, . . . ,m0such that Φ(·, i0, ·) is C1 in [0, T ]× S and Φ(·, j0, ·) is continuous in [0, T ]× S for all

165

Page 175: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

j0 6= i0. Then we have:

limh→0

1

h[(Tt,t+hΦ(t+ h, ·, ·))(i0, x)− Φ(t, i0, x)]

= ∂tΦ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) + (1−

m−1∑k=1

xk)

m−1∑k=1

∂xkΦ(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+

m−1∑i,j=1

∂xjΦ(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x) +∑j0 6=i0

[Φ(t, j0, x)− Φ(t, i0, x)]q0(t, i0, j0, α0, x)

.

We now prove the subsolution property. Let θ be a function defined on [0, T ]× S

such that (V0(·, i0, ·)−θ) attains a maximum at (t, x) and V0(t, i0, x) = θ(t, x). Define

the function Φ on [0, T ]×1, . . . ,m0×S by Φ(·, i0, ·) := θ and Φ(·, j0, ·) := V0(·, j0, ·)

for j0 6= i0. Then clearly Φ ≥ V0, which implies:

(Tt,sΦ(s, ·, ·))(i0, x) ≥ (Tt,sV0(s, ·, ·))(i0, x).

By the Dynamic Programming Principle and the fact that Φ(t, i0, x) = V0(t, i0, x),

we have:

limh→0

1

h[(Tt,sΦ(s, ·, ·))(i0, x)− Φ(t, i0, x)] ≥ 0.

Then applying the lemma we obtain the desired inequality. The viscosity property

for supersolution can be checked in exactly the same way.

For later reference, we state the comparison principle for the HJB equation we

just derived. Again, its proof is postponed to the appendix of this chapter.

Theorem 5.4.4. (Comparison Principle) Let us assume that the feedback function

φ is Lipschitz, and let w (resp. v) be a viscosity subsolution (resp. supersolution) of

the equation (5.16). Then we have w ≤ v.

We now turn to the deviating minor agent’s optimization problem assuming that

the major player uses the feedback function φ0 and all the other minor players use

166

Page 176: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

the feedback function φ. We define the value function:

V φ0,φ(t, i, i0, x) := infα↔φ∈A

E[∫ T

tfφ0,φ(s,Xs, X

0s , µs)ds+ g(XT , X

0T , µT )|Xt = i,X0

t = i0, µt = x

],

(5.19)

where the Markov process (Xt, X0t , µt)0≤t≤T has infinitesimal generator Gφ0,φ,φ. In

line with the analysis of the major player’s problem, we can show that V φ0,φ is the

unique viscosity solution to a coupled system of PDEs.

Theorem 5.4.5. Assume that for all i ∈ 1, . . . ,m and i0 ∈ 1, . . . ,m0, the map-

ping (t, x) → V φ0,φ(t, i, i0, x) is continuous on [0, T ] × S. Then V φ0,φ is a viscosity

solution to the system of PDEs:

0 = ∂tv(t, i, i0, x)

+ infα∈A

f(t, i, α, i0, φ0(t, i0, x), x) +

∑j 6=i

[v(t, j, i0, x)− v(t, i, i0, x)]q(t, i, j, α, i0, φ0(t, i0, x), x)

+∑j0 6=i0

[v(t, i, j0, x)− v(t, i, i0, x)]qφ00 (t, i0, j0, x) +m−1∑i,j=1

∂xjv(t, i, i0, x)xiqφ0,φ(t, i, j, i0, x)

+(1−m−1∑k=1

xk)m−1∑k=1

∂xkv(t, i, i0, x)qφ0,φ(t,m, k, i0, x),

0 = g(i, i0, x)− v(T, i, i0, x).

(5.20)

Moreover, if the feedback functions φ0 and φ are Lipschitz, then the above system of

PDEs satisfies the comparison principle.

It turns out that the value functions V φ0 and V φ0,φ are Lipschitz in (t, x). To

establish this regularity property and estimate the Lipschitz constants, we need to

first study the regularity of the value functions for the finite player games, and control

the convergence in the regime of large games. We will state these results in Section

5.7, where we deal with the propagation of chaos and highlight more connections

between finite player games and mean field games.

We conclude this section with a result which we will use frequently in the sequel.

To state the result, we denote by Jφ0,φ0 the expected cost of the major player when it

167

Page 177: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

uses the feedback function φ0 and the minor players all use the feedback function φ.

Put differently:

Jφ0,φ0 (t, i0, x) := E[∫ T

tfφ00 (s,X0

s , µs)ds+ g0(X0T , µT )|X0

t = i0, µt = x

], (5.21)

where the Markov process (X0s , µs)t≤s≤T has infinitesimal generator Gφ0,φ0 . Then

by definition, we have V φ0 (t, i0, x) = infα0↔φ0 J

φ0,φ0 (t, i0, x). Similarly, we denote by

Jφ0,φ,φ the expected cost of the deviating minor player when it uses the feedback

function φ, while the major player uses the feedback function φ0 and all the other

minor players use the same feedback function φ:

Jφ0,φ,φ(t, i, i0, x) := E[∫ T

t

fφ0,φ(s,Xs, X0s , µs)ds+ g(XT , X

0T , µT )|Xt = i,X0

t = i0, µt = x

](5.22)

where the Markov process (Xs, X0s , µs)t≤s≤T has infinitesimal generator Gφ0,φ,φ.

Proposition 5.4.6. If the feedback functions φ0, φ and φ are Lipschitz, then Jφ0,φ0

and Jφ0,φ,φ are respectively continuous viscosity solutions of the PDEs (5.23) and

(5.24): 0 = [Gφ0,φ0 v0](t, i0, x) + fφ00 (t, i0, x),

0 = g0(i0, x)− v0(T, i0, x),

(5.23)

0 = [Gφ0,φ,φv](t, i, i0, x) + fφ0,φ(t, i, i0, x),

0 = g(i, i0, x)− v(T, i, i0, x).

(5.24)

moreover, the PDEs (5.23) and (5.24) satisfy the comparison principle.

Proof. The continuity of Jφ0,φ0 and Jφ0,φ,φ follows from the fact that (X0t , µt) and

(Xt, X0t , µt) are Feller processes, which we have shown in Theorem 5.3.7 and Theorem

5.3.8. The viscosity property can be shown using the exact same technique as in the

proof of Theorem 5.4.1. Finally, the comparison principle is a consequence of the

168

Page 178: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Lipschitz property of f0, f , q0, q, φ0, φ and φ, and it can be proved by slightly

modifying the proof of Theorem 5.4.4.

5.5 Existence of Nash Equilibria

In this section, we prove existence of Nash equilibria when the minor player’s jump

rates and cost functions do not depend upon the major player’s control. We work

under the following assumption:

Assumption 5.5.1. Assumption 5.3.1 and Assumption 5.3.2 are in force. In addi-

tion, the transition rate function q and the cost function f for the minor player do

not depend upon the major player’s control α0 ∈ A0.

The following assumptions will guarantee the existence of optimal strategies for

both the major player and the deviating minor player.

Assumption 5.5.2. For all i0 = 1, . . . ,m0, (t, x) ∈ [0, T ] × P and v0 ∈ Rm0, the

function α0 → f0(t, i0, α0, x) +∑

j0 6=i0(v0[j0] − v0[i0])q0(t, i0, j0, α0, x) has a unique

maximizer in A0 denoted as a0(t, i0, x, v0). Additionally, a0 is Lipschitz in (t, x, v0)

for all i0 = 1, . . . ,m0 with common Lipschitz constant Lα0

Assumption 5.5.3. For all i = 1, . . . ,m, i0 = 1, . . . ,m0, (t, x) ∈ [0, T ] × S and

v ∈ Rm×m0, the function α → f(t, α, i, i0, x) +∑

j 6=i(v[j, i0] − v[i, i0])q(t, i, j, α, i0, x)

has a unique minimizer in A denoted as α(t, i, i0, x, v). Additionally, α is Lipschitz

in (t, x, v) for all i0 = 1, . . . ,m0 and i = 1, . . . ,m with common Lipschitz constant

Lα.

Remark 5.5.4. Assumptions 5.5.2 and 5.5.3 are essentially about the existence of

a unique minimizer of the Hamiltonian for both the major and minor player’s op-

timization problem. It is very similar to Assumption 3.1.7 in Chapter 3, where we

study finite state mean field games without the major player. In Remark 3.1.8 and

169

Page 179: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Assumption 3.1.9, we provided a set of sufficient conditions, namely the convexity of

the cost functions and the linearity of the transition rate, which guarantees the exis-

tence of a unique Lipschitz-continuous minimizer of the Hamiltonian. We can easily

adapt these conditions to the current context, so that Assumptions 5.5.2 and 5.5.3

hold.

Proposition 5.5.5. Under Assumptions 5.5.1, 5.5.2 & 5.5.3, we have:

(i) For any Lipschitz feedback function φ for the deviating minor player, the best

response φ∗0(φ) of the major player exists and is given by:

φ∗0(φ)(t, i0, x) = α0(t, i0, x, Vφ

0 (t, ·, x)), (5.25)

where α0 is the minimizer defined in Assumption 5.5.2 and V φ0 is the value function

of the major player’s optimization problem.

(ii) For any Lipschitz feedback function φ0 for the major player and φ for the other

minor players, the best response φ∗(φ0, φ) of the deviating minor player exists and is

given by:

φ∗(φ0, φ)(t, i, i0, x) = α(t, i, i0, x, Vφ0,φ(t, ·, ·, x)), (5.26)

where α is the minimizer defined in Assumption 5.5.3 and V φ0,φ is the value function

of the deviating minor player’s optimization problem.

Proof. Consider the expected total cost Jφ∗0(φ),φ0 of the major player when all of the

minor players use the feedback function φ and the major player uses the strategy given

by the feedback function φ∗0(φ) defined by (5.25). Also consider V φ0 , the value function

of the major player’s optimization problem. By the definition of φ∗0(φ) and the PDE

(5.16), we see that V φ0 is a viscosity solution of the PDE (5.23) with φ0 = φ∗0(φ) and

φ = φ in Proposition 5.4.6. To be able to use the comparison principle, we need to

show that φ0 = φ∗0(φ) and φ are Lipschitz. Indeed the Lipschitz property follows from

Assumption 5.5.2 and Corollary 5.7.8 (see Section 5.7). Now since Jφ∗0(φ),φ0 is another

170

Page 180: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

viscosity solution for the same PDE, we conclude that Jφ∗0(φ),φ0 = V φ

0 = infα0↔φ0 Jφ0,φ0

and hence the optimality of φ∗0(φ). Likewise we can show that φ∗(φ0, φ) is the best

response of the deviating minor player.

In order to show that the Nash equilibrium is actually given by a couple of Lips-

chitz feedback functions, we need an additional assumption on the regularity of the

value functions.

Assumption 5.5.6. There exist two constants Lφ0 , Lφ, such that for all Lφ0-Lipschitz

feedback function φ0 and Lφ-Lipschitz feedback function φ, V φ0 is (Lφ0/Lα0 − 1)-

Lipschitz and V φ0,φ is (Lφ/Lα − 1)-Lipschitz.

The above assumption holds, for example, when the horizon of the game is suffi-

ciently small. We shall provide more details (see Remark 5.7.9 below) after we reveal

important connections between finite player games and mean field games in Section

5.7. We now state and prove the result on the existence of Nash equilibria.

Theorem 5.5.7. Under Assumptions 5.5.1, 5.5.2, 5.5.3 & 5.5.6, there exists a Nash

equilibrium in the sense that there exists Lipschitz feedback functions φ0 and φ such

that:

[φ0, φ] = [φ∗0(φ),φ∗(φ0, φ)].

Proof. We apply Schauder’s fixed point theorem. To this end, we need to: (i) specify

a Banach space V containing the admissible feedback functions (φ0, φ) as elements,

and a relatively compact convex subset K of V; (ii) show that the mapping R :

[φ0, φ]→ [φ∗0(φ),φ∗(φ0, φ)] is continuous and leaves K invariant (i.e. R(K) ⊂ K).

(i) Define C0 as the collection of A0 - valued functions φ0 on [0, T ]×1, . . . ,m0× S

such that (t, x)× φ0(t, i0, x) is continuous for all i0, C as the collection of A - valued

functions φ on [0, T ]× 1, . . . ,m × 1, . . . ,m0 × S such that (t, x)× φ(t, i, i0, x) is

171

Page 181: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

continuous for all i, i0, and set V := C0 ×C. For all (φ0, φ) ∈ V, we define the norm:

‖(φ0, φ)‖ := max

sup

i0,t∈[0,T ],x∈S|φ0(t, i0, x)|, sup

i,i0,t∈[0,T ],x∈S|φ(t, i, i0, x)|

.

It is easy to check that (V, ‖·‖) is a Banach space. Next, we define K as the collection

of elements in V such that the mappings (t, x) → φ0(t, i0, x) are L0−Lipschitz and

(t, x)→ φ(t, i, i0, x) are L−Lipschitz in (t, x) for all i0 = 1, . . . ,m0 and i = 1, . . . ,m,

where L0, L are specified in Assumption 5.5.6. Clearly K is convex. Now consider

the family (φ0(·, i0, ·))(φ0,φ)∈K of functions defined on [0, T ] × S. Thanks to the Lip-

schitz property, we see immediately that the family is equicontinuous and pointwise

bounded. Therefore by the Arzela-Ascoli theorem, the family is compact with respect

to the uniform norm. Repeating this argument for all i, i0 we see that K is compact

under the norm ‖ ·‖. moreover, thanks to Assumptions 5.5.2, 5.5.3 & 5.5.6, we obtain

easily that K is stable by R.

(ii) It remains to show that R is a continuous mapping. We use the following lemma:

Lemma 5.5.8. Let (φn0 , φn) be a sequence in K converging to (φ0, φ) in ‖ · ‖, and

denote by V n0 and V n the value functions of the major and the deviating minor players,

respectively, associated with (φn0 , φn). Then V n

0 and V n converge uniformly to V0 and

V , respectively, where V0 and V are the value functions of the major player and the

deviating minor player associated with (φ0, φ).

The proof of the lemma uses standard arguments from the theory of viscosity

solutions. We give it in the appendix. Now, the continuity of the mapping R follows

readily from Lemma 5.5.8, Proposition 5.5.5 and Assumption 5.5.2 & 5.5.3. This

completes the proof.

172

Page 182: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

5.6 The Master Equation for Nash Equilibria

If a Nash equilibrium exists and is given by feedback functions φ0 for the major player

and φ for the minor players, these functions should also be equal to the respective

minimizers of the Hamiltonians in the HJB equations of the optimization problems.

This informal remark leads to a system of coupled PDEs with terminal conditions

specified at t = T , which we expect to hold if the equilibrium exists. Now the natural

question to ask is: if this system of PDEs has a solution, does this solution provide a

Nash equilibrium? The following result provides a verification argument:

Theorem 5.6.1. (Verification Argument) Assume that there exists two function φ0 :

[0, T ] × 1, . . . ,m0 × S 3 (t, i0, x) → φ0(t, i0, x) ∈ R and φ : [0, T ] × 1, . . . ,m ×

1, . . . ,m0 × S 3 (t, i, i0, x) → φ(t, i, i0, x) ∈ R such that the system of PDEs in

(v0, v):

0 = [Gφ0,φ0 v0](t, i0, x) + f0(t, i0, φ0(t, i0, x), x),

v0(T, i0, x) = g0(i0, x),

0 = [Gφ0,φ,φv](t, i, i0, x) + f(t, i, φ(t, i, i0, x), i0, φ0(t, i0, x), x),

v(T, i, i0, x) = g(i, i0, x),

(5.27)

admits a classical solution (V0, V ) (i.e. the solution are C1 in t and x). Assume in

addition that:

φ0(t, i0, x) = α0(t, i0, x, V0(t, ·, x)),

φ(t, i, i0, x) = α(t, i, i0, x, V (t, ·, ·, x)).

(5.28)

Then φ0 and φ form a Nash equilibrium and V0(0, X00 , µ0) and V (0, X0, X

00 , µ0) are

the equilibrium expected costs of the major and minor players.

Proof. We show that φ0 = φ∗0(φ) and φ = φ∗(φ0, φ). Notice first that φ0 and φ

are Lipschitz strategies due to the regularity of V0 and V , Assumption 5.5.2 and

Assumption 5.5.3.

173

Page 183: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Consider the major player’s optimization problem where we let φ = φ and denote

by V φ0 the corresponding value function. Then since V0 is a classical solution to (5.27)

and because of (5.28), we deduce that V0 is a viscosity solution to the HJB equation

(5.16) associated with the value function V φ0 . By uniqueness of the viscosity solution,

we conclude that V0 = V φ0 .

On the other hand, if we denote by J φ0,φ0 the expected cost function of the major

player when it uses the feedback function φ0 and all the minor players use strategy

φ, then the fact that V0 is a classical solution to (5.27) implies that V0 is also a

viscosity solution. Then by Proposition 5.4.6 we have J φ0,φ0 = V0 and therefore

J φ0,φ0 = V φ0 = infα0↔φ0 J

φ0,φ0 . This means that φ0 is the best response of the major

player to the minor players using the feedback function φ.

For the optimization problem of the deviating minor player, we use the same

argument based on the uniqueness of solution of PDE to obtain J φ0,φ,φ = V = V φ0,φ =

infα↔φ Jφ0,φ,φ. This implies that φ is the deviating minor player’s best response to

the major player using feedback function φ0 and the rest of the minor players using

φ. We conclude that φ0 and φ form the desired fixed point for the best response

map.

It is important to keep in mind that the above verification argument of the master

equation does not speak to the problem of existence of Nash equilibria. However, it

provides a convenient way to compute numerically the equilibrium via the solution

of a coupled system of first-order PDEs.

174

Page 184: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

5.7 Propagation of Chaos and Approximate Nash

Equilibria

In this section we show that in the (N+1)-player game (see the description in Section

5.1), when the major player and each minor player apply the respective equilibrium

strategies in the mean field game, the system is in an approximate Nash equilibrium.

To uncover this link, we first revisit the (N + 1)-player game. We show that for a

certain strategy profile, the total expected cost of an individual player in the finite

player game converges to that of the mean field game. Our argument is largely similar

to the one used in proving the convergence of the numerical scheme for viscosity

solutions. One crucial intermediate result we use here is the gradient estimate for the

value functions of the (N+1)-player game. Similar results were proved in Gomes et al.

[2013] for discrete state mean field games without a major player. As a byproduct

of the proof, we can also conclude that the value function of the mean field game

is Lipschitz in the measure argument. In the rest of this section, we assume that

Assumption 5.5.1 is in force.

5.7.1 Back to the (N + 1)-Player Game

In this section, we focus on the game with a major player and N minor players. We

show that both the expected costs of individual players and the value functions of

the players’ optimization problems can be characterized by coupled systems of ODEs,

and their gradients are bounded by some constant independent of N . Such a gradient

estimate will be crucial in establishing results on propagation of chaos, as well as the

regularity of the value functions for the limiting mean field game.

We start from the major player’s optimization problem. Consider a strategy

profile where the major player chooses a Lipschitz feedback function φ0 and all the N

minor players choose the same Lipschitz feedback function φ. Recall that the process

175

Page 185: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

comprising the major player’s state and the empirical distribution of the states of

the minor players, i.e. (X0,Nt , µNt ), is a finite state continuous-time Markov chain

in the space 1, . . . ,m0 × SN , where SN := 1N

(k1, . . . , km−1)|∑

i ki ≤ N, ki ∈ N.

Its infinitesimal generator Gφ0,φ0,N was given by (5.10). The total expected cost to the

major player is given by:

Jφ0,φ0,N (t, i0, x) := E[∫ T

tfφ00 (s,X0,N

s , µNs )ds+ g0(X0,NT , µNT )| X0,N

t = i0, µNt = x

],

and the value function of the major player’s optimization problem by:

V φ0,N (t, i0, x) := inf

α0↔φ0∈A0

Jφ0,φ0,N (t, i0, x).

Despite the notation, Jφ0,φ0,N can be viewed as a function defined on [0, T ] with values

given by vectors indexed by (i0, x). The following result shows that Jφ0,φ0,N is charac-

terized by a coupled system of ODEs.

Proposition 5.7.1. Let φ0 (resp. φ) be a Lipschitz feedback function for the major

(resp. minor) player. Then Jφ0,φ0,N is the unique classical solution of the system of

ODEs:

0 = θ(t, i0, x) + fφ00 (t, i0, x) +∑

j0,j0 6=i0

(θ(t, j0, x)− θ(t, i0, x))qφ00 (t, i0, j0, x)

+∑

(i,j),j 6=i

(θ(t, i0, x+1

Neij)− θ(t, i0, x))Nxiq

φ0,φ(t, i, j, i0, x),

0 = θ(t, i0, x)− g0(i0, x).

(5.29)

Proof. The existence and uniqueness of the solution to (5.29) is an easy consequence

of the Lipschitz property of the functions f0, q, q0, φ0, φ and the Cauchy-Lipschitz

Theorem. The fact that Jφ0,φ0,N is a solution to (5.29) follows from Dynkin’s formula.

We state without proof the similar result for V φ0,N .

176

Page 186: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Proposition 5.7.2. If Assumptions 5.5.1, 5.5.2 & 5.5.3 hold and φ is a Lipschitz

strategy, then V φ0,N is the unique classical solution of the system of ODEs:

0 = θ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +∑

j0,j0 6=i0

(θ(t, j0, x)− θ(t, i0, x))q0(t, i0, j0, α0, x)

+

∑(i,j),j 6=i

(θ(t, i0, x+1

Neij)− θ(t, i0, x))Nxiq(t, i, j, φ(t, i, i0, x), i0, x),

0 = θ(t, i0, x)− g0(i0, x).

(5.30)

The following estimates for Jφ0,φ0,N and V φ0,N will play a crucial role in proving

convergence to the solution of the mean field game. Their proofs are postponed to

the appendix.

Proposition 5.7.3. For any Lipschitz feedback function φ0 for the major player and

Lipschitz feedback function φ for the minor player, there exists a constant L only

depending on T and the Lipschitz constants and bounds of φ0, φ, q0, q, f0, g0 such that

for all N > 0, (t, i0, x) ∈ [0, T ] × 1, . . . ,m0 × SN and j, k ∈ 1, · · · ,m, j 6= k, we

have:

|Jφ0,φ0,N (t, i0, x)| ≤ ‖g0‖∞ + T‖f0‖∞, (5.31)

|Jφ0,φ0,N (t, i0, x+1

Nejk)− Jφ0,φ0,N (t, i0, x)| ≤ L

N. (5.32)

Proposition 5.7.4. For each Lipschitz feedback function φ with Lipschitz constant

Lφ, there exist constants C0, C1, C2, C3, C4 > 0 only depending on m0, m, the Lipschitz

constants and bounds of q0, q, f0, g0, such that for all N > 0, (t, i0, x) ∈ [0, T ] ×

1, . . . ,m0 × SN and j, k ∈ 1, · · · ,m, j 6= k, we have:

|V φ0,N (t, i0, x+

1

Nejk)− V φ

0,N (t, i0, x)| ≤ C0 + C1T + C2T2

Nexp[(C3 + C4Lφ)T ]. (5.33)

We now turn to the problem of the deviating minor player. We consider a strategy

profile where the major player uses a feedback function φ0, the first (N − 1) minor

177

Page 187: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

players use a feedback function φ and the remaining (de facto the deviating) minor

player uses the feedback function φ. We recall that (XNt , X

0,Nt , µNt ) is a Markov

process with infinitesimal generator Gφ0,φ,φN defined as in (5.13). We are interested in

the deviating minor player’s expected cost:

Jφ0,φ,φN (t, i, i0, x)

:= E[ ∫ T

tfφ0,φ(s,XN

s , X0,Ns , µNs )ds+ g(XN

T , X0,NT , µNT )| XN

t = i,X0,Nt = i0, µ

Nt = x

],

as well as the value function of the deviating minor player’s optimization problem:

V φ0,φN (t, i, i0, x) := sup

α↔φ∈AJφ0,φ,φN (t, i, i0, x).

In full analogy with Propositions 5.7.3 and Proposition 5.7.4, we state the following

results without proof.

Proposition 5.7.5. For any Lipschitz feedback function φ0 for the major player and

Lipschitz feedback function φ, φ for the minor player, there exists a constant L only

depending on T and the Lipschitz constants and bounds of φ0, φ, φ, q0, q, f, g such that

for all N > 0, (t, i0, x) ∈ [0, T ] × 1, . . . ,m0 × SN and j, k ∈ 1, · · · ,m, j 6= k, we

have:

|Jφ0,φ,φN (t, i, i0, x)| ≤ ‖g‖∞ + T‖f‖∞, (5.34)

|Jφ0,φ,φN (t, i, i0, x+1

Nejk)− Jφ0,φ,φN (t, i, i0, x)| ≤ L

N. (5.35)

Proposition 5.7.6. There exist constants D0, D1, D2, D3, D4, D5 > 0 depending only

on m0, m, the Lipschitz constants and bounds of q0, q, f, g such that for all Lipschitz

feedback functions φ0 and φ with Lipschitz constants Lφ0 and Lφ respectively, and for

178

Page 188: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

all N > 0, (t, i0, x) ∈ [0, T ]×1, . . . ,m0×SN and j, k ∈ 1, · · · ,m, j 6= k, we have:

|V φ0,φ0,N (t, i, i0, x+

1

Nejk)− V φ0,φ

0,N (t, i, i0, x)|

≤D0 +D1T +D2T

2 +D3Lφ0T

Nexp[(D4 +D5Lφ))T ].

(5.36)

5.7.2 Propagation of Chaos

We now prove two important limiting results. They are related to the propagation

of chaos in the sense that they identify the limiting behavior of an individual when

interacting with the mean field. First, we prove uniform convergence of the value func-

tions of the individual players’ optimization problems. Combined with the gradient

estimates proven in the previous subsection, this establishes the Lipschitz property of

the value functions in the mean field limit. Second, we prove that the expected costs

of the individual players in the (N + 1) - player game converge to their mean field

limits at the rate (N−1/2). This will help us show that the Nash equilibrium of the

mean field game provides approximate Nash equilibria for the finite player games.

Theorem 5.7.7. For all Lipschitz strategies φ0, φ, we have:

supt,i0,x|V φ

0 (t, i0, x)− V φ0,N (t, i0, x)| → 0, N → +∞, (5.37)

supt,i,i0,x

|V φ0,φ(t, i, i0, x)− V φ0,φN (t, i, i0, x)| → 0, N → +∞. (5.38)

Proof. We only provide a proof for (5.37), as (5.38) can be shown in the exact the

same way.

(i) Fix a Lipschitz strategy φ for the minor players. To simplify the notation, we

set vN := V 0,Nφ . Notice that (t, i0, x) → vN(t, i0, x) is only defined on [0, T ] ×

1, . . . ,m0 × SN , so our first step is to extend the domain of vN(t, i0, x) to [0, T ] ×

1, . . . ,m0 × S. This can be done by considering the linear interpolation of vN .

more specifically, for any x ∈ S, we denote xk, k = 1, . . . , 2m−1 the 2m−1 closest

neighbors of x in the set SN . There exists αk, k = 1, . . . , 2m−1 positive constants

179

Page 189: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

such that x =∑2m−1

k=1 αkxk. We then define the extension, still denoted as vN , to be

vN(t, i0, x) :=∑2N−1

k=1 αkvN(t, i0, xk). It is straightforward to verify that vN is contin-

uous in (t, x), Lipschitz in x uniformly in (t, i0), and C1 in t. Using the boundedness

and Lipschitz continuity of vN , f0, q0, q and φ, we obtain a straightforward estimation:

L

N≥∣∣∣∣vN (t, i0, x) + inf

α0∈A0

f0(t, i0, α0, x) +

∑j0,j0 6=i0

(vN (t, j0, x)− vN (t, i0, x))q0(t, i0, j0, α0, x)

+∑

(i,j),j 6=i

(vN (t, i0, x+1

Neij)− vN (t, i0, x))Nxiq(t, i, j, φ(t, i, i0, x), i0, x)

∣∣∣∣,(5.39)

L

N≥ |vN(t, i0, x)− g0(i0, x)|. (5.40)

where the constant L only depends on the bounds and Lipschitz constants of f0, q0,

q and φ.

(ii) Now let us denote v(t, i0, x) := lim sup∗ vN(t, i0, x) and v(t, i0, x) := lim inf∗ vN(t, i0, x)

(see Section 5.8.3 for definitions of the operators lim sup∗ and lim inf∗). We show

that v and v are a viscosity subsolution and a viscosity supersolution of the HJB

equation (5.16) of the major player, respectively. Recall that we assume now that

q does not depend on α. Then since V 0φ is also a viscosity solution to (5.16), the

comparison principle allows us to conclude that v(t, i0, x) = v(t, i0, x) = V 0φ and the

uniform convergence follows by standard arguments.

(iii) It remains to show that v is a viscosity subsolution to the PDE (5.16). The

proof of v being a viscosity supersolution can be done in exactly the same way.

Let θ be a smooth function and (t, i0, x) ∈ [0, T ] × 1, . . . ,m0 × S be such that

(t, x) → v(t, i0, x) − θ(t, x) has a maximum at (t, x) and v(t, i0, x) = θ(t, x). Then

by Lemma 6.1 in Crandall et al. [1992], there exist sequences Nn → +∞, tn → t,

xn → x such that for each n, the mapping (t, x) → vNn(t, i0, x) − θ(t, x) attains a

maximum at tn, xn and δn := vNn(tn, i0, xn) − θ(tn, xn) → 0. Instead of extracting a

180

Page 190: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

subsequence, we may assume that vNn(tn, ·, xn)→ (r1, . . . , rm0), where rj0 ≤ v(t, j0, x)

and ri0 = v(t, i0, x).

For t = T , v(t, i0, x) ≤ g0(i0, x) follows easily from (5.40). Now assume that t < T .

Instead of extracting a subsequence, we may assume that tn < T for all n. Then by

maximality we have ∂tθ(tn, xn) = ∂tvNn(tn, xn). Again by maximality, we have for all

i, j = 1, . . . ,m, i 6= j:

vNn(tn, i0, xn +1

Nnei,j)− vNn(tn, i0, xn) ≤ θ(tn, i0, xn +

1

Nnei,j)− θ(tn, i0, xn).

Injecting the above inequalities into the estimation (5.39) and using the postivity of

q, we obtain:

− L

Nn≤ ∂tθ(tn, i0, xn)

+ infα0∈A0

f0(tn, i0, α0, xn) +

∑j0,j0 6=i0

(vNn(tn, j0, xn)− vNn(tn, i0, xn))q0(tn, i0, j0, α0, xn)

+

∑(i,j),j 6=i

[θ(tn, i0, xn +1

Nneij)− θ(tn, i0, xn)]Nn(xn)iq(tn, i, j, φ(tn, i, i0, xn), i0, xn).

Taking the limit in n we obtain:

0 ≤ ∂tθ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +

∑j0,j0 6=i0

(rj0 − ri0)q0(t, i0, j0, α0, x)

+∑

(i,j),j 6=i

[1(j 6= m)∂xjθ(t, i0, x)− 1(i 6= m)∂xiθ(t, i0, x)]xiq(t, i, j, φ(t, i, i0, x), i0, x).

Now since q0 is positive and rj0 ≤ v(t, j0, x) for all j0 6= i0 and ri0 = v(t, i0, x) we

have:

infα0∈A0

f0(t, i0, α0, x) +

∑j0,j0 6=i0

(rj0 − ri0)q0(t, i0, j0, α0, x)

≤ infα0∈A0

f0(t, i0, α0, x) +

∑j0,j0 6=i0

(v(t, j0, x)− v(t, i0, x))q0(t, i0, j0, α0, x)

.

The desired inequality for the viscosity subsolution follows immediately. This com-

pletes the proof.

181

Page 191: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

As an immediate consequence of the uniform convergence and the gradient esti-

mates for the value functions V φ0,N and V φ0,φ

N , we have:

Corollary 5.7.8. Under Assumption 5.5.1, 5.5.2 & 5.5.3, for all Lipschitz feed-

back functions φ0 and φ with Lipschitz constants Lφ0 and Lφ, respectively, the value

functions V φ0 and V φ0,φ are Lipschitz in (t, x). More specifically, there exist strictly

positive constants B, Ci, i = 0, . . . , 4, Di, i = 0, . . . , 5 that only depend on the bounds

and Lipschitz constants of f , f0, g0, g, q and q0 such that:

|V φ0 (t, i0, x)− V φ

0 (s, i0, y)|

≤ B|t− s|+ (C0 + C1T + C2T2) exp((C3 + C4Lφ)T )‖x− y‖,

(5.41)

|V φ0,φ(t, i, i0, x)− V φ0,φ(s, i, i0, y)|

≤ B|t− s|+ (D0 +D1T +D2T2 +D3Lφ0T ) exp((D4 +D5Lφ)T )‖x− y‖.

(5.42)

Proof. The Lipschitz property in x is an immediate consequence of Theorem 5.7.7,

Proposition 5.7.4 and Proposition 5.7.6. To prove the Lipschitz property on t, we

remark that for each N , V φ0,N and V φ0,φ

N are Lipschitz in t, uniformly in N . Indeed

V φ0,N is a classical solution of the system of ODEs (5.30). Then it is clear that d

dtV φ

0,N

is bounded by the bounds of f0, g0, q0 and q. We deduce that V φ0,N is Lipschitz in t

with a Lipschitz constant that only depends on m,m0 and the bounds of f0, g0, q0

and q. By convergence of V φ0,N , we conclude that V φ

0 is also Lipschitz in t and shares

the Lipschitz constant of V φ0,N . The same argument applies to V φ0,φ.

Remark 5.7.9. From Corollary 5.7.8, we see that Assumption 5.5.6 holds when T

is sufficiently small. Indeed we can first choose Lφ0 > La(1 + maxB,C0) and

Lφ > Lb(1 + maxB,D0) and then choose T sufficiently small, so that the Lipschitz

constant of V φ0 is smaller than (Lφ0/La − 1) and the Lipschitz constant of Vφ0,φ is

smaller than (Lφ/Lb − 1).

182

Page 192: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We now state our second result on the propagation of chaos: the expected cost of

an individual player in the (N + 1)-player game converges to the expected cost in the

mean field game at a rate of N−1/2.

Theorem 5.7.10. There exists a constant L depending only on T and the Lipschitz

constants of φ0, φ, φ, f , f0, g, g0, q and q0 such that for all N > 0, t ≤ T , x ∈ SN ,

i = 1, . . . ,m and i0 = 1, . . . ,m0, we have:

|Jφ0,φ0,N (t, i0, x)− Jφ0,φ0 (t, i0, x)| ≤L/√N,

|Jφ0,φ,φN (t, i, i0, x)− Jφ0,φ,φ(t, i, i0, x)| ≤L/√N.

Our proof is based on standard techniques from the convergence rate analysis of

numerical schemes for viscosity solutions of PDEs, c.f. Briani et al. [2012] and Barles

and Souganidis [1991] for example. The key step of the proof is the construction

of a smooth subsolution and a smooth supersolution of the PDEs (5.23) and (5.24)

that characterize Jφ0,φ0 and Jφ0,φ,φ, respectively (see Proposition 5.4.6). We construct

these solutions by mollifying an extended version of Jφ0,φ0,N . Then we derive the bound

by using the comparison principle. In the following, we detail the proof for the

convergence rate of the major player’s expected cost. The case of the generic minor

player can be dealt with in exactly the same way.

Since Jφ0,φ0,N (t, i, x) is only defined for x ∈ SN , in order to mollify Jφ0,φ0,N , we need to

first construct an extension of Jφ0,φ0,N defined for all x ∈ O for an open set O containing

S. To this end, we consider the following system of ODEs:

0 = θ(t, i0, x) + f0(t, i0, φ0(t, i0, x), x) +∑

j0,j0 6=i0

(θ(t, j0, x)− θ(t, i0, x))q0(t, i0, j0, φ0(t, i0, x), x)

+∑

(i,j),j 6=i

(θ(t, i0, x+1

Neij)− θ(t, i0, x))N maxxi, 0q(t, i, j, φ(t, i, i0, x), i0, φ0(t, i0, x), x),

0 = θ(T, i0, x)− g0(i0, x).

183

Page 193: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Here φ0, φ, f0, g0 and q0 are, respectively, extensions of φ0, φ, f0, g0 and q0 from

x ∈ S to x ∈ RN−1, which are Lipschitz in x. The following is proved using the same

arguments as for Proposition 5.7.1.

Lemma 5.7.11. The system of ODEs (5.29) admits a unique solution vN defined in

[0, T ]× 1, . . . ,m0 × Rm−1. moreover we have:

(i) vN(t, i0, x) is Lipschitz in x uniformly in t and i0 and the Lipschitz constant only

depends on T and the Lipschitz constants of φ0, φ, f0, g0 and q0.

(ii) vN(t, i0, x) = Jφ0,φ0,N (t, i0, x) for all x ∈ SN .

To construct smooth super and sub solutions, we use a family of mollifiers ρε

defined by ρε(x) := ρ(x/ε)/εN−1, where ρ is a smooth and positive function with

compact support in the unit ball of Rm−1 and satisfying∫Rm−1 ρ(x)dx = 1. For ε > 0,

we define vεN as the mollification of vN on [0, T ]× 1, . . . ,m0 × S:

vεN (t, i0, x) :=

∫y∈Rm−1

vN (t, i0, x− y)ρε(y)dy.

Using the Lipschitz property of φ0, φ, f0, g0 and q0 and straightforward estimates on

the mollifier ρε, we obtain the following properties on vεN .

Lemma 5.7.12. vεN is C∞ in x and C1 in t. moreover, there exists a constant C

that depends only on T and the Lipschitz constants of φ0, φ, f0, g0 and q0 such that

for all i0 = 1, . . . ,m0, i = 1, . . . ,m, t ≤ T and x, y ∈ S, the following estimations

hold:

|[Gφ0,φ0,N vεN ](t, i0, x) + f0(t, i0, φ0(t, i0, x), x)| ≤ Lε, (5.43)

|vεN (T, i0, x)− g0(i0, x)| ≤ Lε, (5.44)

|∂xivεN (t, i0, x)− ∂xivεN (t, i0, y)| ≤ L

ε‖x− y‖. (5.45)

184

Page 194: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

We are now ready to prove Theorem 5.7.10. We construct viscosity super and sub

solutions by adjusting vεN with a linear function on t. Then the comparison principle

allows us to conclude.

Proof. (of Theorem 5.7.10) We denote by L a generic constant that only depends on

T and the Lipschitz constants of φ0, φ,f0, g0 and q0. Using (5.43), (5.44) and (5.45)

in Lemma 5.7.12, we obtain:

|[Gφ0,φ0 vεN ](t, i0, x) + f0(t, i0, φ0(t, i0, x), x)| ≤ L(ε+

1

),

|vεN (T, i0, x)− g0(i0, x)| ≤ Lε.

Next we define:

v±N (t, i0, x) := vεN (t, i0, x)± [L(ε+1

Nε)(T − t) + Lε].

Since v−N and v+N are smooth, the above estimation immediately implies that v−N

and v+N are viscosity sub and super solutions of the PDE (5.23), respectively. By

Proposition 5.4.6, Jφ0,φ0 is a continuous viscosity solution to the PDE (5.23). Then

by the comparison principle we have v−N ≤ Jφ0,φ0 ≤ v+N , which implies:

|vεN(t, i0, x)− Jφ0,φ0 (t, i0, x)| ≤ L(ε+1

Nε)(T − t) + Lε.

Now using the property of the mollifier and Lemma 5.7.11, we have for all t ≤ T ,

i0 = 1, . . . ,m0 and x ∈ SN :

|vεN(t, i0, x)− Jφ0,φ0,N (t, i0, x)| = |vεN(t, i0, x)− vN(t, i0, x)| ≤ Lε.

The desired results follow by combining the above inequalities and choosing ε =

1/√N .

185

Page 195: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

5.7.3 Approximate Nash Equilibria

The following is an immediate consequence of the above propagation of chaos results.

Theorem 5.7.13. Assume that the Mean Field Game attains Nash equilibrium when

the major player uses a Lipschitz feedback function φ0 and all the minor players uses

a Lipschitz feedback function φ. Denote L0 the Lipschitz constant for φ0 and φ. Then

for any L ≥ L0, the strategy profile (α0, α1, . . . , αN) defined by α0t := φ0(t,X0,N

t , µNt )

and αnt := φ(t,Xn,Nt , X0,N

t , µNt ) gives an approximative Nash equilibrium within all

the L−Lipschitz strategies. More specifically, there exist constants C > 0 and N0 ∈ N

depending on L0 such that for any N ≥ N0, any L-Lipschitz feedback function φ0 for

major player and any L-Lipschitz feedback function φ, we have:

J0,N(φ0, φ, . . . , φ) ≤ J0,N(φ0, φ, . . . , φ)) + C/√N,

JN(φ0, φ, . . . , φ) ≤ JN(φ0, φ, . . . , φ, . . . φ) + C/√N.

5.8 Appendix

5.8.1 Proof of Lemma 5.4.3

Recall the dynamic programming operator defined in (5.18). Let Φ be a mapping

on [0, T ] × 1, . . . ,m0 × S such that Φ(·, i0, ·) is C1 in [0, T ] × S and Φ(·, j0, ·) is

continuous in [0, T ]× S for j0 6= i0. We are going to evaluate the following limit:

limh→0

Ih := limh→0

1

h[Tt,t+hΦ(t+ h, ·, ·)](i0, x)− Φ(t, i0, x) .

(i) Let us first assume that Φ(·, j0, ·) is C1 in [0, T ]×S for all j0. Consider a constant

control α0. Then by the definition of the operator Tt,t+h we have:

Ih ≤1

h

E[∫ t+h

tf0(u,X0

u, α0, µu)du+ Φ(t+ h,X0t+h, µt+h)|X0

t = i0, µ = x

]− Φ(t, i0, x)

.

186

Page 196: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Using the infinitesimal generator of the process (u,X0u, µu), the right hand side of

the above inequality has a limit. Taking the limit and then the infimum over α, we

obtain:

lim suph→0

Ih ≤ ∂tΦ(t, i0, x)

+ infα0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

∂xjΦ(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

∂xkΦ(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+∑j0 6=i0

[Φ(t, j0, x)− Φ(t, i0, x)]q0(i0, j0, α0, x)

.

On the other hand, for all h > 0, there exists a control φh0 such that:

E[∫ t+h

tf0(u,X0

u, φh0(u,X0

u, µu), µu)du+ Φ(t+ h,X0t+h, µt+h)

∣∣∣∣X0t = i0, µ = x

]≤ Tt,t+hΦ(t+ h, ·, ·))(i0, x) + h2.

This implies:

Ih ≥1

h

E[∫ t+h

tf0(u,X0

u, φh0(u,X0

u, µu), µu)du+ Φ(t+ h,X0t+h, µt+h)

∣∣∣∣X0t = i0, µ = x

]− Φ(t, i0, x)

− h.

Since Φ(·, j0, ·) is C1 for all j0, this can be further written using the infinitesimal

generator:

Ih ≥1

hE[∫ t+h

tf0(u,X0

u, φh0(u,X0

u, µu), µu) + [G0φh0 ,φ

Φ](u,X0u, µu)du|X0

t = i0, µ = x

]− h.

Taking the supremum over the control and applying the dominated convergence

theorem, we obtain:

lim infh→0

Ih

187

Page 197: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

≥ ∂tΦ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

∂xjΦ(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

∂xkΦ(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+∑j0 6=i0

[Φ(t, j0, x)− Φ(t, i0, x)]q0(i0, j0, α0, x)

.

This proves the lemma for Φ such that Φ(·, j0, ·) is C1 for all j0.

(ii) Now take a continuous mapping Φ and only assume that Φ(·, i0, ·) is C1. Applying

the Weierstrass approximation theorem, for any ε > 0, there exists a C1 function φεj0

on [0, T ]× S for all j0 6= i0 such that:

sup(t,x)∈[0,T ]×S

|Φ(t, j0, x)− φεj0(t, x)| ≤ ε.

Define Φε(t, j0, x) := φεj0(t, x) + ε for j0 6= i0 and Φε(t, i0, x) := Φ(t, i0, x). Then we

have Φε ≥ Φ. By the monotonicity of the operator Tt,t+h we have:

1

h[Tt,t+hΦε(t+ h, ·, ·)](i0, x)− Φε(t, i0, x) ≥ 1

h[Tt,t+hΦ(t+ h, ·, ·)](i0, x)− Φ(t, i0, x) := Ih.

Now applying the results from step (i), we obtain:

lim suph→0

Ih

≤ ∂tΦε(t, i0, x) + inf

α0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

∂xjΦε(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)

m−1∑k=1

∂xkΦε(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+∑j0 6=i0

[Φε(t, j0, x)− Φε(t, i0, x)]q0(t, i0, j0, α0, x)

= ∂tΦ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

∂xjΦ(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

∂xkΦ(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

188

Page 198: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

+∑j0 6=i0

[Φε(t, j0, x)− Φ(t, j0, x) + Φ(t, j0, x)− Φ(t, i0, x)]q0(t, i0, j0, α0, x)

≤ ∂tΦ(t, i0, x) + infα0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

∂xjΦ(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

∂xkΦ(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+∑j0 6=i0

[Φ(t, j0, x)− Φ(t, i0, x)]q0(t, i0, j0, α0, x)

+ εL.

The last equality is due to the fact that q0 is bounded and |Φε(t, j0, x)−Φ(t, j0, x)| ≤ ε.

We can write a similar inequality for lim infh→0 Ih. Then tending ε to 0 yields the

desired result.

5.8.2 Proof of Theorem 5.4.4

In this section, we present the proof of the comparison principle for the HJB equation

associated with the major player’s optimization problem. The arguments used in this

proof can be readily applied to prove the uniqueness of the solution to the deviating

minor player’s HJB equation (5.20) (c.f. Theorem 5.4.5). The same argument can also

be used to prove the uniqueness result for equations (5.23) and (5.24) (c.f. Proposition

5.4.6).

Let v and w be, respectively, a viscosity subsolution and a viscosity supersolution

to equation (5.16). Our objective is to show v(t, i0, x) ≤ w(t, i0, x) for all 1 ≤ i0 ≤ m0,

x ∈ S and t ∈ [0, T ].

189

Page 199: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

(i) Without loss of generality, we may assume that v is a viscosity subsolution of:

0 = − η + ∂tv0(t, i0, x)

+ infα0∈A0

f0(t, i0, α0, x) +

∑j0 6=i0

[v0(t, j0, x)− v0(t, i0, x)]q0(t, i0, j0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

∂xkv0(t, i0, x)q(t,m, k, φ(t,m, i0, x), i0, α0, x)

+m−1∑i,j=1

∂xjv0(t, i0, x)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

,

0 = v0(T, i0, x)− g0(i0, x), ∀x ∈ S,

(5.46)

where η > 0 is a small parameter. Indeed, we may consider the function vη(t, i, x) :=

v(t, i, x)−η(T−t). Then it is easy to see that vη is a viscosity subsolution to the above

equation. If we can prove vη ≤ w, then tending η to 0 yields v ≤ w. In the following,

we will only consider the subsolution v to equation (5.46) and the supersolution w to

equation (5.16), and prove that v ≤ w.

(ii) For ε > 0 and 1 ≤ i0 ≤ m0, consider the function Γi0,ε defined on [0, T ]2 × S2:

Γi0,ε(t, s, x, y) := v(t, i0, x)− w(s, i0, y)− 1

ε|t− s|2 − 1

ε‖x− y‖2,

where ‖ · ‖ is the euclidian norm on R(m0−1). Since Γi0,ε is a continuous function

on a compact set, it attains the maximum denoted as Ni0,ε. Denote by (t, s, x, y)

the maximizer (which obviously depends on ε and i0, but for simplicity, we suppress

this dependency in the notation). We show that for all 1 ≤ i0 ≤ m0, there exists a

sequence εn → 0 and the corresponding maximizer (tn, sn, xn, yn) such that

(tn, sn, xn, yn)→ (t, t, x, x), where (t, x) := arg sup(t,x)∈[0,T ]×S

v(t, i0, x)−w(t, i0, x), (5.47a)

1

εn|tn − sn|2 +

1

εn‖xn − yn‖2 → 0, (5.47b)

190

Page 200: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Ni0,ε → Ni0 := sup(t,x)∈[0,T ]×S

v(t, i0, x)− w(t, i0, x). (5.47c)

Indeed, for any (t, x) ∈ [0, T ]× S, we have v(t, i0, x)− w(t, i0, x) = Γi0,ε(t, t, x, x) ≤

Ni0,ε. Taking the supremum, we obtain Ni0 ≤ Ni0,ε and therefore:

1

ε|t− s|2 +

1

ε‖x− y‖2 ≤ v(t, i0, x)− w(s, i0, y)−Ni0 ≤ 2L−Ni0 .

The last inequality comes from the fact that v and w are bounded on the compact

set [0, T ]×S. It follows that |t− s|2 +‖x− y‖2 → 0. Now since the sequence (t, s, x, y)

(indexed by ε) is in a compact set, we can extract a subsequence εn → 0 such that

(tn, sn, xn, yn)→ (t, t, x, x). We have the following inequality:

Ni ≤ v(tn, i0, xn)− w(sn, i0, yn)− 1

εn|tn − sn|2 −

1

εn‖xn − yn‖2

= Ni0,εn ≤ v(tn, i0, xn)− w(sn, i0, yn).

Noticing that v(tn, i0, xn)−w(sn, i0, yn)→ v(t, i0, xn)−w(t, i0, xn) ≤ Ni0 , we deduce

that Ni0 = v(t, i0, xn)−w(t, i0, xn) which implies (5.47a). (5.47b) and (5.47c) follows

easily by taking the limit in n in the above inequality.

(iii) Now we prove the comparison principle. Using the notation introduced in step

(ii), we need to prove Ni0 ≤ 0 for all 1 ≤ i0 ≤ m0. Assume that there exists

1 ≤ i0 ≤ m0 such that:

Ni0 = sup1≤j0≤m0

Nj0 > 0.

We now work towards a contradiction. Without loss of generality we assume that

Ni0 > Nj0 for all j0 6= i0. We then consider the subsequence (εn, tn, sn, xn, yn) →

(0, t, t, x, x) with regard to i0 constructed in step (ii), for which (5.47a), (5.47b) and

(5.47c) are satisfied. Since v(t, i0, x) − w(t, i0, x) = Ni0 > 0, we have t 6= T . Instead

of extracting a subsequence, we may assume that tn 6= T and sn 6= T for all n ≥ 0.

191

Page 201: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

moreover for any j0 6= i0, we have:

v(t, j0, x)− w(t, j0, x) ≤ Nj < Ni = v(t, i0, x)− w(t, i0, x).

Since v(tn, j0, xn) − w(sn, j0, yn) → v(t, j0, x) − w(t, j0, x), instead of extracting a

subsequence, we can assume that for all j0 6= i0 and n ≥ 0:

v(tn, j0, xn)− w(sn, j0, yn) ≤ v(tn, i0, xn)− w(sn, i0, yn). (5.48)

In the following we suppress the index n for the sequence (εn, tn, sn, xn, yn) for sake

of simplicity of notation. By definition of the maximizer, for any (t, x) ∈ [0, T ] × S,

we have:

v(t, i0, x)−w(s, i0, y)− 1

ε‖x− y‖2− 1

ε|t− s|2 ≤ v(t, i0, x)−w(s, i0, y)− 1

ε‖x− y‖2− 1

ε|t− s|2.

Therefore v(·, i0, ·)− φ attains a maximum at (t, x) where:

φ(t, x) :=1

ε‖x− y‖2 +

1

ε|t− s|2, ∂tφ(t, x) =

2

ε(t− s), ∇φ(t, x) =

2

ε(x− y). (5.49)

Similarly w(·, i0, ·)− ψ attains a minimum at (s, y) where:

ψ(t, x) := −1

ε‖x− x‖2 − 1

ε|t− t|2 ∂tψ(s, y) =

2

ε(t− s) ∇ψ(s, y) =

2

ε(x− y) (5.50)

Since s 6= T and t 6= T , the definition of the viscosity solution, along with equation

(5.49) and (5.50) gives the following inequalities:

η ≤ 2

ε(t− s) + inf

α0∈A0

f0(t, i0, α0, x) +

m−1∑i,j=1

2

ε(xj − yj)xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)

+ (1−m−1∑k=1

xk)m−1∑k=1

2

ε(xk − yk)q(t, m, k, φ(t, m, i0, x), i0, α0, x)

+∑j0 6=i0

[v(t, j0, x)− v(t, i0, x)]q0(t, i0, j0, α0, x)

,

192

Page 202: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

0 ≥ 2

ε(t− s) + inf

α0∈A0

f0(s, i0, α0, y) +m−1∑i,j=1

2

ε(xj − yj)yiq(s, i, j, φ(s, i, i0, y), i0, α0, y)

+ (1−m−1∑k=1

yk)

m−1∑k=1

2

ε(xk − yk)q(s,m, k, φ(s,m, i0, y), i0, α0, y)

+∑j0 6=i0

[w(s, j0, y)− w(s, i0, y)]q0(s, i0, j0, α0, y)

.

Subtracting the above two inequalities, we obtain:

0 < η ≤ I1 + I2 + I3, (5.51)

where the three terms I1, I2 and I3 will be dealt with in the following. For I1 we have:

I1 := supα0∈A0

f0(t, i0, α0, x)− f0(s, i0, α0, y)

+ supα0∈A0

∑j0 6=i0

[v(t, j0, x)− v(t, i0, x)]q0(t, i0, j0, α0, x)

−∑j0 6=i0

[w(s, j0, y)− w(s, i0, y)]q0(s, i0, j0, α0, y)

≤ supα0∈A0

f0(t, i0, α0, x)− f0(s, i0, α0, y)

+ supα0∈A0

∑j0 6=i0

[v(t, j0, x)− v(t, i0, x)](q0(t, i0, j0, α0, x)− q0(s, i0, j0, α0, y))

+ supα0∈A0

∑j0 6=i0

[v(t, j0, x)− w(s, j0, y)− v(t, i0, x) + w(s, i0, y)]q0(s, i0, j0, α0, y)

≤L(|t− s|+ ‖x− y‖) + 2CL(|t− s|+ ‖x− y‖) + 0.

In the last inequality we use (5.48) and the fact that q0(s, i0, j0, α0, y) ≥ 0 for j0 6= i0. We

also use the Lipschitz property of f0 and q0. Now in light of (5.47a), we obtain I1 → 0 as

ε→ 0. Now turning to I2:

I2 := supα0∈A0

m−1∑i,j=1

2

ε(xj − yj) [xiq(t, i, j, φ(t, i, i0, x), i0, α0, x)− yiq(s, i, j, φ(s, i, i0, y), i0, α0, y)]

≤ supα0∈A0

m−1∑i,j=1

2

ε|(xj − yj)(xi − yi)q(t, i, j, φ(t, i, i0, x), i0, α0, x)|

193

Page 203: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

+ supα0∈A0

m−1∑i,j=1

2

ε|yi(xj − yj)(q(t, i, j, φ(t, i, i0, α0, x), i0, α0, x)− q(s, i, j, φ(s, i, i0, y), i0, α0, y))|

≤m−1∑i,j=1

2

εC|xj − yj ||xi − yi|+

m−1∑j=1

2

ε|xj − yj |L(m− 1)(|t− s|+ ‖x− y‖),

where in the last inequality we used the Lipschitz property of q and φ uniformly in α, as

well as the boundedness of the function q. It follows that I2 ≤ C 1ε (|t− s|

2 + ‖x− y‖2) and

by (5.47b) we see that I2 → 0 as ε→ 0. Finally, we deal with I3, which is defined by:

I3 := supα0∈A0

(1−

m−1∑k=1

xk)

m−1∑k=1

2

ε(xk − yk)q(t, m, k, φ(t, m, i0, x), i0, α0, x)

− (1−m−1∑k=1

yk)m−1∑k=1

2

ε(xk − yk)q(s,m, k, φ(s,m, i0, y), i0, α0, y)

.

Using a similar estimation as for I2, we obtain I3 ≤ C 1ε (|t− s|

2 + ‖x− y‖2). Therefore by

tending ε to 0 in the inequality (5.51), we obtain a contradiction. This completes the proof.

5.8.3 Proof of Lemma 5.5.8

The main tool we use for the proof is the theory of limit operation on viscosity solutions.

We refer the reader to Crandall et al. [1992] for an introductory presentation of limit op-

eration on viscosity solutions to non-linear second order PDEs. Here we adapt the results

established therein to the case of a coupled system of non-linear first order PDEs, which

is the HJB equation we are interested in throughout the paper. Let O be some locally

compact subset of Rd and (Fn)n≥0 be a sequence of functions defined on 1, . . . ,m × O.

We define the limit operator lim sup∗ and lim inf∗ as follow:

lim sup∗Fn(i, x) := limn→+∞,ε→0

supFk(i, y)|k ≥ n, ‖y − x‖ ≤ ε,

lim inf∗Fn(i, x) := limn→+∞,ε→0

infFk(i, y)|k ≥ n, ‖y − x‖ ≤ ε.

194

Page 204: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Intuitively, we expect that the limit of a sequence of viscosity solutions solves the limiting

PDE. It turns out that this is the proper definition of the limit operation, as is stated in

the following lemma:

Lemma 5.8.1. Let O be a locally compact subset of Rd, (un)n≥0 be a sequence of continuous

functions defined on 1, . . . ,m × O and Hn be a sequence of functions defined on [0, T ]×

1, . . . ,m ×O×Rm ×Rd, such that for each n, un is a viscosity solution to the system of

PDEs Hn(t, i, x, u, ∂tu(·, i, ·),∇u(·, i, ·)) = 0, ∀1 ≤ i ≤ m, in the sense of Definition 5.4.2.

Assume that for each i, x, d, p, the mapping u → H(t, i, x, u, d, p) is non-decreasing in the

j-th component of u, for all j 6= i. Then u∗ := lim sup∗ un (resp. u∗ := lim inf∗ un) is a

viscosity subsolution (resp. supersolution) to H∗(t, i, x, u, ∂tu(i, ·),∇u(i, ·)) = 0,∀1 ≤ i ≤ m

(resp. H∗(t, i, x, u, ∂tu(·, ·, i, ·),∇u(·, ·, i, ·)) = 0,∀1 ≤ i ≤ m).

The proof requires the definition of viscosity subsolution and supersolution based on

the notion of first-order subjet and superjet (see Definition 2.2 in Crandall et al. [1992]).

Let u be a continuous function defined on [0, T ]×1, . . . ,m×O. We define the first order

superjet of u on (t, i, x) to be:

J +u(t, i, x)

:=

(d, p) ∈ R× Rd : u(s, i, y) ≤ u(t, i, x) + d(s− t) + (y − x)p+ o(|t− s|+ ‖y − x‖), (s, y)→ (t, x)

Then it can be showed that u is a viscosity subsolution (resp. supersolution) to the equation:

H(t, i, x, u, ∂tu(i, ·),∇u(i, ·)) = 0,

if only if for all (t, i, x) and (d, p) ∈ J +u(t, i, x) (resp. (d, p) ∈ J −u(t, i, x)), we have

H(t, i, x, u, d, p) ≥ 0 (resp. H(t, i, x, u, d, p) ≤ 0). We now give the detail of the proof of

Lemma 5.8.1.

Proof. Fix (t, i, x) ∈ [0, T ]×1, . . . ,m×O and let (d, p) ∈ J +u∗(t, i, x). We want to show

that:

H∗(t, i, x, u∗(t, ·, x), d, p) ≥ 0

195

Page 205: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where u∗(t, ·, x) represents the m−dimensional vector [u∗(t, k, x)]1≤k≤m. By Lemma 6.1 in

Crandall et al. [1992], there exists a sequence nj → +∞, xj ∈ O, (dj , pj) ∈ J +unj (tj , i, xj)

such that:

(tj , xj , unj (tj , i, xj), dj , pj)→ (t, x, u∗(t, i, x), d, p), j → +∞.

Since un is viscosity subsolution to Hn = 0, we have:

Hnj (tj , i, xj , unj (tj , ·, xj), dj , pj) ≥ 0.

Now let us denote:

Sun,ε(t, k, x) := supuj(s, k, y),max(|s− t|, ‖y − x‖) ≤ ε, j ≥ n.

Then we have u∗(t, i, x) = limn→+∞,ε→0 Sun,ε(t, i, x). Let us fix δ > 0. Then there exists

ε0 > 0 and N0 > 0 such that for all ε ≤ ε0 and j ≥ N0, we have:

|Sunj ,ε(t, k, x)− u∗(t, k, x)| ≤ δ, ∀k 6= i.

Moreover, there exists N > 0 such that for all j ≥ N ,

‖(tj , xj , unj (tj , i, xj), dj , pj)− (t, x, u∗(t, i, x), d, p)‖ ≤ δ ∧ ε0.

Then for any j ≥ N0 ∨ N , we have ‖(tj , xj) − (t, x)‖ ≤ ε0, and by definition of Suε,n we

deduce that unj (tj , k, xj) ≤ Sunj ,ε0(t, k, x) for all k 6= i. By the monotonicity property of

Hnj , we have:

Hnj (tj , i, xj , Sunj ,ε0

(t, 1, x), . . . , Sunj ,ε0(t, i−1, x), unj (tj , i, xj), Sunj ,ε0

(t, i+1, x), . . . , Sunj ,ε0(t,m, x), dj , pj) ≥ 0.

Now in the above inequality, all the arguments of Hnj except i are located in a ball of

radius δ centered on the point (t, x, u∗(t, ·, x), d, p). Thus, we have:

SHnj ,δ(t, i, x, u∗(t, ·, x), d, p) ≥ 0,

196

Page 206: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

where we have defined:

SHn,δ(t, i, x, u, d, p) := supHj(s, i, y, v, e, q), ‖(s, y, v, e, q)− (t, x, u, d, p)‖ ≤ ε, j ≥ n.

We have just proved that for any δ > 0, there exists m > 0 such that for any j ≥ m, we

have SHnj ,δ(t, i, x, u∗(t, ·, x), d, p) ≥ 0. Since we have:

H∗(t, i, x, , u∗(t, ·, x), d, p) = limn→+∞,ε→0

SHnj ,δ(t, i, x, u∗(t, ·, x), d, p),

we deduce that H∗(t, i, x, , u∗(t, ·, x), d, p) ≥ 0.

Now going back to the proof of Lemma 5.5.8, we consider a converging sequence of

elements (φn0 , φn) → (φ0, φ) in K, where we have defined K to be the collection of major

and minor player’s controls (φ0, φ) that are L−Lipschitz in (t, x). We denote V n0 (resp. V0)

the value function of the major player’s control problem associated with the controls φn

(resp. φ). We also use the notation V ∗0 := lim supV n0 and V0∗ := lim inf V n

0 . For all n ≥ 0,

we define the operator Hn:

Hn0 (t, i0, x, u, d, p) := d+ inf

α0∈A0

f0(t, i0, α0, x) +

∑j0 6=i0

(uj0 − ui0)q0(t, i0, j0, α0, x)

+ (1−m∑k=1

xk)m−1∑k=1

pkq(t,m, k, φn(t,m, i0, x), i0, α0, x)

+

m−1∑k,l=1

plxkq(t, k, l, φn(t, k, i0, x), i0, α0, x)

, if t < T,

Hn0 (T, i0, x, u, d, p) := g0(i0, x)− ui0 .

Then V n0 is viscosity solution to the equation Hn

0 = 0. It is clear to see that the operator

Hn0 satisfies the monotonicity condition in Lemma 5.8.1. To evaluate H∗0 := lim sup∗Hn

0

and H0∗ := lim inf∗Hn0 , we remark that for each 1 ≤ i0 ≤ m0, the sequence of functions

(t, x, u, d, p)→ Hn(t, i0, x, u, d, p) is equicontinuous. Indeed, this is due to the fact that the

sequence φn is equicontinuous and the function q is Lipschitz. Therefore H∗0 and H0∗ are

197

Page 207: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

simply the limit in the pointwise sense when t < T . When t = T , the boundary condition

needs to be taken into account. The following computation is straightforward to verify:

Lemma 5.8.2. Define the operator H0 as:

H0(t, i0, x, u, d, p) := d+ infα0∈A0

f0(t, i0, α0, x) +

∑j0 6=i0

(uj0 − ui0)q0(t, i0, j0, α0, x)

+ (1−m∑k=1

xk)

m−1∑k=1

pkq(t,m, k, φ(t,m, i0, x), i0, α0, x)

+m−1∑k,l=1

plxkq(t, k, l, φ(t, k, i0, x), i0, α0, x)

, ∀t ≤ T.

Then we have:

H∗0 (t, i0, x, u, d, p) = H0∗(t, i0, x, u, d, p) = H0(t, i0, x, u, d, p), if t < T,

H∗0 (T, i0, x, u, d, p) = max(g0(i0, x)− ui0), H0(T, i0, x, u, d, p),

H0∗(T, i0, x, u, d, p) = min(g0(i0, x)− ui0), H0(T, i0, x, u, d, p).

From the proof of Theorem 5.4.1, we see immediately that V0 is a viscosity subsolution

(resp. supersolution) of H∗0 = 0 (resp. H0∗ = 0). By Lemma 5.8.2, V ∗0 (resp. V0∗) is a

viscosity subsolution of H∗0 = 0 (resp. supersolution of H0∗ = 0). Indeed, following exactly

the proof of Theorem 5.4.4, we can show that a viscosity supersolution of H0∗ = 0 is greater

than a viscosity subsolution of H∗0 = 0. By definition of lim sup∗ and lim inf∗, we have

V ∗0 ≥ V0∗. It follows that V0 ≤ V0∗ ≤ V ∗0 ≤ V0, and therefore we have lim sup∗ V n0 =

lim inf∗ Vn

0 = V0. Then we obtain the uniform convergence of V n0 to V0 following Remark

6.4 in Crandall et al. [1992].

5.8.4 Proof of Propositions 5.7.3 & 5.7.4

We use similar techniques as in Gomes et al. [2013] where the author provides gradient

estimates for N-player games without major player. Let us first remark that the system of

198

Page 208: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

ODEs (5.29) can be written in the following form:

−θm(t) = fm(t) +∑m′ 6=m

am′m(t)(θm′ − θm), θm(T ) = bm,

where we denote the index m := (i0, x) ∈ 1, . . . ,m0 × SN and we notice that am′m ≥ 0

for all m′ 6= m. This can be further written in the compact form:

− θ(t) = f(t) +m(t)θ, θ(T ) = b, (5.52)

where m is a matrix indexed by m, with all off-diagonal entries being positive and the sum

of every row equals 0. Define ‖ · ‖ to be the uniform norm of a vector: ‖b‖ := maxm |bm|.

Instead of proving (i) in Proposition 5.7.3, we prove a more general result, which is a

consequence of Lemma 4 and Lemma 5 in Gomes et al. [2013].

Lemma 5.8.3. Let φ be a solution to the ODE (5.52). Assuming f is bounded, we have:

‖θ(t)‖ ≤∫ T

t‖f(s)‖ds+ ‖b‖.

Proof. For any t ≤ s ≤ T we denote the matrix K(t, s) as the solution of the following

system:

dK(t, s)

dt= −m(t)K(t, s), K(s, s) = I,

where I stands for the m by m identity matrix. Now for θ, we clearly have:

−θ(s) = f(s) +m(s)θ(s) ≤ ‖f(s)‖e+m(s)θ(s),

where we denote e to be the vector with all the components equal to 1. Then using Lemma

5 in Gomes et al. [2013], we have:

−K(t, s)θ(s) ≤ K(t, s)m(t)θ(s) + ‖f(s)‖K(t, s)e.

199

Page 209: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Note that K(t, s)θ(s)+K(t, s)m(t)θ(s) = ddsK(t, s)θ(s). We integrate the above inequality

between t and T to obtain:

θ(t) ≤ K(t, T )b+

∫ T

t‖f(s)‖K(t, s)e ds.

Now using Lemma 4 in Gomes et al. [2013], we have ‖K(t, T )b‖ ≤ ‖b‖ and ‖K(t, T )e‖ ≤

‖e‖ = 1. This implies that:

maxm

θm(t) ≤ ‖b‖+

∫ T

t‖f(s)‖ds.

Indeed starting from the inequality θ(s) ≥ −‖f(s)‖e + m(s)φ(s) and going through the

same steps we obtain:

minm

θm(t) ≥ −‖b‖ −∫ T

t‖f(s)‖ds.

The desired inequality follows.

Now we turn to the proof of Proposition 5.7.3. Let θ be the unique solution to the

system of ODEs (5.29). Recall the notation eij := 1(j 6= m)ej −1(i 6= m)ei. For any k 6= l,

define z(t, i0, x, k, l) := θ(t, i0, x + 1N ekl) − θ(t, i0, x). Then z(t, ·, ·, ·, ·) can be viewed as a

vector indexed by i0, x, k and l. Substracting the ODEs satisfied by θ(t, i0, x+ 1N ekl) and

θ(t, i0, x), we obtain that z solves the following system of ODEs:

−z(t, i0, x, k, l) =F (t, i0, x, k, l) +∑

j0,j0 6=i0

(z(t, j0, x, k, l)− z(t, i0, x, k, l))qφ00 (t, i0, j0, x)

+∑

(i,j),j 6=i

(z(t, i0, x+1

Neij , k, l)− z(t, i0, x, k, l))Nxiqφ(t, i, i0, x),

z(T, i0, x, k, l) =g0(i0, x+1

Nekl)− g0(i0, x),

(5.53)

where we have defined the F to be:

F (t, i0, x, k, l)

200

Page 210: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

:= fφ00 (t, i0, x+1

Nekl)− fφ00 (t, i0, x)

+∑

j0,j0 6=i0

[θ(t, j0, x+1

Nekl)− θ(t, i0, x+

1

Nekl)][q

0φ0(t, i0, j0, x+

1

Nekl)− q0

φ0(t, i0, j0, x)]

+∑

(i,j),j 6=i

[θ(t, i0, x+1

Neij +

1

Nekl)− θ(t, i0, x+

1

Nekl)]

×N [(x+1

Nekl)iqφ(t, i, j, i0, x+

1

Nekl)− xiqφ(t, i, j, i0, x)].

Then by the uniform bound provided in Lemma 5.8.3, together with the Lipschitz property

of f0, g0, q0, q, α, and β, we have:

‖F (t)‖ ≤ L

N+ 2(T‖f0‖∞ + ‖g0‖∞)

L

N+ ‖z(t)‖ · L, ‖z(T )‖ ≤ L

N,

where L is a generic Lipschitz constant. Now we are ready to apply Lemma 5.8.3 to (5.53).

We have for all t ≤ T :

‖z(t)‖ ≤ L

N+

∫ T

t(C

N+ L‖z(s)‖)ds,

where C is a constant only depending on T, L, ‖f0‖∞, ‖g0‖∞. Finally Gronwall’s inequality

allows us to conclude.

The proof of Proposition 5.7.4 is similar. We consider the solution θ to the ODE

(5.30) and we keep the notation z(t, i0, x, k, l) as before. For v ∈ Rm0 , x ∈ S, t ≤ T and

i0 = 1, . . . ,m0, denote:

h0(t, i0, x, v) := infα0∈A0

f0(t, α0, i0, x) +∑j0 6=i0

(vj0 − vi0)q0(t, i0, j0, α0, x).

Then for all x, y ∈ S, u, v ∈ Rm0 , using the Lipschitz property of f0 and q0 and the

boundedness of q0, we have:

|h0(t, i0, x, v)−h0(t, i0, y, u)| ≤ L|x−y|+2(m0−1) max‖u‖, ‖v‖‖x−y‖+Cq2(m0−1)‖v−u‖.

(5.54)

201

Page 211: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Subtracting the ODEs satisfied by θ(t, i0, x+ 1N ekl) and θ(t, i0, x), we obtain that z solves

the following system of ODEs:

−z(t, i0, x, k, l) =F (t, i0, x, k, l) +∑

(i,j),j 6=i

(z(t, i0, x+1

Neij , k, l)− z(t, i0, x, k, l))Nxiqφ(t, i, j, i0, x),

z(T, i0, x, k, l) =g0(i0, x+1

Nekl)− g0(i0, x),

where F is given by:

F (t, i0, x, k, l) :=h0(t, i0, x+1

Nekl, θ(t, ·, x+

1

Nekl))− h0(t, i0, x, θ(t, ·, x))

+∑

(i,j),j 6=i

[θ(t, i0, x+1

Neij +

1

Nekl)− θ(t, i0, x+

1

Nekl)]

×N [(x+1

Nekl)iq

φ(t, i, j, i0, x+1

Nekl)− xiqφ(t, i, j, i0, x)].

By the estimation in equation (5.54), the uniform bound provided in Lemma 5.8.3, together

with the Lipschitz property of f0, g0, q0, q, φ0, and φ, we have:

‖F (t)‖ ≤ 1

N[L+2(m0−1)(T‖f0‖∞+‖g0‖∞)]+2Cq(m0−1)‖z(t)‖+m(m−1)(LφL+Cq)‖z(t)‖.

We apply Lemma 5.8.3 to obtain:

‖z(t)‖ ≤ L

N+

∫ T

t[m(m− 1)(LφL+ Cq) + 2Cq(m0 − 1)]‖z(s)‖ds

+

∫ T

t

1

N[L+ 2(m0 − 1)(T‖f0‖∞ + ‖g0‖∞)]ds

:=C0 + C1T + C2T

2

N+

∫ T

t(C3 + C4Lφ)‖z(s)‖ds.

We then use Gronwall’s inequality to conclude.

202

Page 212: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Chapter 6

Conclusion

In this final chapter, we provide a few directions for future research on finite state mean

field games. We would like to warn the reader that these are some very preliminary ideas,

and their implementations are potentially prone to insurmountable technical difficulties.

From Finite State Space to Countable State Space?

In the mean field game models we have presented in this thesis so far, we have assumed

that the states of the players belong to a finite space. The first natural and meaningful

extension we can consider is to formulate the games in an infinite and countable space. This

is certainly the case in the information percolation model in Duffie et al. [2009] (see Section

1.2 for a detailed presentation of the model). If we use the analytical approach to study

the Nash equilibria, we should still be able to derive the Hamilton-Jacobi-Bellman equation

characterizing the optimality of the control and the Kolmogorov equation characterizing

the flow of measures for players’ states. However, since we now have a coupled system

of countably many ODEs, we need to reconsider the topological argument leading to the

existence of the solutions. On the other hand, most of the arguments in the probabilistic

approach will remain valid in the case of countable state space, provided that we can take

care of issues regarding the integrability of the state process. This means that we need to

carefully consider the assumptions on the transition rate functions.

203

Page 213: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

If we managed to extend the mean field game to any discrete space, not only will we

be able to handle a wider range of applications, we will be able to develop new numerical

methods for diffusion-based mean field game. It is a well known numerical algorithm in

stochastic optimal control of diffusion that we can approximate the Wiener processes by

random walks, or use the quantification to approximate the diffusion process by a stochastic

process living on a discrete grid. Therefore, one interesting question to ask here is whether

the equilibrium of a diffusion-based mean field game can be approximated by that of a

discrete state mean field game where the state process is the corresponding discrete approx-

imation. If the answer is yes, can we also have the convergence of the master equations

of the approximate finite state mean field games to the master equation of the stochastic

differential mean field game?

Dynamic Programming Principle for Finite State Principal Agent Problems?

In Chapter 4, we have shown that the principal agent problem in a finite state space

can be formulated as a McKean-Vlasov optimal control problem (see equation (4.28) and

Theorem 4.2.5). We have also shown in Section 4.3 that with the additional assumptions of

linearity of the transition rate, convex cost function and linear utility function of the minor

players, the optimal control problem is Markovian and can be reduced to a deterministic

control problem on the Kolmogorov equation. A natural question to ask is if the Markovian

structure is still preserved with a more general set of assumptions. Another related question

is: can we use the dynamic programming principle to tackle the optimization problem

(4.28)? Indeed, dynamic programming principle for McKean-Vlasov control problems was

developed for the diffusion controlled processes in Pham and Wei [2017]. However, the

authors dealt with the optimal control of the strong solution of SDE in which the probability

measure and the Wiener process are both fixed, whereas the situation we deal with here

is an optimal control problem in the weak formulation, meaning that the control impacts

both the probability measure and the martingale driving the process. If we can establish a

specific version of dynamic programming principle applicable to the control problem (4.28),

we can derive the HJB equation and compute the optimal contract in a more general setup.

204

Page 214: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Probabilistic Formulation for Finite State Mean Field Games with Major and

Minor Players?

In Chapter 5, we applied the analytical method to study the finite state mean field

game with major and minor players. In particular, we described the dynamics of the states

via the infinitesimal generators and we tackled the optimization problems of the players

by applying the dynamic programming principle. This allowed us to obtain the master

equation characterizing the Nash equilibria. However, in order to obtain a well-defined

dynamics for the players’ dynamics in the mean field limit, we had to make some strong

assumptions on the information structure of the players. As a result, we were only able

to establish the existence of Nash equilibria for a certain set of Markovian strategies and

a small duration of the game. Indeed, as pointed out in Carmona and Wang [2017], for

mean field games with major and minor players, different assumptions on the information

structure and the admissible strategies lead to different Nash equilibria. This makes us

wonder: can we improve the results of the existence of Nash equilibria by switching to a

less restrictive information structure? One possible angle to attack the problem is to extend

the weak formulation in Chapter 3 to the mean field game with a major player. Another

possible way, as suggested by Cecchin and Fischer [2017], is to use SDEs driven by Poisson

random measures to model the state processes of the players. No matter which approach

we take, the success of these attempts hinges on whether we will be able to obtain a well-

defined and tractable dynamics for the state processes of the players in the mean field game

limit.

205

Page 215: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Bibliography

Yves Achdou, Fabio Camilli, and Italo Capuzzo-Dolcetta. Mean field games: numericalmethods for the planning problem. SIAM Journal on Control and Optimization, 50(1):77–109, 2012.

Yves Achdou, Francisco J Buera, Jean-Michel Lasry, Pierre-Louis Lions, and BenjaminMoll. Partial differential equation models in macroeconomics. Phil. Trans. R. Soc. A,372(2028):20130397, 2014.

Saran Ahuja. Wellposedness of mean field games with common noise under a weak mono-tonicity condition. SIAM Journal on Control and Optimization, 54(1):30–48, 2016.

Charalambos D. Aliprantis and Kim Border. Infinite dimensional analysis: A Hitchhiker’sGuide. Springer, 2006.

Mohamad Aziz and Peter E Caines. A mean field game computational methodology fordecentralized cellular network optimization. IEEE Transactions on Control Systems Tech-nology, 25(2):563–576, 2017.

Guy Barles and Panagiotis E Souganidis. Convergence of approximation schemes for fullynonlinear second order equations. Asymptotic analysis, 4(3):271–283, 1991.

Erhan Bayraktar and Asaf Cohen. Analysis of a finite state many player game using itsmaster equation. arXiv preprint arXiv:1707.02648, 2017.

Chiara Benazzoli, Luciano Campi, and Luca Di Persio. Mean-field games with controlledjumps. 03 2017. URL https://arxiv.org/abs/1703.01919.

Alain Bensoussan, Jens Frehse, and Sheung Chi Phillip Yam. The master equation in meanfield theory. Journal de Mathematiques Pures et Appliquees, 103(6):1441–1474, 2015.

Alain Bensoussan, MHM Chau, and SCP Yam. Mean field games with a dominating player.Applied Mathematics &amp; Optimization, 74(1):91–128, 2016a.

Alain Bensoussan, KCJ Sung, Sheung Chi Phillip Yam, and Siu-Pang Yung. Linear-quadratic mean field games. Journal of Optimization Theory and Applications, 169(2):496–529, 2016b.

Ariela Briani, Fabio Camilli, and Hasnaa Zidani. Approximation schemes for monotonesystems of nonlinear second order partial differential equations: convergence result anderror estimate. Differential Equations and Applications, 4:297–317, 2012.

206

Page 216: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Pierre Cardaliaguet, Jean-Michel Lasry, Pierre-Louis Lions, Alessio Porretta, et al. Longtime average of mean field games. Networks and Heterogeneous Media, 7(2):279–301,2012.

Pierre Cardaliaguet, Francois Delarue, Jean-Michel Lasry, and Pierre-Louis Lions. Themaster equation and the convergence problem in mean field games. arXiv preprintarXiv:1509.02505, 2015.

Rene Carmona and Francois Delarue. Probabilistic analysis of mean-field games. SIAMJournal on Control and Optimization, 51(4):2705–2734, 2013.

Rene Carmona and Francois Delarue. The master equation for large population equilibri-ums. In Stochastic Analysis and Applications 2014, pages 77–128. Springer, 2014.

Rene Carmona and Francois Delarue. Probabilistic Theory of Mean Field Games withApplications I-II. Springer, 2017.

Rene Carmona and Christy V. Graves. Jet lag recovery: Synchronization of circadianoscillators as a mean field game. 04 2018. URL https://arxiv.org/abs/1804.04729.

Rene Carmona and Daniel Lacker. A probabilistic weak formulation of mean field gamesand applications. The Annals of Applied Probability, 25(3):1189–1231, 2015.

Rene Carmona and Peiqi Wang. Finite state mean field games with major and minorplayers. 2016.

Rene Carmona and Peiqi Wang. An alternative approach to mean field game with ma-jor and minor players, and applications to herders impacts. Applied Mathematics andOptimization, 76(1):5–27, 2017.

Rene Carmona and Peiqi Wang. Finite-state contract theory with a principal and a field ofagents. In Revision, 2018a.

Rene Carmona and Peiqi Wang. A probabilistic approach to extended finite state meanfield games. Submitted, 2018b.

Rene Carmona and Xiuneng Zhu. A probabilistic approach to mean field games with majorand minor players. The Annals of Applied Probability, 26(3):1535–1580, 2016.

Rene Carmona, Francois Delarue, and Aime Lachapelle. Control of mckean–vlasov dynamicsversus mean field games. Mathematics and Financial Economics, 7(2):131–166, 2013a.

Rene Carmona, Jean-Pierre Fouque, and Li-Hsien Sun. Mean field games and systemic risk.2013b.

Rene Carmona, Francois Delarue, et al. Forward–backward stochastic differential equationsand controlled mckean–vlasov dynamics. The Annals of Probability, 43(5):2647–2700,2015.

Rene Carmona, Francois Delarue, Daniel Lacker, et al. Mean field games with commonnoise. The Annals of Probability, 44(6):3740–3803, 2016.

207

Page 217: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Rene Carmona, Francois Delarue, and Daniel Lacker. Mean field games of timing andmodels for bank runs. Applied Mathematics and Optimization, 76(1):217–260, 2017.

Alekos Cecchin and Markus Fischer. Probabilistic approach to finite state mean field games.arXiv preprint arXiv:1704.00984, 2017.

Patrick Chan and Ronnie Sircar. Bertrand and cournot mean field games. Applied Mathe-matics and Optimization, 71(3):533–569, 2015.

Jean-Francois Chassagneux, Dan Crisan, and Francois Delarue. Classical solutions to themaster equation for large population equilibria. arXiv preprint arXiv:1411.3009, 2014.

Jean-Francois Chassagneux, Dan Crisan, and Francois Delarue. Numerical method forfbsdes of McKean-Vlasov type. arXiv preprint arXiv:1703.02007, 2017.

Samuel N. Cohen and Robert J. Elliott. Solutions of backward stochastic differen- tialequations on Markov chains. Commun. Stoch. Anal., (2):251–262., 2008.

Samuel N. Cohen and Robert J. Elliott. Comparisons for backward stochastic differ-ential equations on Markov chains and related no-arbitrage conditions. Ann. Appl.Probab., 20(1):267–311, 02 2010. doi: 10.1214/09-AAP619. URL https://doi.org/10.1214/09-AAP619.

Michael G Crandall, Hitoshi Ishii, and Pierre-Louis Lions. User’s guide to viscosity solutionsof second order partial differential equations. Bulletin of the American mathematicalsociety, 27(1):1–67, 1992.

Jaksa Cvitanic, Dylan Possamaı, and Nizar Touzi. Dynamic programming approach toprincipal-agent problems. arXiv preprint, 10 2015. URL https://arxiv.org/abs/1510.07111.

Jaksa Cvitanic, Dylan Possamaı, and Nizar Touzi. Moral hazard in dynamic risk manage-ment. Management Science, 63(10):3328–3346, 2016.

Josu Doncel, Nicolas Gast, and Bruno Gaujal. Mean-field games with explicit interactions.2016.

Darrell Duffie, Semyon Malamud, and Gustavo Manso. Information percolation with equi-librium search dynamics. Econometrica, 77(5):1513–1574, 2009.

Romuald Elie, Thibaut Mastrolia, and Dylan Possamaı. A tale of a principal and manymany agents. arXiv preprint arXiv:1608.05226, 2016.

R. J. Elliott, L. Aggoun, and J. B. Moore. Hidden Markov Models: Estimation and Control.Number 29 in Applications of Mathematics. Springer, New York, 1995.

Stewart N Ethier and Thomas G Kurtz. Markov processes: characterization and conver-gence, volume 282. John Wiley &amp; Sons, 2009.

Aleksei Fedorovich Filippov. Differential equations with discontinuous righthand sides: con-trol systems, volume 18. Springer Science and Business Media, 2013.

208

Page 218: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Wendell H Fleming and Halil Mete Soner. Controlled Markov processes and viscosity solu-tions, volume 25. Springer Science &amp; Business Media, 2006.

Nicolas Fournier and Arnaud Guillin. On the rate of convergence in wasserstein distance ofthe empirical measure. Probability Theory and Related Fields, 162(3-4):707–738, 2015.

G. Georganopoulos. Sur l’approximation des fonctions continues par des fonctions lips-chitziennes. C. R. Acad. Sci. Paris, 264(7):319–321, 1967.

Diogo A Gomes and Edgard Pimentel. Time-dependent mean-field games with logarithmicnonlinearities. SIAM Journal on Mathematical Analysis, 47(5):3798–3812, 2015.

Diogo A Gomes, Joana Mohr, and Rafael Rigao Souza. Continuous time finite state meanfield games. Applied Mathematics &amp; Optimization, 68(1):99–143, 2013.

Diogo A Gomes, Edgard A Pimentel, and Hector Sanchez-Morgado. Time-dependent mean-field games in the subquadratic case. Communications in Partial Differential Equations,40(1):40–76, 2015.

Diogo A Gomes, Edgard Pimentel, and Hector Sanchez-Morgado. Time-dependent mean-field games in the superquadratic case. ESAIM: Control, Optimisation and Calculus ofVariations, 22(2):562–580, 2016.

Olivier Gueant. Existence and uniqueness result for mean field games with congestion effecton graphs. Applied Mathematics and Optimization, 72(2):291–303, 2015.

Martin F Hellwig and Klaus M Schmidt. Discrete–time approximations of the holmstrom–milgrom brownian–motion model of intertemporal incentive provision. Econometrica, 70(6):2225–2264, 2002.

Bengt Holmstrom and Paul Milgrom. Multitask principal-agent analyses: Incentive con-tracts, asset ownership, and job design. Journal of Law, Economics, &amp; Organization,7:24–52, 1991.

Ying Hu and Shige Peng. A stability theorem of backward stochastic differential equationsand its application. Comptes Rendus de l’Academie des Sciences-Series I-Mathematics,324(9):1059–1064, 1997.

Minyi Huang. Large-population lqg games involving a major player: the nash certaintyequivalence principle. SIAM Journal on Control and Optimization, 48(5):3318–3353,2010.

Minyi Huang, Roland P Malhame, Peter E Caines, et al. Large population stochasticdynamic games: closed-loop McKean-Vlasov systems and the nash certainty equivalenceprinciple. Communications in Information and Systems, 6(3):221–252, 2006.

Marius Iosifescu. Finite Markov processes and their applications. Courier Corporation,2014.

Jean Jacod and Jean Memin. Sur un type de convergence intermediaire entre la convergenceen loi et la convergence en probabilite. In Seminaire de Probabilites XV 1979/80, pages529–546. Springer, 1981.

209

Page 219: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Jean Jacod and Albert N Shiryaev. Skorokhod topology and convergence of processes. InLimit theorems for stochastic processes, pages 288–347. Springer, 1987.

Sebastian Jaimungal and Mojtaba Nourian. Mean-field game strategies for a major-minoragent optimal execution problem. Available at SSRN, 2578733, 2015.

Arman C Kizilkale and Roland P Malhame. Mean field based control of power systemdispersed energy storage devices for peak load relief. In Decision and Control (CDC),2013 IEEE 52nd Annual Conference on, pages 4971–4976. IEEE, 2013.

VN Kolokoltsov and Alain Bensoussan. Mean-field-game model for botnet defense in cyber-security. Applied Mathematics and Optimization, 74(3):669–692, 2016.

Aime Lachapelle and Marie-Therese Wolfram. On a mean field game approach modelingcongestion and aversion in pedestrian crowds. Transportation research part B: method-ological, 45(10):1572–1589, 2011.

Aime Lachapelle, Julien Salomon, and Gabriel Turinici. Computation of mean field equilib-ria in economics. Mathematical Models and Methods in Applied Sciences, 20(04):567–588,2010.

Daniel Lacker. Mean field games via controlled martingale problems: existence of Markovianequilibria. Stochastic Processes and their Applications, 125(7):2856–2894, 2015.

Jean-Michel Lasry and Pierre-Louis Lions. Jeux a champ moyen. i–le cas stationnaire.Comptes Rendus Mathematique, 343(9):619–625, 2006a.

Jean-Michel Lasry and Pierre-Louis Lions. Jeux a champ moyen. ii–horizon fini et controleoptimal. Comptes Rendus Mathematique, 343(10):679–684, 2006b.

Jean-Michel Lasry and Pierre-Louis Lions. Mean field games. Japanese journal of mathe-matics, 2(1):229–260, 2007.

Jean-Michel Lasry, Pierre-Louis Lions, and Olivier Gueant. Application of mean field gamesto growth theory. 2008.

Pierre-Louis Lions. College de france course on mean-field games, 2007. URL https://www.college-de-france.fr/site/pierre-louis-lions/course-2006-2007.htm.

Zhongjing Ma, Duncan Callaway, and Ian Hiskens. Optimal charging control for plug-inelectric vehicles. In Control and Optimization Methods for Electric Smart Grids, pages259–273. Springer, 2012.

Francois Meriaux, Samson Lasaulce, and Hamidou Tembine. Stochastic differential gamesand energy-efficient power control. Dynamic Games and Applications, 3(1):3–23, 2013.

Radu Miculescu. Approximation of continuous functions by lipschitz functions. Real Anal-ysis Exchange, 26(1):449–452, 2000.

Holger M Muller. The first-best sharing rule in the continuous-time principal-agent problemwith exponential utility. Journal of Economic Theory, 79(2):276–280, 1998.

210

Page 220: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

John F. Nash. Equilibrium points in n-person games. Proceedings of the national academyof sciences, 36(1):48–49, 1950.

Son Luu Nguyen and Minyi Huang. Linear-quadratic-gaussian mixed games withcontinuum-parametrized minor players. SIAM Journal on Control and Optimization,50(5):2907–2937, 2012a.

Son Luu Nguyen and Minyi Huang. Mean field lqg games with mass behavior responsiveto a major player. In Decision and Control (CDC), 2012 IEEE 51st Annual Conferenceon, pages 5792–5797. IEEE, 2012b.

Mojtaba Nourian and Peter E Caines. ε-nash mean field game theory for nonlinear stochas-tic dynamical systems with major and minor agents. SIAM Journal on Control andOptimization, 51(4):3302–3331, 2013.

Mojtaba Nourian, Peter E Caines, and Roland P Malhame. Mean field analysis of controlledcucker-smale type flocking: Linear analysis and perturbation equations. IFAC ProceedingsVolumes, 44(1):4471–4476, 2011.

Marcel Nutz. A mean field game of optimal stopping. arXiv preprint arXiv:1605.09112,2016.

Huyen Pham and Xiaoli Wei. Dynamic programming for optimal control of stochasticmckean–vlasov dynamics. SIAM Journal on Control and Optimization, 55(2):1069–1101,2017.

Alessio Porretta. On the planning problem for the mean field games system. DynamicGames and Applications, 4(2):231–256, 2014.

Philip E Protter. Stochastic differential equations. In Stochastic Integration and DifferentialEquations, pages 249–361. Springer, 2005.

Yuliy Sannikov. A continuous-time version of the principal-agent problem. The Review ofEconomic Studies, 75(3):957–984, 2008.

Yuliy Sannikov. Contracts: The theory of dynamic principal-agent relationships and thecontinuous-time approach. In 10th World Congress of the Econometric Society, 2012.

Heinz Schattler and Jaeyoung Sung. The first-order approach to the continuous-timeprincipal–agent problem with exponential utility. Journal of Economic Theory, 61(2):331–371, 1993.

Min Shen and Gabriel Turinici. Liquidity generated by heterogeneous beliefs and costlyestimations. Networks and Heterogeneous Media, 7(2):349–361, 2012.

Alexander Sokol and Niels Richard Hansen. Exponential martingales and changes of mea-sure for counting processes. Stochastic analysis and applications, 33(5):823–843, 2015.

Leonardo Stella, Fabio Bagagiolo, Dario Bauso, and Giacomo Como. Opinion dynamics andstubbornness through mean-field games. In Decision and Control (CDC), 2013 IEEE52nd Annual Conference on, pages 2519–2524. IEEE, 2013.

211

Page 221: Finite State Mean Field Games - Princeton University€¦ · discrete state space. This thesis tackles nite state mean eld games. In the rst part of the thesis, we develop a probabilistic

Jaeyoung Sung. Linearity with project selection and controllable diffusion rate incontinuous-time principal-agent problems. The RAND Journal of Economics, pages 720–743, 1995.

Jan Swart and Anita Winter. Markov processes: theory and examples. Technical report,Mimeo, https://www. uni-due. de/hm0110/Markovprocesses/sw20. pdf, 2013.

Cedric Villani. Optimal transport: old and new, volume 338. Springer Science and BusinessMedia, 2008.

212