controlled sequential monte carlo arxiv:1708 · 2017. 9. 5. · controlled sequential monte carlo...

31
Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS & CSIRO), George Deligiannidis (KCL) and Arnaud Doucet (Oxford) SMC Workshop 2017 Jeremy Heng Controlled SMC 1 / 31

Upload: others

Post on 23-Jun-2021

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Controlled Sequential Monte Carlo

arXiv:1708.08396

Jeremy Heng

Department of Statistics, Harvard UniversityJoint work with Adrian Bishop (UTS & CSIRO), George Deligiannidis (KCL) and Arnaud

Doucet (Oxford)

SMC Workshop 2017

Jeremy Heng Controlled SMC 1/ 31

Page 2: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Feynman-Kac path measure

Consider a non-homogenous Markov chain (Xt)t∈[0:T ] on (X,X )with law

Q(dx0:T ) = µ(dx0)T∏

t=1

Mt(xt−1, dxt)

Given positive bounded potential functions (Gt)t∈[0:T ], defineFeynman-Kac path measure

P(dx0:T ) = G0(x0)T∏

t=1

Gt(xt−1, xt)Q(dx0:T )Z−1

where Z := EQ

[

G0(X0)∏T

t=1 Gt(Xt−1,Xt)]

The quantities{

µ, (Mt)t∈[1:T ], (Gt)t∈[0:T ]

}

depend on the specificapplication

Applications of interest: static models and state space models

Jeremy Heng Controlled SMC 2/ 31

Page 3: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Sequential Monte Carlo methods

SMC methods simulate an interacting particle system of sizeN ∈ N

At time t = 0 and particle n ∈ [1 : N ]sample X

n0 ∼ µ;

sample ancestor index An0 ∼ R

(

G0(X10 ), . . . ,G0(X

N0 )

)

For time t ∈ [1 : T ] and particle n ∈ [1 : N ]

sample Xnt ∼ Mt(X

Ant−1

t−1, ·);

sample ancestor index Ant ∼ R

(

Gt(XA1t−1

t−1,X 1

t ), . . . ,Gt(XANt−1

t−1,XN

t )

)

Jeremy Heng Controlled SMC 3/ 31

Page 4: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Sequential Monte Carlo methods

Particle approximation of P

PN =1

N

N∑

n=1

δXn0:T

where X n0:T is obtained by tracing ancestral lineage of particle X n

T

Unbiased estimator of Z

ZN =

{

1

N

N∑

n=1

G0(Xn0 )

}

T∏

t=1

{

1

N

N∑

n=1

Gt(XAn

t−1t−1 ,X n

t )

}

Convergence properties of PN and ZN as N → ∞ are nowwell-understood

However quality of approximation can be inadequate for practicalchoices of N

Performance crucially depends on discrepancy between P and Q

Jeremy Heng Controlled SMC 4/ 31

Page 5: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Twisted path measures

Consider change of measure prescribed by positive and boundedfunctions ψ = (ψt)t∈[0:T ]

Refer to ψ as an admissible policy and denote set of alladmissible policies as Ψ

Given a policy ψ ∈ Ψ, define ψ-twisted path measure of Q as

Qψ(dx0:T ) = µψ(dx0)T∏

t=1

Mψt (xt−1, dxt)

where

µψ(dx0) :=µ(dx0)ψ0(x0)

µ(ψ0), Mψ

t (xt−1, dxt) :=Mt(xt−1, dxt)ψt(xt−1, xt)

Mt(ψt)(xt−1),

for t ∈ [1 : T ]

Jeremy Heng Controlled SMC 5/ 31

Page 6: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Twisted path measures

Given ψ ∈ Ψ, we have

P(dx0:T ) = Gψ0 (x0)T∏

t=1

Gψt (xt−1, xt)Qψ(dx0:T )Z

−1

where

Gψ0 (x0) :=µ(ψ0)G0(x0)M1(ψ1)(x0)

ψ0(x0),

Gψt (xt−1, xt) :=Gt(xt−1, xt)Mt+1(ψt+1)(xt )

ψt(xt−1, xt), t ∈ [1 : T − 1],

GψT(xT−1, xT ) :=

GT (xT−1, xT )

ψT (xT−1, xT ),

are the twisted potentials associated with Qψ

Note Z = EQψ

[

Gψ0 (X0)∏T

t=1 Gψt (Xt−1,Xt)

]

by construction

Jeremy Heng Controlled SMC 6/ 31

Page 7: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Twisted SMC methods

Assume policy ψ ∈ Ψ is such that:sampling µψ and (Mψ

t )t∈[1:T ] feasible

evaluating (Gψt )t∈[0:T ] tractable

Construct ψ-twisted SMC method as standard SMC applied to{

µψ, (Mψt )t∈[1:T ], (G

ψt )t∈[0:T ]

}

Particle approximation of P and Z

Pψ,N =1

N

N∑

n=1

δXn

0:T, Zψ,N =

{

1

N

N∑

n=1

Gψ0 (X n0 )

}

T∏

t=1

{

1

N

N∑

n=1

Gψt (XAn

t−1t−1 ,X n

t )

}

Jeremy Heng Controlled SMC 7/ 31

Page 8: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Optimal policies

A policy with constant functions recover standard SMC method

Consider an iterative scheme to refine policies

Given current policy ψ ∈ Ψ, twisting Qψ further with policy φ ∈ Ψresults in a twisted path measure (Qψ)φ

Note that (Qψ)φ = Qψ·φ where ψ · φ = (ψt · φt)t∈[0:T ]

Choice of φ is guided by the following optimality result

Jeremy Heng Controlled SMC 8/ 31

Page 9: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Optimal policies

Proposition

For any ψ ∈ Ψ, under the policy φ∗ = (φ∗t )t∈[0:T ] defined recursively as

φ∗T (xT−1, xT ) = GψT(xT−1, xT ),

φ∗t (xt−1, xt) = Gψt (xt−1, xt)Mψt+1(φ

t+1)(xt ), t ∈ [T − 1 : 1],

φ∗0(x0) = Gψ0 (x0)Mψ1 (φ

1)(x0),

the refined policy ψ∗ := ψ · φ∗ satisfies:(i) P = Qψ

;(ii) Zψ

∗,N = Z almost surely for any N ∈ N.

Jeremy Heng Controlled SMC 9/ 31

Page 10: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Optimal policies

Refer to φ∗ as the optimal policy w.r.t. Qψ

The refined policy ψ∗ = ψ · φ∗ is the optimal policy w.r.t. Q

ψ∗-twisted potentials

Gψ∗

0 (x0) = Z , Gψ∗

t (xt−1, xt) = 1, t ∈ [1 : T ]

Under ψ∗-twisted SMC method

Zψ∗,N

t =

{

1

N

N∑

n=1

Gψ∗

0 (X n0 )

}

t∏

k=1

{

1

N

N∑

n=1

Gψ∗

k(X

An

t−1

k−1 ,X nk )

}

= Z

for all t ∈ [0 : T ]

Jeremy Heng Controlled SMC 10 / 31

Page 11: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Optimal policies

The connection to Kullback-Leibler optimal control is given by

Proposition

The functions V ∗t := − logφ∗t , t ∈ [0 : T ] are the optimal value functions

of the KL control problem

infφ∈Φ

KL(

(Qψ)φ|P)

where Φ := {φ ∈ Ψ : KL((Qψ)φ|P) < ∞}.

The following is a characterization of φ∗ in a specific setting

Proposition

For any policy ψ ∈ Ψ such that the corresponding twisted potentials(Gψt )t∈[0:T ] and transition densities of (Mψ

t )t∈[1:T ] are log-concave ontheir domain of definition, then the optimal policy φ∗ = (φ∗t )t∈[0:T ] w.r.t.Qψ is a sequence of log-concave functions.

Jeremy Heng Controlled SMC 11 / 31

Page 12: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Dynamic programming recursions

Simplify notation by defining the Bellman operators (Qψt )t∈[0:T−1]

Rewrite the backward recursion defining φ∗ = (φ∗t )t∈[0:T ] as

φ∗T = GψT,

φ∗t = Qψt φ

t+1, t ∈ [T − 1 : 0]

where

Qψt (ϕ)(x , y) = Gψt (x , y)Mψ

t+1(ϕ)(y)

It will be convenient to view Qψt : L2(νψ

t+1) → L2(νψt ) where

νψ0 := µψ , νψt (dx , dy) := ηψt−1(dx)M

ψt (x , dy)

Need to approximate this recursion in practice

Jeremy Heng Controlled SMC 12 / 31

Page 13: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Approximate projections

Given probability measure ν and function class F ⊂ L2(ν),

Define (logarithmic) projection of f onto F as

Pν f = exp

(

− arg minϕ∈F

∥ϕ− (− log f )∥2L2(ν)

)

, for − log f ∈ L2(ν)

Since V ∗t = − logφ∗t this corresponds to learning associated value

functions (more stable numerically)

A practical implementation replaces ν with a Monte Carloapproximation νN

Define approximate (F, ν)-projection as

Pν,N f = exp

(

− arg minϕ∈F

∥ϕ− (− log f )∥2L2(νN )

)

Jeremy Heng Controlled SMC 13 / 31

Page 14: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Approximate dynamic programming

To use output of ψ-twisted SMC to learn optimal φ∗

Define

νψ,N0 =1

N

N∑

n=1

δXn0, νψ,Nt =

1

N

N∑

n=1

δ(X

Ant−1

t−1 ,Xnt

), t ∈ [1 : T ],

which are consistent approximations of (νψt )t∈[0:T ]

Given function class Ft ⊂ L2(νψt ), denote approximate(Ft , ν

ψt )-projection by Pψ,Nt

Approximate backward recursion defining φ∗ = (φ∗t )t∈[0:T ] by

φ̂T = Pψ,NT

GψT,

φ̂t = Pψ,Nt Qψt φ̂t+1, t ∈ [T − 1 : 0]

This is the approximate dynamic programming (ADP) algorithmfor finite horizon control problems (Bertsekas and Tsitsiklis, 1996)

Jeremy Heng Controlled SMC 14 / 31

Page 15: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Policy refinement

Construct iterative algorithm: Controlled SMC

Initialization: set ψ(0) as constant one functionsFor iterations i ∈ [0 : I − 1]:

run ψ(i)-twisted SMC;perform ADP with SMC output to obtain policy φ̂(i+1);construct refined policy ψ(i+1) = ψ(i)

· φ̂(i+1).

At iteration i = I : run ψ(I )-twisted SMC

Jeremy Heng Controlled SMC 15 / 31

Page 16: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Controlled SMC

0 2 4 6 8 10 12 14 16 18 200

10

20

30

40

50

60

70

80

90

100

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

Iteration 4

Iteration 5

0 2 4 6 8 10 12 14 16 18 20

-140

-120

-100

-80

-60

-40

-20

0

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

Iteration 4

Iteration 5

Figure: Illustration on logistic regression example.

Jeremy Heng Controlled SMC 16 / 31

Page 17: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Approximate dynamic programming

We obtain error bounds like

Eψ,N∥φ̂t − φ∗t ∥L2(νψt ) ≤T∑

s=t

Cψt−1,s−1e

ψ,Ns , t ∈ [0 : T ]

where Cψt,s are stability constants of Bellman operators and eψ,Nt

are errors of approximate projections

As N → ∞, one expects φ̂ to converge to φ̃ = (φ̃t )t∈[0:T ], definedby the idealized ADP algorithm

φ̃T = PψTGψT,

φ̂t = Pψt Qψt φ̃t+1, t ∈ [T − 1 : 0],

where Pψt is the exact (Ft , νψt )-projection

We establish a LLN and CLT in the case where (Ft)t∈[0:T ] are givenby a linear basis functions

Jeremy Heng Controlled SMC 17 / 31

Page 18: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Policy refinement

Residuals of logarithmic projections in ADP

εψt := log φ̂t −(

logGψt − logMψt+1(φ̂t+1)

)

Related to twisted potentials of refined policy ψ · φ̂ via

logGψ·φ̂t = −εψt

If we twist Qψ·φ̂ further by a policy ζ̂ ∈ Ψ, logarithmic projections inADP are

− log ζ̂t := arg minϕ∈Ft

∥ϕ− (εψt − logMψ·φ̂t+1 (ζ̂t+1))∥

L2(νψ·φ̂,Nt )

where (νψ·φ̂,Nt )t∈[0:T ] are defined using output of (ψ · φ̂)-twisted SMC

Jeremy Heng Controlled SMC 18 / 31

Page 19: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Policy refinement

Beneficial to have an iterative scheme to construct more refinedpolicies

Allows repeated least squares fitting of residuals – in the spirit ofL2-boosting methods

Ft ={

ϕ(xt ) = xTt Atxt + xTt bt + ct : (At , bt , ct) ∈ Sd × Rd × R}

0 2 4 6 8 10 12 14 16 18 20-5

0

5

10

15

20

25

Iteration 1

Iteration 2

Iteration 3

Iteration 4

Iteration 5

0 2 4 6 8 10 12 14 16 18 20-35

-30

-25

-20

-15

-10

-5

0

5

10

15

Iteration 1

Iteration 2

Iteration 3

Iteration 4

Iteration 5

Figure: Coefficients estimated at each iteration of controlled SMC.

Jeremy Heng Controlled SMC 19 / 31

Page 20: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Iterated ADP

Want to understand the behaviour of policy ψ(I ) as I → ∞

Equipped Ψ with a metric ρ

Write iterating ADP as iterated random function FNU(ψ) = ψ · φ̂,

where φ̂ is ADP approximation with N particles

Iterating FN defines a Markov chain (ψ(I ))I∈N on Ψ

Under regularity conditions, it converges to a unique invariantdistribution π

Write iterating ADP with exact projections as F (ψ) = ψ · φ̃, where φ̃is idealized ADP approximation

If we assume additionally that

ρ(FNU (ψ),F (ψ) ≤ OP(N

−1/2)

for all ψ ∈ Ψ then

Eπ [ρ(ψ,ϕ∗)] ≤ O(N−1/2)

where ϕ∗ is a fixed point of F

Jeremy Heng Controlled SMC 20 / 31

Page 21: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Iterated ADP

14.5 15 15.5 16 16.5 17 17.5 18-1.5

-1

-0.5

0

0.5

1

1.5

14.5 15 15.5 16 16.5 17 17.5 18-3

-2

-1

0

1

2

3

410

-3

5 6 7 8 9 10 11 12 13 14 15-18

-16

-14

-12

-10

-8

-6

146 148 150 152 154 156 158-0.01

-0.008

-0.006

-0.004

-0.002

0

0.002

0.004

0.006

0.008

0.01

Figure: Illustrating invariant distribution of coefficients.

Jeremy Heng Controlled SMC 21 / 31

Page 22: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Log-Gaussian Cox point process

Example from Møller et al. (1998)

Dataset: 126 Scots pine saplings in a natural forest in Finland

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Figure: Locations of 126 Scots pine saplings in square plot of 10 × 10m2.

Jeremy Heng Controlled SMC 22 / 31

Page 23: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Log-Gaussian Cox point process

Discretize into a 30 × 30 regular grid, so d = 900 here

Posterior distribution

η(dx) = N (x ;µ0,Σ0)∏

m∈[1:30]2

exp (xmym − a exp(xm))Z−1

Geometric path: ηt(dx) = N (x ;µ0,Σ0)ℓ(x , y)λtZ−1t ,

0 = λ0 < · · · < λT = 1

Set µ = N (µ0,Σ0) and Mt as unadjusted Langevin algorithm (ULA)targeting ηt

Function classes

Ft ={

ϕ(xt−1, xt) = xTt Atxt + xTt bt + ct − (λt − λt−1) log ℓ(xt−1, y)

: At diagonal, bt ∈ Rd , ct ∈ R}

, t ∈ [1 : T ]

Jeremy Heng Controlled SMC 23 / 31

Page 24: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Log-Gaussian Cox point process

Parameterization provides good approximation of optimal policy

0 2 4 6 8 10 12 14 16 18 200

10

20

30

40

50

60

70

80

90

100

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

0 2 4 6 8 10 12 14 16 18 20

50

100

150

200

250

300

350

400

450

500

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

Figure: Effective sample size (left) and normalizing constant estimation(right) when performing inference on the Scots pine dataset.

Jeremy Heng Controlled SMC 24 / 31

Page 25: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

Log-Gaussian Cox point process

Comparison to AIS with MALA moves

cSMC: N = 4096 particles, I = 3 iterations, T = 20

AIS uses 5 times more particles for fair comparison

Variance of marginal likelihood estimates are 280 times smaller

AIS cSMC

480

482

484

486

488

490

492

494

496

498

Figure: Marginal likelihood estimates obtained by each algorithm over 100independent repetitions.

Jeremy Heng Controlled SMC 25 / 31

Page 26: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

Measurements collected from a neuroscience experiment(Temereanca et al., 2008)

0 500 1000 1500 2000 2500 30000

2

4

6

8

10

12

14

State space model:

µ = N (0, 1),

Mt(xt−1, dxt) = N(

xt ;αxt−1,σ2)

dxt ,

Gt(xt ) = B (yt ;M ,κ(xt)) ,

where M = 30, T = 2999 and κ(u) := (1 + exp(−u))−1, for u ∈ R

Jeremy Heng Controlled SMC 26 / 31

Page 27: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

Function classes:Ft =

{

ϕ(xt ) = atx2t + btxt + ct : (at , bt , ct) ∈ R3

}

, t ∈ [0 : T ],

Parameterization provides good approximation of optimal policy

0 500 1000 1500 2000 25000

10

20

30

40

50

60

70

80

90

100

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

0 500 1000 1500 2000 2500

-3000

-2500

-2000

-1500

-1000

-500

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

Figure: Effective sample size (left) and normalizing constant estimation(right) when performing inference on the neuroscience model.

Jeremy Heng Controlled SMC 27 / 31

Page 28: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

Estimated policies capturing abrupt changes in the data

0 500 1000 1500 2000 2500 30000

1

2

3

4

5

6

7

8

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

0 500 1000 1500 2000 2500 30000

5

10

15

20

25

30

35

40

Uncontrolled

Iteration 1

Iteration 2

Iteration 3

Figure: Coefficients estimated by the controlled SMC sampler at eachiteration when performing inference on the neuroscience model.

Jeremy Heng Controlled SMC 28 / 31

Page 29: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

(Left) Comparison to bootstrap particle filter (BPF)

(Right) Comparison to forward filtering and backward smoother(FFBS) for functional x0:T )→ Mκ(x0:T )

0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2

-9

-8.5

-8

-7.5

-7

-6.5

-6

-5.5

-5

BPF

cSMC

0 500 1000 1500 2000 2500 3000-4

-3.5

-3

-2.5

-2

-1.5

-1

-0.5

FFBS

cSMC

Figure: Relative variance of marginal likelihood estimates (left) andestimates of smoothing expectation (right).

Jeremy Heng Controlled SMC 29 / 31

Page 30: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

Bayesian inference for parameters θ = (α,σ2) within particlemarginal Metropolis-Hastings (PMMH)

cSMC and BPF to produce unbiased estimates of marginal likelihood

0.988 0.99 0.992 0.994 0.996 0.998 10

50

100

150

200

250

300

350

400

450

BPF

cSMC

0.06 0.07 0.08 0.09 0.1 0.11 0.12 0.13 0.14 0.150

5

10

15

20

25

30

35

40

45

50

BPF

cSMC

Figure: Posterior density estimates based on 100, 000 samples.

Jeremy Heng Controlled SMC 30 / 31

Page 31: Controlled Sequential Monte Carlo arXiv:1708 · 2017. 9. 5. · Controlled Sequential Monte Carlo arXiv:1708.08396 Jeremy Heng Department of Statistics, Harvard University Joint work

A model from neuroscience

Autocorrelation function (ACF) of each PMMH chain

ESS improvement roughly 10 times for parameter α and 5 times forparameter σ2

0 10 20 30 40 50 60 70 80 90 100-0.2

0

0.2

0.4

0.6

0.8

1

BPF

cSMC

0 10 20 30 40 50 60 70 80 90 100-0.2

0

0.2

0.4

0.6

0.8

1

BPF

cSMC

Figure: Autocorrelation functions of two PMMH chains.

Jeremy Heng Controlled SMC 31 / 31