nonlinear data assimilation part i: particle filters seminars... · why nonlinear data...

45
Nonlinear data assimilation Part I: Particle filters Melanie Ades, Data Assimilation Research Centre (DARC) KAUST, Saudi Arabia, February 2014

Upload: others

Post on 20-May-2020

17 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Nonlinear data assimilationPart I: Particle filters

Melanie Ades, Data Assimilation Research Centre (DARC)KAUST, Saudi Arabia, February 2014

Page 2: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Why nonlinear data assimilation?

The complexity of models are increasing:

Precipitation

Particle filters are a fully nonlinear data assimilation method, but why are nonlinear schemes required?

Page 3: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Why nonlinear data assimilation?

The complexity of models are increasing:

Precipitation

Results of variational or Kalman filter data assimilation

Negative precipitation

Particle filters are a fully nonlinear data assimilation method, but why are nonlinear schemes required?

Page 4: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Why nonlinear data assimilation?The resolution of models is also increasing:

Page 5: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Particle filters

P(vj+1|Yj) =

Rn

P(vj+1|vj)P(vj |Yj)dvj

P(vj+1|Yj+1) =P(yj+1|vj+1)P(vj+1|Yj)

P(yj+1|Yj)

Prediction:

Analysis:

Filtering:

P(vj |Yj) ≈N�

n=1

w(n)j δ(vj − v(n)j )

Particle filtering:

P(vj |Yj)

v(n)j

Page 6: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

=

Rn

P(vj+1|vj)Q(vj+1|vj , Yj+1)

Q(vj+1|vj , Yj+1)P(vj |Yj)dvj

Particle filters

{v(n)j , w(n)j }Nn=1 �→ {v(n)j+1, w

(n)j+1}

Nn=1

How to determine the new positions and weights of particles at time j+1 given the weights and positions at time j?

Prediction:

P(vj+1|Yj) =

Rn

P(vj+1|vj)P(vj |Yj)dvj

Page 7: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

This new particle is then re-weighted according to the analysis formulae of filtering to give

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

w(n)j+1 =

�w(n)j+1��N

n=1 �w(n)j+1

P(vj+1|Yj+1) ≈ PN (vj+1|Yj+1) :=N�

n+1

w(n)j+1δ(vj+1 − v(n)j+1)

Particle filters

2)

3) The weights are normalised so that they sum to one

Leading to:

v(n)j+1 ∼ Q(vj+1|v(n)j , Yj+1)

1)

Overall framework, given {v(n)j , w(n)j }Nn=1 �→ {v(n)j+1, w

(n)j+1}

Nn=1

Sample the new position of each particle by sampling from a proposal probability distribution

Page 8: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Different particle filters

The key two elements that differ between particle filters are:

v(n)j+1 ∼ Q(v(n)j+1|v(n)j , Yj+1)

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

proposal distribution

associated weight update

Page 9: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

time j

PN (vj |Yj) :=

P(vj |Yj) ≈N�

n=1

w(n)j δ(vj − v(n)j )

v(n)j

Standard particle filter - proposal density

P(vj+1|vj) ∝ exp

�−1

2

���Σ− 12 (vj+1 −Ψ(vj))

���2�

Ψ(v(n)j )v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

proposal distributionQ(vj+1|vj , Yj+1) ≡ P(vj+1|vj)

Page 10: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

=⇒Q(v(n)j+1|v(n)j , Yj+1) ≡ P(v(n)j+1|v

(n)j )

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

P(v(n)j+1|v(n)j )

Standard particle filter - weight update

associated weight update

�w(n)j+1 ∝ w(n)

j P(yj+1|v(n)j+1)

Page 11: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Standard particle filter - weight update

P(yj+1|v(n)j+1) ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2�

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

yj+1

Page 12: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Standard particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 13: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Old weights are multiplied by the new weights

Standard particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

�w(n)j+1 ∝ w(n)

j P(yj+1|v(n)j+1)

Page 14: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 15: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 16: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 17: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 18: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

v(n)j+1 = Ψ(v(n)j ) + ξ(n)j

Page 19: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

A simple resampling scheme

1. Put all the weights after each other on the unit interval:

2. Draw a random number from the uniform distribution over [0,1/N], in this case (with 10 members) over [0,1/10].

3. Put that number on the unit interval: its end point is the first member drawn.

4. Add 1/N to the end point: the new end point is our second member drawn. Repeat this until N new members are obtained,

5. In our example we choose m1, m2, m3, m5 twice, m6 twice, m8 twice and m10 and loose m4, m7 and m9 due to small weights

0 1w1 w2 w3 w4 w5 w6 w7 w8 w9 w10

0 1

m1 m2 m3 m5 m5 m6 m6 m8 m8 m10

Page 20: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Sequential Importance Resampling (SIR) particle filter

1.

2.

3.

4.

5.

6.

Set j = 0 and PN (v0|Y0) = P(v0)

Draw v(n)j ∼ PN (vj |Yj) (resample)

Set w(n)j = 1/N, n = 1, . . . , N

i.e. Set �v(n)j+1 = Ψ(v(n)j ) + ξ(n)j , ξ(n)j ∼ N(0,Σ)

Draw �v(n)j+1 ∼ P(�vj+1|v(n)j )

Calculate w(n)j+1 = P(yj+1|v(n)j+1)/

�N�

n=1

P(yj+1|v(n)j+1)

where P(yj+1|v(n)j+1) ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2�

j + 1 �→ j and return to step 2

Page 21: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

truth run from which the observations are generated

Lorenz 63 system of equations - 20 particles

particles

x-variable

y-variable

z-variable

Page 22: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

−2 −1 0 1 2 3 4 5 6 7 80

0.02

0.04

0.06

0.08

0.1

0.12

0.14

0.16

0.18

0.2

x−variable

Frequency

- observation

Lorenz 63 system of equations

Model prior

Page 23: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

−2 −1 0 1 2 3 4 5 6 7 80

0.02

0.04

0.06

0.08

0.1

0.12

0.14

0.16

0.18

0.2

x−variable

Frequency

- observation

Lorenz 63 system of equations

Model prior (blue) and posterior (red)

Page 24: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

−2 −1 0 1 2 3 4 5 6 7 8x−variable

green truth posterior pdf comes from a 1000 particle run of the standard particle filter

Lorenz 63 system of equations

posterior (red) compared to truth (green)

Page 25: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

truth run from which the observations are generated particles

x-variable

y-variable

z-variable

Lorenz 63 system of equations - 20 particles

Page 26: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Standard Particle Filter fails - filter degeneracy

−8 −6 −4 −2 0 2 4 6 8 100

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

x−variable

Frequency

- observation- truth

Model prior

Page 27: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Standard Particle Filter fails - filter degeneracy

−8 −6 −4 −2 0 2 4 6 8 100

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

x−variable

Frequency

- observation- truth

Model prior (blue) and posterior (red)

Page 28: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

−8 −6 −4 −2 0 2 4 6 8 10x−variable

Standard Particle Filter fails - filter degeneracy

posterior (red) compared to truth (green)

Page 29: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

High dimensional systems

Barotropic vorticity:

256 by 256 grid - 65,536 variables

Doubly periodic boundary conditions

Semi-langrangian time stepping scheme

Twin experiments

Observations every 50 time steps - decorrelation time of 42

32 particles

2

q(xj|xj−1, y

n) ∼ N(f(xj−1) + ‘nudge’, Q̂)

xj = f(xj−1) +K(yn −H(xj−1)) + β̂j

p(xn) =p(xn|xn−1)

q(xn|xn−1, yn)....

p(x1|x0)

q(x1|x0, yn)q(xn|xn−1

, yn)....q(x1|x0

, yn)p(x0)

=1

N

N�

i=1

p(xn|xn−1)

q(xn|xn−1, yn)....

p(x1|x0)

q(x1|x0, yn)δ(xn − x

ni )

=1

N

N�

i=1

wpriori δ(xn − x

ni )

Dq

Dt=

∂q

∂t+ u.∇q = 0

Page 30: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Mean of SIR filter fails to capture truth

True model state

• Every variable is observed

• 1200 time steps

Mean of particles

Page 31: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

50 100 150 200 250

50

100

150

200

250

5

4

3

2

1

0

1

2

3

4

5

50 100 150 200 250

50

100

150

200

250

5

4

3

2

1

0

1

2

3

4

5

SIR Filter - filter degeneracy is evident

True model state Mean of particles

1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 10

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Frequency Mean of ensemble generated with

standard particle filter compared to true model state at time step 50

Page 32: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

�w(1)j+1 ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(1)j+1))

���2�

= exp(−0.005M)

�w(2)j+1 ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(2)j+1))

���2�

= exp(−0.02M)

A closer look at the weights

Assume particle 1 is at 0.1 standard deviations s of M independent observations. Then its weight will be:

Assume particle 2 is at 0.2 standard deviations s of M independent observations. Then its weight will be:

Page 33: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

�w(2)j+1

�w(1)j+1

= exp(−0.03) ≈ 0.1

�w(2)j+1

�w(1)j+1

= exp(−15) ≈ 3× 10−7

A closer look at the weights

The ratio of the weights is:

So for M=2 the ratio of the two weights is:

But for M=1000 the ratio of the two weights is:

Conclusion: the number of independent observations is responsible for the degeneracy in particle filters

�w(2)j+1

�w(1)j+1

= exp(−0.015M)

Page 34: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Different particle filters

Proposal densities can be chosen to try and reduce this variance in the weights.

v(n)j+1 ∼ Q(v(n)j+1|v(n)j , Yj+1)

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

proposal distribution

associated weight update

Page 35: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Q(vj+1|v(n)j , Yj+1) ≡ P(vj+1|v(n)j , yj+1)

P(vj+1|v(n)j ) ∝ exp

�−1

2

���Σ− 12 (vj+1 −Ψ(v(n)j ))

���2�

v(n)j+1 = Ψ(v(n)j ) + ΣHT

�HΣHT + Γ

�−1�yj+1 −HΨ(v(n)j )

�+ ζ(n)j

ζ(n)j ∼ N(0, P ), P−1 = Σ−1 +H

TΓ−1H

Optimal proposal densityproposal distribution

for:

P(yj+1|v(n)j+1) ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2�

this equates to (assuming a linear observation operator):

This is similar to Kalman gain matrix K but using model error rather than prior error

Page 36: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Q(v(n)j+1|v(n)j , Yj+1) = P(v(n)j+1|v

(n)j , yj+1) =

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

P(yj+1|v(n)j )

�w(n)j+1 ∝ w(n)

j P(yj+1|v(n)j )

Optimal proposal density - weight update

associated weight update

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

= w(n)j P(yj+1|v(n)j )

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

= w(n)j P(yj+1|v(n)j )

Linked to the use of the Kalman gain, the choice of new model state is the maximum a-posterior of these two distributions

weight does not depend on new model state

Page 37: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Optimal proposal density - weight update

v(n)j+1 = Ψ(v(n)j ) + ΣHT

�HΣHT + Γ

�−1�yj+1 −HΨ(v(n)j )

�+ ζ(n)j

v(n)j

weight is the maximum weight it is possible for a particle to achieve given its position

sample of random error from stated distribution has no effect on the weight

The variance in the weights is therefore the variance in the maximum weight it is possible for each particle to achieve

P(yj+1|v(n)j ) ∝ exp

�−1

2

���(HΣHT + Γ)−12 (yj+1 −HΨ(v(n)j ))

���2�

Page 38: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

1.

2.

3.

4.

5.

6.

Optimal proposal density

Set j = 0 and PN (v0|Y0) = P(v0)

Draw v(n)j ∼ PN (vj |Yj) (resample)

Set w(n)j = 1/N, n = 1, . . . , N

j + 1 �→ j and return to step 2

Draw �v(n)j+1 ∼ P(�vj+1|v(n)j , yj+1)

i.e. Set v(n)j+1 = Ψ(v(n)j ) + ΣHT�HΣHT + Γ

�−1�yj+1 −HΨ(v(n)j )

�+ ζ(n)j

Calculate w(n)j+1 = P(yj+1|v(n)j )/

�N�

n=1

P(yj+1|v(n)j )

where P(yj+1|v(n)j ) ∝ exp

�−1

2

���(HΣHT + Γ)−12 (yj+1 −HΨ(v(n)j ))

���2�

Page 39: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Optimal proposal density - variance of weightsSNYDER: PARTICLE FILTERS

0 0.2 0.4 0.6 0.8 10

200

400

max wi

occu

rrenc

es Nx = 160

0 0.2 0.4 0.6 0.8 10

200

400

max wi

Nx = 160

0 0.2 0.4 0.6 0.8 10

200

400

occu

rrenc

es Nx = 40

0 0.2 0.4 0.6 0.8 10

200

400Nx = 40

0 0.2 0.4 0.6 0.8 10

200

400

occu

rrenc

es Nx = 10

Standard proposal

0 0.2 0.4 0.6 0.8 10

200

400Nx = 10

Optimal proposal

Figure 2: Histograms of the maximum weight from 103 simulations of the system (17), using eitherthe standard proposals (left column) or the optimal proposal (right column). The weights come froma single update step for observations yk, with xik!1 drawn from N(0,I). In the case of the optimalproposal with Nx = 10 and the standard proposal witht Nx = 160, the first and last bin, respectively,have greater than 400 occurrences.

The optimal proposal distribution involves a Kalman-filter update of the forecast from xk!1 given obser-vations yk.

The new weights satisfy

wik ! exp!

!12|yk!xik|2

"

, (20)

for the standard proposal and

wik ! exp

#

!|yk!axik!1|2

2(1+q2)

$

. (21)

The arguments of the exponentials in (20) and (21) have expected values that grow linearly with Ny.Thus, when Ny is large, a unit change in the arguments will produce increasingly dramatic changes inwik, leading to a situation in which one or a few realizations of xik (or xik!1 in the case of the optimalproposal) produce weights that are much larger than all others. For specified Ny, however, the argumentof the exponential in (21) will be less than that in (20) with high probability: the denominator is alwayslarger and axik!1 will usually be closer to yk than xik is, since xik is affected by the system noise. Thissuggests, correctly, that the optimal proposal will reduce the problem of degeneracy.

These points are illustrated in Fig. 2, which shows histograms of the maximum weight from simulationsof a single update step with each of the proposals. The ensemble size is fixed, Ne = 103, and thedimension of the state and observation vectors varies, Nx = 10, 40, 160. As Nx increases, maximumweights close to unity become more frequent with either proposal – this is the degeneracy problemor, in the terminology of Snyder et al. (2008), the collapse of the weights. At each Nx, however, thedegeneracy is less pronounced when using the optimal proposal.

6 ECMWF Seminar on Data assimilation for atmosphere and ocean, 6 - 9 September 2011

Optimal proposal improves on SIR filter, but filter degeneracy still occurs in high dimensional systems

(‘Particle filters, the ‘optimal’ proposal and high dimensional systems’ Snyder, 2012)

Page 40: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

v(n)j+1 = Ψ(v(n)j ) +Kj+1(yj+1 −H(Ψ(v(n)j ))) + (I −Kj+1H)ξ(n)j +Kj+1η

(n)j+1}

= µ(n)j+1

Q = (I −Kj+1H)Σ(I −Kj+1H)T +Kj+1ΓKj+1

Q(v(n)j+1|v(n)j , Yj+1) ∝ exp

�−1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

Ensemble Kalman Filter as proposal

proposal distribution

EnKF:

�v(n)j+1 = Ψ(v(n)j ) + ξ(n)j , ξ(n)j ∼ N(0,Σ)

v(n)j+1 = (I −Kj+1H)�v(n)j+1 +Kj+1y

(n)j+1

y(n)j+1 = yj+1 + η(n)j+1, η(n)j+1 ∼ N(0,Γ)

prediction

analysis

observation perturbation }

Deterministic Stochastic

Page 41: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Kj+1 = �Cj+1HT�H �Cj+1H

T + Γ�−1

Ensemble Kalman Filter as proposal

Kalman gain:

�Cj+1 =1

N − 1

N�

n=1

(�v(n)j+1 − �mj+1)(�v(n)j+1 − �mj+1)T

�mj+1 =1

N

N�

n=1

�v(n)j+1

Approximated using the ensemble

So the proposal used to update each particle actually depends on all the other particle positions. However, for an infinitely large ensemble of particles, the Kalman gain will depend only on the system of model equations via the prediction step.

Page 42: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

P(v(n)j+1|v(n)j ) ∝ exp

�−1

2

���Σ− 12 (v(n)j+1 −Ψ(v(n)j ))

���2�

�w(n)j+1 ∝ w(n)

j exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2− 1

2

���Σ− 12 (v(n)j+1 −Ψ(v(n)j ))

���2+

1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

Q(v(n)j+1|v(n)j , Yj+1) ∝ exp

�−1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

Ensemble Kalman Filter as proposal - weight update

�w(n)j+1 = w(n)

j

P(yj+1|v(n)j+1)P(v(n)j+1|v

(n)j )

Q(v(n)j+1|v(n)j , Yj+1)

= w(n)j P(yj+1|v(n)j )

No longer have a simplification in the weights (as in the optimal proposal density) and so need to directly calculate the different constituent parts of the weight for each particle

P(yj+1|v(n)j+1) ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2�

�w(n)j+1 ∝ w(n)

j exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2− 1

2

���Σ− 12 (v(n)j+1 −Ψ(v(n)j ))

���2+

1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

Page 43: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Ensemble Kalman Filter as proposal

1.

2.

3.

4.

5.

6.

Set j = 0 and PN (v0|Y0) = P(v0)

Draw v(n)j ∼ PN (vj |Yj) (resample)

Set w(n)j = 1/N, n = 1, . . . , N

j + 1 �→ j and return to step 2

v(n)j+1 = Ψ(v(n)j ) +Kj+1(yj+1 −H(Ψ(v(n)j ))) + (I −Kj+1H)ξ(n)j +Kj+1η

(n)j+1

Set

Calculate w(n)j+1 = �w(n)

j+1/

�N�

n=1

�w(n)j+1

where �w(n)j+1 ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2− 1

2

���Σ− 12 (v(n)j+1 −Ψ(v(n)j ))

���2+

1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

where �w(n)j+1 ∝ exp

�−1

2

���Γ− 12 (yj+1 − h(v(n)j+1))

���2− 1

2

���Σ− 12 (v(n)j+1 −Ψ(v(n)j ))

���2+

1

2

���Q− 12 (v(n)j+1 − µ(n)

j+1)���2�

Weighted EnKF now ensures the true posterior is represented but unfortunately filter degeneracy

still occurs.

Page 44: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

SIR particle filter

Summary

Optimal proposal density

Weighted ensemble Kalman filter

• The proposal density is the model transition density and so the model equations are used to propagate each particle forward in time• The weight is calculated based on the likelihood, so the distance of each particle to the observation

• The new model state is sampled from a proposal which is the density of possible new model states given the old model state and the observation• The weight is then calculated based on the maximum weight a particle could achieve given the old model state and the observation

• The new model state is generated using the Ensemble Kalman equations• The weight then must be calculated directly through all three constituent parts

Although the optimal proposal density improves on the SIR filter, all three schemes still suffer from filter degeneracy. This is because of the difficulty in sampling from the high probability region of the posterior in high dimensions with large numbers of independent observations.

Page 45: Nonlinear data assimilation Part I: Particle filters seminars... · Why nonlinear data assimilation? The complexity of models are increasing: Precipitation Results of variational

Doucet AN, Freitas D, Gordon N. 2001. Sequential Monte-Carlo Methods inPractice. Springer-Verlag

Robert CP, Cassela G. 2004. Monte-Carlo Statistical Methods. Springer-Verlag

Doucet AS, Godsill S, Andrieu C. 2000. On sequential Monte Carlo sampling

methods for Bayesian filtering. Statistics and Computing 10: 197-208

Snyder C, Bengtsson T, Bickel P, Anderson J. 2008. Obstacles to high-

dimensional particle filtering. Monthly Weather Review 136: 4629-4640

Papadakis N, Memin E, Cuzol A, Gengembre N. 2010. Data assimilationwith the weighted ensemble Kalman filter. Tellus A 62: 673-697

Van Leeuwen PJ. 2009. Particle filtering in geophysical systems. MonthlyWeather Review 137: 4089-4114

References

Optimal proposal density

SIR particle filter

Weighted ensemble Kalman filter