data assimilation theory ctcd data assimilation workshop nov 2005

15
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005 Sarah Dance

Upload: maxine

Post on 21-Jan-2016

72 views

Category:

Documents


0 download

DESCRIPTION

Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005. Sarah Dance. Data assimilation is often treated as a black box algorithm. OUT Analysis. IN Observations and a priori information. (apologies to Rube Goldberg). - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005

Sarah Dance

Page 2: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

(apologies to Rube Goldberg)

Data assimilation is often treated as a black box algorithm

OUTAnalysis

INObservations

anda priori

information

BUT, understanding and developing what goes on inside the box is crucial !!

Page 3: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Some DARC DA Theory Group Projects

Nonlinear assimilation techniques

Convergence of 4D-Var

Reduced order modelling

Stochastic Processes in DA

Treatment of observation error correlations

Background error modelling and balance

Multiscale DA Phase errors and rearrangement theory

Model errors and bias correction

EnKF and bias

Information content Observation targeting

Page 4: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Formulations of the Ensemble Kalman Filter and Bias

MSc thesis by David Livings, supervised by Sarah Dance and Nancy

Nichols

Page 5: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Outline

• Bayesian state estimation and the Kalman Filter

• The EnKF• Bias and the EnKF• Conclusions

Page 6: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

e.g. Suppose xk = M xk-1+

M is linear, the prior and model noise are Gaussian

P(xk-1) ~ N(xb, P) ~ N(0, Q)

Then P(xk |xk-1) ~N(Mxb, MPMT+Q)

Prediction (between observations)

Page 7: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

At an observation we use Bayes rule

)(xp

)|()()|( xyxyx ppp

Prior Background error distribution

)|( xypLikelihood of observationsObservation error pdf

Bayes rule

Page 8: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Bayes rule illustrated

)(xp)|( xyp

Page 9: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Bayes rule illustrated (cont)

)|( xyp )(xp

)|()()|( xyxyx ppp

Page 10: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

The Kalman Filter

• Use prediction equation and Bayes rule • Assume linear models (forecast and observation)• Assume Gaussian statistics

Kalman filter BUT• Models are nonlinear• Evolving large covariance matrices is

expensive (106 x 106 in meteorology)• So use an ensemble (Monte Carlo idea)

Page 11: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

=

=

=

Page 12: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

N=10, Perfect observations

Red ensemble mean

Blue ensemble std.

Error bars indicate obs std.

Results with ETKF (old formulation) and Peter Lynch’s swinging spring model

Ensemble statistics not consistent with the truth!

Page 13: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Bias and the EnKF

• Many EnKF algorithms, can be put into a “square root” framework.

• Define an ensemble perturbation matrix:

x1x

2x

3x4x

5x

So, by definition of the ensemble mean

01XX

1

1

Page 14: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

The mean of the ensemble is updated separately.

Ensemble perturbations are updated as

where T is a (non-unique) square root of an update equation.

Thus, for consistency,

David discovered that not all implementations preserve this property.

We have now found nec. and suff. conditions for consistency.

Square-root ensemble updates

0T1X1X fa

Page 15: Data Assimilation Theory  CTCD Data Assimilation Workshop Nov 2005

Consequences

• The ensemble will be biased

• The size of the ensemble spread will be too small

• Filter divergence is more likely to occur !

• Care must be taken in algorithm choice and implementation