optimized localization and hybridization to lter ensemble ......introduction linear ltering joint...

84

Upload: others

Post on 22-May-2020

20 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Optimized localization and hybridizationto �lter ensemble-based covariances

Benjamin Ménétrier and Tom AulignéNCAR - Boulder - Colorado

Roanoke - 06/04/2015

Acknowledgement: AFWA

Page 2: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 3: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 4: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 5: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 6: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 7: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 8: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 9: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Context:

• DA often relies on forecast error covariances.

• This matrix can be sampled from an ensemble of forecasts.

• Sampling noise arises because of the limited ensemble size.

• Question: how to �lter this sampling noise?

Usual methods:

• Covariance localization→ tapering with a localization matrix

• Covariance hybridization→ linear combination with a static covariance matrix

1

Page 10: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 11: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 12: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 13: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 14: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 15: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 16: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 17: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Introduction

Questions:

1. Can localization and hybridization be considered together?

2. Is it possible to optimize localization and hybridizationcoe�cients objectively and simultaneously?

The method should:

• use data from the ensemble only.

• be a�ordable for high-dimensional systems.

3. Is hybridization always improving the accuracy of forecasterror covariances?

2

Page 18: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Outline

Introduction

Linear �ltering of sample covariances

Joint optimization of localization and hybridization

Results

Conclusions

3

Page 19: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Outline

Introduction

Linear �ltering of sample covariances

Joint optimization of localization and hybridization

Results

Conclusions

4

Page 20: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

An ensemble of N forecasts {x̃bp} is used to sample B̃:

B̃ =1

N−1

N

∑p=1

δ x̃b(δ x̃b

)T

where: δ x̃bp

= x̃bp−〈x̃b〉 and 〈x̃b〉=

1

N

N

∑p=1

x̃bp

Asymptotic behavior: if N → ∞ , then B̃→ B̃?

In practice, N < ∞ ⇒ sampling noise B̃e = B̃− B̃?

Theory of sampling error:

E[B̃2ij

]=

N(N−3)

(N−1)2E[B̃?2ij

]− 1

(N−1)(N−2)E[B̃iiB̃jj

]+

N2

(N−1)2(N−2)E[Ξ̃ijij

]

5

Page 21: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

An ensemble of N forecasts {x̃bp} is used to sample B̃:

B̃ =1

N−1

N

∑p=1

δ x̃b(δ x̃b

)T

where: δ x̃bp

= x̃bp−〈x̃b〉 and 〈x̃b〉=

1

N

N

∑p=1

x̃bp

Asymptotic behavior: if N → ∞ , then B̃→ B̃?

In practice, N < ∞ ⇒ sampling noise B̃e = B̃− B̃?

Theory of sampling error:

E[B̃2ij

]=

N(N−3)

(N−1)2E[B̃?2ij

]− 1

(N−1)(N−2)E[B̃iiB̃jj

]+

N2

(N−1)2(N−2)E[Ξ̃ijij

]

5

Page 22: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

An ensemble of N forecasts {x̃bp} is used to sample B̃:

B̃ =1

N−1

N

∑p=1

δ x̃b(δ x̃b

)T

where: δ x̃bp

= x̃bp−〈x̃b〉 and 〈x̃b〉=

1

N

N

∑p=1

x̃bp

Asymptotic behavior: if N → ∞ , then B̃→ B̃?

In practice, N < ∞ ⇒ sampling noise B̃e = B̃− B̃?

Theory of sampling error:

E[B̃2ij

]=

N(N−3)

(N−1)2E[B̃?2ij

]− 1

(N−1)(N−2)E[B̃iiB̃jj

]+

N2

(N−1)2(N−2)E[Ξ̃ijij

]

5

Page 23: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

An ensemble of N forecasts {x̃bp} is used to sample B̃:

B̃ =1

N−1

N

∑p=1

δ x̃b(δ x̃b

)T

where: δ x̃bp

= x̃bp−〈x̃b〉 and 〈x̃b〉=

1

N

N

∑p=1

x̃bp

Asymptotic behavior: if N → ∞ , then B̃→ B̃?

In practice, N < ∞ ⇒ sampling noise B̃e = B̃− B̃?

Theory of sampling error:

E[B̃2ij

]=

N(N−3)

(N−1)2E[B̃?2ij

]− 1

(N−1)(N−2)E[B̃iiB̃jj

]+

N2

(N−1)2(N−2)E[Ξ̃ijij

]

5

Page 24: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

An ensemble of N forecasts {x̃bp} is used to sample B̃:

B̃ =1

N−1

N

∑p=1

δ x̃b(δ x̃b

)T

where: δ x̃bp

= x̃bp−〈x̃b〉 and 〈x̃b〉=

1

N

N

∑p=1

x̃bp

Asymptotic behavior: if N → ∞ , then B̃→ B̃?

In practice, N < ∞ ⇒ sampling noise B̃e = B̃− B̃?

Theory of sampling error:

E[B̃2ij

]=

N(N−3)

(N−1)2E[B̃?2ij

]− 1

(N−1)(N−2)E[B̃iiB̃jj

]+

N2

(N−1)2(N−2)E[Ξ̃ijij

]5

Page 25: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)Localization by L + hybridization with B

Covariance matrix

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 26: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)

Localization by L + hybridization with B

Covariance matrix

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 27: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)Localization by L + hybridization with B

Covariance matrix

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 28: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)Localization by L + hybridization with B

Covariance matrix

B̂h = (βe)2 L◦ B̃+ (β

c)2 B

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 29: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)Localization by L + hybridization with B

Covariance matrix

B̂h = (βe)2 L︸ ︷︷ ︸

Gain Lh

◦ B̃+ (βc)2 B︸ ︷︷ ︸

Offset

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 30: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Linear �ltering of sample covariances

Localization by L (Schur product)

Covariance matrix

B̂ = L◦ B̃

Increment

δxe =1√N−1

N

∑p=1

δ x̃bp◦(L1/2vα

p

)Localization by L + hybridization with B

Covariance matrix

B̂h = (βe)2 L︸ ︷︷ ︸

Gain Lh

◦ B̃+ (βc)2 B︸ ︷︷ ︸

Offset

Increment

δx = βe

δxe + βc B1/2vc

Localization + hybridization = linear �ltering of B̃

Lh and βc have to be optimized together

6

Page 31: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Outline

Introduction

Linear �ltering of sample covariances

Joint optimization of localization and hybridization

Results

Conclusions

7

Page 32: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 33: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 34: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 35: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 36: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 37: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

Step 1: optimizing the localization only, without hybridization

Goal: to minimize the expected quadratic error:

e = E[‖ L◦ B̃︸ ︷︷ ︸

Localized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

(1)

Light assumptions:

• The unbiased sampling noise B̃e = B̃− B̃? is not correlatedwith the asymptotic sample covariance matrix B̃?.

• The two random processes generating the asymptotic B̃? andthe sample distribution are independent.

An explicit formula for the optimal localization L is given inMénétrier et al. 2015 (Montly Weather Review).

8

Page 38: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

This formula of optimal localization L involves:

• the ensemble size N

• the sample covariance B̃

• the sample fourth-order centered moment Ξ̃

Lij

=(N−1)2

N(N−3)

− N

(N−2)(N−3)

E[Ξ̃ijij

]E[B̃2ij

]+

N−1

N(N−2)(N−3)

E[B̃iiB̃jj

]E[B̃2ij

]

9

Page 39: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 1

This formula of optimal localization L involves:

• the ensemble size N

• the sample covariance B̃

• the sample fourth-order centered moment Ξ̃

Lij

=(N−1)2

N(N−3)

− N

(N−2)(N−3)

E[Ξ̃ijij

]E[B̃2ij

]+

N−1

N(N−2)(N−3)

E[B̃iiB̃jj

]E[B̃2ij

]9

Page 40: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 2

Step 2: optimizing localization and hybridization together

Goal: to minimize the expected quadratic error

eh = E[‖ Lh ◦ B̃+ (β

c)2B︸ ︷︷ ︸Localized / hybridized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

Same assumptions as before.

Result of the minimization: a linear system in Lh and (βc)2

Lhij

= Lij−

E[B̃ij

]E[B̃2ij

]Bij

(βc)2 (2a)

(βc)2 =

∑ijB

ij

(1−Lh

ij

)E[B̃ij

]∑

ijB2

ij

(2b)

10

Page 41: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 2

Step 2: optimizing localization and hybridization together

Goal: to minimize the expected quadratic error

eh = E[‖ Lh ◦ B̃+ (β

c)2B︸ ︷︷ ︸Localized / hybridized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

Same assumptions as before.

Result of the minimization: a linear system in Lh and (βc)2

Lhij

= Lij−

E[B̃ij

]E[B̃2ij

]Bij

(βc)2 (2a)

(βc)2 =

∑ijB

ij

(1−Lh

ij

)E[B̃ij

]∑

ijB2

ij

(2b)

10

Page 42: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 2

Step 2: optimizing localization and hybridization together

Goal: to minimize the expected quadratic error

eh = E[‖ Lh ◦ B̃+ (β

c)2B︸ ︷︷ ︸Localized / hybridized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

Same assumptions as before.

Result of the minimization: a linear system in Lh and (βc)2

Lhij

= Lij−

E[B̃ij

]E[B̃2ij

]Bij

(βc)2 (2a)

(βc)2 =

∑ijB

ij

(1−Lh

ij

)E[B̃ij

]∑

ijB2

ij

(2b)

10

Page 43: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Joint optimization: step 2

Step 2: optimizing localization and hybridization together

Goal: to minimize the expected quadratic error

eh = E[‖ Lh ◦ B̃+ (β

c)2B︸ ︷︷ ︸Localized / hybridized B̃

− B̃?︸︷︷︸Asymptotic B̃

‖2]

Same assumptions as before.

Result of the minimization: a linear system in Lh and (βc)2

Lhij

= Lij−

E[B̃ij

]E[B̃2ij

]Bij

(βc)2 (2a)

(βc)2 =

∑ijB

ij

(1−Lh

ij

)E[B̃ij

]∑

ijB2

ij

(2b)

10

Page 44: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Hybridization bene�ts

Comparison of:

• B̂ = L◦ B̃, with an optimal L minimizing e

• B̂h = Lh ◦ B̃+ (βc)2 B, with optimal Lh and β

c minimizing eh

We can show that:

eh− e =−(βc)2∑

ij

B2ijVar(B̃ij

)E[B̃2ij

]︸ ︷︷ ︸≤0

(3)

With optimal parameters, whatever the static B:Localization + hybridization is better than localization alone

11

Page 45: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Hybridization bene�ts

Comparison of:

• B̂ = L◦ B̃, with an optimal L minimizing e

• B̂h = Lh ◦ B̃+ (βc)2 B, with optimal Lh and β

c minimizing eh

We can show that:

eh− e =−(βc)2∑

ij

B2ijVar(B̃ij

)E[B̃2ij

]︸ ︷︷ ︸≤0

(3)

With optimal parameters, whatever the static B:Localization + hybridization is better than localization alone

11

Page 46: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Hybridization bene�ts

Comparison of:

• B̂ = L◦ B̃, with an optimal L minimizing e

• B̂h = Lh ◦ B̃+ (βc)2 B, with optimal Lh and β

c minimizing eh

We can show that:

eh− e =−(βc)2∑

ij

B2ijVar(B̃ij

)E[B̃2ij

]︸ ︷︷ ︸≤0

(3)

With optimal parameters, whatever the static B:Localization + hybridization is better than localization alone

11

Page 47: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Outline

Introduction

Linear �ltering of sample covariances

Joint optimization of localization and hybridization

Results

Conclusions

12

Page 48: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Practical implementation

An ergodicity assumption is required to estimate the statisticalexpectations E in practice:

• whole domain average,

• local average,

• scale dependent average,

• etc.

→ This assumption is independent from earlier theory.

Localization Lh and hybridization coe�cient βc can be computed:

• from the ensemble at each assimilation window,

• climatologically from an archive of ensembles.

13

Page 49: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Practical implementation

An ergodicity assumption is required to estimate the statisticalexpectations E in practice:

• whole domain average,

• local average,

• scale dependent average,

• etc.

→ This assumption is independent from earlier theory.

Localization Lh and hybridization coe�cient βc can be computed:

• from the ensemble at each assimilation window,

• climatologically from an archive of ensembles.

13

Page 50: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Practical implementation

An ergodicity assumption is required to estimate the statisticalexpectations E in practice:

• whole domain average,

• local average,

• scale dependent average,

• etc.

→ This assumption is independent from earlier theory.

Localization Lh and hybridization coe�cient βc can be computed:

• from the ensemble at each assimilation window,

• climatologically from an archive of ensembles.

13

Page 51: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Practical implementation

An ergodicity assumption is required to estimate the statisticalexpectations E in practice:

• whole domain average,

• local average,

• scale dependent average,

• etc.

→ This assumption is independent from earlier theory.

Localization Lh and hybridization coe�cient βc can be computed:

• from the ensemble at each assimilation window,

• climatologically from an archive of ensembles.

13

Page 52: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Practical implementation

An ergodicity assumption is required to estimate the statisticalexpectations E in practice:

• whole domain average,

• local average,

• scale dependent average,

• etc.

→ This assumption is independent from earlier theory.

Localization Lh and hybridization coe�cient βc can be computed:

• from the ensemble at each assimilation window,

• climatologically from an archive of ensembles.

13

Page 53: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Experimental setup

• WRF-ARW model, large domain, 25 km-resolution, 40 levels

• Initial conditions randomized from a homogeneous static B

• Reference and test ensembles (1000 / 100 members)

• Forecast ranges: 12, 24, 36 and 48 h

Temperature at level 7 (∼ 1 km above ground), 48 h-range forecasts

Standard-deviation (K) Correlations functions

14

Page 54: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Experimental setup

• WRF-ARW model, large domain, 25 km-resolution, 40 levels

• Initial conditions randomized from a homogeneous static B

• Reference and test ensembles (1000 / 100 members)

• Forecast ranges: 12, 24, 36 and 48 h

Temperature at level 7 (∼ 1 km above ground), 48 h-range forecasts

Standard-deviation (K) Correlations functions

14

Page 55: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Experimental setup

• WRF-ARW model, large domain, 25 km-resolution, 40 levels

• Initial conditions randomized from a homogeneous static B

• Reference and test ensembles (1000 / 100 members)

• Forecast ranges: 12, 24, 36 and 48 h

Temperature at level 7 (∼ 1 km above ground), 48 h-range forecasts

Standard-deviation (K) Correlations functions

14

Page 56: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Experimental setup

• WRF-ARW model, large domain, 25 km-resolution, 40 levels

• Initial conditions randomized from a homogeneous static B

• Reference and test ensembles (1000 / 100 members)

• Forecast ranges: 12, 24, 36 and 48 h

Temperature at level 7 (∼ 1 km above ground), 48 h-range forecasts

Standard-deviation (K) Correlations functions

14

Page 57: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Experimental setup

• WRF-ARW model, large domain, 25 km-resolution, 40 levels

• Initial conditions randomized from a homogeneous static B

• Reference and test ensembles (1000 / 100 members)

• Forecast ranges: 12, 24, 36 and 48 h

Temperature at level 7 (∼ 1 km above ground), 48 h-range forecasts

Standard-deviation (K) Correlations functions

14

Page 58: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

15

Page 59: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

15

Page 60: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Localization length-scale:

15

Page 61: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Hybridization coe�cients for zonal wind:

15

Page 62: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Impact of the hybridization:

• B̃?is estimated with the reference ensemble

• Expected quadratic errors e and ehare computed

Error reduction from e to eh for 25 members

Zonal wind Meridian wind Temperature Speci�c humidity

4.5 % 4.2 % 3.9 % 1.7 %

→ Hybridization with B improves the accuracy of theforecast error covariance matrix

15

Page 63: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Impact of the hybridization:

• B̃?is estimated with the reference ensemble

• Expected quadratic errors e and ehare computed

Error reduction from e to eh for 25 members

Zonal wind Meridian wind Temperature Speci�c humidity

4.5 % 4.2 % 3.9 % 1.7 %

→ Hybridization with B improves the accuracy of theforecast error covariance matrix

15

Page 64: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Impact of the hybridization:

• B̃?is estimated with the reference ensemble

• Expected quadratic errors e and ehare computed

Error reduction from e to eh for 25 members

Zonal wind Meridian wind Temperature Speci�c humidity

4.5 % 4.2 % 3.9 % 1.7 %

→ Hybridization with B improves the accuracy of theforecast error covariance matrix

15

Page 65: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Impact of the hybridization:

• B̃?is estimated with the reference ensemble

• Expected quadratic errors e and ehare computed

Error reduction from e to eh for 25 members

Zonal wind Meridian wind Temperature Speci�c humidity

4.5 % 4.2 % 3.9 % 1.7 %

→ Hybridization with B improves the accuracy of theforecast error covariance matrix

15

Page 66: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Localization and hybridization

• Optimization of the horizontal localization Lh

horand of the

hybridization coe�cient βc at each vertical level.

• Static B = horizontal average of B̃

• Impact of the hybridization:

• B̃?is estimated with the reference ensemble

• Expected quadratic errors e and ehare computed

Error reduction from e to eh for 25 members

Zonal wind Meridian wind Temperature Speci�c humidity

4.5 % 4.2 % 3.9 % 1.7 %

→ Hybridization with B improves the accuracy of theforecast error covariance matrix

15

Page 67: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Outline

Introduction

Linear �ltering of sample covariances

Joint optimization of localization and hybridization

Results

Conclusions

16

Page 68: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 69: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 70: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 71: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 72: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 73: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 74: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 75: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Conclusions

1. Localization and hybridization are two joint aspects of thelinear �ltering of sample covariances.

2. We have developed a new objective method to optimizelocalization and hybridization coe�cients together:

• Based on properties of the ensemble only

• A�ordable for high-dimensional systems

• Tackling the sampling noise issue only

3. If done optimally, hybridization always improves the accuracyof forecast error covariances.

Ménétrier, B. and T. Auligné: Optimized Localization andHybridization to Filter Ensemble-Based CovariancesMonthly Weather Review, 2015, accepted

17

Page 76: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 77: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 78: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 79: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 80: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 81: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 82: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 83: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Page 84: Optimized localization and hybridization to lter ensemble ......Introduction Linear ltering Joint optimization Results Conclusions Linear ltering of sample covariances Localization

Introduction Linear �ltering Joint optimization Results Conclusions

Perspectives

Already done in the paper:

• Extension to vectorial hybridization weights:

δx = βe ◦δxe +βc ◦δxc

→ Requires the solution of a nonlinear system A(Lh,βc) = 0,performed by a bound-constrained minimization.

• Heterogeneous optimization: local averages over subdomains

• 3D optimization: joint computation of horizontal and verticallocalizations, and hybridization coe�cients

To be done:

• Tests in a cycled quasi-operational con�guration

• Extension of the theory to account for systematic errors in B̃?

(theory is ready, tests are underway...)

18

Thank you for your attention!

Any question?