+ All Categories
Home > Documents > Some ideas for Ensemble Kalman...

Some ideas for Ensemble Kalman...

Date post: 17-Nov-2020
Category:
Upload: others
View: 2 times
Download: 0 times
Share this document with a friend
44
Some ideas for Ensemble Some ideas for Ensemble Kalman Filter Kalman Filter Former students Former students and and Eugenia Kalnay Eugenia Kalnay UMCP UMCP Acknowledgements: UMD Chaos-Weather Group : Brian Hunt, Istvan Szunyogh, Ed Ott and Jim Yorke, Kayo Ide, and students Former students : Shu-Chih Yang, Takemasa Miyoshi, Junjie Liu, Hong Li, Chris Danforth Also: Malaquías Peña, Matteo Corazza, Pablo Grunman, DJ Patil, Ji-Sun Kang, Debra Baker, Steve Greybush, Tamara Singleton, Matt Hoffman, Steve Penny, Elana Fertig.
Transcript
Page 1: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Some ideas for EnsembleSome ideas for EnsembleKalman FilterKalman Filter

Former students Former students andand Eugenia KalnayEugenia KalnayUMCPUMCP

Acknowledgements:UMD Chaos-Weather Group: Brian Hunt, Istvan Szunyogh, EdOtt and Jim Yorke, Kayo Ide, and studentsFormer students: Shu-Chih Yang, Takemasa Miyoshi, Junjie Liu,Hong Li, Chris DanforthAlso: Malaquías Peña, Matteo Corazza, Pablo Grunman, DJPatil, Ji-Sun Kang, Debra Baker, Steve Greybush, TamaraSingleton, Matt Hoffman, Steve Penny, Elana Fertig.

Page 2: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Some ideas for EnsembleSome ideas for EnsembleKalman FilterKalman Filter

Basic idea:Basic idea:

• We are in competition with 4D-Var…• We take advantage of ideas and propertiesthat were developed for 4D-Var and show thatit can be done easier, cheaper and better withour EnKF, the LETKF (Hunt et al., 2007)• And we don’t need adjoint, priors, etc.

Page 3: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Some ideas to improve LETKFSome ideas to improve LETKFWe can adapt ideas developed within 4D-VarWe can adapt ideas developed within 4D-Var: No-cost smoother (Kalnay et al, Tellus)

Accelerating the spin-up: Running in place(Kalnay and Yang, QJRMS, submitted)

“Outer loop” and nonlinearities (Yang and Kalnay)

Forecast sensitivity to observations (Liu andKalnay, QJRMS, 2008)

Coarse analysis resolution interpolatingweights (Yang, Kalnay, Hunt, Bowler, QJ submitted)

Low-dimensional model bias correction (Li,Kalnay, Danforth, Miyoshi, MWR, submitted)

Page 4: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Local Ensemble Transform Kalman Filter(Ott et al, 2004, Hunt et al, 2004, 2007)

• Model independent(black box)• Obs. assimilatedsimultaneously at eachgrid point• 100% parallel: very fast•No adjoint needed• 4D LETKF extension

(Start with initial ensemble)

LETKFObservationoperator

Model

ensemble analyses

ensemble forecasts

ensemble“observations”

Observations

Page 5: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Perform data assimilation in a local volume, choosing observations

The state estimate is updated at thecentral grid red dot

Localization based on observations

Page 6: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Perform data assimilation in a local volume, choosing observations

The state estimate is updated at thecentral grid red dot

All observations (purple diamonds)within the local region are assimilated

Localization based on observations

The LETKF algorithm can be described in a single slide!

Page 7: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Local Ensemble Transform Kalman Filter (Local Ensemble Transform Kalman Filter (LETKFLETKF))

Forecast step:Analysis step: construct

Locally: Choose for each grid point the observations to be used, andcompute the local analysis error covariance and perturbations inensemble space:

Analysis mean in ensemble space:and add to to get the analysis ensemble in ensemble space

The new ensemble analyses in model space are the columns of . Gathering the grid point analyses forms the new

global analyses. Note that the the output of the LETKF are analysisweights and perturbation analysis matrices of weights . Theseweights multiply the ensemble forecasts.

x

n,k

b= M

nx

n!1,k

a( )Xb

= x1

b ! xb | ... | xK

b ! xb"# $%;

yi

b= H (x

i

b); Y

n

b= y

1

b ! yb | ... | yK

b ! yb"# $%

!Pa= K !1( )I + YbT

R!1Y

b"# $%!1;W

a= [(K !1) !Pa ]1/2

Xn

a= X

n

bW

a+ x

b

wa= !P

aY

bTR

!1(yo! y

b )

Wa

Globally:

wa

Wa

Page 8: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

The 4D-LETKF produces an analysis in terms ofweights of the ensemble forecast members at theanalysis time tn, giving the trajectory that best fits allthe observations in the assimilation window.

Analysis time

Page 9: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

No-cost LETKF smoother ( ): apply at tn-1 the sameweights found optimal at tn. It works for 3D- or 4D-LETKF

The 4D-LETKF produces an analysis in terms ofweights of the ensemble forecast members at theanalysis time tn, giving the trajectory that best fits allthe observations in the assimilation window.

Page 10: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

No-cost LETKF smoothertest on a QG model: It really works!

“Smoother” reanalysis

LETKF Analysisxna= xn

f+ Xn

fwn

aLETKF analysis

at time n

Smoother analysis at time n-1

!xn!1a

= xn!1f

+ Xn!1

fwn

a

This very simple smoother allows us to go backand forth in time within an assimilation widow!!

Page 11: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

““Running in placeRunning in place”” to spin-up fasterto spin-up fasterKalnay and Yang (2008)Kalnay and Yang (2008)

• 4D-Var spins-up faster than EnKF because it is a smoother: itkeeps iterating until it fits the observations within theassimilation window as well as possible• EnKF spins-up fast if starting from a “good” initial state, e.g.,3D-Var, but needs also an ensemble representing the “errors ofthe day”• In a severe storm where radar observations start with thestorm, there is little real time to spin-up• Caya et al. (2005): “EnKF is eventually better than 4D-Var”(but it is too late to be useful, it misses the storm).• Jidong Gao, (pers. comm. 2007): spin-up is the main obstaclefor the use of EnKF for storm prediction.

Page 12: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Can we use the dataCan we use the data more than once?more than once?

• Hunt et al., 2007: The background term representsthe evolution of the maximum likelihood trajectorygiven all the observations in the past

• After the analysis a similar relationship is valid:

• From here one can derive the linear KF equations• Also the rule: “Use the data once and then discard it”

x ! x

n

b"#

$%

T

Pn

b( )!1

x ! xn

b"#

$% + y

n

o ! Hnx"

#$%

T

Rn

!1( ) yn

o ! Hnx"

#$% = x ! x

n

a"#

$%

T

Pn

a( )!1

x ! xn

a"#

$% + c '

yj

o ! HjM

tn

,tj

x"#$

%&'

T

Rj

!1y

j

o ! HjM

tn

,tj

x"#$

%&'=

j=1

n!1

( x ! xn

b"#

%&

T

Pn

b( )!1

x ! xn

b"#

%& + c

Page 13: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Can we use the data Can we use the data more than once?more than once?

• The rule: “Use the data once and then discard it” (Ideet al., 1997) makes sense when the analysis/forecastsare the most likely given all the past data, not when westart from scratch.• We propose “Running in place” until we extract themaximum information form the observations.• We need

– 4D-LETKF (Hunt et al, 2004) to use all the observationswithin an assimilation window at their right time– A No-Cost Smoother (Kalnay et al., 2007b)– An appropriate iterative scheme

Page 14: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

““Running in PlaceRunning in Place””• EnKF is a sequential data assimilation system where, after thenew data is used at the analysis time, it should be discarded…

• only if the previous analysis and the new background are themost likely states given the past observations.

• If the system has converged after the initial spin-up all theinformation from past observations is already included inthe background.

• During the spin-up we should use the observationsrepeatedly if we can extract extra information. But we shouldavoid overfitting the observations

Page 15: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Running in Place algorithm (1)Running in Place algorithm (1)

• Cold-start the EnKF from any initial ensemble meanand random perturbations at t0, and integrate the initialensemble to t1. The “running in place” loop with n=1, is:

Page 16: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Running in Place algorithm (2)Running in Place algorithm (2)

a) Perform a standard EnKF analysis and obtain the analysisweights at tn, saving the mean square observations minusforecast (OMF) computed by the EnKF.b) Apply the no-cost smoother to obtain the smoothed analysisensemble at tn-1 by using the same weights obtained at tn.c) Perturb the smoothed analysis ensemble with a small amountof random Gaussian perturbations, similar to additive inflation.d) Integrate the perturbed smoothed ensemble to tn. If theforecast fit to the observations is smaller than in the previousiteration according to some criterion, go to a) and performanother iteration. If not, let and proceed to the nextassimilation window.

tn!1

" tn

Page 17: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Running in Place algorithm (notes)Running in Place algorithm (notes)

Notes:c) Perturb the smoothed analysis ensemble with a small amount

of random Gaussian perturbations, a method similar toadditive inflation.

This perturbation has two purposes:1) Avoid reaching the same analysis as before, and2) Encourage the ensemble to explore new unstable directions

d) Convergence criterion: if

with do another iteration. Otherwise go to the nextassimilation window.

OMF2(iter) !OMF

2(iter +1)

OMF2(iter)

> "

! ! 5%

Page 18: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Results with a QG modelResults with a QG model

Spin-up depends on initial perturbations, but RIP works welleven with random perturbations.It becomes as fast as 4D-Var(blue). RIP takes only 2-4 iterations.

Page 19: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Results with a QG modelResults with a QG model

LETKF spin-up from random perturbations: 141 cycles. With RIP: 46 cyclesLETKF spin-up from 3D-Var perts. 54 cycles. With RIP: 37 cycles4D-Var spin-up using 3D-Var prior: 54 cycles.

LETKF

Random initial ensemble

LETKF

B3DV initial ensemble

LETKF,

Random initial

ensemble

Variational

No RIP With RIP No RIP With RIP Fixed 10

iterations RIP

3D-Var

B3DV

4D-Var

0.05B3DV

Spin-up:

DA cycles

to reach

5% error

141 46 54 37 37 44 54

RMS error

(x10-2

) 0.5 0.54 0.5 0.52 1.16 1.24 0.54

Page 20: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Discussion of spin-up accelerationDiscussion of spin-up acceleration

• Number of iterations during spin-up: 2-4,computationally acceptable

• We could use the weights interpolation of Yang etal. (2008b) and run in place only where “the actionis”.

• There are many applications where a fast spin-upis important.

• It could also help to handle the initial bias in biasedsystems, a very difficult problem (Miyoshi, pers.comm.)

Page 21: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Nonlinearities and Nonlinearities and ““outer loopouter loop””

• The main disadvantage of EnKF is that it cannot handlenonlinear (non-Gaussian) perturbations and therefore needsshort assimilation windows.

•• It doesnIt doesn’’t have the important outer loopt have the important outer loop so important in 3D-so important in 3D-Var and 4D-Var (DaSilva, pers. Var and 4D-Var (DaSilva, pers. commcomm. 2006). 2006)

Lorenz -3 variable model (Kalnay et al. 2007a Tellus), RMSanalysis error

4D-Var LETKFWindow=8 steps 0.31 0.30 (linear window)Window=25 steps 0.53 0.66 (nonlinear window)

Long windows + Pires et al. => 4D-Var wins!

Page 22: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

““Outer loopOuter loop”” in 4D-Var in 4D-Var

Page 23: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Comparison of ensemble-based and variational-baseddata assimilation schemes in a Quasi-Geostrophic model.

3D-VarHybrid (3DVar+20 BVs)

12-hour 4D-VarLETKF (40 ensemble)24-hour 4D-Var

EnKF does not handle well long windows because ensembleperturbations become non-Gaussian. 4D-Var simply iteratesand produces a more accurate control. We can imitate thiswith the “outer loop” idea for LETKF.

Page 24: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Nonlinearities and Nonlinearities and ““outer loopouter loop””Outer loop: do the same as 4D-Var, and use the final

weights to correct only the mean initial analysis,keeping the initial perturbations. Repeat the analysisonce or twice. It centers the ensemble on a moreaccurate nonlinear solution.

Miyoshi pointed out that Jaszwinski (1970) suggested this in a footnote!!!!!

Lorenz -3 variable model RMS analysis error

4D-Var LETKF LETKF+outer loop

Window=8 steps 0.31 0.30 0.27Window=25 steps 0.53 0.66 0.48

Running in place further reduces RMS from 0.48 to 0.39!

Page 25: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Estimation of observation impactwithout adjoint in an ensemble

Kalman filter

Junjie Liu and Eugenia Kalnay

Page 26: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Background

The adjoint method proposed by Langland and Baker (2004) and Zhu andGelaro (2007) quantifies the reduction in forecast error for each individualobservation source

The adjoint method detects the observations which make the forecast worse.

The adjoint method requires adjoint model which is difficult to get.

AIRS shortwave 4.180 µm

AIRS shortwave 4.474 µm

AIRS longwave 14-13 µm

AMSU/A

Page 27: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Objective and outline

Objective

Propose an ensemble sensitivity method to calculate observationimpact without using adjoint model.

Outline

Illustrate and derive the ensemble sensitivity method;

With Lorenz-40 variable model, compare the ensemble sensitivitymethod with adjoint method in

the ability to represent the actual error reduction;

the ability to detect the poor quality observations.

Summary

Page 28: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Schematic of the observation impact on the reduction offorecast error

The only difference between and is the assimilation of observations at 00hr.

Observation impact on the reduction of forecast error:

(Adapted from Langlandand Baker, 2004)

et |0 = xt |0f! xt

a

et |!6 = xt |!6f

! xta

et |0

et |!6

J =1

2(e

t |0

Tet |0! e

t |!6

Tet |!6)

analysis t

et |!6

et |0

-6hr 00hr

OBS.

Page 29: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

J =1

2(e

t |0

Tet |0! e

t |!6

Tet |!6)Euclidian cost function:

The ensemble sensitivity method

v0= y

0

o! h(x

0|!6

b)

J = v0,!J

!v0

Cost function as function of obs. increments:

Page 30: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

J =1

2(e

t |0

Tet |0! e

t |!6

Tet |!6)Euclidian cost function:

The ensemble sensitivity method

!J

!v0

= !K0

TXt |"6

fT#$ %& et |"6 + Xt |"6f !K

0v0

#$ %&

The sensitivity of cost function with respect to the assimilated observations:

v0= y

0

o! h(x

0|!6

b)

J = v0,!J

!v0

Cost function as function of obs. Increments:

With this formula we can predict the impact of observations on the forecasts!

Page 31: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Ability to detect the poor quality observation

Like adjoint method, ensemble sensitivity method can detect the observationpoor quality (11th observation location)

The ensemble sensitivity method has a stronger signal when the observation hasnegative impact on the forecast.

Observation impact from LB (red) and from ensemble sensitivity method (green)

Larger random error Biased observation case

Page 32: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Summary for forecast sensitivity to obs.

• Derived a formula to calculate the observation impact based on theensemble without using the adjoint model which usually is not available.

• The results based on Lorenz-40 variable model show that ensemblesensitivity method without using adjoint model gives results similar to adjointmethod .

• Like adjoint method, ensemble sensitivity method can detect theobservation which either has larger random error or has bias. Under suchconditions, the ensemble sensitivity method has stronger and more accuratesignal.

• It provides a powerful tool to check the quality of the observations.

Page 33: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

• In EnKF the analysis is a weighted average of the forecast ensemble• We performed experiments with a QG model interpolating weights

compared to analysis increments.• Coarse grids of 11%, 4% and 2% interpolated analysis points.

1/(3x3)=11% analysis grid

Coarse analysis with interpolated weightsYang et al (2008)

Page 34: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

• Weights vary on very large scales: they interpolate well.• Interpolated weights are obtained even for data void areas.

Coarse analysis with interpolated weights

Page 35: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Analysis increments

With increment interpolation, the analysis is OK only with 50%analysis coverage

With weight interpolation, there is almost no degradation!EnKF maintains balance and conservation properties

Page 36: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Impact of coarse analysis on accuracy

With increment interpolation, the analysis degradesWith weight interpolation, there is no degradation,

the analysis is actually better!

Page 37: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Model error: comparison ofModel error: comparison of methodsmethodsto correct model bias and inflationto correct model bias and inflation

Hong Li, Chris Danforth, Takemasa Miyoshi,and Eugenia Kalnay

Page 38: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Model error: If we assume a perfect model in EnKF,Model error: If we assume a perfect model in EnKF,we underestimate the analysis errors (Li, 2007)we underestimate the analysis errors (Li, 2007)

imperfect modelimperfect model(obs from NCEP- NCAR(obs from NCEP- NCARReanalysis NNR)Reanalysis NNR)

perfect SPEEDY modelperfect SPEEDY model

Page 39: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

— Why is EnKF vulnerable to model errors ?

In the theory of Extended Kalmanfilter, forecast error is represented bythe growth of errors in IC and themodel errors.

However, in ensemble Kalman filter,error estimated by the ensemblespread can only represent the firsttype of errors.

Pif!

1

k "1(xi

f

i=1

K

# " xf)(xi

f" x

f)T

1 11a a

i i

f a T

i i! !

!= +

x xP M P M Q

Ens mean

Ens 1

Ens 2

Truth

Imperfect model

Forecasterror

The ensemble spread is ‘blind’to model errors

Page 40: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

imperfect modelperfect model

Low Dimensional Method to correct the bias (Danforth et al, 2007)combined with additive inflation

We compared several methods to handlebias and random model errors

Page 41: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

Discussion: 4D-Var Discussion: 4D-Var vsvs. EnKF . EnKF ““warwar””• We should be able to adopt some simple strategies to

capture the advantages of 4D-Var:– Smoothing and running in place– A simple outer loop to deal with nonlinearities– Adjoint sensitivity without adjoint model– Coarse resolution analysis without degradation– Correcting model bias (Baek et al, 2006, Danforth et al, 2007, Li et al.

submitted).– The correction of model bias combined with additive inflation gives the

best results

• It seems like there is nothing that 4D-Var can do that EnKFcannot do as well, usually simpler, cheaper and better.

• There is a WMO Workshop in Buenos Aires this November on “4D-Varand EnKF intercomparisons” and a 2-week intensive course on dataassimilation for Latin American scientists (140 applications!)

Page 42: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

• Generate a long time series of model forecast minus reanalysisfrom the training period

2.3 Low-dim method (Danforth et al, 2007: Estimating and correcting globalweather model error. Mon. Wea. Rev, J. Atmos. Sci., 2007)

t=0t=6hr

model

fx

xtruth

Bias removal schemes (Low Dimensional Method)

e

hrx6

We collect a large number of estimated errors and estimate from them bias, etc.

NNR

NNR

NNR

Time-meanmodel bias

Diurnalmodel error

Statedependentmodel error

m

M

m

mnl

L

l

ln

t

n

a

n

t

n

f

n

f

n MM febxxxx !!==

++++++"="=

1

,

1

,111 )()( #$%

fx

Forecast errordue to error in IC

NNRNNR

Page 43: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

model error = b + !n,l

l=1

2

" el+ #

n,m

m=1

10

" fm

Low-dimensional method

Include Bias, Diurnal and State-Dependent model errors:

Having a large number of estimated errors allows toestimate the global model error beyond the bias

Page 44: Some ideas for Ensemble Kalman Filter4dvarenkf.cima.fcen.uba.ar/course/download/Kalnay_8_Revision.pdfKalman Filter Basic idea: • We are in competition with 4D-Var… • We take

SPEEDY 6 hr model errors against NNR (diurnal cycle)

1987 Jan 1~ Feb 15

Error anomalies

• For temperature at lower-levels, in additionto the time-independent bias, SPEEDY hasdiurnal cycle errors because it lacks diurnalradiation forcing

Leading EOFs for 925 mb TEMP

pc1pc2

e

hr

e

hri

e

hrxxx 66

'

)(6 !=


Recommended