+ All Categories
Home > Documents > Comparison of Ensemble Data Assimilation methods for the ...

Comparison of Ensemble Data Assimilation methods for the ...

Date post: 18-Dec-2021
Category:
Upload: others
View: 2 times
Download: 0 times
Share this document with a friend
31
Sequential filtering for the shallow water equations model 1 Comparison of Ensemble Data Assimilation methods for the shallow water equations model in the presence of nonlinear observation operator M. Jardak, a * I. M. Navon b and M. Zupanski c a Laboratoire de M´ et´ orologie Dynamique/CNRS ´ Ecole Normale Sup´ erieure, Paris, France b Department of Scientific Computing, Florida State University, Tallahassee, FL 32306-4120, USA c Cooperative Institute for Research in the Atmosphere, Colorado State University, 1375 Campus Deliver, Fort Collins, CO 80523-1375, USA * Correspondence to: M.Jardak, Laboratoire de M´ et´ orologie Dynamique/CNRS ´ Ecole Normale Sup´ erieure, Paris, France, E-mail:[email protected] A new comparison of three frequently used sequential data assimilation methods illuminating their strengths and weaknesses in the presence of linear and nonlinear observation operators is presented. The ensemble Kalman filter (EnKF), the particle filter (PF) and the maximum likelihood ensemble filter (MLEF) methods were implemented and the spectral shallow water equations model in spherical geometry model was employed using the Rossby-Haurwitz Wave no. 4 test case as initial condition. Numerical tests conducted reveal that all three methods perform satisfactory in the presence of linear observation operator for 10 to 15 days model integration, whereas the EnKF, even with the Evensen fixture [Evensen 03] for the nonlinear observation operator failed in all tested metrics. The particle filter and the hybrid filter MLEF both performed satisfactorily in the presence of highly nonlinear observation operators with a slight advantage in terms of CPU time to the MLEF method. Copyright c 2010 Royal Meteorological Society Key Words: Data Assimilation, EnKF, PF, MLEF, shallow water equations Received . . . Citation: ... Copyright c 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 223 (2010) Prepared using qjrms4.cls
Transcript

Sequential filtering for the shallow water equations model 1

Comparison of Ensemble Data Assimilation methods for the

shallow water equations model in the presence of nonlinear

observation operator

M. Jardak,a ∗I. M. Navon b and M. Zupanski c

a Laboratoire de Metorologie Dynamique/CNRS Ecole Normale Superieure, Paris, France

b Department of Scientific Computing, Florida State University, Tallahassee, FL 32306-4120, USA

c Cooperative Institute for Research in the Atmosphere, Colorado State University, 1375 Campus Deliver, Fort Collins,

CO 80523-1375, USA

∗Correspondence to: M.Jardak, Laboratoire de Metorologie Dynamique/CNRS Ecole Normale Superieure, Paris, France,

E-mail:[email protected]

A new comparison of three frequently used sequential data assimilation

methods illuminating their strengths and weaknesses in the presence of linear

and nonlinear observation operators is presented. The ensemble Kalman filter

(EnKF), the particle filter (PF) and the maximum likelihood ensemble filter

(MLEF) methods were implemented and the spectral shallow water equations

model in spherical geometry model was employed using the Rossby-Haurwitz

Wave no. 4 test case as initial condition. Numerical tests conducted reveal that

all three methods perform satisfactory in the presence of linear observation

operator for 10 to 15 days model integration, whereas the EnKF, even with the

Evensen fixture [Evensen 03] for the nonlinear observation operator failed in all

tested metrics. The particle filter and the hybrid filter MLEF both performed

satisfactorily in the presence of highly nonlinear observation operators with a

slight advantage in terms of CPU time to the MLEF method. Copyright c© 2010

Royal Meteorological Society

Key Words: Data Assimilation, EnKF, PF, MLEF, shallow water equations

Received . . .

Citation: . . .Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Quarterly Journal of the Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

1. Introduction

Sequential data assimilation fuses observations of the

current (and possibly, past) state of a system with

predictions from a mathematical model (the forecast) to

produce an analysis, providing ”the best” estimate of the

current state of the system. Central to the concept of

sequential estimation data assimilation is the propagation

of flow dependent probability density function (pdf) given

an estimate of the initial pdf.

In sequential estimation, the analysis and forecasts

can be viewed as probability distributions. The analysis

step is an application of the Bayes theorem. Advancing

the probability distribution in time, for the general case is

done by the Chapman-Kolmogorov equation, but since it is

unrealistically expensive, various approximations operating

on representations of the probability distributions are

used instead. If the probability distributions are normal,

they can be represented by their mean and covariance,

which gives rise to the Kalman filter (KF). However,

due to the high computational and storage overheads

it requires, an approximation based on Monte-Carlo

ensemble calculations has been proposed by [Evensen

1994], [Evensen and Van Leeuwen 1996], [Burgers et al.

1998] and [Houtekamer and Mitchell 1998]. The method

is essentially a Monte-Carlo approximation of the Kalman

filter which avoids evolving the covariance matrix of the

state vector. A second type of EnKF filter consists of the

class of square root filters of [Anderson and Anderson 2003]

see also [Bishop et al. 2001].

Variants of the EnKF such as Kalman ensemble square

root filters (KSRF), the singular evolutive extended Kalman

(SEEK) filter, and the less common singular evolutive

interpolated Kalman (SEIK) filter can be found in the paper

of [Tippett et al. 2003], see also the paper of [Nerger et al.

2005] where the aforementioned filters were reviewed and

compared.

The PF methods,also known as sequential Monte-

Carlo (SMC) methods or Bayesian filters. The SMC

methods are an efficient means for tracking and forecasting

dynamical systems subject to both process and observation

noise. Applications include robot tracking, video or audio

analysis, and general time series analysis see [Doucet et

al. 2001]. These methods utilize a large number N of

random samples named particles to represent the posterior

probability distributions. The particles are propagated

over time using a combination of sequential importance

sampling and resampling steps. Resampling for PF is used

to avoid the problem of degeneracy of this algorithm that is,

avoiding the situation where all but one of the importance

weights are close to zero. The performance of the PF

algorithm can be crucially affected by a judicious choice

of a resampling method. See [Arulampalam et al. 2002]

for a listing of the most used resampling algorithms. The

PF suffers from ”the curse of dimensionality” requiring

computations that increase exponentially with dimension

as pointed out by Silverman [Silverman 1986]. This

argument was enhanced and amplified by the recent work of

[Bengtsson et al. 2008] and [Bickel et al. 2008] and finally

explicitly quantified by [Synder et al. 2008]. They indicated

that unless the ensemble size is greater than exp(τ2/2),

where τ2 is the variance of the observation log-likelihood,

the PF update suffers from a ”collapse” in which with

high probability only few members are assigned a posterior

weight close to one while all other members have vanishing

small weights. This issue becomes more acute as we move

to higher spatial dimensions.

The MLEF filter of [Zupanski 2005]; [Zupanski and

Zupanski 2006] is a hybrid between the 3-D variational

method and the EnKF. It maximizes the likelihood of

a posterior probability distribution, thus its name. The

MLEF belongs to the class of deterministic ensemble

filters, since no perturbed observations are employed. As

in variational and ensemble data assimilation methods, the

cost function is derived using Gaussian probability density

function framework. Like other ensemble data assimilation

algorithms, the MLEF produces an estimate of the analysis

uncertainty ( e.g. analysis error covariance)

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 3

In this paper data assimilation experiments are

performed and compared using the ensemble Kalman filter

EnKF, the particle filter PF and the Maximum Likelihood

Ensemble Filter MLEF. These methods were tested on a

spectral shallow water equations model using the Rossby-

Haurwitz wave no 4 test case for both linear and nonlinear

observation operators.

The spectral shallow water equations model in

spherical geometry of Williamson [Williamson 1992] and

[Jakob et al.1995] is used to generate the true solution

and forecasts. To improve EnKF analysis errors and

avoid ensemble errors that generate spurious corrections,

a covariance localization investigated by [Houtekamer and

Mitchell 1998,2001] is incorporated. To improve both the

analysis and the forecast results, the forecast ensemble

solutions are inflated from the mean as suggested in

[Anderson and Anderson 1999] and reported in [Hamill et

al. 2001].

Since the resampling is a crucial step for (PF) method,

the systematic, multinomial and the residual resampling

methods [Arulampalam et al. 2002],[Doucet et al. 2001] and

[Nakano and al. 2007] were tested.

The paper is structured as follows. Section 2

presents the spectral shallow-water equations in spherical

coordinates model and the numerical methods used for its

resolution. In section 3 we present algorithmic details of

each of the data assimilation methods considered. In section

4 we present the numerical results obtained and discuss

them for both the linear and nonlinear observation operators

in several commonly used error metrics. In particular the

Talagrand diagrams ( skill or score test) see [Wilks 2005]

, the early work of Murphy [ Murphy 1971 and 1993] and

[Talagrand et al. 1997 ] and the interpretation of [Hamill

2001], and the root mean square error (rmse) are employed

to detect ensemble members spread, while the rmses are

used to track filter divergence [Houtekamer 2005]

Finally section 5 is reserved for summary and

conclusions.

2. Shallow-Water equations in spherical geometry

The shallow water equations are a set of hyperbolic

partial differential equations that describe the flow below

a pressure surface in a fluid.

The equations are derived from depth-integrating the

Navier-Stokes equations, they rely primarily on the

assumptions of constant density and hydrostatic balance.

The shallow-water equations in spherical geometry are

given by

∂u

∂t+

u

a cos θ

∂u

∂λ+v

a

∂u

∂θ− tan θ

avu− fv = − g

a cos θ

∂h

∂λ

∂v

∂t+

u

a cos θ

∂v

∂λ+v

a

∂v

∂θ+

tan θ

au2 + fu = −g

a

∂h

∂θ

∂h

∂t+

u

a cos θ

∂h

∂λ+v

a

∂h

∂θ+

h

a cos θ

[∂u

∂λ+∂(cos θ)

∂θ

]= 0

where V = u~i+ v~j is the horizontal velocity vector

( with respect to the surface of the sphere), gh is the free

surface geopotential, h is the free surface height, g is the

gravity acceleration. f = 2Ω sin θ is the Coriolis parameter,

Ω is the angular velocity of the earth. θ denotes the angle of

latitude, µ = sin θ is the longitude. λ the longitude,and a is

the radius of the earth.

One of the major advances in meteorology was the use

by [Rossby 1939] of the barotropic vorticity equation with

the β -plane approximation to the spherical of the Earth, and

the deduction of solutions reminiscent of some large scale

waves in the atmosphere. These solutions have become

known as Rossby waves see [Haurwitz 1940 ] While the

shallow water do not have corresponding analytic solutions

they are expected to evolve in a similar way as the above R-

H equations which explains why they have been widely used

to test shallow water numerical models since the seminal

paper of [Phillips 1959]. Following the research work of

[Hoskins 1973] Rossby -Haurwitz waves with zonal wave

numbers less or equal to 5 are believed to be stable. This

makes the R-H zonal wave no 4 a suitable candidate for

assessing accuracy of numerical schemes as was evident

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

4 M.Jardak et al.

from its being chosen as a test case by [Williamson et al.

1992] and by a multitude of other authors. It has been

numerically shown that the R-H wave no 4 breaks down

into more turbulent behavior after long term numerical

integration as recently discovered by [Thuburn and Li 2000]

and also by [Smith and Dritschel 2006].

The initial velocity field for the Rossby-Haurwitz wave

is defined as

u = aω cosφ+ aK cosr−1 φ(r sin2 φ− cos2 φ)

cos(rλ)

v = −aKr cosr−1 φ sinφ sin(rλ)

(1)

The initial height field is defined as,

h = h0 +a2

g[A(φ) +B(φ) cos(rλ) + C(φ) cos(2rλ)]

(2)

where the variables A(φ), B(φ), C(φ) are given by

A(φ) =ω

2(2Ω + ω) cos2 φ+

1

4k2

cos2r φ[(r + 1) cos2 φ+ (2r2 − 2r − 2)− 2r2 cos2 φ]

B(φ) =2(Ω + ω)k

(r + 1)(r + 2)

cosr φ[(r2 + 2r + 2)− (r + 1)2 cos2 φ]

C(φ) =1

4k2 cos2r φ[(r + 1) cos2 φ− (r + 2)]

In here, r represents the wave number,h0 is the height at the

poles. The strength of the underlying zonal wind from west

to east is given by ω and k controls the amplitude of the

wave.

3. Sequential Bayesian Filter- theoretical setting

This section applies to the three sequential data assimilation

methods discussed herein. One also can see the survey of

Diard [Diard et al. 2003] and the overview of [Clappe et al.

2007]. The sequential Bayesian filter is a large numberN of

random samples advanced in time by a stochastic evolution

equation, to approximate the probability densities. In order

to analyze and make inference about the dynamic system at

least a model equation along with an observation operator

are required. Generically, stochastic filtering problem is a

dynamic system that assumes the form

xt = f(t,xt−1,vt)

yt = h(t,xt,nt)

(3)

The first equation of (3) is the state equation or the system

model, the second represents the observation equation.

The vectors xt and yt are respectively the state and the

observation vectors. The state and observation noises are

represented by vt and nt respectively. The discrete-time

counterpart of the system (3) reads

xk = f(xk−1,vk−1)

yk = h(xk,nk)

(4)

The deterministic mapping fk : Rnx × Rnd −→ Rnx is a

possibly non-linear function of the state xk−1, vk, k ∈ N

is an independent identically distributed (i.i.d) process noise

sequence, nx, nd are dimensions of the state and the process

noise vectors, respectively. Likewise, the deterministic

mapping hk : Rnx × Rnn −→ Rnz is a possibly non-

linear function, nk, k ∈ N is an i.i.d. observation noise

sequence, and nx, nn are dimensions of the state and

observation noise vectors, respectively.

Let y1:k denote the set of all available observations

yi up to time t = k, (i.e. y1:k = yi|i = 1, · · · , k). From

a Bayesian point of view, the problem is to recursively

calculate some degree of belief in the state xk at time

t = k, taking different values, given the data y1:k up to

the time t = k. Equivalently, the Bayesian solution would

be to calculate the pdf p(xk|y1:k). This density will

encapsulate all the information about the state vector xk

that is contained in the observations y1:k and the prior

distribution for xk.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 5

Assume the required pdf p(xk−1|y1:k−1) at time k − 1 is

available. The prediction stage uses the state equation (4) to

obtain the prior pdf of the state variable at time k via the

Chapman-Kolmogorov equation

p(xk|y1:k−1) =

∫p(xk|xk−1)p(xk−1|y1:k−1)dxk−1,

(5)

where the probabilistic model of the state evolution,

p(xk|xk−1) is readily obtained from the state equation (4)

and the known statistics of vk−1.

As an observation yk becomes available at time t = k,

the prior pdf could be updated via the Bayes rule

p(xk|y1:k) =p(zk|xk)p(xk|y1:k−1)

p(yk|y1:k−1), (6)

where the normalizing constant

p(yk|y1:k−1) =

∫p(yk|xk)p(xk|y1:k−1)dxk. (7)

depends on the likelihood function p(yk|xk), defined by

the measurement or observation equation (4) and the known

statistics of nk.

The relations (5) and (6) form the basis for the

optimal Bayesian solution. This recursive propagation of the

posterior density is only a conceptual solution. One cannot

generally obtain an analytical solution. Solutions exist only

in a very restrictive set of cases like that of the Kalman

filters for instance (namely, if fk and hk are linear and both

vk and nk are Gaussian). We turn now to describe in detail

each of the three used filters.

3.1. Particle Filters

Particle filters see [Doucet et al. 2000, Doucet et al. 2001,

Arulampalam et al. 2002 , Berliner and Wikle 2007A]

and recently the review paper of [Van Leeuwen 2009]

approximate the posterior densities by population of states.

These states are called ”particles”. Each of the particles

has an assigned weight, and the posterior distribution can

then be approximated by a discrete distribution which has

support on each of the particles. The probability assigned to

each particle is proportional to its weight. See for instance

[Metropolis and Ulam 1944], [Gordon et al., 1993] [Doucet

et al. 2001] and [Stuart 2010].

The different (PF) algorithms differ in the way that the

population of particles evolves and assimilates the incoming

observations. A major drawback of particle filters is that

they suffer from sample degeneracy after a few filtering

steps.

In this 2-D plus time problem the common remedy

is to resample the prior pdf whenever the weights focus

on few members of the ensemble. Here we use several

strategies such as Systematic Resampling (SR), Residual

Resampling (RR) and or the Bayesian bootstrap filter of

Gordon et al. [Gordon 1993] see also [Berliner and Wikle

2007A and 2007B]. Multinomial Resampling (MR). The

SR algorithm generates a population of equally weighted

particles to approximate the posterior at some time k. This

population of particles is assumed to be an approximate

sample from the true posterior at that time instant.

The PF algorithm proceeds as follows:

• Initialization: The filter is initialized by drawing a

sample of size N from the prior pdf.

• Filtering:

Preliminaries: Assume that xik−1i=1,··· ,N is a

population of N particles, approximately distributed

as in an independent sample from p(xk−1|y1:k−1)

Prediction: Sample N values, q1k, · · · , qNk ,

from the distribution of vk. Use these to

generate a new population of particles,

x1k|k−1,x

2k|k−1, · · · ,x

Nk|k−1 via the equation

xik|k−1 = fk(xik−1,vik) (8)

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

6 M.Jardak et al.

Filtering: Assign each xik|k−1, a weight qik. This

weight is calculated by

qik =p(yk|xik|k−1)∑Nj=1 p(yk|x

jk|k−1)

(9)

This defines a discrete distribution which, for

i ∈ 1, 2, · · · , N, assigns probability mass qik to

element xik|k−1

• Resampling: Resample independently N times, with

replacement, from the distribution obtained in the

filtering stage. The resulting particles, xiki=1,··· ,N ,

form an approximate sample from p(xk|y1:k) .

The method outlined above can be justified as follows.

If the particles at time t = k − 1 were an i.i.d sample

from the posterior at time t = k − 1, then the predictive

stage just produces an i.i.d. sample from the prior at time

t = k. The filtering stage can be viewed as an importance

sampling approach to generate an empirical distribution

which approximates the posterior.

The proposal density is just the prior p(xk|y1:k−1),

and as a result of Bayes formula, we obtain

p(xk|y1:k−1,yk) ∝ p(xk|y1:k−1)p(yk|xk), (10)

the weights are proportional to the likelihood p(yk|xk).

As N tends to infinity, the discrete distribution which

has probability mass qi at point xik|k−1, converges weakly

to the true posterior. The resampling step is crucial to

particle filters. It is used to generate equally weighted

particles aimed at avoiding the problem of degeneracy of

the algorithm, that is, avoiding the situation that all but

one of the weights are close to zero. The resampling step

modifies the weighted approximate density p(xk|yk) to

an unweighted density p(xk|yk) by eliminating particles

having low importance weights and by multiplying particles

having highly importance weights. Formally:

p(xk|yk) =

N∑i=1

qiδ(xk − xki) (11)

is replaced by

p(xk|yk) =

N∑i=1

1

Nδ(xk − xk

?) =

N∑i=1

niNδ(xk − xk

i)

(12)

where ni is the number of copies of particle xki in the

new set of particles xk?. For extensive review of particle

filtering see[Van Leeuwen 09].

3.2. The Ensemble Kalman Filter

The ensemble Kalman filter (EnKF) was first proposed by

Evensen [Evensen 1994] and further developed by [Burgers

et al. 1998] and [Evensen 2003, Evensen 2007]. For the

current status and the potential of the EnKF we refer to

[Kalnay 2009]. It is related to particle filters in the context

that a particle is identical to an ensemble member. EnKF

is a sequential filter method, which means that the model

is integrated forward in time and, whenever observations

are available, these are used to reinitialize the model

before the integration continues. The EnKF originated as a

version of the Extended Kalman Filter (EKF) of Jazwinski

[Jazwinski 1970] and Bucy[Bucy 1965] for large problems.

The classical KF [Kalman 60] method is optimal in the

sense of minimizing the variance only for linear systems

and Gaussian statistics.

Similar to the particle filter method, the EnKF stems

from a Monte Carlo integration of the Fokker-Planck

equation governing the evolution of the pdf that describes

the prior, forecast, and error statistics. In the analysis step,

each ensemble member is updated according to the KF

scheme and replaces the covariance matrix by the sample

covariance computed from the ensemble. However, the

EnKF presents two potential problems namely:

1) Even though the EnKF uses full non-linear dynamics to

propagate the forecast error statistics, the EnKF assumes

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 7

that all probability distributions involved are Gaussian.

2) The updated ensemble preserves only the first two

moments of the posterior.

Let p(x) denote the Gaussian prior probability density

distribution of the state vector x with mean µ and covariance

Q

p(x) ∝ exp(−1

2(x− µ)TQ−1(x− µ)

)We assume the data y to have a Gaussian pdf with

covariance R and mean Hx, where H is the so-called the

observation matrix, is related to h of equation (4), and where

the value Hx assumes what the value of the data y would

be in absence of observation errors. Then p(y|x) is the

probability (likelihood) of the observations y given that the

truth is given by the model state x and is of the form

p(y|x) ∝ exp

(−1

2(y −Hx)TR−1(y −Hx)

).

According to the Bayes theorem the posterior probability

density follows from the relation

p(x|y) ∝ p(y|x)p(x). (13)

There are many variants of implementing the EnKF

of various computational efficiency and in what follows

we employ standard formulation of the EnKF for linear

and nonlinear observation operators with covariance

localization. See [Evensen 1994, Burgers et al. 1998,

Mandel 2006, Mandel 2007 and Lewis et al. 2006], also see

[Nerger et al. 2005] and Sakov and Oke [Sakov 2008] The

implementation of the standard EnKF may be divided into

three steps, as follows:

• Setting and matching

Define the ensemble

X = [x1, · · · ,xN ] (14)

be an nx ×N matrix whose columns are a sample

from the prior distribution. N being the number of

the ensemble members.

Form the ensemble mean

X = X · 1N , (15)

where 1N ∈ RN×N is the matrix where each element

is equal to 1.

Define the ensemble perturbation matrixX ′ and set

the Rnx×nx ensemble covariance matrix C

X ′ = X − 1

NX , (16)

C =X ′X ′T

N − 1, (17)

• Sampling

Generate

Y = [y1, · · · ,yN ] (18)

be an ny ×N matrix whose columns are a replica of

the measurement vector y plus a random vector from

the normal distribution N (0,R).

Form the Rny×ny measurement error covariance

R =YYt

N − 1, (19)

• Updating Obtain the posterior X p by the linear

combinations of members of the prior ensemble

X p = X + CHT (HCHT +R)−1(Y −HX ). (20)

The matrix

K = CHT (HCHT +R)−1 (21)

is the Kalman gain matrix. Since R is always positive

definite( i.e. covariance matrix), the inverse (HCHT +

R)−1 exists.

An easy computation shows that the mean and covariance

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

8 M.Jardak et al.

of the posterior or updated ensemble are given by

X p = X p +K [y − (HX p + d)] , (22)

with the vector d results from the affine measurement

relation

h(x) = Hx + d. (23)

The covariance of the posterior obeys

Cp = C − K[HCHT +R

]KT , (24)

In the case of nonlinear observation operators, a

modification to the above algorithm is advised. As advised

by Evensen [Evensen 2003] and following his notation, let

x the augmented state vector made of the state vector and

the predicted observation vector (nonlinear in this case).

x =

x

H(x)

. (25)

Define the linear observation operator H by

H

x

y

= y (26)

and carry out the steps of the EnKF formulation in

augmented state space x and H instead of x and H.

Superficially, this technique appears to reduce the nonlinear

problem to the previous linear observation operator case.

However, whilst the augmented problem, involving linear

observation problem, is a reasonable way of formulating the

EnKF, it is not as well-founded as the linear case, which can

be justified as an approximation to the exact and optimal KF.

To prevent the occurrence of filter divergence usually

due to the background-error covariance estimates from

small number of ensemble members as pointed out in

[Houtekamer and Mitchell 1998], the use of covariance

localization was suggested. Mathematically, the covariance

localization increases the effective rank of the background

error covariances. See the work of [ Gaspari and

Cohn 1999] also [Hamill and Snyder 2000, 2006] and

[Ehrendorfer 2007]. The covariance localization consists of

multiplying point by point the covariance estimate from

the ensemble with a correlation function that is 1.0 at

the observation location and zero beyond some prescribed

distance. Mathematically, to apply covariance localization,

the Kalman gain

K = CHT (HCHT +R)−1

is replaced by a modified gain

K = [ρ C]HT (H [ρ C]HT +R)−1 (27)

where ρ denotes the Schur product ( The Schur product

of matrices A and B is a matrix D of the same dimension,

where dij = aijbij) of a matrix S with local support with

the covariance model generated by the ensemble. Various

correlation matrices have been used. For the present model,

we used the usual Gaussian correlation function

ρ(D) = exp[−[D

l]2], (28)

where l is the correlation length here l = 200 length units.

In addition, the additive covariance inflation of [Anderson

and Anderson 1999] with the inflation factor r = 1.001 has

been employed.

3.3. The Maximum Likelihood Ensemble Filter

The Maximum Likelihood Ensemble Filter (MLEF)

proposed by Zupanski [Zupanski 2005], and Zupanski and

Zupanski [Zupanski 2006] is a hybrid filter combining

the 3-D variational method with the EnKF. It maximizes

the likelihood of posterior probability distribution which

justifies its name. The method comprises three steps, a

forecast step that is concerned with the evolution of the

forecast error covariances, an analysis step based on solving

a non-linear cost function and an updating step.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 9

• Forecasting:

It consists of evolving the square root analysis

error covariance matrix through the ensembles. The

starting point is from the evolution equation of

the discrete Kalman filter described in Jazwinski

[Jazwinski 1970]

P kf =Mk−1,kPkaMT

k−1,k +Qk−1, (29)

where P (k)f is the forecast error covariance matrix at

time k,Mk−1,k is the linearized forecast model (e.g.,

Jacobian) from time k − 1 to time k, and Qk−1 is the

model error matrix which is assumed to be normally

distributed. Since P k−1a is positive matrix for any k,

equation (29) could be factorized and written as

P kf =

(Pkf )1/2︷ ︸︸ ︷(

Mk−1,k(P ka )1/2)(Mk−1,k(P ka )1/2

)T+Qk−1

where (P ka )1/2 is of the form

(P ka )1/2 =

pk(1,1) p

k(2,1) · · · p

k(N,1)

pk(1,2) pk(2,2) · · · p

k(N,2)

pk(1,n) pk(2,n) · · · p

k(N,n)

, (30)

as usual N is the number of ensemble members and

n the number of state variables. The lower case pki,j

are obtained by calculating the square root of (P ka ).

Using equation(30), the square root forecast error

covariance matrix (P kf )1/2 can then be expressed as

(P kf )1/2 =

bk(1,1) b

k(2,1) · · · b

k(N,1)

bk(1,2) bk(2,2) · · · b

k(N,2)

bk(1,n) bk(2,n) · · · b

k(N,n)

, (31)

where for each 1 ≤ i ≤ N

bki =

bk(i,1)

bk(i,2)

...

bk(i,n)

=Mk−1,k

xk1 + pk(i,1)

xk2 + pk(i,2)

...

xkn + pk(i,n)

Mk−1,k

xk1

xk2...

xkn

.

The vector xk =(xk1x

k2 · · ·xkn

)Tis the analysis state

from the previous assimilation cycle. which is found

from the posterior analysis pdf as presented in

[Lorenc 1986].

• Analyzing:

The analysis step for the MLEF involves solving

a non-linear minimization problem. As in Lorenc

[Lorenc 1986], the associated cost function is defined

in terms of the forecast error covariance matrix and is

given as

J (x) = 12 (x− xb)

T(P kf

)−1

(x− xb)+

12 [y − h(x)]

T R−1 [y − h(x)]

(32)

where y is the vector of observations, h is the non-

linear observation operator, R is the observational

error covariance matrix and xb is a background state

given by

xb =Mk−1,k(xk) +Qk−1. (33)

Through a Hessian preconditioner we introduce the

change of variable

(x− xb) = (P kf )1/2(I +O)−T/2ξ (34)

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

10 M.Jardak et al.

where ξ is vector of control variables, O is referred

to as the observation information matrix and I is the

identity matrix. The matrix O is provided by

O = (P kf )T/2HTR−1H(P kf )T/2 =

(R−1/2H(P kf )1/2)T (R−1/2H(P kf )1/2)

here H is the Jacobian matrix of the non-linear

observation operator h evaluated at the background

state xb.

Let Z the matrix defined by

Z =

z(1,1) z(2,1) · · · z(N,1)

z(1,2) z(2,2) · · · z(N,2)

z(1,n) z(2,n) · · · z(N,n)

,

and let zi assumes the form

zi =

z(i,1)

z(i,2)

...

z(i,n)

= R−1/2Hbki .

From the following approximations

zi ≈ R−1/2[h(x + bki )− h(x)

], (35)

and

O ≈ ZZT . (36)

one can use an eigenvalue decomposition of the

of symmetric positive definite matrix I +O to

calculate the inverse square root matrix necessary to

the updating step. It is worth mentioning that the

approximation (35) is not necessary and a derivation

of the MLEF not involving (35) has been recently

developed in Zupanski et al. [Zupanski 08]

• Updating:

The final point about MLEF is to update the square

root analysis error covariance matrix. In order to

estimate the analysis error covariance at the optimal

point, the optimal state xopt minimizing the cost

function J given by (32)is substituted

(P ka )T/2 = (P kf )T/2(I +O(xopt))−T/2. (37)

4. Numerical results

4.1. Model set-up

As in [Williamson 1992,1997,2007] and [Jakob et al. 1995],

the grid representation for any arbitrary variable φ is related

to the following spectral decomposition

φ(λ, µ) =

M∑m=−M

N(m)∑n=|m|

φm,nPm,n(µ)eimλ, (38)

where Pm,n(µ)eimλ are the spherical harmonic functions

[Boyd 01]. Pm,n(µ) stands for the Legendre polynomial.M

is the highest Fourier wavenumber included in the east-west

representation, N(m) is the highest degree of the associated

Legendre polynomials for longitudinal wavenumber m.

The coefficients of the spectral representation (38) are

determined by

φm,n =

∫ 1

−1

1

∫ 2π

0

φ(λ, µ)e−imλPm,n(µ)dλdµ (39)

The inner integral represents a Fourier transform,

φm(µ) =1

∫ 2π

0

φ(λ, µ)e−imλdλ (40)

which is evaluated using a fast Fourier transform routine.

The outer integer is evaluated using Gaussian quadrature on

the transform grid.

φm,n =

J∑j=1

φm(µ)Pm,n(µ)ωj , (41)

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 11

where µj denotes the Gaussian grid points in the meridional

direction and ωj is the Gaussian weight at point µj .

The meridional grid points are located at the Gaussian

latitudes θj , which are the J roots of the Legendre

polynomial Pj(sin θj) = 0. The number of grid points in

the longitudinal and meridional directions are determined so

as to allow the unaliased representation of quadratic terms,

I ≥ 3M + 1

J ≥ 3N + 1

2

where N is the highest wavenumber retained in the

latitudinal Legendre representation N = maxN(m) = M .

The pseudo spectral method, also known as the spectral

transform method in the geophysical community, of [Orszag

1969, 1970] and [Eliassen et al. 1970] has been used to

tackle the nonlinearity.

In conjunction with the spatial discretization described

before, the time discretization, two semi-implicit time

steps have been used for the initialization. Because of the

hyperbolic type of the shallow water equations, the centered

leapfrog scheme

φk+1m,n − φk−1

m,n

2∆t= F(φkm,n) (42)

has been invoked for the subsequent time steps. After

the leapfrog time-differencing scheme is used to obtain

the solution at t = (k + 1)∆t, a slight time smoothing is

applied to the solution at time k∆t (Asselin filter) .

φkm,n = φkm,n + α[φk+1m,n − 2φkm,n + φk−1

m,n

](43)

replacing the solution at time k. It reduces the amplitude of

different frequencies ν by a factor 1− 4α sin2(ν∆t2 ).

4.2. EnKF set-up and numerical results

In all our numerical experiments a time step of ∆t = 600

sec has been employed. The observations were provided at a

frequency consisting of one set of observations every 36∆t.

4.2.1. Impact of the number of observations and the

number of ensemble members

The impact of the number of observations follows the

findings of [ Fletcher and Zupanski 2008]. In fact after

attaining a number of observations threshold, all the root

mean square errors rmses for all the fields ( namely the

components of the velocity and geopotential) coincide. In

our test the number of the observations threshold was

attained at around 912 observations at each observation

time. The observational grid is a subset of the rhomboidal

spectral truncation (48x38) and consists of observations

being distributed at every grid point in the longitude and

every 2 points in the latitude, as presented in figure 1.

The impact of the number of ensemble members on

the results has also been examined. We have conducted

several experiments with different numbers of ensemble

members, namely, 50, 100, 200 and 300 ensemble members,

respectively. Our results reveal that employing only 100

ensemble members was sufficient to successfully perform

the data assimilation and that using higher number of

ensemble members had no impact on the ensuing results.

4.2.2. Linear observation operator case - description of

the EnKF results

Figure 2 provides an overview of the unperturbed shallow

water solution using the Rossby-Haurwitz wave no.4 as

initial conditions after 10 days of time integration.

In figure 3 we present a 1% random perturbation

around the mean of the geopotential field. This perturbed

field along with the perturbed velocity divergence and vor-

ticity fields will serve as initialization for the observations.

In figure 4 we present an overview of the results

obtained using 10 days EnKF data assimilation time

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

12 M.Jardak et al.

integration. The linear observation operator H(u) = u has

been employed, 100 ensemble members were used with 1%

random perturbation around the mean accounting for the

standard deviation for the model error covariance matrix.

In the presence of 1% and 3% random perturbations

around the mean respectively, the RH wave no. 4 test

case preserves its global characteristic shape for more than

30 days of data assimilation time integration. This shows

the resilience of the RH wave no. 4 test case to random

perturbation in the presence of linear observation operator.

4.2.3. Nonlinear observation operator case - description

of the EnKF results

Here we are employing H(u) = u2 as a representative of

a nonlinear observation operator ( we have also tested

H(u) = u4 results not shown) The number of observation

points is 912 ( their distribution in space and time being

identical to the linear observation operator EnKF case).

The EnKF for the nonlinear observation operator H(u) =

u2 starts to manifest a deterioration at a round 5 days

of data assimilation time integration. By 10 days of data

assimilation time integration a full filter divergence takes

place as it can be seen in the rmses as well as in the

corresponding Talagrand diagram. Figure 5 displays the

corresponding Talagrand diagram also known as the rank

score. Contrary to the linear observation operator Talagrand

diagram where a uniform ensemble repartition is observed,

the Talagrand diagram for the nonlinear case displays a

tendency to a weakly U-shaped ensemble repartition. This

result characterizes the onset of filter divergence. In figure

6 we present the true, the EnKF analysis for one ensemble

member and the expected analysis geopotential fields for

5 days data assimilation time integration. This illustration

confirms the validity of the Talagrand diagram. As a matter

of fact, a deterioration of the typical shape of the RH wave

no.4 test case is noticeable.

Figure 7 regroups the Talagrand diagrams and root

mean square errors rmses corresponding to different values

of the random perturbation around the mean, namely

1%, 3% and 6% perturbations. The Talagrand diagrams

display now a non symmetrical U shape characteristic of

filter divergence as well as bias. An increase in the values

of the frequencies follows the increase in the perturbation

percentage. The rmses of the geopotential field, presented in

the same figure, increase accordingly with the perturbation.

In figures 8 and 9 we present the true, the EnKF

analysis for one ensemble member and the expected

analysis geopotential forecast and analysis geopotential

fields for 10 days EnKF window of data assimilation

with 1% and 3% perturbation around the mean. The

same nonlinear observation operator H(u) = u2 has been

employed. The results obtained point to a total loss of shape

and symmetry deterioration of the RH wave no. 4 test case.

The purpose of the this section is not to establish the

onset of filter divergence, rather to show that the fixture

of Evenson [Evenson 2003] is not suitable for nonlinear

observation operator. New efforts to mitigate the problem

of high nonlinearity in EnKF have been undertaken. The

recent work of Kalnay and collaborators, [Kalnay priv.

comm. 2008], [Kalnay and Yang 2008] and [Kalnay and

Yang 2010] suggests an approach involving outer loops of

minimization, borrowed from variational data assimilation

practice. They introduce an empirical convergence criterion

within a nonstandard iterative minimization scheme.

Common to both mentioned algorithms is that they treat

each iteration of minimization as a quadratic sub-problem

and thus rely on the use of traditional Kalman filter linear

analysis update equation in each iteration.

4.3. PF set-up and numerical results

As in the previous (EnKF) section, all our numerical

experiments employ a time step of ∆t = 600sec. The

observations are sampled once every 6 hours i.e. at a

frequency consisting of one observation every 36 time steps.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 13

4.3.1. The impact of number of particles and of the

resampling strategy

In view of recent results in the literature ( such as

[Van Leeuwen 2009] [Snyder et al. 2008] ) the issue of

choice of an adequate number of particles is of paramount

importance. In order to avoid filter degeneracy we follow

the suggestion of Van-Leeuwen [Van Leeuwen 2009] as

to the choice of suitable number of particles, that should

be not smaller than the number of degrees of freedom of

the simulated system. We tested three cases namely, 1000,

1500 and 1900 particle filters (compare to 48× 38 modes

in the latitude-longitude directions for the spherical-spectral

shallow-water model )

In as far as PF resampling methods are concerned,

we tested three resampling techniques namely, residual,

systematic and multinomial resampling,respectively. All

of them performed equally well with a slight edge

to the systematic resampling technique in as far as

computational efficiency is concerned see [Doucet et al.

2001, Arulampalam et al. 2002 and Van Leeuwen 2009]. A

novel feature introduced here is the equivalent of Talagrand

diagrams, where instead of taking the ensemble members

we took the particles. Then we iterated over the number

of bins (1900 particles, 51 bins). Similar to the ensemble

framework, this diagram could be a good tool to detect

systematic flaws of a particle filter prediction system.

4.3.2. Linear observation operator case

The linear observation operator assuming the formH(u) =

u has been employed for the particle filter case and we do

not show the related results. However, we observe almost

identical results to those obtained for the EnKF linear

observation operator case, in fact the R-H wave no 4 test

case preserves its global characteristic shapes with minor

changes occurring in the center of the recirculation zones at

around 10 days of data assimilation time integration.

4.3.3. Non-linear observation operator case

We consider the impact of nonlinear observation operator

for the particle filter (PF) when H(u) = u2. As displayed

in figure 10, where the true, PF analysis for one particle

geopotential fields for 1% and 3% perturbation around

the mean are depicted. After 10 days data assimilation

time integration,the R-H wave no 4 test case preserves

its global characteristic shape and no distortion nor cell

pattern breakdown can be detected. These results should be

contrasted to our EnKF findings of filter divergence for the

same data assimilation time integration.

A part from the bias ( i.e. the first and last bins) of

the U-shaped diagrams shown in figure 11, we have

uniformity of the frequencies repartition indicates that the

probability distribution has been well sampled. In fig 12

we present the time evolution of the rmses after 10 days

of data assimilation time integration of PF for 1% and

3% perturbation around the mean. The striking result is

that the rmses variation is now bounded in contrast to the

EnKf case where the increase in the corresponding rmses

is quasi-exponential. The effect of degree of nonlinearity

of the observation operator in the PF case has been also

examined. As figure 13 shows that a stronger nonlinearity

(i.eH(u) = exp(u) ) has no substantial effect on the rmses.

Indeed, the rmses envelope is not affected despite presence

of oscillations characteristic of the PF filter.

4.4. MLEF set-up and results

Only 30 ensemble members were sufficient for carrying

out the MLEF runs, while the number of cycles was set to

70. The iterative nonlinear minimization method employed

for cost function minimization was the Flecher-Reeves

nonlinear conjugate gradient algorithm. In each of the

MLEF data assimilation cycles, 5 iterations are performed

to obtain the analysis. The standard deviation of each of the

Gaussian state and observation noise was taken to be 1%

random perturbation around the mean. No localization nor

inflation were used within the ensemble part of this method.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

14 M.Jardak et al.

We observe similar results to those obtained for the

EnKF and PF in linear observation operator case, namely

that the global characteristic shapes of the filtered solution

are preserved. This feature is presented in figure

When H(u) = u2, we observe additional aspects in

the divergence and the vorticity fields as it can be seen

in figure after 10 days of model integration with data

assimilation. The striking resemblance between the PF and

MLEF results manifests the successful implementation of

both filters. This result is further confirmed as the degree

of nonlinearity of the observation operators was increased

fromH(u) = u2 toH(u) = eu. Our numerical experiments

are presented in in figure 14. As presented in figure 15,

we observe that a part from the bias ( i.e. the first and last

bins) of the U-shaped diagrams shown ,we have uniformity

of the frequencies repartition indicates that the probability

distribution has been well sampled. This is in agreement

with our PF findings. The rmses presented in figure 16 are

bounded and the nonlinearity of the observation operator

has no effect on the rmses, consequently the MLEF filter

predicts well the true solution independently of the linearity

of the observation operator.

5. Summary and conclusions

In this paper, a comparison was carried out for the three

most popular ensemble data assimilation methods (EnsDA)

for a spherical nonlinear shallow water equations model

discretized via a spectral model and tested for the Rossby

Haurwitz wave no. 4.

The Monte Carlo version of EnKF, the particle filter PF with

several resampling strategies and finally as a representative

of hybrid filters, the maximum likelihood filter MLEF were

tested in the presence of both linear and highly nonlinear

observation operators. The nonlinear observation operators

replicate and stand as a proxy for more realistic processes

such as cloud, aerosol and precipitation processes, as well

as remote sensing (e.g., satellite and radar) observations.

While the typical EnsDA analysis equation assumes

linearity and Gaussianity for EnKF, (thus fundamentally

preventing this EnsDA from extracting maximum

information from nonlinear observations) we have also

tested a Bayesian filter, the particle filter, whose analysis

equation is nonlinear as well as the MLEF hybrid filter,

which accommodates nonlinear observation operators as

well as non-Gaussianity.

In the case of linear observation operator results for the

three above mentioned sequential data assimilation filters

leads to the conclusion, that the aforementioned methods

are comparable and yield satisfactory results. Amongst the

different error metrics used to assess filter divergence, we

chose to retain the Talgrand diagram (known also as Rank

Score) , the usual rmse statistics. Results of Talagrand

diagram for EnKF in the presence of nonlinear operators

show that filter divergence onset occurs at about 5 days of

data assimilation using the R-H wave no.4 test case. We

conclude that EnKF with Evensen’s fixture for nonlinearity

fails to converge for nonlinear observation operators . In

contrast the other two filters (PF and MLEF) handled well

the issue of nonlinearity of the observation operators as well

as non-Gaussianity, as expected. The three error metrics

retained support the results obtained.

These results are of course a function of our particular

implementation of the EnKF, PF and MLEF, respectively

and are a function of the number of particles (number of

ensemble members) the number of observations as well

as the particular model tested. In as far as computational

efficiency, the MLEF was found to be faster than the EnKF,

while the PF filter proved to be the most time consuming in

as far as the shallow water equations model with nonlinear

observation operator is concerned.

The EnKF version for the nonlinear observation oper-

ator H(u) = u2 is displaying a significant discrepancy

between the true and filtered EnKF solutions. is consid-

erable. We join here many authors, [Jardak et al. 2010],

[Nakano et al. 2007] and [Apte et al. 2008A,2008B] to

cite but a few, to draw the conclusion that even with

the modification suggested by [Evensen 2003], the EnKF

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 15

filter diverges in the case of nonlinear observation oper-

ator applied here to the shallow water equations model.

In conclusion, we can only compare between the PF and

the MLEF filters in the presence of nonlinear observation

operators. Both the PF and the MLEF filters exhibit good

performance in terms of results of sequential data assimi-

lation as evidenced by the rmses and the divergence statis-

tics. for the tests conducted with the nonlinear observation

operators the MLEF method appears to have an edge in

terms of CPU time over the PF method method with SR

resampling. Indeed, the PF results were obtained using

1000-1900 particles, while MLEF obtained similar results

for only 30-300 ensembles.

Acknowledgements

The research of Prof. I.M. Navon and Dr. M. Jardak was

supported by the National Science Foundation (NSF), grant

ATM-03727818. The authors also acknowledge the support

by NASA Modeling, Analysis, and Prediction Program

under Award NNG06GC67G. Prof. Navon acknowledges

the National Science Foundation (NSF), grant ATM-

0931198. Dr. Zupanski acknowledges the National Science

Foundation (NSF), grant ATM-0930265. Finally, Dr.

Jardak is grateful to W.H.Hanya and M.O.Nefysa for the

continuous help.

References

[1] D. L. Williamson, J. B. Drake, J. J. Hack, R. Jacob

and P. N. Swarztrauber.1992. A standard test set for

numerical approximations to shallow-water equations

in spherical geometry. J. Comput. Phys 102: 211–224.

[2] D. L. Williamson. 1997. Climate Simulations with

a spectral semi-Lagrangian model with linear

grid,” Numerical Methods in Atmospheric and

Ocean Modelling. Canadian Meteorological and

Oceanographic Society 279–292.

[3] D. L. Williamson. 2007. The Evolution of Dynamical

Cores for Global Atmospheric Models. Journal of the

Meteorological Society of Japan 85B: 241–269.

[4] R. Jakob-Chien, J. J. Hack and D. L. Williamson.1995.

Spectral transform solutions to the shallow water test

set. J. Comput. Phys 119: 164–187 .

[5] J. P. Boyd , 2001. Chebyshev and Fourier Spectral

Methods, Second Edition (Revised). Dover Publishing

[6] M. L. Berliner and C. K. Wikle. 2007. A Bayesian

tutorial for data assimilation. Physica D. 30: 1–16.

[7] M. L. Berliner and C. K. Wikle. 2007. Approximate

importance sampling Monte Carlo for data assimila-

tion. Physica D. 30: 37-49

[8] A. Doucet, A. , S. Godsill, S. and C. Andrieu. 2000.

On Sequential Monte Carlo sampling methods for

Bayesian filtering. Statistics and Computing. 10 (3):

197–208.

[9] A. Doucet, J. F. G. de Freitas and N. J. Gordon. 2001.

An introduction to sequential Monte Carlo methods.

Sequential Monte Carlo Methods in Practice, Doucet,

A. , de Freitas, J.F.G and Gordon, N.J (eds.) Springer-

Verlag, New York

[10] M. S. Arulampalam, S. Maskell, N. Gordon and

T. Clapp. 2002. A Tutorial on Particle Filters for

Online Nonlinear/Non-Gaussian Bayesian Tracking.

IEEE transactions on signal processing. Vol(150)

no.2: 174–188.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

16 M.Jardak et al.

0 50 100 150 200 250 300 350

−80

−60

−40

−20

0

20

40

60

80

longitude

latit

ude

grid points

observation locations

Figure 1. The locations of the grid and observation points: observations at every 4 grid points in longitude and every 2 grid points in latitude

[11] N. J. Gordon, D. J. Salmond and A. F. M. Smith. 1993.

Novel approach to nonlinear non-Gaussian Bayesian

state estimate. IEEE Proc. F. 140: 107–113.

[12] G. Kitagawa. 1996. Monte Carlo filter and smoother

for non-Gaussian nonlinear state space models.

Journal of Computational and Graphical Statistics

5,1: 1–25.

[13] J. S. Liu and R. Chen. 1998. Sequential Monte

Carlo Methods for Dynamic Systems Journal of

the American Statistical Association Vol.93, No.442:

1032-1044.

[14] R.E. Kalman. 1960. A new approach to linear filtering

and prediction problems. Transaction of the ASME -

Journal of Basic Engineering. Series D, 82: 35–45.

[15] G. Evensen. 1994 Sequential data assimilation in non-

linear quasi-geostrophic model using Monte Carlo

methods to forecast error statistics J. Geophys Res, 99

(C5), 10 14310 162.

[16] G. Evensen and P. J. Van Leeuwen. 1996. Assimilation

of Geosat altimeter data for Agulhas Current using the

Ensemble Kalman filter with quasi-geostrophic model.

Monthly Weather Review, 124(1): 85-96.

[17] G. Burgers, P. J. Van Leeuwen and G. Evensen.

1998. Analysis scheme in the ensemble Kalman Filter.

Monthly Weather Review 126: 1719-1724.

[18] G. Evensen. 2003. The Ensemble Kalman Filter:

theoretical formulation and practical implementation.

Ocean Dynamics, 53: 343-367.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 17

0 100 200 300

−50

0

50

longitude

latitu

de

velocity field

−10

0

10

20

30

40

50

longitude

latitu

de

geopotential

0 100 200 300

−50

0

50

8500

9000

9500

10000

10500

longitude

latitu

de

divergence

0 100 200 300

−50

0

50

−1

−0.5

0

0.5

1x 10

−6

longitude

latitu

de

vorticity

0 100 200 300

−50

0

50

−1.5

−1

−0.5

0

0.5

1

1.5x 10

−4

Figure 2. Non-perturbed shallow water solution R-H wave 4 case: velocity, geopotential, divergence and vorticity fields after 10 days of time integration

[19] G. Evensen. 2007. Data Assimilation: The ensemble

Kalman filter. Springer, Berlin

[20] J. Mandel. 2007. A brief tutorial on the ensemble

Kalman filter. UCDHSC/CCM Report No.242.

[21] J. Mandel. 2006. Efficient Implementation of the

Ensemble Kalman Filter. CCM Report 231.

[22] A. H.Jazwinski. 1970. Stochastic processes and

filtering theory. Academic Press, New York

[23] R. S. Bucy. 1965. Recurrent sets. Ann. Math. Statist.

36: 535–545.

[24] J. L. Anderson and S. L. Anderson. 1999. A Monte

Carlo Implementation of the Nonlinear Filtering

Problem to Produce Ensemble Assimilations and

Forecasts. Mon. Wea. Rev. 127: 2741–2785.

[25] J. L. Anderson. 2003. A local least squares framework

for ensemble filtering. Monthly Weather Review 131:

634–642.

[26] C. H. Bishop, B. J. Etherton and S. J. Majumdar.

2001. Adaptive sampling with the ensemble transform

Kalman filter. Part I: Theoretical aspects. Monthly

Weather Review 129(3): 420–436.

[27] M. K.,Tippett, J. L. Anderson, C. H. Bishop and T.M.

Hamill. 2003. Ensemble square root filters. Monthly

Weather Review 131: 1485–1490.

[28] J. D. Hol , T. B. Schon and F. Gustafsson. 2006. On

Resampling Algorithms for Particle Filters. Nonlinear

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

18 M.Jardak et al.

longitude

latit

ude

0 50 100 150 200 250 300 350

−80

−60

−40

−20

0

20

40

60

80

8000

8500

9000

9500

10000

10500

Figure 3. EnKF filter linear observation operator case : initial perturbation geopotential field

Statistical Signal Processing Workshop, Cambridge,

United Kingdom.

[29] N. Metropolis, and S. Ulam. 1944. The Monte Carlo

Method. J. Amer. Stat. Assoc. 44: 335–341.

[30] J. L. Lewis, S. Lakshivarahan and S. K. Dhall.

2006. Dynamic Data Assimilation: A least squares

approach. Cambridge University Press, in the series

on Encyclopedia of Mathematics and its Applications

Volume 104

[31] E. N. Lorenz. 1963. Deterministic nonperiodic flow. J.

Atmos. Sci., 20:130–141 .

[32] E. N. Lorenz. 1969. The predictability of a flow which

possesses many scales of motion. Tellus 21: 289–307.

[33] S. E. Cohn. 1997. An introduction to estimation

theory. Journal of the Meteorological Society of Japan

Vol.75, No.1B 257–288

[34] X. Xiong, I. M. Navon and B. Uzunoglu. 2006. A

Note on the Particle Filter with Posterior Gaussian

Resampling Tellus 58A: 456–460.

[35] P. Houtekamer and H. Mitchell. 1998. Data assimila-

tion using an ensemble Kalman filter technique.Mon.

Wea. Rev. 126: 796–811.

[36] P. L. Houtekamer and H. L. Mitchell. 2001. A

Sequential Ensemble Kalman Filter for Atmospheric

Data Assimilation. Mon.Wea.Rev.129: 123–137.

[37] P. L. Houtekamer, H. L. Mitchell, G. Pellerin, M.

Buehner, M. Charron, L. Spacek and B. Hansen. 2005.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 19

0 100 200 300

−50

0

50

longitude

latit

ude

velocity field

−20

0

20

40

60

longitude

latit

ude

geopotential

0 100 200 300

−50

0

50

8500

9000

9500

10000

10500

longitude

latit

ude

divergence

0 100 200 300

−50

0

50

−8

−6

−4

−2

0

2

4

6

x 10−6

longitude

latit

ude

vorticity

0 100 200 300

−50

0

50

−1.5

−1

−0.5

0

0.5

1

1.5x 10

−4

Figure 4. EnKF filter linear observation operator case : velocity, geopotential, divergence and vorticity fields after 10 days of DA time integration. 100ensemble members and 1% random perturbation around the mean

A Sequential Ensemble Kalman Filter for Atmospheric

Data Assimilation Mon.Wea.Rev.133: 604–620.

[38] S. L. Dance. 2004. Issues in high resolution limited

area data assimilation for quantitative precipitation

forecasting. Physica D 196: 1–27.

[39] G. Kotecha and P. M. Djuric. 2003. Gaussian particle

filtering IEEE Trans. Signal Processing 51: 2592–

2601.

[40] D. E. Goldberg. 1989. Genetic algorithms in search,

optimization and machine learning Addison-Wesley,

Reading.

[41] B. W. Silverman. 1986. Density Estimation for

Statistics and Data Analysis. Chapman and Hall

[42] T. Bengtsson, P. Bickel, and B. Li. 2008. Curse-of-

dimensionality revisited:Collapse of the particle filter

in very large scale systems. Probability and Statistics:

Essays in Honor of David A. Freedman Vol. 2, 316–

334 .

[43] P. Bickel, B. Li and T. Bengtsson. 2008. Sharp

Failure Rates for the Bootstrap Particle Filter in High

Dimensions. IMS Collections: Pushing the Limits of

Contemporary Statistics: Contributions in Honor of

Jayanta K. Ghosh Vol.3: 318-329 .

[44] C. Snyder, T. Bengtsson, P. Bickel, and J.L. Anderson.

2008 Obstacles to high-dimensional particle filtering

Mon. Wea. Rev. Vol 136, No12 ,4629–4640.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

20 M.Jardak et al.

5 10 15 20 25 30 35 40 45 500

50

100

150

200

bins

freq

uenc

y

Talagrand diagram

0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5300

320

340

360

380rm

seφ

time(days)

Figure 5. EnKF filter non-linear observation operator caseH(u) = u2: time evolution of the geopotential rmse and Talagrand diagram after 5 days dataassimilation of time integration

[45] D. Zupanski. 1997. A general weak constraint

applicable to operational 4DVAR data assimilation

systems. Mon.Wea.Rev. 125: 2274-2292

[46] M. Zupanski. 2005. Maximum Likelihood Ensemble

Filter. Part 1: the theoretical aspect. Mon.Wea.Rev.

133: 1710–1726.

[47] D. Zupanski, M. Zupanski. 2006. Model error

estimation employing ensemble data assimilation

approach Mon.Wea.Rev. 134: 137–1354.

[48] M. Zupanski, I. M. Navon and D. Zupanski. 2008.

The Maximum Likelihood Ensemble Filter as a

non-differentiable minumization algorithm Q.J.Roy.

Meteor.Soc. 134: 1039–1050

[49] M. Jardak, I. M. Navon and M. Zupanski. 2010.

Comparison of Sequential data assimilation methods

for the Kuramoto-Sivashinsky equation. International

Journal for Numerical Methods in Fluids 136 Issue 4,

374–402

[50] A. C. Lorenc. 1986. Analysis methods for numerical

weather prediction Q. J. Roy. Meteor. Soc. 112: 1177-

1194 .

[51] T. M. Hamill, J. S. Whitaker and C. Snyder. 2001.

Distance-Dependent Filtering of Background Error

Covariance Estimates in an Ensemble Kalman Filter

Mon.Wea.Rev.,129: 2776–2790 .

[52] T. M. Hamill and C. Snyder. 2000. A Hybrid

Ensemble Kalman Filter-3D Variational Analysis

Scheme. Mon.Wea.Rev.128: 2905–2919 .

[53] T. M. Hamill. 2006. Ensemble-based atmospheric

data assimilation. Predictability of Weather and

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 21

longitude

latit

ude

true

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

longitude

latit

ude

EnkF analysis geopotential field for one ensemble member

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

longitude

latit

ude

EnKF expected analysis geopotential field

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

Figure 6. EnKF filter non-linear observation operator case H(u) = u2: true, the EnKF analysis for one ensemble member and the expected analysisgeopotential fields after 5 days of data assimilation time integration. 100 ensemble members and 1% random perturbation around the mean

Climate, Cambridge Press, 124–156 .

[54] T.M. Hamill. 2001. Interpretation of rank histogram

for verifying ensemble forecasts. Mon. Weather Rev.

129(3): 550–560

[55] L. Nerger, ,W. Hiller, J. Schrater. 2005. A comparison

of error subspace Kalman filters. Tellus A,57 (5): 715–

735

[56] G. Gaspari, S. E. Cohn. 1999. Construction of

correlation functions in two and three dimensions. Q.

J. Roy. Meteor Soc. Vol 125 Issue 554:723-757 .

[57] S. Nakano, G. Ueno and T. Higuchi T. 2007. Merging

particle filter for sequential data assimilation.

Nonlinear Processes in Geophysics 14: 395409

[58] A. Apte , C. K. R. T. Jones, A. M. Stuart and J. Voss.

2008. Data assimilation: Mathematical and statistical

perspectives. Int. J. Numer. Meth. Fluids 56: 1033–

1046

[59] A. Apte , C. K. R. T. Jones and A. M. Stuart. 2008.

Bayesian approach to Lagrangian data assimilation.

Tellus A,60: 336–347

[60] R. K. Smith and D. G. Dritschel. 2006. Revisiting

the Rossby–Haurwitz wave test case with contour

advection. Journal of Computational Physics 217,

Issue 2, 473–484

[61] P. J. van Leeuwen. 2009. Particle filtering in

geophysical systems. Mon.Wea.Rev. 137: Issue 12,

4089–4114

[62] E. Kalnay. 2010 Ensemble Kalman Filter: Current

Status and Potential Data Assimilation: Making Sense

of Observations William Lahoz, Boris Khattatov,

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

22 M.Jardak et al.

10 20 30 40 500

100

200

300

400

500

bins

freq

uenc

y

Talagrand diagram for 1% perturbation

10 20 30 40 500

50

100

150

200

250

300

350

bins

freq

uenc

y

Talagrand diagram for 3% perturbation

10 20 30 40 500

50

100

150

bins

freq

uenc

y

Talagrand diagram for 6% perturbation

0 2 4 6 8 100

200

400

600

800

1000

time(days)

rmse

φ

1% perturbation3% perturbation6% perturbation

Figure 7. EnKF filter non-linear observation operator caseH(u) = u2 and for different random perturabtion around the mean: Talagrand diagrams andtime evolution of root mean square error for the geopotential field for 10 days of time integration

Richard Menard (Editors) Springer

[63] P. Sakov and P. Oke. 2008. Implications of the Form of

the Ensemble Transformation in the Ensemble Square

Root Filters.Mon.Wea.Rev.136: 1042–1052

[64] B. Haurwitz. 1940. The motion of atmospheric

disturbances on the spherical earth. J. Mar. Res. 3

254–267

[65] J. Thuburn and Y. Li. 2000. Numerical simulations of

Rossby-Haurwitz waves. Tellus A 52 (2): 181–189

[66] B. J. Hoskins. 1973. Stability of the Rossby-Haurwitz

wave. Quart. J. Roy. Meteorol. Soc.99: 723–745

[67] N.A. Phillips. 1959. Numerical integration of the

primitive equations on the hemisphere. Mon. Wea.

Rev.87: 333–345

[68] C. G. Rossby and collaborators. 1939. Relation

between variations in the intensity of the zonal

circulation of the atmosphere and the displacements of

the semi-permanent centres of action. J. Marine Res.

2: 38–55

[69] C. A. Pires, O. Talagrand and M. Bocquet. 2010. Diag-

nosis and impacts of non-Gaussianity of innovations

in data assimilation. Physica D 239: 1701–1717

[70] M. Bocquet. 2008. Issues of nonlinearity and non-

gaussianity Contribution to the WWRP/THORPEX.

workshop on 4D-Var and ensemble Kalman filter

inter-comparisons Buenos Aires

[71] A.H. Murphy. 1973. A New Vector Partition of the

Probability Score. Journal of Applied Meteorology

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 23

longitude

latit

ude

true

0 50 100 150 200 250 300 350

−50

0

50

longitude

latit

ude

One ensemble member of the EnKF analysis

0 50 100 150 200 250 300 350

−50

0

50

longitude

latit

ude

Expected EnKF analysis

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

8000

9000

10000

8000

9000

10000

Figure 8. EnKF filter non-linear observation operator case H(u) = u2: true, the EnKF analysis for one ensemble member and the expected analysisgeopotential fields after 10 days of data assimilation time integration, 100 ensemble members and 1% random perturbation around the mean

vol. 12, Issue 4: 595–600

[72] A.H. Murphy. 1973. What is a good forecast? An essay

on the nature of goodness in weather forecasting.

Weather and forecasting. 8: 281–293

[73] D.S. Wilks. 1995. Statistical Methods in the Atmos-

pheric Sciences. Academic Press, New York

[74] O. Talagrand, R. Vautard, and B. Strauss. 1997. Evalu-

ation of probabilistic prediction systems. Proceedings,

ECMWF Workshop on Predictability. ECMWF. 1–25

[75] E. Kalnay and S. C Yang. 2010. Accelerating the

spin-up of Ensemble Kalman Filtering. Quart. J. Roy.

Meteor. Soc. in press

[76] J. Diard, P. Bessiere and E. Mazer. 2003. A survey of

probabilistic models, using the bayesian programming

methodology as unifying framework. The second Inter-

national Conference on Computational Intelligence,

Robotics and Autonomous Systems, Singapore

[77] O. Cappe, S.J. Godsill and E. Moulines. 2007. An

Overview of Existing Methods and Recent Advances

in Sequential Monte Carlo. Proceedings of the IEEE

95: 899–924

[78] A.M Stuart. 2010. Inverse Problems: A Bayesian

Prespective. Acta Numerica Vol 19: 451–559

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

24 M.Jardak et al.

longitude

latit

ude

true

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

longitude

latit

ude

One ensemble member of the EnKF analysis

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

longitude

latit

ude

Expected EnKF analysis

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

Figure 9. EnKF filter non-linear observation operator case H(u) = u2: true, the EnKF analysis for one ensemble member and the expected analysisgeopotential fields after 10 days of data assimilation time integration, 100 ensemble members and 3% random perturbation around the mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 25

longitude

latit

ude

true geopotential field

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

longitude

latit

ude

PF geopotential field, run with 3% perturbation around the mean

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

11000

longitude

latit

ude

PF geopotential field, run with 1% perturbation around the mean

0 50 100 150 200 250 300 350

−50

0

50

8000

9000

10000

Figure 10. PF filter non-linear observation operator caseH(u) = u2: true, and PF geopotential field with 1% and 3% random perturbation around themean after 10 days of time integration. 1900 particles

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

26 M.Jardak et al.

5 10 15 20 25 30 35 40 45 500

50

100

150

200

250

300

350

bins

freq

uenc

y

Talagrand diagram for 3% random perturbation

5 10 15 20 25 30 35 40 45 500

100

200

300

400

500

600

bins

freq

uenc

y

Talagrand diagram for 1% random perturbation

Figure 11. PF filter non-linear observation operator caseH(u) = u2: Talagrand diagrams after 10 days of time integration, 1900 particles for 1% and3% random perturbation around the mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 27

0 1 2 3 4 5 6 7 8 9 10500

520

540

560

580

600

620

time(days)

rmse

φ

1% perturbation around the mean3% perturbation around the mean

Figure 12. PF filter non-linear observation operator caseH(u) = u2: time evolution of the geopotential rmse 1% and 3% random perturbation aroundthe mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

28 M.Jardak et al.

0 1 2 3 4 5 6 7 8 9 10500

505

510

515

520

525

530

535

540

545

time(days)

rmse

φ

H(u) = u2

H(u) = exp(u)

Figure 13. PF non-linear observation operatorH(u) = u2 andH(u) = exp(u) case: time evolution of the geopotential rmse 1% random perturbationaround the mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 29

longitude

latit

ude

Analysis geopotential field H(u) = u2

5 10 15 20 25 30 35 40 45

5

10

15

20

25

30

35

Analysis geopotential field H(u) = exp(u)

latit

ude

longitude

5 10 15 20 25 30 35 40 45

5

10

15

20

25

30

35

8500

9000

9500

10000

10500

8500

9000

9500

10000

10500

Figure 14. MLEF filter nonlinear observation operatorH(u) = u2 andH(u) = exp(u) case : geopotential field, after 10 days of DA time integration.1% random perturbation around the mean.

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

30 M.Jardak et al.

5 10 15 20 25 30 35 40 45 500

50

100

150

200

bins

freq

uenc

y

Talagrand diagram for H(u) = exp(u)

5 10 15 20 25 30 35 40 45 500

50

100

150

200

bins

freq

uenc

y

Talagrand diagram for H(u) = u2

Figure 15. MLEF filter nonlinear observation operatorH(u) = u2 andH(u) = exp(u) case : Talagrand diagrams, after 10 days of DA time integration.1% random perturbation around the mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls

Sequential filtering for the shallow water equations model 31

0 1 2 3 4 5 6 7 8 9 100

100

200

300

400

500

600

time(days)

rmse

φ

H(u) =exp(u)

H(u) = u2

Figure 16. MLEF filter nonlinear observation operator H(u) = u2 and H(u) = exp(u) case: time evolution of the geopotential rmse. 1% randomperturbation around the mean

Copyright c© 2010 Royal Meteorological Society Q. J. R. Meteorol. Soc. 00: 2–23 (2010)

Prepared using qjrms4.cls


Recommended