Top Banner
Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of Economics London Mathematical Society - EPSRC Durham Symposium Mathematics of Data Assimilation
36

Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Apr 01, 2015

Download

Documents

Amani McGregor
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Better Data Assimilation through Gradient Descent

Leonard A. Smith, Kevin Judd and Hailiang Du

Centre for the Analysis of Time Series London School of Economics

London Mathematical Society - EPSRC Durham Symposium

Mathematics of Data Assimilation

Page 2: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Outline

Perfect model scenario (PMS)

GD method GD is NOT 4DVAR Results compared with Ensemble KF

Imperfect model scenario (IPMS)

GD method with stopping criteria GD is NOT WC4DVAR Results compared with Ensemble KF

Conclusion & Further discussion

Page 3: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Experiment Design (PMS)

Page 4: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Ensemble techniques

Generate ensemble directly, e.g. Particle Filter, Ensemble Kalman Filter

Generate ensemble from perturbations of a reference trajectory, e.g. SVD on 4DVAR

Gradient Descent (GD) Method

K Judd & LA Smith (2001) Indistinguishable States I: The Perfect Model Scenario, Physica D 151: 125-141.

Page 5: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Gradient Descent (Shadowing Filter)

Page 6: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Gradient Descent (Shadowing Filter)

5s

0s

4s

)( 5sF

Page 7: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Gradient Descent (Shadowing Filter)

Page 8: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Gradient Descent (Shadowing Filter)

Page 9: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Gradient Descent (Shadowing Filter)

Page 10: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

GD is NOT 4DVAR

Difference in cost function

Noise model assumption

Observational noise model 4DVAR cost function

GD cost function not depend on noise model

Assimilation window

4DVAR dilemma: difficulties of locating the global minima with long assimilation

window

losing information of model dynamics and observations without long window

Page 11: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Methodology

Page 12: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Form ensemble

Obs

t=0

Reference trajectory

GD result

Page 13: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Form ensemble

t=0Candidate trajectories

Sample the local space

Perturb observations and run GD

Page 14: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Form ensemble

t=0Ensemble trajectory

Draw ensemble members according to likelihood

Page 15: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Form ensemble

Obs

t=0Ensemble trajectory

Page 16: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Ensemble members in the state space

Compare ensemble members generated by Gradient Descent method and Ensemble Adjustment Kalman Filter method in the state space.

Low dimensional example to visualize, higher dimensional results later.

Page 17: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Ikeda Map, Std of observational noise 0.05, 512 ensemble

members

Page 18: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Evaluate ensemble via Ignorance

The Ignorance Score is defined by:

where Y is the verification.

Ikeda Map and Lorenz96 System, the noise model is N(0, 0.4) and

N(0, 0.05) respectively. Lower and Upper are the 90 percent

bootstrap resampling bounds of Ignorance score

Ensemble->p(.)

Page 19: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Imperfect Model Scenario

Page 20: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Toy model-system pairs

Ikeda system:

Imperfect model is obtained by using the truncated polynomial, i.e.

Page 21: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Toy model-system pairs

Lorenz96 system:

Imperfect model:

Page 22: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Insight of Gradient Descent

Define the implied noise to be

and the imperfection error to be

Page 23: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Insight of Gradient Descent

5s0s

4s

)( 5sf

w0

Page 24: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Insight of Gradient Descent

w

Page 25: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Insight of Gradient Descent

0w

Page 26: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Statistics of the pseudo-orbit as a function of the number of Gradient Descent iterations for both higher dimension Lorenz96 system-model pair experiment (left) and low dimension Ikeda system-model pair experiment (right).

Implied noise

Imperfection error

Distance from

the “truth”

Page 27: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

GD with stopping criteria

GD minimization with “intermediate” runs produces more consistent pseudo-orbits

Certain criteria need to be defined in advance to decide when to stop or how to tune the number of iterations.

The stopping criteria can be built by testing the consistency between implied noise and the noise model

or by minimizing other relevant utility function

Page 28: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Imperfection error vs model error

Model error Imperfection error

Obs Noise level: 0.01

Not accessible!

Page 29: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Imperfection error vs model error

Imperfection error

Obs Noise level: 0.002 Obs Noise level: 0.05

Page 30: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

GD vs WC4DVAR

WC4DVAR Model error

assumption

GDModel error

estimates

Page 31: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Forming ensemble

Apply the GD method on perturbed observations.

Apply the GD method on perturbed pseudo-orbit.

Apply the GD method on the results of other data assimilation methods. Particle filter?

Page 32: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Imperfect model experiment: Ikeda system-model pair, Std of

observational noise 0.05, 1024 EnKF ensemble members, 64 GD ensemble members

Page 33: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Evaluate ensemble via Ignorance

The Ignorance Score is defined by:

where Y is the verification.

Ikeda system-model pair and Lorenz96 system-model pair, the noise model is N(0, 0.5) and N(0, 0.05) respectively. Lower and Upper are the 90 percent bootstrap resampling bounds of Ignorance score

Systems Ignorance Lower Upper

EnKF GD EnKF GD EnKF GD

Ikeda -2.67 -3.62 -2.77 -3.70 -2.52 -3.55

Lorenz96

-3.52 -4.13 -3.60 -4.18 -3.39 -4.08

Page 34: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Conclusion Methodology of applying GD for data assimilation in

PMS is demonstrated outperforms the 4DVAR and Ensemble Kalman filter methods

Outside PMS, mmethodology of applying GD for data assimilation with a stopping criteria is introduced and shown to outperform the WC4DVAR and Ensemble Kalman filter methods.

Applying the GD method with a stopping criteria also produces informative estimation of model error.

No data assimilation without dynamics.

Page 35: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.

Thank you!

[email protected]

Centre for the Analysis of Time Series:http://www2.lse.ac.uk/CATS/home.aspx

Page 36: Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of.