dynamic data-driven adaptive observations in data ... · introduction and motivation table of...
TRANSCRIPT
Dynamic Data-Driven Adaptive Observationsin Data Assimilation forMultiscale Systems
AFOSR DDDAS Program Review Meeting, Dayton OH
PI: N. Sri NamachchivayaRyne Beeson, Hoong Chieh Yeong,
Nicolas Perkowski∗ and Peter Sauer†
Department of Aerospace Engineering & Information Trust Institute
University of Illinois at Urbana-Champaign
Urbana, Illinois, USA
September 7th, 2017
∗Applied Mathematics, Humboldt-Universitat zu Berlin†Electrical and Computer Engineering, University of Illinois at Urbana-Champaign
Beeson (University of Illinois) September 7th, 2017 1 / 28
Introduction and Motivation
Research Objectives
(I) Develop efficient and robust methods to produce lower-dimensional recursivenonlinear filtering equations driven by the observations; particle filters for theintegration of observations with the simulations of large-scale complex systems.
(II) Develop an integrated framework that combines the ability to dynamically steer themeasurement process, extracting useful information, with nonlinear filtering forinference and prediction of large scale complex systems.
Beeson (University of Illinois) September 7th, 2017 2 / 28
Introduction and Motivation
Table of Contents
1 Introduction and Motivation
2 Reduced Order Dynamic Data Assimilation
3 The Nudged Particle Filter Method
4 Adaptive Observations - Dynamically Steering the Measurement Process
5 Summary
6 Reporting
7 References
Beeson (University of Illinois) September 7th, 2017 3 / 28
Introduction and Motivation
Motivating Problems and Characteristics
Problems are,
(i) Multiscale
(ii) Chaotic
(iii) High dimensional
(iv) Sparse observations
(v) Ability for sensor selection, placement and control - adaptive observation
(vi) Sensors correlated to environment
Motivating problems,
(a) Weather prediction and forecasting
(b) Detection and tracking of contaminants in the environment(e.g. chemical and radioactive)
Beeson (University of Illinois) September 7th, 2017 4 / 28
Introduction and Motivation
Estimation and prediction in Earth (climate) system models
Figure: NCAR Community Climate System Model [1]
land/vegetation sea ice
atmosphere
coupler
driver
ocean
t
i
m
e
Figure: Coupling components in climate model[NCAR] [1]
Beeson (University of Illinois) September 7th, 2017 5 / 28
Introduction and Motivation
Simple multiscale example: Lorenz-Maas model
Coupled equations [2], [3], [4] :
dρx
dt= −ρyLz + (ρx + f ′ρy)ρz − ρx
dρy
dt= ρxLz − (f ′ρx − ρy)ρz − ρy
+ k1q + k2(x − k3ρy)
dρz
dt= −ρ2
x − ρ2y −µρz
ε2
ε3
dLz
dt= −Lz − k4x
ε2dxdt
= −y2 − z2 − ax + aF0
+ ε1(k3ρy − x)
ε2dydt
= xy − bxz − y + G
ε2dzdt
= bxy + xz − z
Figure: Coupled Lorenz 1984 atmosphere – Maas2004 ocean models
Beeson (University of Illinois) September 7th, 2017 6 / 28
Introduction and Motivation
Simple multiscale example: Lorenz-Maas model
Coupled equations [2], [3], [4] :
ddt
X = b(X, Lz, Z1)
εddt
Lz = −Lz − k4Z1
ε2 ddt
Z = f0(Z) + εf1(X2, Z1),
where ε is O(10−2).
Figure: Coupled Lorenz 1984 atmosphere – Maas2004 ocean models
Beeson (University of Illinois) September 7th, 2017 6 / 28
Introduction and Motivation
Estimation and prediction in Earth (climate) system models
Sensor Data(MOCHA,dropsondes,drifters)
Data assimilation methods(EnKF, 3D-,4D-Var, OI)
GCMs
Climate phenomena
description of
estimate/prediction
Weather phenomena can be studied through estimation/prediction using GCMs.
GCMs can be improved using data assimilation results.
Filtering theory provides rigorous approach to quantifying probabilisticinformation - as opposed to methods such as 3D-Var, 4D-Var, and OI.
Beeson (University of Illinois) September 7th, 2017 7 / 28
Introduction and Motivation
Estimation and prediction in Earth (climate) system models
Sensor Data(MOCHA,dropsondes,drifters)
Data assimilation methods(EnKF, 3D-,4D-Var, OI)
GCMs
Climate phenomena
description of
estimate/prediction
Weather phenomena can be studied through estimation/prediction using GCMs.
GCMs can be improved using data assimilation results.
Filtering theory provides rigorous approach to quantifying probabilisticinformation - as opposed to methods such as 3D-Var, 4D-Var, and OI.
Beeson (University of Illinois) September 7th, 2017 7 / 28
Introduction and Motivation
Mobile Platforms, Adaptive Observations, and Correlated Noise
Figure: NCAR Globe, 500km SectorsFigure: UAV Flight Controls [5]
1 Sensor Selection / Placement2 Sensor Control3 Mobile Platforms are Embedded in Signal Environment→ Correlated Noise4 Require Efficient and Robust Filtering Methods for Multiscale Correlated Case
Beeson (University of Illinois) September 7th, 2017 8 / 28
Reduced Order Dynamic Data Assimilation
Table of Contents
1 Introduction and Motivation
2 Reduced Order Dynamic Data Assimilation
3 The Nudged Particle Filter Method
4 Adaptive Observations - Dynamically Steering the Measurement Process
5 Summary
6 Reporting
7 References
Beeson (University of Illinois) September 7th, 2017 9 / 28
Reduced Order Dynamic Data Assimilation
Multiscale Correlated Noise Problem Setup
Let (Ω,F, Ft>0,Q) be a probability space upon which the following SDEs are defined:
dXεt =
[b(Xε
t , Zεt ) +
1ε
b1(Xεt , Zε
t )
]dt + σ(Xε
t , Zεt )dWt Xε
0 = x
dZεt =
1ε2 f (Xε
t , Zεt )dt +
1ε
g(Xεt , Zε
t )dVt Zε0 = z
dYεt = h(Xε
t , Zεt )dt + αdWt + βdVt + γdUt Yε
0 = 0
= h(Xεt , Zε
t )dt + dBt
1 Wt, Vt and Ut are independent standard BM under Q2 0 < ε 1 is the time-scale separation3 w.l.o.g., let α2 + β2 + γ2 = 1 and define the standard BM
Bt ≡ αWt + βVt + γUt
Beeson (University of Illinois) September 7th, 2017 10 / 28
Reduced Order Dynamic Data Assimilation
The Nonlinear Filter
The objective in filtering theory is to obtain a solution for the normalized conditionalmeasure - the filter,
Normalized Conditional Measure
πεt (ϕ(X
εt , Zε
t )) ≡ EQ [ϕ(Xεt , Zε
t ) | Yεt ] ,
where ϕ(Xεt , Zε
t ) is an integrable function and
Yεt ≡ σ(Yε
s − Yε0 | s ∈ [0, t]).
1 Density equivalent of πεt satisfies a high dimensional SPDE: “Curse of
Dimensionality”.2 If ϕ = ϕ(Xε
t ) and Xεt ⇒ X0
t as ε→ 0, does there exists πεt → π0
t ?3 Proof and insight will enable improvement of nonlinear filtering algorithms for
multiscale correlated noise case.4 Xε
t ⇒ X0t does not imply πε
t → π0t
Beeson (University of Illinois) September 7th, 2017 11 / 28
Reduced Order Dynamic Data Assimilation
The Nonlinear Filter
The objective in filtering theory is to obtain a solution for the normalized conditionalmeasure - the filter,
Normalized Conditional Measure
πεt (ϕ(X
εt , Zε
t )) ≡ EQ [ϕ(Xεt , Zε
t ) | Yεt ] ,
where ϕ(Xεt , Zε
t ) is an integrable function and
Yεt ≡ σ(Yε
s − Yε0 | s ∈ [0, t]).
1 Density equivalent of πεt satisfies a high dimensional SPDE: “Curse of
Dimensionality”.2 If ϕ = ϕ(Xε
t ) and Xεt ⇒ X0
t as ε→ 0, does there exists πεt → π0
t ?3 Proof and insight will enable improvement of nonlinear filtering algorithms for
multiscale correlated noise case.4 Xε
t ⇒ X0t does not imply πε
t → π0t
Beeson (University of Illinois) September 7th, 2017 11 / 28
Reduced Order Dynamic Data Assimilation
The Nonlinear Filter
The objective in filtering theory is to obtain a solution for the normalized conditionalmeasure - the filter,
Normalized Conditional Measure
πεt (ϕ(X
εt , Zε
t )) ≡ EQ [ϕ(Xεt , Zε
t ) | Yεt ] ,
where ϕ(Xεt , Zε
t ) is an integrable function and
Yεt ≡ σ(Yε
s − Yε0 | s ∈ [0, t]).
1 Density equivalent of πεt satisfies a high dimensional SPDE: “Curse of
Dimensionality”.2 If ϕ = ϕ(Xε
t ) and Xεt ⇒ X0
t as ε→ 0, does there exists πεt → π0
t ?3 Proof and insight will enable improvement of nonlinear filtering algorithms for
multiscale correlated noise case.4 Xε
t ⇒ X0t does not imply πε
t → π0t
Beeson (University of Illinois) September 7th, 2017 11 / 28
Reduced Order Dynamic Data Assimilation
The Nonlinear Filter
The objective in filtering theory is to obtain a solution for the normalized conditionalmeasure - the filter,
Normalized Conditional Measure
πεt (ϕ(X
εt , Zε
t )) ≡ EQ [ϕ(Xεt , Zε
t ) | Yεt ] ,
where ϕ(Xεt , Zε
t ) is an integrable function and
Yεt ≡ σ(Yε
s − Yε0 | s ∈ [0, t]).
1 Density equivalent of πεt satisfies a high dimensional SPDE: “Curse of
Dimensionality”.2 If ϕ = ϕ(Xε
t ) and Xεt ⇒ X0
t as ε→ 0, does there exists πεt → π0
t ?3 Proof and insight will enable improvement of nonlinear filtering algorithms for
multiscale correlated noise case.4 Xε
t ⇒ X0t does not imply πε
t → π0t
Beeson (University of Illinois) September 7th, 2017 11 / 28
Reduced Order Dynamic Data Assimilation
Mathematical Tools and Proof of Convergence Approach
1 Introduce an unnormalized conditional measure
Unnormalized Conditional Measure ρεt
ρεt (ϕ)
ρεt (1)=EPε
[ϕ(Xε
t , Zεt )Dε
t
∣∣∣Yεt
]EPε
[Dε
t
∣∣∣Yεt
] = EQ [ϕ(Xεt , Zε
t ) | Yεt ] = π
εt (ϕ)
2 Introduce function valued dual process, vε, satisfying a BSPDE3 Ansatz v0, ρ0,π0
4 Asymptotic expansion of vε = v0 +ψ+ R; ψ the corrector, and R the remainder5 Utilize homogenization estimates [6], estimates for BSPDE [7] and Feynman-Kac
representations with FBDSDE [8] to prove convergence,
Dual Convergence Implies Filter Convergence
E[∣∣ρε,x
T (ϕ) − ρ0T(ϕ)
∣∣p]6 ∫R2E[∣∣vε
0 (x, z) − v00(x)
∣∣p]QXε0 ,Zε
0(dx, dz)
Beeson (University of Illinois) September 7th, 2017 12 / 28
Reduced Order Dynamic Data Assimilation
Mathematical Tools and Proof of Convergence Approach
1 Introduce an unnormalized conditional measure
Unnormalized Conditional Measure ρεt
ρεt (ϕ)
ρεt (1)=EPε
[ϕ(Xε
t , Zεt )Dε
t
∣∣∣Yεt
]EPε
[Dε
t
∣∣∣Yεt
] = EQ [ϕ(Xεt , Zε
t ) | Yεt ] = π
εt (ϕ)
2 Introduce function valued dual process, vε, satisfying a BSPDE3 Ansatz v0, ρ0,π0
4 Asymptotic expansion of vε = v0 +ψ+ R; ψ the corrector, and R the remainder5 Utilize homogenization estimates [6], estimates for BSPDE [7] and Feynman-Kac
representations with FBDSDE [8] to prove convergence,
Dual Convergence Implies Filter Convergence
E[∣∣ρε,x
T (ϕ) − ρ0T(ϕ)
∣∣p]6 ∫R2E[∣∣vε
0 (x, z) − v00(x)
∣∣p]QXε0 ,Zε
0(dx, dz)
Beeson (University of Illinois) September 7th, 2017 12 / 28
Reduced Order Dynamic Data Assimilation
Mathematical Tools and Proof of Convergence Approach
1 Introduce an unnormalized conditional measure
Unnormalized Conditional Measure ρεt
ρεt (ϕ)
ρεt (1)=EPε
[ϕ(Xε
t , Zεt )Dε
t
∣∣∣Yεt
]EPε
[Dε
t
∣∣∣Yεt
] = EQ [ϕ(Xεt , Zε
t ) | Yεt ] = π
εt (ϕ)
2 Introduce function valued dual process, vε, satisfying a BSPDE3 Ansatz v0, ρ0,π0
4 Asymptotic expansion of vε = v0 +ψ+ R; ψ the corrector, and R the remainder5 Utilize homogenization estimates [6], estimates for BSPDE [7] and Feynman-Kac
representations with FBDSDE [8] to prove convergence,
Dual Convergence Implies Filter Convergence
E[∣∣ρε,x
T (ϕ) − ρ0T(ϕ)
∣∣p]6 ∫R2E[∣∣vε
0 (x, z) − v00(x)
∣∣p]QXε0 ,Zε
0(dx, dz)
Beeson (University of Illinois) September 7th, 2017 12 / 28
Reduced Order Dynamic Data Assimilation
Mathematical Tools and Proof of Convergence Approach
1 Introduce an unnormalized conditional measure
Unnormalized Conditional Measure ρεt
ρεt (ϕ)
ρεt (1)=EPε
[ϕ(Xε
t , Zεt )Dε
t
∣∣∣Yεt
]EPε
[Dε
t
∣∣∣Yεt
] = EQ [ϕ(Xεt , Zε
t ) | Yεt ] = π
εt (ϕ)
2 Introduce function valued dual process, vε, satisfying a BSPDE3 Ansatz v0, ρ0,π0
4 Asymptotic expansion of vε = v0 +ψ+ R; ψ the corrector, and R the remainder5 Utilize homogenization estimates [6], estimates for BSPDE [7] and Feynman-Kac
representations with FBDSDE [8] to prove convergence,
Dual Convergence Implies Filter Convergence
E[∣∣ρε,x
T (ϕ) − ρ0T(ϕ)
∣∣p]6 ∫R2E[∣∣vε
0 (x, z) − v00(x)
∣∣p]QXε0 ,Zε
0(dx, dz)
Beeson (University of Illinois) September 7th, 2017 12 / 28
Reduced Order Dynamic Data Assimilation
Mathematical Tools and Proof of Convergence Approach
1 Introduce an unnormalized conditional measure
Unnormalized Conditional Measure ρεt
ρεt (ϕ)
ρεt (1)=EPε
[ϕ(Xε
t , Zεt )Dε
t
∣∣∣Yεt
]EPε
[Dε
t
∣∣∣Yεt
] = EQ [ϕ(Xεt , Zε
t ) | Yεt ] = π
εt (ϕ)
2 Introduce function valued dual process, vε, satisfying a BSPDE3 Ansatz v0, ρ0,π0
4 Asymptotic expansion of vε = v0 +ψ+ R; ψ the corrector, and R the remainder5 Utilize homogenization estimates [6], estimates for BSPDE [7] and Feynman-Kac
representations with FBDSDE [8] to prove convergence,
Dual Convergence Implies Filter Convergence
E[∣∣ρε,x
T (ϕ) − ρ0T(ϕ)
∣∣p]6 ∫R2E[∣∣vε
0 (x, z) − v00(x)
∣∣p]QXε0 ,Zε
0(dx, dz)
Beeson (University of Illinois) September 7th, 2017 12 / 28
The Nudged Particle Filter Method
Table of Contents
1 Introduction and Motivation
2 Reduced Order Dynamic Data Assimilation
3 The Nudged Particle Filter Method
4 Adaptive Observations - Dynamically Steering the Measurement Process
5 Summary
6 Reporting
7 References
Beeson (University of Illinois) September 7th, 2017 13 / 28
The Nudged Particle Filter Method
Particle Filters - Discrete Time [11]
true path
prediction
prior
observation
Continuous signal, discrete observations:
dXt = b(Xt)dt + σ(Xt)dWt and Ytk = h(Xtk) + Btk
Beeson (University of Illinois) September 7th, 2017 14 / 28
The Nudged Particle Filter Method
Particle Filters - Discrete Time [11]
true path
prediction
prior
observation
1 xi ∈ RmNsi=1, an ensemble of particles.
2 wi, normalized weights:∑Ns
i=1 wi = 1.
Approximation of posterior distribution at time tk
πk(x|y0:k) ≈Ns∑i=1
wikδ(x − xi
k)
Beeson (University of Illinois) September 7th, 2017 14 / 28
The Nudged Particle Filter Method
Particle Filters - Discrete Time [11]
true path
prediction
prior
observation
Sequential Importance Sampling - SIS
πk(x|y0:k) ∝ ψ(x), xik ∼ q(x), then wi
k ∝ψ(x)q(x)
ψ, can be evaluated
q, easy to draw samples from
Beeson (University of Illinois) September 7th, 2017 14 / 28
The Nudged Particle Filter Method
Particle Filters - Discrete Time [11]
true path
prediction
prior
observation
Weights Update
wik+1 ∝
u(yk+1|xik+1)u(x
ik+1|x
ik)
q(xik+1|x
ik)
wik
Typically (for simplicity) choose: q(xik+1|x
ik) = u(xi
k+1|xik)
Nudged Particle Filter
Choose q(xik+1|x
ik) in an intelligent, but flexible hands-off manner
Beeson (University of Illinois) September 7th, 2017 14 / 28
The Nudged Particle Filter Method
Particle Filters - Discrete Time [11]
true path
prediction
prior
observation
δt
Δtt t+Δt
ave. windowtrans.
Heterogenous Multiscale Method (HMM) for Homogenized Hybrid PF (HHPF),
dX0t = b(X0
t )dt + σ(X0t )dWt and Ytk = h(X0
tk) + Btk
Doeblin Condition
For every fixed x, the solution Zxt of
dZxt = f (x, Zx
t )dt + g(x, Zxt )dVt
is ergodic and converges rapidly to its unique stationary distribution µx.
Beeson (University of Illinois) September 7th, 2017 14 / 28
The Nudged Particle Filter Method
Nudging of particles
Standard Par)cle filter
Not very efficient !
Par$cle filter with proposal transi$on
density
Continuous signal, discrete observations:
dXt = b(Xt)dt + σ(Xt)dWt and Ytk = h(Xtk) + Btk
Nudge particles:
dXit =
(b(Xi
t) + uit
)dt + σ(Xi
t)dWt, t ∈ (tk, tk+1).
Beeson (University of Illinois) September 7th, 2017 15 / 28
The Nudged Particle Filter Method
Multiscale Lorenz ’96 Model [13], [14]
1 Mid-Latitude Atmospheric Dynamics2 Linear Dissipation3 External Forcing F4 Quadratic Advection-Like Terms
(Conserve Total Energy)5 Chaotic for a wide range of F, hx, hz
X1
X2
X3
X4 X5
X6
X7
X8
Zj,1
dXkt = (Xk−1
t (Xk+1t − Xk−2
t ) − Xkt + F +
hx
J
J∑j=1
Zk,jt )dt k = 1, . . . , K,
dZk,jt =
1ε
(Zk,j+1
t (Zk,j−1t − Zk,j+2
t ) − Zk,jt + hzXk
t
)dt j = 1, . . . , J.
Beeson (University of Illinois) September 7th, 2017 16 / 28
The Nudged Particle Filter Method
Multiscale Lorenz ’96 Model [13], [14]
1 Mid-Latitude Atmospheric Dynamics2 Linear Dissipation3 External Forcing F4 Quadratic Advection-Like Terms
(Conserve Total Energy)5 Chaotic for a wide range of F, hx, hz
X1
X2
X3
X4 X5
X6
X7
X8
Zj,1
dXkt = (Xk−1
t (Xk+1t − Xk−2
t ) − Xkt + F +
hx
J
J∑j=1
Zk,jt )dt + σxdWk
t , k = 1, . . . , K,
dZk,jt =
1ε
(Zk,j+1
t (Zk,j−1t − Zk,j+2
t ) − Zk,jt + hzXk
t
)dt +
1√εσzdVj
t, j = 1, . . . , J.
Beeson (University of Illinois) September 7th, 2017 16 / 28
The Nudged Particle Filter Method
Nudged HHPF (HHPFc) on Lorenz ’96: 36 slow, 360 fast
Figure: PF, Observations every 36 hours Figure: Nudged HHPF, Observations every 72hours
Legend:
Truth (Top 3 Plots), Error (Bottom Plot); Filter mean with 1 std and 2 std
Beeson (University of Illinois) September 7th, 2017 17 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Table of Contents
1 Introduction and Motivation
2 Reduced Order Dynamic Data Assimilation
3 The Nudged Particle Filter Method
4 Adaptive Observations - Dynamically Steering the Measurement Process
5 Summary
6 Reporting
7 References
Beeson (University of Illinois) September 7th, 2017 18 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Information TheoryMotivation and Sensor Placement / Control
Figure: NCAR Globe, 500km SectorsFigure: UAV Flight Controls [5]
Improved posterior yields better prior for next observation cycle (e.g. predictionor forecasting)
Information theory provides general tool for describing improvement inknowledge (uncertain) of random variables
A useful ‘metric’ is Kullback-Leibler divergence - for filtering, expectation of‘distance’ between posterior and prior over all possible observations
Beeson (University of Illinois) September 7th, 2017 19 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Basic Tools in Information Theory
Shannon Entropy
Shannon entropy, an absolute entropy,
H(X) ≡ −
∫X
p(x) log p(x)µ(dx)
quantifies the information content of a random variable. It can be interpreted as howmuch uncertainty there is about the random variable.
Entropy of Normal Random Variable
If X ∼ N(ν,Σ), thenH(X) = log((2πe)d|Σ|)/2,
where | · | will denote the determinant and d is the dimension of Σ ∈ Rd×d.
Conditional Entropy
H(X|Y) ≡ −
∫X×Y
p(x, y) log p(x|y)µ(dx, dy)
Beeson (University of Illinois) September 7th, 2017 20 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Maximization of Kullback-Leibler Divergence
Definition (Kullback-Leibler Divergence (DKL))
A relative entropy that quantifies the ’distance’ between two densities. Given densitiesp and q, their KL divergence is defined as:
DKL(p||q) ≡∫X
p(x) log (p(x)/q(x))µ(dx)
If p is the actual density for a random variable X, then DKL(p||q) can be interpreted asthe loss of information due to using q instead of p.
Discrete Time Objective Functional
J(uk|y0:k−1) =
∫Y
DKL(p(xk|y0:k; uk) || p(xk|y0:k−1; uk))p(yk|y0:k−1; uk)dyk
=...
= H∣∣y0:k−1
(Xk) − H∣∣y0:k−1
(Xk|Yk; uk)
Beeson (University of Illinois) September 7th, 2017 21 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Maximization of Kullback-Leibler Divergence
Definition (Kullback-Leibler Divergence (DKL))
A relative entropy that quantifies the ’distance’ between two densities. Given densitiesp and q, their KL divergence is defined as:
DKL(p||q) ≡∫X
p(x) log (p(x)/q(x))µ(dx)
If p is the actual density for a random variable X, then DKL(p||q) can be interpreted asthe loss of information due to using q instead of p.
Discrete Time Objective Functional
J(uk|y0:k−1) =
∫Y
DKL(p(xk|y0:k; uk) || p(xk|y0:k−1; uk))p(yk|y0:k−1; uk)dyk
=...
= H∣∣y0:k−1
(Xk) − H∣∣y0:k−1
(Xk|Yk; uk)
Beeson (University of Illinois) September 7th, 2017 21 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Information Theoretic Cost Functionals
Maximization of Kullback-Leibler Divergence
Definition (Kullback-Leibler Divergence (DKL))
A relative entropy that quantifies the ’distance’ between two densities. Given densitiesp and q, their KL divergence is defined as:
DKL(p||q) ≡∫X
p(x) log (p(x)/q(x))µ(dx)
If p is the actual density for a random variable X, then DKL(p||q) can be interpreted asthe loss of information due to using q instead of p.
Discrete Time Objective Functional
J(uk|y0:k−1) =
∫Y
DKL(p(xk|y0:k; uk) || p(xk|y0:k−1; uk))p(yk|y0:k−1; uk)dyk
=...
= H∣∣y0:k−1
(Xk) − H∣∣y0:k−1
(Xk|Yk; uk)
Beeson (University of Illinois) September 7th, 2017 21 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
Table of Contents
1 Introduction and Motivation
2 Reduced Order Dynamic Data Assimilation
3 The Nudged Particle Filter Method
4 Adaptive Observations - Dynamically Steering the Measurement Process
5 Summary
6 Reporting
7 References
Beeson (University of Illinois) September 7th, 2017 22 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
nVortex FlowfieldModel
1 Deterministic vortexdynamics simulates theEuler equations
2 The first random pointvortex method tosimulate viscousincompressible flow wasintroduced in [15]
J ≡[
0 1−1 0
],
dXi,t =1
2π
n∑k=1
Γk
‖Xk,t − Xi,t‖22J(Xk,t − Xi,t)dt +
√σxdWi,t, Xi,0 = x ∈ R2,
Beeson (University of Illinois) September 7th, 2017 23 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
A Controllable Tracer for Adaptive Observations
Assume that tracer is controllable,
dXi,t = fi(Xt) + bi(ui,t) +√σxdBi,t and ui,t ∈ U
Ytk = h(Xtk) + Btk
where U is some admissible control set.
How one might control the tracer so as to best improve the filtering process?
One approach, formulation of an optimal control problem; specifically in terms ofinformation theoretic quantities.
Beeson (University of Illinois) September 7th, 2017 24 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
PF Implementation and Result
Figure: PF with no controlFigure: Controlled PF with RHC over 10observation steps
Beeson (University of Illinois) September 7th, 2017 25 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
PF Implementation and Result
Figure: PF, no control - posterior entropyFigure: PF, control with RHC over 10 observationsteps - posterior entropy
* Cost Function shown is: −H(X|Y).
Beeson (University of Illinois) September 7th, 2017 25 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
PF Implementation and Result
Figure: PF, no control - Vortex-1 x-stateFigure: PF, control with RHC 10 observationsteps - Vortex-1 x-state
Figure: x-state shown in top figure, while RMSE shown in bottom
Beeson (University of Illinois) September 7th, 2017 26 / 28
Adaptive Observations - Dynamically Steering the Measurement Process Sensor Control
PF Implementation and Result
Figure: PF, no control - Vortex-1 y-stateFigure: PF, control with RHC 10 observationsteps - Vortex-1 y-state
Figure: x-state shown in top figure, while RMSE shown in bottom
Beeson (University of Illinois) September 7th, 2017 26 / 28
Summary
Objectives:
(I) Develop an integrated framework that combines the ability to dynamically steer themeasurement process, extracting useful information, with nonlinear filtering forinference and prediction of large scale complex systems.
(II) Develop efficient and robust methods to produce lower-dimensional recursivenonlinear filtering equations driven by the observations; particle filters for theintegration of observations with the simulations of large-scale complex systems.
Presented:
(i) Use of powerful mathematical techniques - homogenization, SPDE, FBDSDE - toderive convergence results of correlated filter in multiscale problems as well asprovide future mechanisms for extension of nudging particle method andinformation flow for the multiscale correlated noise case.
(ii) Introduced framework by which to breakdown adaptive observation problem intohierarchy of sensor selection / placement and sensor control problems.
(iii) Described preliminary algorithms using information theoretic cost functionals todrive the sensor placement and control problems - demonstrations on test bedproblems.
Beeson (University of Illinois) September 7th, 2017 27 / 28
Reporting
Journal Articles:Namachchivaya, N. Sri; Random dynamical systems: addressing uncertainty,nonlinearity and predictability; Meccanica, (51, 2975-2995); 2016https://link.springer.com/article/10.1007%2Fs11012-016-0570-4
Lingala, N., Namachchivaya, N. Sri, et al.; Random perturbations of a periodically drivennonlinear oscillator: escape from a resonance zone; Nonlinearity, (30, 4, 1376); 2017http://iopscience.iop.org/article/10.1088/1361-6544/aa5dc7/meta
Yeong, H. C., et al. Particle Filters with Nudging in Multiscale Chaotic Systems: withApplication to the Lorenz-96 Atmospheric Model; Submitted to ZAMM, Journal ofApplied Mathematics and Mechanics
Beeson, R., et al., Dynamic Data-Driven Adaptive Observations in a Vortex Flowfield; InPreparation to European Journal of Applied Mathematics
Conference Proceedings:Beeson, R., et al., Dynamic Data-Driven Adaptive Observations in a Vortex Flowfield; 9thEuropean Nonlinear Dynamics Conference; Budapest Hungary; June 2017
Yeong, H.C., et al. Particle Filters with Nudging in Multiscale Chaotic Systems: withApplication to the Lorenz-96 Atmospheric Model; Budapest Hungary; June 2017
Lingala, N., Namachchivaya, N. Sri, et al.; Random perturbations of a periodically drivennonlinear oscillator: escape from a resonance zone; SIAM Conference on DynamicalSystems 2017, Snowbird UtahBeeson (University of Illinois) September 7th, 2017 28 / 28
References
[WBC09] W. M. Washington, L. Buja, and A. Craig. “The computational future for climate and Earth systemmodels: on the path to petaflop and beyond”. In: Phil. Trans. Royal Soc. A: Mathematical,Physical and Engineering Sciences 367.1890 (2009), pp. 833–846.
[Lor84] E. N. Lorenz. “Irregularity: A Fundamental Property of the Atmosphere”. In: Tellus 36A (1984),pp. 98–110.
[Maa04] Leo R. M. Maas. “Theory of basin scale dynamics of a stratified rotating fluid”. In: Surveys inGeophysics 25 (2004), pp. 249–279.
[Van03] Lennart Van Veen. “Overturning and wind driven circulation in a low-order ocean-atmospheremodel”. In: Dynamics of Atmospheres and Oceans 37 (2003), pp. 197–221.
[EA15] Ahmed EA. “Modeling of a Small Unmanned Aerial Vehicle”. In: 9 (Apr. 2015), pp. 413–421.
[PV03] E. Pardoux and A. Y. Veretennikov. “On Poisson equation and diffusion approximation 2”. In:Ann. Probab. 31.3 (2003), pp. 1166–1192.
[Roz90] Boris L. Rozovskii. Stochastic Evolution System: Linear Theory and Aplications to Non-linearFiltering. Dordrecht: Kluwer Academic Publishers, 1990, p. 315.
[PP94] Etienne Pardoux and Shige Peng. “Backward Doubly Stochastic Differential Equations andSystems of Quasilinear SPDEs”. In: Probab. Theory Related Fields 98 (1994), pp. 209–227.
[Zak69] M. Zakai. “On the optimal filtering of diffusion processes”. In: Probability Theory and RelatedFields 11.3 (1969).
[Par79] Etienne Pardoux. “Stochastic partial differential equations and filtering of diffusion processes”.In: Stochastics 3 (1979), pp. 127–167.
[GSS93] N. J. Gordon, D. J. Salmond, and A. F. M. Smith. “Novel approach to nonlinear/non-GaussianBayesian state estimation”. In: IEEE Proceedings F 140.2 (1993), pp. 107–113.
Beeson (University of Illinois) September 7th, 2017 28 / 28
References
[Fle82] Wendell H. Fleming. “PLogarithmic transformations and stochastic control”. In: Advances inFiltering and Optimal Stochastic Control: Proceedings of the IFIP-WG 7/1 Working ConferenceCocoyoc, Mexico, February 1–6, 1982. Berlin, Heidelberg: Springer Berlin Heidelberg, 1982,pp. 131–141. isbn: 978-3-540-39517-1. doi: 10.1007/BFb0004532. url:http://dx.doi.org/10.1007/BFb0004532.
[Lor95] E. N. Lorenz. Predictability: A problem partly solved. 1995.
[MTV01] A.J. Majda, I. Timofeyev, and E. Vanden-Eijnden. “A Mathematical Framework for StochasticClimate Models”. In: Comm. Pure Appl. Math. 54 (2001), pp. 891–974.
[Cho73] A. J. Chorin. “Numerical study of slightly viscous flow”. In: Journal of Fluid Mechanics 57.4(1973), pp. 785–796.
[CT06] T. M. Cover and J. A. Thomas. Elements of Information Theory. NJ, USA: Wiley Series inTelecommunications and Signal Processing, John Wiley & Sons Inc., 2006.
Beeson (University of Illinois) September 7th, 2017 28 / 28