data assimilation for high-dimensional systems...data assimilation for high-dimensional systems...
TRANSCRIPT
![Page 1: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/1.jpg)
Data Assimilation for High-DimensionalSystems
− Challenges, algorithms and opportunities
Wei KangU.S. Naval Postgraduate School
2018 IMA Workshop
![Page 2: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/2.jpg)
Collaboration
• Liang Xu (NRL)
• Sarah King (NRL)
• Kai Sun (UTK)
• Junjian Qi (UCF)
• Isaac Kaminer (NPS)
• Qi Gong (UCSC)
• Lucas Wilcox (NPS)
• Arthur J. Krener (NPS)
• Randy Boucher (Army)
• Data assimilation
• Numerical weatherprediction
• Partial observability
• Power systems
• Swarms of autonomousvehicles
• Numerical algorithms
![Page 3: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/3.jpg)
High-Dimensional Systems
Numerical Weather Prediction- ECMWF Global Model
ECMWF Global Model: 8× 107
model variables are updated ev-ery 12 hours using 1.3×107 ob-servations
Power System- Eastern Interconnection
EI: A largest electrical gridin North America. A simpli-fied model has more than 25Kbuses, 28K lines, 8K transform-ers, 1, 000 generators.
![Page 4: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/4.jpg)
High-Dimensional Systems
ChallengesThe scalability of algorithms is limited by several factors: computationalload, I/O overhead and required memory size, degree of parallelism, andpower consumption.
Covariance Matrixdimension vs RAM size
102 103 104 105 106 107 108
Dimension
10-5
100
103
106
108
Me
mo
ry (
GB
)
GB
TB
PB
Quantitative ChangeBecomes Qualitative
![Page 5: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/5.jpg)
Data Assimilation
Liang Xu - NRL
Numerical Weather Prediction
Observation(y)
Background(xb)
DATA ASSIMILATION
Analysis(xa)
FORECASTMODEL Forecast
![Page 6: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/6.jpg)
Data Assimilation
System Model
xk+1 = M(xk ) + ηk , xk ∈ Rn, ηk ∼ model uncertainty,Qyk = H(xk ) + δk , yk ∈ Rm, δk ∼ sensor noise,R
Linearizationxk+1 = Mkxk + · · ·yk = Hkxk + · · ·
Both n and m are large. In daily operations, only a small part of sensordata is used.
![Page 7: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/7.jpg)
4D-Var
Cost function
J (x0) =1
2(x0−xb)T (Pb
0 )−1(x0−xb)+1
2
N∑i=0
(yi−H(xi ))TR−1(yi−H(xi ))
4D-Var estimation{min
xa0
J (xa0 )
xak =M(xa
k−1), k = 1, 2, · · · ,N
The trajecotry, xa0 , x
a1 , · · · , xa
N , is called an analysis; xb is the initialestimate, or background; Pb
0 is a positive definite matrix (fixed).
![Page 8: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/8.jpg)
4D-Var
A linear solution
xak = xb
k + gk , 0 ≤ k ≤ N
g = PHT(HPHT + R
)−1(y − Hxb)
g =
g0...gN
,P = (Pij )Ni ,j=0,H = diag(H0, · · · ,HN),R, y , xb, · · ·
Or define
z =(HPHT + R
)−1(y − Hxb)
Thenxa
k = xbk + gk , 0 ≤ k ≤ N
g = PHT zHPHT z + Rz = (y − Hxb)
![Page 9: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/9.jpg)
4D-Var
A 4D-Var algorithm
fN+1 = 0for k = N : 0
fk = MTk fk+1 + HT zk ,
g0 = Pb0 f0
for k = 0 : Ngk+1 = Mkgk ,
• Mkgk and MTk fk are computed using a tangent linear model and
an adjoint model.
• The equation HPHT + Rzk = (y − Hxb) is solved using conjugategradient algorithm.
![Page 10: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/10.jpg)
4D-Var
• 4D-Var has been widely used in today’s numerical weatherprediction (NWP). It was adopted shortly after the introduction of3D-Var.
• Historically 3D-Var and 4D-Var are inspired by optimal control inwhich a cost function is minimized (Sasaki (1958), Lorenc (1986)).
• It is an effective method to provide estimation results with anaffordable computational load.
• The method does not provide information about error covariance.
• It requires the development and maintenance of tangent linearmodels and adjoint models.
![Page 11: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/11.jpg)
EnKF
Ensemble KF
Ensemble x ik|k , 1 ≤ i ≤ Nens Nens << n
Forecast x ik+1|k =M(x i
k|k ) + ηk
y ik+1|k = H(x i
k|k )
xk+1 = 1Nens
Nens∑i=1
x ik+1|k
yk+1 = 1Nens
Nens∑i=1
y ik+1|k
![Page 12: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/12.jpg)
EnKF
Ensemble KF (Cont.)
Analysis ∆X = 1√Nens−1
([x1k+1|k · · · xNens
k+1|k
]− xk+11
)n × Nens
∆Y = 1√Nens−1
([y1
k+1|k · · · yNens
k+1|k
]− yk+11
)m × Nens
K = ∆X∆Y T (∆Y∆Y T + R)−1 n ×mxk+1|k+1 = xk+1 + K (yk+1 − yk+1)
Update D = I −∆Y T (∆Y∆Y T + R)−1∆Y Nens × Nens
Pk+1|k+1 = ∆XD∆XT[x1k+1|k+1 x2k+1|k+1 · · · xNens
k+1|k+1
]= xk+1|k+1 + ∆X (
√(Nens − 1)D)T
![Page 13: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/13.jpg)
EnKF
• EnKF does not require tangent linear model and adjoint model
• It contains partial information about error statistics
• Undersampling and rank deficiency
• Filter divergence
• Inbreeding - systematically underestimate the analysis errorcovariance
• Spurious correlations - covariance between state components thatare not physically related
![Page 14: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/14.jpg)
EnKF
Localization - components physically far a way are uncorrelated.
Set Pij = 0 for i and j far awaySuppose y is a part of state variables at i1, · · · , im. Let ρ be a sparsematrix that defines the sparsity of P. Define ρxy ∈ Rn×m andρyy ∈ Rm×m
Pxyst = ρis it ,P
yyst = ρis it
Then
Kloc = ρxy . ∗ (∆X∆Y T )(ρyy . ∗ (∆Y∆Y T ) + R
)−1Inflation - rescale ∆X and ∆Y by a small factor.
![Page 15: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/15.jpg)
Sparsity-based Filters
Sparsity-based filters: The goal is to avoid rank deficiency, providemore error covariance information, and achieve granularity control foroptimal parallelism.
A variety of parallel computing architectures are available; and newtechnologies are being developed rapidly.
• Multi-core CPU
• General-purpose GPU
• Clusters or massively parallel computing
• Grid computing
• Application-specific integrated curcuits
• ......
![Page 16: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/16.jpg)
Sparsity-based Filters
Sparsity based methods
• Approximately sparse error covariance
Nsp = maximum number of nonzero entries in columns
Ii (P) = indices of nonzero entries in the ith-column
• Component-based numerical model
M(x spk ; I) or Mcomp
I = indices of entries to be evaluated
![Page 17: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/17.jpg)
Sparsity-based Filters
A progressive approach
AssumeMkPkM
Tk = Pk + ∆Pk+1
To estimate ∆Pk+1, assume
Mk+1 = I + ∆Mk
xk+1 =M(xk ) = xk + ∆(xk )
ThenMkPkM
Tk = (I + ∆Mk )Pk (I + ∆MT
k )
= Pk + ∆MkPk + (∆MkPk )T + · · ·≈ (M(xk + δPk )−M(xk )) /δ
+ (M(xk + δPk )−M(xk ))T /δ − Pk
![Page 18: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/18.jpg)
Sparsity-based Filters
Prograssive KF
Background xk|k and Pspk|k (sparse covariance approximation)
Forecast xk+1|k =M(xk|k )yk+1|k = H(xk|k )
Pspk+1|k =
(Mcomp(x sp
k|k + δPspk|k )−Mcomp(x sp
k|k ))/δ
+(Mcomp(x sp
k|k + δPspk|k )−Mcomp(x sp
k|k ))T
/δ
−Pspk|k + Q
Analysis K = Pspk+1|kH
Tk+1(Hk+1P
spk+1|kH
Tk+1 + R)−1
Pspk+1|k+1 = (I − KHk+1)Psp
k+1|kxk+1|k+1 = xk+1|k + K (yk+1 − yk+1|k )
![Page 19: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/19.jpg)
Sparsity-based Filters
Computational loadProgressive KF number of modelFull model components evaluation
M(xk )M (xk + δPk (:, i)) (n + 1)nNp
i = 1, 2, · · · , n Np- progressive steps
Progressive KFComponent-based model
M(xk )M (xk + δPk (:, i), Ii (P)) (Nsp + 1)nNp
i = 1, 2, · · · , nEnsemble KF
M(x ik )
i = 1, 2, · · · ,Nens Nensn
![Page 20: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/20.jpg)
Sparsity-based Filters
Avoid rank deficiency and achieve granularity control
Ensemble
XXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXX
Sparse Covariance
X X XX X X XX X X X X0 0 0 0 X X0 0 0 0 0 X X0 0 0 0 0 0 X X0 0 0 0 0 0 0 X X0 0 0 0 0 0 0 0 X X0 0 0 0 0 0 X X X X X0 0 0 0 0 0 0 X X X X X0 0 0 0 0 0 0 0 X X X X X0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 X X X
ØP has full rankØNsp is a variableØTasks can be grouped in different size:
Coarse-grained, medium-grained, fine-grainedØP localization is straightforward
Rank:Nens
Size: n x Nens
![Page 21: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/21.jpg)
Sparsity-based Filters
Lorenz-96 model
dxi
dt= (xi+1 − xi−2)xi−1 − xi + F , i = 1, 2, · · · ,m
xm+1 = x1
Discretization - 4th-order RK
xk =M(xk−1)
∆t = 0.025
F = 8
m = 400 5 10 15 20 25
t
-6
-4
-2
0
2
4
6
8
10
12
x
Lorenz-96 Chaotic Trajectory
![Page 22: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/22.jpg)
Sparsity-based Filters
A comparisonN = 1000 initial states in [−1 1] - uniform distribution.Nfilter = 4000 filter stepsm = 20 measurement locationsR = I
Filter Size CMPT
EVAL
EnKF Nens = 10 400
P-KF Nsp = 7
Np = 1 320
P-KF Nsp = 11
Np = 2 480x2
P-KF Nsp = 11
Np = 3 480x3 EnKF PKF PKF PKF UKF
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
RM
SE
EnKF vs Progressive-KF
![Page 23: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/23.jpg)
Sparsity-based Filters
Unscented KF (UKF)
σ-points x ik|k , 0 ≤ i ≤ 2n
x0k|k = xk|k
Forecast x ik+1|k =M(x i
k|k ), y ik+1|k = H(x i
k|k ), 0 ≤ i ≤ 2n
xk+1 =2n∑
i=0
wixik+1|k , yk+1 =
2n∑i=0
wiyik+1|k
Pk+1|k =2n∑
i=0
wi ∆x ik+1(∆x i
k+1)T + Q
∆x ik+1 = x i
k+1|k − xk+1
w0 = κn+κ , wi = κ
2(n+κ)
![Page 24: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/24.jpg)
Sparsity-based Filters
UKF (Cont.)
Analysis Pyy =2n∑
i=0
wi ∆y ik+1(∆y i
k+1)T + R, ∆yk+1 = y ik+1|k − yk+1
Pxy =2n∑
i=0
wi ∆x ik+1(∆y i
k+1)T
KPyy = Pxy
xk+1|k+1 = xk+1 + K (yk+1 − yk+1)
Update Pk+1|k+1 = Pk+1|k − K (Pxy )T
x ik+1|k+1 = xk+1|k+1 +
√(n + κ)Pk+1|k+1, i = 1, 2, · · · , n
x ik+1|k+1 = xk+1|k+1 −
√(n + κ)Pk+1|k+1, i = n + 1, · · · , 2n
![Page 25: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/25.jpg)
Sparsity-based Filters
Sparsity of square root matrix
Theorem (S. Toledo). If P is a symmetric positive definite matrix. Theamount of storage for a Cholesky decomposition of P is O(n + 2η(P)),where η(P) is the number of nonzero entries in P.
Assumption: The sparsity patterns of P and√P are known - I (P),
I (√P).
![Page 26: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/26.jpg)
Sparsity-based Filters
Sparse UKF
Sparse x0k|k = xk|kσ-points σi , Ii (sparsity index) 1 ≤ i ≤ n
Forecast x0k+1|k =M(x0k|k ),
x ik+1|k =Mcomp(x0k|k + σi ), x i+n
k+1|k =Mcomp(x0k|k − σi )
y ik+1|k = H(x i
k+1|k.Iix0k+1|k ), 1 ≤ i ≤ 2n
xk+1 =2n∑
i=0
wi (xik+1|k.Ii
x0k+1|k ), yk+1 =2n∑
i=0
wiyik+1|k
Pspk+1|k =
2n∑i=0
wi
(∆x i
k+1(∆x ik+1)T
)sp+ Q
w0 = κn+κ , wi = κ
2(n+κ) , ∆x ik+1 = x i
k+1|k.Iix0k+1|k − xk+1
x sp1 .I x2 - merging operation.
![Page 27: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/27.jpg)
Sparsity-based Filters
Sparse UKF (Cont.)
Analysis Pyy =2n∑
i=0
wi ∆y ik+1(∆y i
k+1)T , ∆y ik+1 = y i
k+1|k − yk+1
Pxy =2n∑
i=0
wi ∆x ik+1(∆y i
k+1)T
KPyy = Pxy
xk+1|k+1 = xk+1 + K (yk+1 − yk+1)
Update Pspk+1|k+1 = Psp
k+1|k −(K (Pxy )T
)sp
σi , Ii ∼√
(n + κ)Pspk+1|k+1, i = 1, 2, · · · , n
![Page 28: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/28.jpg)
Sparsity-based Filters
A comparisonN = 1000 initial states in [−1 1] - uniform distribution.Nfilter = 4000 filter stepsm = 20 measurement locationsR = I
Filter Size CMPT
EVAL
EnKF Nens = 10 400
S-UKF Nsp = 7 640
S-UKF Nsp = 11 960
EnKF S-UKF S-UKF UKF
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
RM
SE
EnKF vs Sparse-UKF
![Page 29: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/29.jpg)
Summary
Progressive KF
• Full rank covariance
• Sparse covariance matrix reduces I/O load and memory size.
• Component-based model reduces computational load.
• Requirement: Pk+1 ≈ Pk + MPk + PkMT
Sparse UKF
• Full rank covariance
• Sparse covariance matrix reduces I/O load and memory size.
• Component-based model reduces computational load.
• Requirement: Cholesky decomposition (additional computationand memory).
![Page 30: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/30.jpg)
Partial Observability
Partial Observability - A quantitative definition
– x - space variable, λ - sensor location, u(x , t) - state trajectory
– y(t) = h(u(x , t), λ) - system output (sensor data)
– Background variation (data assimilation cost function)
J(u, δu, λ) = δuTP1δu + ||y(·, λ; u + δu)− y(·, λ; u)||P2
Definition. Let ε > 0 be a positive number. The observabilityambiguity, ρ, is defined as
ρ2 = maxδu||P(δu)||W
subject to: ||J(uB , δu, λ)|| ≤ εsystem model of u(x , t)
P(δu) ∈W (subspace for estimation)
The ratio, ρ/ε, is called the unobservability index.
![Page 31: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/31.jpg)
Partial Observability
Empirical Gramian
• Let {wi} be an orthonormal basis in W . Let ∆y(t) be thevariation of the output. The Gramian is defined as follows
G = 〈wi ,wj〉P1 + 〈∆yi (t),∆yj (t)〉P2 .
• Let σmin be the smallest eigenvalue of G then 1/√σmin is a first
order approximation of the unobservability index ρ/ε.
• The large scale and high dimensions in NWP make it less desirableor even impossible to make the entire state space observable. Afinite number of modes is enough to provide an accurateapproximation.
![Page 32: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/32.jpg)
Shallow Water Equations (SWEs)
Simplest atmospheric flow model
∂u
∂t+ u
∂u
∂x+∂φ
∂x+ g
∂P
∂x= 0
∂φ
∂t+ u
∂φ
∂x+ φ
∂u
∂x= 0
– x- horizontal distance, u-horizontal velocity, p-fluid depth
– g -gravitational constant, P(x)-height of the obstacle
– φ = gp-geopotential.
0 1 2 3 4 5 6 7 8 9 10
0
0.02
0.04
0.06
0.08
0.1
0.12
0.14
0.16
0.18
0.2
0 1 2 3 4 5 6 7 8 9 100
1
2
3
0.198
0.1985
0.199
0.1995
0.2
0.2005
0.201
0.2015
0.202
t
x
Initial conditions Solution
![Page 33: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/33.jpg)
Algorithm for Optimal Sensor Placement
Optimal sensor location - maximizing the unobservability index
maxσmin(λ)
subject to: λlower < λ < λupper
where σmin is the smallest eigenvalue of G .
1 Initialize λ0, P1, P2
2 Approximate the current solution to the maximization problem bycomputing the smallest eigenvalue σ of the gramian matrix usingthe TLM.
3 Update λ using BFGS step
mk (λ) = σ(λk ) +
(dσ
dλ k
)T
(λ− λk ) +1
2(λ− λk )TBk (λ− λk )
where B is the pseudo-Hessian.
4 Repeat until converges
![Page 34: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/34.jpg)
Parameters
Nλ = 6 six sensorsNcoef = 5 number of coefficients for each u, φL = 10 length of x intervalT = 2.3 time intervalNt = 250ρ = .01R = 1.6e − 3 variance in sensor dataσb
u = 5e − 5 used to compute Lc
σbφ = .02 used to compute Lc
L−1c =
(σb
u L 0
0 σbφL
)weight matrix P1
where L = I + γ−1(
c4
2∆x4(Lxx )2
), Lxx =
−2 1 0 · · · 0 11 −2 1 · · · 0 0...
. . .
0 11 0 · · · 1 −2
![Page 35: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/35.jpg)
Monte Carlo Numerical Experiments
0 1 2 3 4 5 6 7 8 9 10
optimalequalrandom 1random 2
ρ/ε RMSE ua(0) RSME of ua RMSE φa(0) RSME of φa
Optimal 9.29 0.3577 0.1174 0.5186 0.1591Equal 13.05 0.4419 0.1484 0.5787 0.1974Rand 1 19.08 0.4323 0.1459 0.5734 0.1951Rand 2 14.17 0.4245 0.1431 0.5671 0.1914
Equal Imp 28.8% 19.1 % 20.9 % 10.4 % 19.4 %Rand 1 Imp 51.3 % 17.3 % 19.5 % 9.6% 18.45 %Rand 2 Imp 34.4 % 15.7 % 17.96 % 8.6 % 16.88 %
The optimal sensor locations improved the estimation accuracy in the 4D-Var dataassimilation for the observed variable φ and the unobserved variable u.
![Page 36: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/36.jpg)
Some Remarks
• Partial observability and estimation for large scale systems havealso been applied to power systems and networked swarms.
• Unobservability index is a worst-case measure of observability.Other quantitative measure exists for various types of systems andsensor networks.
• User-knowledge and observability?
• Estimation of variables observable in a zero measure set?
![Page 37: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/37.jpg)
Some Remarks
• Partial observability and estimation for large scale systems havealso been applied to power systems and networked swarms.
• Unobservability index is a worst-case measure of observability.Other quantitative measure exists for various types of systems andsensor networks.
• User-knowledge and observability?
• Estimation of variables observable in a zero measure set?
![Page 38: Data Assimilation for High-Dimensional Systems...Data Assimilation for High-Dimensional Systems Challenges, algorithms and opportunities Wei Kang U.S. Naval Postgraduate School 2018](https://reader034.vdocuments.us/reader034/viewer/2022051512/602f743bd0f87a6b00654850/html5/thumbnails/38.jpg)
Thank you!