empirical localization of observation impact in ensemble filters jeff anderson image/dares thanks to...
TRANSCRIPT
Empirical Localization of Observation Impact in Ensemble Filters
Jeff AndersonIMAGe/DAReS
Thanks to Lili Lei, Tim Hoar, Kevin Raeder, Nancy Collins, Glen Romine, Chris Snyder, Doug Nychka
5th EnKF Workshop, 23 May 2012 1
For an observation y and state variable x;Increments for N ensemble samples of x are:
Where is a sample regression coefficient,and is a localization.
Traditionally , but here there is no upper bound.
5th EnKF Workshop, 23 May 2012 2
Definition of Localization
5th EnKF Workshop, 23 May 2012 3
Empirical Localization
Have output from an OSSE.Know prior ensemble and truth for each state variable.
5th EnKF Workshop, 23 May 2012 4
Empirical Localization
Have output from an OSSE.Know prior ensemble and truth for each state variable.Can get truth & prior ensemble for any potential observations.
5th EnKF Workshop, 23 May 2012 5
Empirical Localization
Estimate localization for set of observations and subset of state variables.e.g. state variables at various horizontal distances from observations.
5th EnKF Workshop, 23 May 2012 6
Empirical Localization
Example: how to localize impact of temperature observations (4 shown) on a U state variable that is between 600 and 800 km distant.
5th EnKF Workshop, 23 May 2012 7
Empirical Localization
Given observational error variance, can compute expected ensemble mean increment for state.Plot this vs prior state truth - ensemble mean.
5th EnKF Workshop, 23 May 2012 8
Empirical Localization
Do this for all state variables in subset.
5th EnKF Workshop, 23 May 2012 9
Empirical Localization
Do this for all state variables in subset.
5th EnKF Workshop, 23 May 2012 10
Empirical Localization
Do this for all state variables in subset.
5th EnKF Workshop, 23 May 2012 11
Empirical LocalizationFind a least squares fit.
Slope is .
Least squares minimizes:
Same as minimizing
Posterior mean
5th EnKF Workshop, 23 May 2012 12
Empirical Localization
Define set of all pairs (y, x) of potential observations and state variable instances in an OSSE.
(A state variable instance is defined by type, location and time).
Choose subsets of this set.
5th EnKF Workshop, 23 May 2012 13
Empirical Localization
Find that minimizes the RMS difference between the posterior ensemble mean for x and the true value over this subset.
This can be computed from the output of the OSSE.
Can then use this localization in a new OSSE for all (y, x) in the subset.
Call the values of localization for all subsets anEmpirical Localization Function (ELF).
5th EnKF Workshop, 23 May 2012 14
Lorenz-96 40-Variable ExamplesAssume all observations are located at a model grid point.
(Easier but not necessary).
Define 40 subsets of (y, x) pairs:x is 20 to the left, 19 to the left, … 1 to the left,colocated, 1 to the right, …, 19 to the right of y.
5th EnKF Workshop, 23 May 2012 15
Computing ELFs
Start with a climatological ensemble.
Do set of 6000-step OSSEs.(only use last 5000 steps).
First has no localization.
Compute ELF from each.
Use ELF for next OSSE.
ELF1
ELF2
ELF3
ELF4
ELF5
No Localization
5th EnKF Workshop, 23 May 2012 16
Evaluation Experiments
Start with a climatological ensemble.
Do 110,000 step assimilation, discard first 10,000 steps.
Adaptive inflation with 0.1 inflation standard deviation.
Many fixed Gaspari-Cohn localizations tested for each case.
Also five ELFs (or should it be ELVEs?).
5th EnKF Workshop, 23 May 2012 17
Case 1: Frequent low-quality obs.
Identity observations.
Error variance 16.
Assimilate every standard model timestep.
5th EnKF Workshop, 23 May 2012 18
Case 1: Frequent low-quality obs.
N=20 Gaspari Cohn (GC) function with smallest time mean prior RMSE.
5th EnKF Workshop, 23 May 2012 19
Case 1: Frequent low-quality obs.
N=20 first ELF is negative for many distances, but minimum localization is 0 when this ELF is used.
5th EnKF Workshop, 23 May 2012 20
Case 1: Frequent low-quality obs.
Subsequent N=20 ELFs are less negative, smoother, closer to best GC.
5th EnKF Workshop, 23 May 2012 21
Case 1: Frequent low-quality obs.
Subsequent N=20 ELFs are less negative, smoother, closer to best GC.
5th EnKF Workshop, 23 May 2012 22
Case 1: Frequent low-quality obs.
Subsequent N=20 ELFs are less negative, smoother, closer to best GC.
5th EnKF Workshop, 23 May 2012 23
Case 1: Frequent low-quality obs.
Subsequent N=20 ELFs are less negative, smoother, closer to best GC.
5th EnKF Workshop, 23 May 2012 24
Case 1: Frequent low-quality obs.
N=20, best GC has half-width 0.2, time mean RMSE of ~1.03.
5th EnKF Workshop, 23 May 2012 25
Case 1: Frequent low-quality obs.
N=20, best GC has half-width 0.2, time mean RMSE of ~1.03.ELFs give RMSE nearly as small as this.
5th EnKF Workshop, 23 May 2012 26
Case 1: Frequent low-quality obs.
N=20, best GC has half-width 0.2, time mean RMSE of ~1.03.ELFs give RMSE nearly as small as this.
5th EnKF Workshop, 23 May 2012 27
Case 1: Frequent low-quality obs.
Similar results for smaller ensemble, N=10. Note larger RMSE, narrower best GC half-width.
5th EnKF Workshop, 23 May 2012 28
Case 1: Frequent low-quality obs.
Similar results for larger ensemble, N=40. Note smaller RMSE, wider best GC half-width.
5th EnKF Workshop, 23 May 2012 29
Case 1: Frequent low-quality obs.
N=40 ELFs have smaller time mean RMSE than best GC.
5th EnKF Workshop, 23 May 2012 30
Case 1: Frequent low-quality obs.
ELFs are nearly symmetric so can ignore negative distances.
5th EnKF Workshop, 23 May 2012 31
Case 1: Frequent low-quality obs.
ELF for smaller ensemble is more compact.
5th EnKF Workshop, 23 May 2012 32
Case 1: Frequent low-quality obs.
ELF for larger ensemble less compact, consistent with GC results.
5th EnKF Workshop, 23 May 2012 33
Case 1: Frequent low-quality obs.
ELFs for even bigger ensembles are broader, but noisier at large distances.
5th EnKF Workshop, 23 May 2012 34
Case 1: Frequent low-quality obs.
ELFs for even bigger ensembles are broader, but noisier at large distances.
5th EnKF Workshop, 23 May 2012 35
Case 1: Frequent low-quality obs.
ELFs for even bigger ensembles are broader, but noisier at large distances.
5th EnKF Workshop, 23 May 2012 36
Case 2: Infrequent high-quality obs.
Identity observations.
Error variance 1.
Assimilate every 12th standard model timestep.
5th EnKF Workshop, 23 May 2012 37
Case 2: Infrequent high-quality obs.
For N=10, all ELF cases have smaller RMSE than best GC.
5th EnKF Workshop, 23 May 2012 38
Case 2: Infrequent high-quality obs.
For N=20, first ELF is worse than best GC; all others better.Best GC gets wider as ensemble size grows.
5th EnKF Workshop, 23 May 2012 39
Case 2: Infrequent high-quality obs.
For N=40, all ELFs have smaller RMSE.
5th EnKF Workshop, 23 May 2012 40
Case 2: Infrequent high-quality obs.
N=10 ELF is non-Gaussian. Has local minimum localization for distance 1.
5th EnKF Workshop, 23 May 2012 41
Case 2: Infrequent high-quality obs.
N=40 ELF is broader; also has local minimum for distance 1.Need a non-gaussian ELF to possibly do better than GC.
5th EnKF Workshop, 23 May 2012 42
Case 3: Integral observations.
Each observation is average of grid point plus its nearest 8 neighbors on both side; total of 17 points.
(Something like a radiance observation.)
5th EnKF Workshop, 23 May 2012 43
Case 3: Integral observations.
Each observation is average of grid point plus its nearest 8 neighbors on both side; total of 17 points.
(Something like a radiance observation.)
Error variance 1.
Assimilate every standard model timestep.
Very low information content:Assimilate 8 of these observations for each grid
point.Total of 320 observations per assimilation time.
5th EnKF Workshop, 23 May 2012 44
Case 3: Integral observations.
ELFs are not very Gaussian. No values close to 1, two peaks at distance +/- 7.
5th EnKF Workshop, 23 May 2012 45
Case 3: Integral observations.
ELFs are not very Gaussian. Best GC is much larger near the observation location.
5th EnKF Workshop, 23 May 2012 46
Case 3: Integral observations.
RMSE is a more complicated function of the GC half-width in this case.
5th EnKF Workshop, 23 May 2012 47
Case 3: Integral observations.
ELFs all have significantly smaller time mean RMSE than best GC.
5th EnKF Workshop, 23 May 2012 48
Case 4: Frequent low-quality obs., imperfect model
Identity observations.
Error variance 16.
Assimilate every standard model timestep.
Truth has forcing F=8 (chaotic).
Ensemble has forcing F=5 (not chaotic).
5th EnKF Workshop, 23 May 2012 49
Case 4: Frequent low-quality obs., imperfect model
These are the localizations for the Case 1 perfect model.
5th EnKF Workshop, 23 May 2012 50
Case 4: Frequent low-quality obs., imperfect model
Best GC is more compact for imperfect model case.Fifth ELF also more compact, but not as close to imperfect GC.
5th EnKF Workshop, 23 May 2012 51
How long an OSSE does this take?
For large localization get good results with O(100) OSSE steps.Errors grow much more quickly for small localizations.
5th EnKF Workshop, 23 May 2012 52
ConclusionsCan get estimates of good localization for any subset of observations and state variables from an OSSE.
If good localizations are non-Gaussian do better than Gaspari Cohn.
When Gaussian, can still be cheaper than tuning half-widths.
Can this be applied to real geophysical models?
How much could real applications be improved? Unclear…
Can localization functions be separable in large models?Loc(time diff) * Loc(horizontal dist.) * Loc(vertical dist.) * Loc(obs_type, state_type)???
5th EnKF Workshop, 23 May 2012 53
Related Activities: Lili Lei Poster
Testing ELFs in global climate model (CAM),
and in WRF regional nested configuration.
Some results look very similar to earlier sampling error correction methods.
Next step, using ELFs in iterated OSSEs.
5th EnKF Workshop, 23 May 2012 54
Empirical Localization Without Knowing Truth
Find that minimizes the RMSE between the posterior ensemble mean for x and observed value of x over the subset of (y, x) pairs.
This can be computed from the output of an assimilation.
Can then use this localization in a new assimilation for all (y, x) in the subset.
BUT, can only compute for pairs of OBSERVED quantities.
Can act as a way to calibrate OSSE results?
5th EnKF Workshop, 23 May 2012 55
Case 1 without knowing truth
Identity observations.
Error variance 16.
Assimilate every standard model timestep.
All state variables are observed, so no problem there.
5th EnKF Workshop, 23 May 2012 56
Case 1 without knowing truth
Using real obs is much noisier for small localization values.Similar to using truth for larger localization values.Could be used to calibrate results from an OSSE for real assimilation use.