Chapter 7 - Ensemble methods.pdf
Chapter 7 - Ensemble methods.pdf
Chapter 7 - Ensemble methods.pdf
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
<strong>Chapter</strong> 7<br />
<strong>Ensemble</strong> <strong>methods</strong>
Each of the following introduces some<br />
error into the modeling process<br />
• Initial conditions<br />
• Lateral and upper boundary conditions<br />
• Specification of the lower-boundary conditions<br />
• Numerical approximations to the equations<br />
• Parameterizations of physical processes<br />
The sum of the errors can be very case/date<br />
dependent
The essence of ensemble prediction<br />
• Perform parallel forecasts or simulations using<br />
different arbitrary choices for the previous<br />
imperfect data or <strong>methods</strong>.<br />
• This “samples the uncertainty space”<br />
associated with the modeling process, to see<br />
how this uncertainty projects onto the<br />
forecast.
Example ensemble forecast – track<br />
prediction for hurricane Katrina, from the ECMWF<br />
ensemble prediction system. The heavy line is from the<br />
deterministic model forecast
Deterministic vs. ensemble prediction<br />
• Without ensemble prediction (deterministic<br />
forecast) – there is one forecast state of the<br />
atmosphere, and the user has no information<br />
about the confidence we should place in it.<br />
– Compared to yesterday’s forecast<br />
– One part of the model grid compared to others<br />
• With ensemble prediction (probabilistic forecasts)<br />
– we get a better forecast of the most probable<br />
conditions, and we get uncertainty information<br />
(“spread” of ensemble members).
Benefit of ensemble method<br />
• 1) The mean of the ensemble of forecasts is generally<br />
more accurate than the forecast from an individual<br />
member, for a large number of forecasts.<br />
• 2) The spread or variance or dispersion of the<br />
ensemble members can be related to the uncertainty<br />
in the ensemble mean.<br />
• 3) The probability distribution (or density) function<br />
(PDF) can provide information about extreme events.<br />
• 4) Quantitative probabilistic products can be moreeffectively<br />
employed in decision-support systems
Historical context<br />
• Since the 1960’s it has been shown that<br />
combining human forecasts from different<br />
forecasters produces a group-mean forecast<br />
that is superior – “consensus forecasting”.<br />
• Forecasting paradigm<br />
– When models agree…good confidence in forecast<br />
– When models disagree…low confidence
Compromises with<br />
ensemble prediction<br />
• Running many parallel models is computationally<br />
expensive.<br />
• Must use coarser horizontal grid increment for<br />
the ensemble, than you would use for a single<br />
deterministic forecast.<br />
• If you increased the grid increment by a factor of<br />
4, how many ensemble members could you run<br />
with the same hardware?
Compromises with<br />
ensemble prediction<br />
• Running many parallel models is computationally<br />
expensive.<br />
• Must use coarser horizontal grid increment for<br />
the ensemble, than you would use for a single<br />
deterministic forecast.<br />
• If you increased the grid increment by a factor of<br />
4, how many ensemble members could you run<br />
with the same hardware?<br />
64
The spread among ensemble<br />
members, and uncertainty<br />
Cyclogenesis?<br />
No Cyclogenesis?<br />
Eight member ensemble with slightly different initial conditions.<br />
• The “x” shows the ensemble mean at each time.<br />
• Circles show the trajectory of individual ensemble members.<br />
• Solid line shows the “trajectory” of the initial mean of the ensemble.
High<br />
predictability<br />
Example of<br />
different spreads<br />
of ensemble<br />
members<br />
London 2-m AGL<br />
temperature forecasts<br />
from ECMWF, initialized<br />
on different dates.<br />
Low<br />
predictability<br />
Different initial<br />
conditions and model<br />
configurations
Another way of looking at day-to-day variability in predictability<br />
that can be captured by ensemble <strong>methods</strong>: Temporal variability<br />
of single model predictive skill - 500-mb heights in NH
Defining the ensemble members<br />
• Initial-condition uncertainty – how should it<br />
be done???<br />
• Physical-process parameterization uncertainty<br />
– again, how?<br />
• Errors in the dynamical core – i.e., numerical<br />
algorithms – how?<br />
• LBC uncertainty (If a ALM) – how?<br />
• Surface BC uncertainty
Some comments<br />
• With sequential data-assimilation systems, where the<br />
model forecast is used as a first-guess in the next cycle,<br />
model error and initial-condition error are closely<br />
related.<br />
• Multi-model ensembles - using completely different<br />
models to construct the ensemble - make sense if the<br />
models are already being run for the same geographic<br />
area.<br />
• Super-ensembles involve running ensembles for<br />
individual models, and then combining the multiple<br />
sets of ensemble members into a single ensemble.
Verification of ensemble predictions<br />
• The ensemble mean – can be verified the same<br />
way as a deterministic forecast.<br />
• The uncertainty information (e.g., 2 nd moment)<br />
– Reliability diagrams
Reliability diagrams – reliability is an<br />
important attribute of ensemble forecasts of<br />
dichotomous events (ones that occur or do not<br />
occur at a grid point or over an area)<br />
Using a large number of archived ensemble forecasts, select the forecasts<br />
where the forecasted probability of occurrence of an event was p f . Of<br />
those events, what was the observed frequency of occurrence p o ? Do this<br />
for all the different thresholds in the forecast probabilities, and plot as<br />
above.
An example of a real reliability diagram<br />
Single-model IC ensemble<br />
Multi-model super-ensemble
Verification of ensemble predictions<br />
• The ensemble mean – can be verified the same<br />
way as a deterministic forecast.<br />
• The uncertainty information (e.g., 2 nd moment)<br />
– Reliability diagrams<br />
– Rank histograms
Rank histograms<br />
• Also called verification rank histograms and<br />
Talagrand diagrams.<br />
• Define the bias of ensemble predictions<br />
• For a specific variable and location of an<br />
observation, take the ensemble forecast for<br />
that variable at that location and rank-order<br />
the forecasts from each of the members.<br />
• Then define the n+1 intervals that are<br />
bounded by the n ordered forecast variables.
Example with 4 ensemble members<br />
and 5 intervals, for variable P<br />
For this location and<br />
time t forecast , the<br />
observed P (X obs ) is<br />
lower than any of<br />
the forecast Ps and<br />
the observation is<br />
thus in I 1 .
• Follow the same process for all other pairs of<br />
observations and forecasts at this time.<br />
• Calculate the total number of observations in<br />
each of the five intervals, or ranks, and plot a<br />
histogram of the frequency.<br />
• This provides a graphical view of how the<br />
ensemble of forecasts relates to the observations.<br />
• A non-uniformity in the distribution reveals<br />
systematic errors in the ensemble.
• Over-forecasting bias – Observations tend to fall in the<br />
lower intervals of the ensemble. Forecast values are<br />
generally too high.<br />
• Under-forecasting bias – Observations tend to fall in<br />
the upper intervals of the ensemble. Forecast values<br />
are generally too low.<br />
• Rank uniformity – Ideal situation<br />
• Under-dispersion – Observations tend to fall in the<br />
upper and lower intervals. Forecasts are too similar to<br />
each other…need more spread.<br />
• Over-dispersion – Observations tend to fall in the<br />
center intervals. Forecasts are too dispersed.
Quantitatively interpreting<br />
• Many approaches<br />
ensemble forecasts<br />
• Example: Democratic voting method – e.g., for<br />
visibility or wind speed at an airport, pick a<br />
relevant value, count the number of ensemble<br />
members above and below the value, and<br />
translate to a probability.<br />
15 m/s<br />
4 forecasts 6 forecasts<br />
Wind speed<br />
Probability of winds exceeding threshold is 60%
Calibration of ensembles<br />
• Calibration is a post-processing step that removes the bias<br />
from the first and possibly the higher moments.<br />
• Calibration is important because it:<br />
– provides greater accuracy in the ensemble mean,<br />
– provides improved estimates of the probabilities of extreme<br />
events,<br />
– represents ensemble spread in terms of quantitative measures<br />
of the uncertainty in the forecast of the ensemble mean.<br />
• A history of observations and ensemble forecasts is needed<br />
to perform the calibration.<br />
• Historical archives of operational forecasts are not ideal for<br />
this. WHY?
Calibration of ensembles<br />
• Calibration is a post-processing step that removes the bias<br />
from the first and possibly the higher moments.<br />
• Calibration is important because it:<br />
– provides greater accuracy in the ensemble mean,<br />
– provides improved estimates of the probabilities of extreme<br />
events,<br />
– represents ensemble spread in terms of quantitative measures<br />
of the uncertainty in the forecast of the ensemble mean.<br />
• A history of observations and ensemble forecasts is needed<br />
to perform the calibration.<br />
• Historical archives of operational forecasts are not ideal for<br />
this. Because the model is continually being updated, and<br />
thus the calibration changes also.<br />
• Ideally, reforecasts with the current version of the model<br />
should be used.
Example skill of calibrated versus<br />
uncalibrated ensembles<br />
• Uncalibrated ensemble – probability of<br />
precipitation occurring over a threshold amount<br />
is calculated based on the number of ensemble<br />
members that produce precip above and below<br />
the threshold (democratic voting method).<br />
skill<br />
no skill<br />
One more day of<br />
skill from calibration
Time lagged ensemble<br />
• Advantage: It is based on deterministic forecasts from<br />
different initial times, that are valid at the same time.<br />
• Not as good as conventional ensemble, but some<br />
benefits have been shown.
Spread of time-lagged ensembles<br />
• Forecasters frequently look at how consistent<br />
forecasts from different cycles.<br />
• If the forecast for a certain time remains the<br />
same for different cycles (different initial<br />
times) – forecast has confidence in the<br />
prediction.<br />
• If the forecast changes for a new cycle, the<br />
forecaster is alerted that there may be<br />
increasing uncertainty.
Short-range ensemble prediction<br />
with high-resolution LAMs<br />
• LBCs may cause excessively small ensemble<br />
dispersion, even when they are perturbed.<br />
• Near-surface processes are sometimes not<br />
very predictable, limiting the usefulness of an<br />
ensemble.<br />
• Methods for generating IC error have been<br />
designed for larger-scale models – it is unclear<br />
how to do this on the mesoscale.
Graphically displaying<br />
ensemble-model products
The ensemble mean<br />
• These fields look like any map of a model<br />
dependent variable, and can be displayed in<br />
the same way
The spread or dispersion<br />
of the ensemble
12 h<br />
Spaghetti plots:<br />
36 h<br />
84 h<br />
5520-m contour<br />
of the 500-mb<br />
height based on<br />
a 17-member<br />
ensemble<br />
forecast by<br />
NCEP
Meteograms<br />
Show one variable<br />
for one point
Probability of exceedance plots<br />
The probability<br />
that near-surface<br />
wind gusts will<br />
exceed 50 m/s at<br />
the 42-hour lead<br />
of a forecast for 26<br />
December 1999,<br />
based on an<br />
ECMWF 50-<br />
member<br />
ensemble
Plots of ensemble variance<br />
19-member<br />
physics ensemble<br />
(solid line) and a<br />
19-member<br />
initial-condition<br />
ensemble<br />
(dashed line)<br />
from MM5, for a<br />
long-lived MCS.
“Stamp maps”
Probability of the dosage exceeding a<br />
threshold – democratic voting method
An “electricity-gram” – an example of a specialized<br />
plot: uncertainty in 10-day forecast of electricity demand<br />
ECMWF ensemble prediction used as input to an energy-demand<br />
model. The middle 50% of predictions falls within the box, and the<br />
whiskers contain all the values
Using probabilistic information from<br />
ensemble predictions<br />
• The cost-loss approach<br />
• Given an uncertain prediction of whether an<br />
event will or will not occur, a decision maker has<br />
the option to either protect against the<br />
occurrence of the wx event, or not to protect.<br />
• Simple decision problem – two actions (protect,<br />
or not) and two outcomes (event occurs, or not)<br />
• Example events – freeze, flood, high winds, heavy<br />
snow
• Decision to protect will have cost C, whether<br />
or not the event occurs.<br />
• A decision to not protect will result in a loss L,<br />
if the event occurs
• Assume that a calibrated ensemble forecast<br />
predicts that the probability of an event occurring<br />
is p.<br />
• The optimal decision will be the one resulting in<br />
the smallest expense.<br />
• If the decision is to protect, expense = C with a<br />
probability of 1.<br />
• If the decision is to not protect, the expense will<br />
be pL.<br />
• Thus, protecting against the risk will result in the<br />
smallest expense when C < pL. Or<br />
C/L < p
• Such an approach to decision making is<br />
straightforward when economic value is<br />
used…it is more problematic when societal or<br />
environmental “values” must be considered.