13.11.2014 Views

Chapter 7 - Ensemble methods.pdf

Chapter 7 - Ensemble methods.pdf

Chapter 7 - Ensemble methods.pdf

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Chapter</strong> 7<br />

<strong>Ensemble</strong> <strong>methods</strong>


Each of the following introduces some<br />

error into the modeling process<br />

• Initial conditions<br />

• Lateral and upper boundary conditions<br />

• Specification of the lower-boundary conditions<br />

• Numerical approximations to the equations<br />

• Parameterizations of physical processes<br />

The sum of the errors can be very case/date<br />

dependent


The essence of ensemble prediction<br />

• Perform parallel forecasts or simulations using<br />

different arbitrary choices for the previous<br />

imperfect data or <strong>methods</strong>.<br />

• This “samples the uncertainty space”<br />

associated with the modeling process, to see<br />

how this uncertainty projects onto the<br />

forecast.


Example ensemble forecast – track<br />

prediction for hurricane Katrina, from the ECMWF<br />

ensemble prediction system. The heavy line is from the<br />

deterministic model forecast


Deterministic vs. ensemble prediction<br />

• Without ensemble prediction (deterministic<br />

forecast) – there is one forecast state of the<br />

atmosphere, and the user has no information<br />

about the confidence we should place in it.<br />

– Compared to yesterday’s forecast<br />

– One part of the model grid compared to others<br />

• With ensemble prediction (probabilistic forecasts)<br />

– we get a better forecast of the most probable<br />

conditions, and we get uncertainty information<br />

(“spread” of ensemble members).


Benefit of ensemble method<br />

• 1) The mean of the ensemble of forecasts is generally<br />

more accurate than the forecast from an individual<br />

member, for a large number of forecasts.<br />

• 2) The spread or variance or dispersion of the<br />

ensemble members can be related to the uncertainty<br />

in the ensemble mean.<br />

• 3) The probability distribution (or density) function<br />

(PDF) can provide information about extreme events.<br />

• 4) Quantitative probabilistic products can be moreeffectively<br />

employed in decision-support systems


Historical context<br />

• Since the 1960’s it has been shown that<br />

combining human forecasts from different<br />

forecasters produces a group-mean forecast<br />

that is superior – “consensus forecasting”.<br />

• Forecasting paradigm<br />

– When models agree…good confidence in forecast<br />

– When models disagree…low confidence


Compromises with<br />

ensemble prediction<br />

• Running many parallel models is computationally<br />

expensive.<br />

• Must use coarser horizontal grid increment for<br />

the ensemble, than you would use for a single<br />

deterministic forecast.<br />

• If you increased the grid increment by a factor of<br />

4, how many ensemble members could you run<br />

with the same hardware?


Compromises with<br />

ensemble prediction<br />

• Running many parallel models is computationally<br />

expensive.<br />

• Must use coarser horizontal grid increment for<br />

the ensemble, than you would use for a single<br />

deterministic forecast.<br />

• If you increased the grid increment by a factor of<br />

4, how many ensemble members could you run<br />

with the same hardware?<br />

64


The spread among ensemble<br />

members, and uncertainty<br />

Cyclogenesis?<br />

No Cyclogenesis?<br />

Eight member ensemble with slightly different initial conditions.<br />

• The “x” shows the ensemble mean at each time.<br />

• Circles show the trajectory of individual ensemble members.<br />

• Solid line shows the “trajectory” of the initial mean of the ensemble.


High<br />

predictability<br />

Example of<br />

different spreads<br />

of ensemble<br />

members<br />

London 2-m AGL<br />

temperature forecasts<br />

from ECMWF, initialized<br />

on different dates.<br />

Low<br />

predictability<br />

Different initial<br />

conditions and model<br />

configurations


Another way of looking at day-to-day variability in predictability<br />

that can be captured by ensemble <strong>methods</strong>: Temporal variability<br />

of single model predictive skill - 500-mb heights in NH


Defining the ensemble members<br />

• Initial-condition uncertainty – how should it<br />

be done???<br />

• Physical-process parameterization uncertainty<br />

– again, how?<br />

• Errors in the dynamical core – i.e., numerical<br />

algorithms – how?<br />

• LBC uncertainty (If a ALM) – how?<br />

• Surface BC uncertainty


Some comments<br />

• With sequential data-assimilation systems, where the<br />

model forecast is used as a first-guess in the next cycle,<br />

model error and initial-condition error are closely<br />

related.<br />

• Multi-model ensembles - using completely different<br />

models to construct the ensemble - make sense if the<br />

models are already being run for the same geographic<br />

area.<br />

• Super-ensembles involve running ensembles for<br />

individual models, and then combining the multiple<br />

sets of ensemble members into a single ensemble.


Verification of ensemble predictions<br />

• The ensemble mean – can be verified the same<br />

way as a deterministic forecast.<br />

• The uncertainty information (e.g., 2 nd moment)<br />

– Reliability diagrams


Reliability diagrams – reliability is an<br />

important attribute of ensemble forecasts of<br />

dichotomous events (ones that occur or do not<br />

occur at a grid point or over an area)<br />

Using a large number of archived ensemble forecasts, select the forecasts<br />

where the forecasted probability of occurrence of an event was p f . Of<br />

those events, what was the observed frequency of occurrence p o ? Do this<br />

for all the different thresholds in the forecast probabilities, and plot as<br />

above.


An example of a real reliability diagram<br />

Single-model IC ensemble<br />

Multi-model super-ensemble


Verification of ensemble predictions<br />

• The ensemble mean – can be verified the same<br />

way as a deterministic forecast.<br />

• The uncertainty information (e.g., 2 nd moment)<br />

– Reliability diagrams<br />

– Rank histograms


Rank histograms<br />

• Also called verification rank histograms and<br />

Talagrand diagrams.<br />

• Define the bias of ensemble predictions<br />

• For a specific variable and location of an<br />

observation, take the ensemble forecast for<br />

that variable at that location and rank-order<br />

the forecasts from each of the members.<br />

• Then define the n+1 intervals that are<br />

bounded by the n ordered forecast variables.


Example with 4 ensemble members<br />

and 5 intervals, for variable P<br />

For this location and<br />

time t forecast , the<br />

observed P (X obs ) is<br />

lower than any of<br />

the forecast Ps and<br />

the observation is<br />

thus in I 1 .


• Follow the same process for all other pairs of<br />

observations and forecasts at this time.<br />

• Calculate the total number of observations in<br />

each of the five intervals, or ranks, and plot a<br />

histogram of the frequency.<br />

• This provides a graphical view of how the<br />

ensemble of forecasts relates to the observations.<br />

• A non-uniformity in the distribution reveals<br />

systematic errors in the ensemble.


• Over-forecasting bias – Observations tend to fall in the<br />

lower intervals of the ensemble. Forecast values are<br />

generally too high.<br />

• Under-forecasting bias – Observations tend to fall in<br />

the upper intervals of the ensemble. Forecast values<br />

are generally too low.<br />

• Rank uniformity – Ideal situation<br />

• Under-dispersion – Observations tend to fall in the<br />

upper and lower intervals. Forecasts are too similar to<br />

each other…need more spread.<br />

• Over-dispersion – Observations tend to fall in the<br />

center intervals. Forecasts are too dispersed.


Quantitatively interpreting<br />

• Many approaches<br />

ensemble forecasts<br />

• Example: Democratic voting method – e.g., for<br />

visibility or wind speed at an airport, pick a<br />

relevant value, count the number of ensemble<br />

members above and below the value, and<br />

translate to a probability.<br />

15 m/s<br />

4 forecasts 6 forecasts<br />

Wind speed<br />

Probability of winds exceeding threshold is 60%


Calibration of ensembles<br />

• Calibration is a post-processing step that removes the bias<br />

from the first and possibly the higher moments.<br />

• Calibration is important because it:<br />

– provides greater accuracy in the ensemble mean,<br />

– provides improved estimates of the probabilities of extreme<br />

events,<br />

– represents ensemble spread in terms of quantitative measures<br />

of the uncertainty in the forecast of the ensemble mean.<br />

• A history of observations and ensemble forecasts is needed<br />

to perform the calibration.<br />

• Historical archives of operational forecasts are not ideal for<br />

this. WHY?


Calibration of ensembles<br />

• Calibration is a post-processing step that removes the bias<br />

from the first and possibly the higher moments.<br />

• Calibration is important because it:<br />

– provides greater accuracy in the ensemble mean,<br />

– provides improved estimates of the probabilities of extreme<br />

events,<br />

– represents ensemble spread in terms of quantitative measures<br />

of the uncertainty in the forecast of the ensemble mean.<br />

• A history of observations and ensemble forecasts is needed<br />

to perform the calibration.<br />

• Historical archives of operational forecasts are not ideal for<br />

this. Because the model is continually being updated, and<br />

thus the calibration changes also.<br />

• Ideally, reforecasts with the current version of the model<br />

should be used.


Example skill of calibrated versus<br />

uncalibrated ensembles<br />

• Uncalibrated ensemble – probability of<br />

precipitation occurring over a threshold amount<br />

is calculated based on the number of ensemble<br />

members that produce precip above and below<br />

the threshold (democratic voting method).<br />

skill<br />

no skill<br />

One more day of<br />

skill from calibration


Time lagged ensemble<br />

• Advantage: It is based on deterministic forecasts from<br />

different initial times, that are valid at the same time.<br />

• Not as good as conventional ensemble, but some<br />

benefits have been shown.


Spread of time-lagged ensembles<br />

• Forecasters frequently look at how consistent<br />

forecasts from different cycles.<br />

• If the forecast for a certain time remains the<br />

same for different cycles (different initial<br />

times) – forecast has confidence in the<br />

prediction.<br />

• If the forecast changes for a new cycle, the<br />

forecaster is alerted that there may be<br />

increasing uncertainty.


Short-range ensemble prediction<br />

with high-resolution LAMs<br />

• LBCs may cause excessively small ensemble<br />

dispersion, even when they are perturbed.<br />

• Near-surface processes are sometimes not<br />

very predictable, limiting the usefulness of an<br />

ensemble.<br />

• Methods for generating IC error have been<br />

designed for larger-scale models – it is unclear<br />

how to do this on the mesoscale.


Graphically displaying<br />

ensemble-model products


The ensemble mean<br />

• These fields look like any map of a model<br />

dependent variable, and can be displayed in<br />

the same way


The spread or dispersion<br />

of the ensemble


12 h<br />

Spaghetti plots:<br />

36 h<br />

84 h<br />

5520-m contour<br />

of the 500-mb<br />

height based on<br />

a 17-member<br />

ensemble<br />

forecast by<br />

NCEP


Meteograms<br />

Show one variable<br />

for one point


Probability of exceedance plots<br />

The probability<br />

that near-surface<br />

wind gusts will<br />

exceed 50 m/s at<br />

the 42-hour lead<br />

of a forecast for 26<br />

December 1999,<br />

based on an<br />

ECMWF 50-<br />

member<br />

ensemble


Plots of ensemble variance<br />

19-member<br />

physics ensemble<br />

(solid line) and a<br />

19-member<br />

initial-condition<br />

ensemble<br />

(dashed line)<br />

from MM5, for a<br />

long-lived MCS.


“Stamp maps”


Probability of the dosage exceeding a<br />

threshold – democratic voting method


An “electricity-gram” – an example of a specialized<br />

plot: uncertainty in 10-day forecast of electricity demand<br />

ECMWF ensemble prediction used as input to an energy-demand<br />

model. The middle 50% of predictions falls within the box, and the<br />

whiskers contain all the values


Using probabilistic information from<br />

ensemble predictions<br />

• The cost-loss approach<br />

• Given an uncertain prediction of whether an<br />

event will or will not occur, a decision maker has<br />

the option to either protect against the<br />

occurrence of the wx event, or not to protect.<br />

• Simple decision problem – two actions (protect,<br />

or not) and two outcomes (event occurs, or not)<br />

• Example events – freeze, flood, high winds, heavy<br />

snow


• Decision to protect will have cost C, whether<br />

or not the event occurs.<br />

• A decision to not protect will result in a loss L,<br />

if the event occurs


• Assume that a calibrated ensemble forecast<br />

predicts that the probability of an event occurring<br />

is p.<br />

• The optimal decision will be the one resulting in<br />

the smallest expense.<br />

• If the decision is to protect, expense = C with a<br />

probability of 1.<br />

• If the decision is to not protect, the expense will<br />

be pL.<br />

• Thus, protecting against the risk will result in the<br />

smallest expense when C < pL. Or<br />

C/L < p


• Such an approach to decision making is<br />

straightforward when economic value is<br />

used…it is more problematic when societal or<br />

environmental “values” must be considered.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!