4 research outputs found
Statistical post-processing of hydrological forecasts using Bayesian model averaging
Accurate and reliable probabilistic forecasts of hydrological quantities like
runoff or water level are beneficial to various areas of society. Probabilistic
state-of-the-art hydrological ensemble prediction models are usually driven
with meteorological ensemble forecasts. Hence, biases and dispersion errors of
the meteorological forecasts cascade down to the hydrological predictions and
add to the errors of the hydrological models. The systematic parts of these
errors can be reduced by applying statistical post-processing. For a sound
estimation of predictive uncertainty and an optimal correction of systematic
errors, statistical post-processing methods should be tailored to the
particular forecast variable at hand. Former studies have shown that it can
make sense to treat hydrological quantities as bounded variables. In this
paper, a doubly truncated Bayesian model averaging (BMA) method, which allows
for flexible post-processing of (multi-model) ensemble forecasts of water
level, is introduced. A case study based on water level for a gauge of river
Rhine, reveals a good predictive skill of doubly truncated BMA compared both to
the raw ensemble and the reference ensemble model output statistics approach.Comment: 19 pages, 6 figure
Post-processing numerical weather prediction ensembles for probabilistic solar irradiance forecasting
In order to enable the transition towards renewable energy sources, probabilistic energy forecasting is of critical importance for incorporating volatile power sources such as solar energy into the electrical grid. Solar energy forecasting methods often aim to provide probabilistic predictions of solar irradiance. In particular, many hybrid approaches combine physical information from numerical weather prediction models with statistical methods. Even though the physical models can provide useful information at intra-day and day-ahead forecast horizons, ensemble weather forecasts from multiple model runs are often not calibrated and show systematic biases. We propose a post-processing model for ensemble weather predictions of solar irradiance at temporal resolutions between 30 min and 6 h. The proposed models provide probabilistic forecasts in the form of a censored logistic probability distribution for lead times up to 5 days and are evaluated in two case studies covering distinct physical models, geographical regions, temporal resolutions, and types of solar irradiance. We find that post-processing consistently and significantly improves the forecast performance of the ensemble predictions for lead times up to at least 48 h and is well able to correct the systematic lack of calibration
Machine learning for total cloud cover prediction
Accurate and reliable forecasting of total cloud cover (TCC) is vital for many areas such as astronomy, energy demand and production, or agriculture. Most meteorological centres issue ensemble forecasts of TCC; however, these forecasts are often uncalibrated and exhibit worse forecast skill than ensemble forecasts of other weather variables. Hence, some form of post-processing is strongly required to improve predictive performance. As TCC observations are usually reported on a discrete scale taking just nine different values called oktas, statistical calibration of TCC ensemble forecasts can be considered a classification problem with outputs given by the probabilities of the oktas. This is a classical area where machine learning methods are applied. We investigate the performance of post-processing using multilayer perceptron (MLP) neural networks, gradient boosting machines (GBM) and random forest (RF) methods. Based on the European Centre for Medium-Range Weather Forecasts global TCC ensemble forecasts for 2002–2014, we compare these approaches with the proportional odds logistic regression (POLR) and multiclass logistic regression (MLR) models, as well as the raw TCC ensemble forecasts. We further assess whether improvements in forecast skill can be obtained by incorporating ensemble forecasts of precipitation as additional predictor. Compared to the raw ensemble, all calibration methods result in a significant improvement in forecast skill. RF models provide the smallest increase in predictive performance, while MLP, POLR and GBM approaches perform best. The use of precipitation forecast data leads to further improvements in forecast skill, and except for very short lead times the extended MLP model shows the best overall performance
Machine learning for total cloud cover prediction
Accurate and reliable forecasting of total cloud cover (TCC) is vital for
many areas such as astronomy, energy demand and production, or agriculture.
Most meteorological centres issue ensemble forecasts of TCC, however, these
forecasts are often uncalibrated and exhibit worse forecast skill than ensemble
forecasts of other weather variables. Hence, some form of post-processing is
strongly required to improve predictive performance. As TCC observations are
usually reported on a discrete scale taking just nine different values called
oktas, statistical calibration of TCC ensemble forecasts can be considered a
classification problem with outputs given by the probabilities of the oktas.
This is a classical area where machine learning methods are applied. We
investigate the performance of post-processing using multilayer perceptron
(MLP) neural networks, gradient boosting machines (GBM) and random forest (RF)
methods. Based on the European Centre for Medium-Range Weather Forecasts global
TCC ensemble forecasts for 2002-2014 we compare these approaches with the
proportional odds logistic regression (POLR) and multiclass logistic regression
(MLR) models, as well as the raw TCC ensemble forecasts. We further assess
whether improvements in forecast skill can be obtained by incorporating
ensemble forecasts of precipitation as additional predictor. Compared to the
raw ensemble, all calibration methods result in a significant improvement in
forecast skill. RF models provide the smallest increase in predictive
performance, while MLP, POLR and GBM approaches perform best. The use of
precipitation forecast data leads to further improvements in forecast skill and
except for very short lead times the extended MLP model shows the best overall
performance.Comment: 24 pages, 7 figure