7,775 research outputs found
A statistical downscaling framework for environmental mapping
In recent years, knowledge extraction from data has become increasingly popular, with many numerical forecasting models, mainly falling into two major categories—chemical transport models (CTMs) and conventional statistical methods. However, due to data and model variability, data-driven knowledge extraction from high-dimensional, multifaceted data in such applications require generalisations of global to regional or local conditions. Typically, generalisation is achieved via mapping global conditions to local ecosystems and human habitats which amounts to tracking and monitoring environmental dynamics in various geographical areas and their regional and global implications on human livelihood. Statistical downscaling techniques have been widely used to extract high-resolution information from regional-scale variables produced by CTMs in climate model. Conventional applications of these methods are predominantly dimensional reduction in nature, designed to reduce spatial dimension of gridded model outputs without loss of essential spatial information. Their downside is twofold—complete dependence on unlabelled design matrix and reliance on underlying distributional assumptions. We propose a novel statistical downscaling framework for dealing with data and model variability. Its power derives from training and testing multiple models on multiple samples, narrowing down global environmental phenomena to regional discordance through dimensional reduction and visualisation. Hourly ground-level ozone observations were obtained from various environmental stations maintained by the US Environmental Protection Agency, covering the summer period (June–August 2005). Regional patterns of ozone are related to local observations via repeated runs and performance assessment of multiple versions of empirical orthogonal functions or principal components and principal fitted components via an algorithm with fully adaptable parameters. We demonstrate how the algorithm can be extended to weather-dependent and other applications with inherent data randomness and model variability via its built-in interdisciplinary computational power that connects data sources with end-users
A review of applied methods in Europe for flood-frequency analysis in a changing environment
The report presents a review of methods used in Europe for trend analysis, climate change projections and non-stationary analysis of extreme precipitation and flood frequency. In addition, main findings of the analyses are presented, including a comparison of trend analysis results and climate change projections. Existing guidelines in Europe on design flood and design rainfall estimation that incorporate climate change are reviewed. The report
concludes with a discussion of research needs on non-stationary frequency analysis for considering the effects of climate change and inclusion in design guidelines.
Trend analyses are reported for 21 countries in Europe with results for extreme precipitation, extreme streamflow or both. A large number of national and regional trend studies have been carried out. Most studies are based on statistical methods applied to individual time series of extreme precipitation or extreme streamflow using the non-parametric Mann-Kendall trend test or regression analysis. Some studies have been reported that use field significance or regional consistency tests to analyse trends over larger areas. Some of the studies also include analysis of trend attribution. The studies reviewed indicate that there is
some evidence of a general increase in extreme precipitation, whereas there are no clear indications of significant increasing trends at regional or national level of extreme streamflow. For some smaller regions increases in extreme streamflow are reported. Several studies from regions dominated by snowmelt-induced peak flows report decreases in extreme streamflow and earlier spring snowmelt peak flows. Climate change projections have been reported for 14 countries in Europe with results for extreme precipitation, extreme streamflow or both. The review shows various approaches for producing climate projections of extreme precipitation and flood frequency based on
alternative climate forcing scenarios, climate projections from available global and regional climate models, methods for statistical downscaling and bias correction, and alternative hydrological models. A large number of the reported studies are based on an ensemble modelling approach that use several climate forcing scenarios and climate model projections in order to address the uncertainty on the projections of extreme precipitation and flood frequency. Some studies also include alternative statistical downscaling and bias correction methods and hydrological modelling approaches. Most studies reviewed indicate an increase in extreme precipitation under a future climate, which is consistent with the observed trend of extreme precipitation. Hydrological projections of peak flows and flood frequency show both positive and negative changes. Large increases in peak flows are reported for some catchments with rainfall-dominated peak flows, whereas a general decrease in flood magnitude and earlier spring floods are reported for catchments with snowmelt-dominated peak flows. The latter is consistent with the observed trends. The review of existing guidelines in Europe on design floods and design rainfalls shows that only few countries explicitly address climate change. These design guidelines are based on climate change adjustment factors to be applied to current design estimates and may
depend on design return period and projection horizon. The review indicates a gap between the need for considering climate change impacts in design and actual published guidelines that incorporate climate change in extreme precipitation and flood frequency. Most of the studies reported are based on frequency analysis assuming stationary conditions in a certain time window (typically 30 years) representing current and future climate. There is a need for developing more consistent non-stationary frequency analysis methods that can account for the transient nature of a changing climate
Probabilistic estimates of future changes in California temperature and precipitation usingstatistical and dynamical downscaling
Sixteen global general circulation models were used to develop probabilistic projections of temperature (T) and precipitation (P) changes over California by the 2060s. The global models were downscaled with two statistical techniques and three nested dynamical regional climate models, although not all global models were downscaled with all techniques. Both monthly and daily timescale changes in T and P are addressed, the latter being important for a range of applications in energy use, water management, and agriculture. The T changes tend to agree more across downscaling techniques than the P changes. Year-to-year natural internal climate variability is roughly of similar magnitude to the projected T changes. In the monthly average, July temperatures shift enough that that the hottest July found in any simulation over the historical period becomes a modestly cool July in the future period. Januarys as cold as any found in the historical period are still found in the 2060s, but the median and maximum monthly average temperatures increase notably. Annual and seasonal P changes are small compared to interannual or intermodel variability. However, the annual change is composed of seasonally varying changes that are themselves much larger, but tend to cancel in the annual mean. Winters show modestly wetter conditions in the North of the state, while spring and autumn show less precipitation. The dynamical downscaling techniques project increasing precipitation in the Southeastern part of the state, which is influenced by the North American monsoon, a feature that is not captured by the statistical downscaling
Variational Downscaling, Fusion and Assimilation of Hydrometeorological States via Regularized Estimation
Improved estimation of hydrometeorological states from down-sampled
observations and background model forecasts in a noisy environment, has been a
subject of growing research in the past decades. Here, we introduce a unified
framework that ties together the problems of downscaling, data fusion and data
assimilation as ill-posed inverse problems. This framework seeks solutions
beyond the classic least squares estimation paradigms by imposing proper
regularization, which are constraints consistent with the degree of smoothness
and probabilistic structure of the underlying state. We review relevant
regularization methods in derivative space and extend classic formulations of
the aforementioned problems with particular emphasis on hydrologic and
atmospheric applications. Informed by the statistical characteristics of the
state variable of interest, the central results of the paper suggest that
proper regularization can lead to a more accurate and stable recovery of the
true state and hence more skillful forecasts. In particular, using the Tikhonov
and Huber regularization in the derivative space, the promise of the proposed
framework is demonstrated in static downscaling and fusion of synthetic
multi-sensor precipitation data, while a data assimilation numerical experiment
is presented using the heat equation in a variational setting
Recommended from our members
Bias adjustment of satellite-based precipitation estimation using artificial neural networks-cloud classification system over Saudi Arabia
Precipitation is a key input variable for hydrological and climate studies. Rain gauges can provide reliable precipitation measurements at a point of observations. However, the uncertainty of rain measurements increases when a rain gauge network is sparse. Satellite-based precipitation estimations SPEs appear to be an alternative source of measurements for regions with limited rain gauges. However, the systematic bias from satellite precipitation estimation should be estimated and adjusted. In this study, a method of removing the bias from the precipitation estimation from remotely sensed information using artificial neural networks-cloud classification system (PERSIANN-CCS) over a region where the rain gauge is sparse is investigated. The method consists of monthly empirical quantile mapping of gauge and satellite measurements over several climate zones as well as inverse-weighted distance for the interpolation of gauge measurements. Seven years (2010–2016) of daily precipitation estimation from PERSIANN-CCS was used to test and adjust the bias of estimation over Saudi Arabia. The first 6 years (2010–2015) are used for calibration, while 1 year (2016) is used for validation. The results show that the mean yearly bias is reduced by 90%, and the yearly root mean square error is reduced by 68% during the validation year. The experimental results confirm that the proposed method can effectively adjust the bias of satellite-based precipitation estimations
- …