90,000 research outputs found
Quantile-based bias correction and uncertainty quantification of extreme event attribution statements
Extreme event attribution characterizes how anthropogenic climate change may
have influenced the probability and magnitude of selected individual extreme
weather and climate events. Attribution statements often involve quantification
of the fraction of attributable risk (FAR) or the risk ratio (RR) and
associated confidence intervals. Many such analyses use climate model output to
characterize extreme event behavior with and without anthropogenic influence.
However, such climate models may have biases in their representation of extreme
events. To account for discrepancies in the probabilities of extreme events
between observational datasets and model datasets, we demonstrate an
appropriate rescaling of the model output based on the quantiles of the
datasets to estimate an adjusted risk ratio. Our methodology accounts for
various components of uncertainty in estimation of the risk ratio. In
particular, we present an approach to construct a one-sided confidence interval
on the lower bound of the risk ratio when the estimated risk ratio is infinity.
We demonstrate the methodology using the summer 2011 central US heatwave and
output from the Community Earth System Model. In this example, we find that the
lower bound of the risk ratio is relatively insensitive to the magnitude and
probability of the actual event.Comment: 28 pages, 4 figures, 3 table
Quantifying the effect of interannual ocean variability on the attribution of extreme climate events to human influence
In recent years, the climate change research community has become highly
interested in describing the anthropogenic influence on extreme weather events,
commonly termed "event attribution." Limitations in the observational record
and in computational resources motivate the use of uncoupled,
atmosphere/land-only climate models with prescribed ocean conditions run over a
short period, leading up to and including an event of interest. In this
approach, large ensembles of high-resolution simulations can be generated under
factual observed conditions and counterfactual conditions that might have been
observed in the absence of human interference; these can be used to estimate
the change in probability of the given event due to anthropogenic influence.
However, using a prescribed ocean state ignores the possibility that estimates
of attributable risk might be a function of the ocean state. Thus, the
uncertainty in attributable risk is likely underestimated, implying an
over-confidence in anthropogenic influence.
In this work, we estimate the year-to-year variability in calculations of the
anthropogenic contribution to extreme weather based on large ensembles of
atmospheric model simulations. Our results both quantify the magnitude of
year-to-year variability and categorize the degree to which conclusions of
attributable risk are qualitatively affected. The methodology is illustrated by
exploring extreme temperature and precipitation events for the northwest coast
of South America and northern-central Siberia; we also provides results for
regions around the globe. While it remains preferable to perform a full
multi-year analysis, the results presented here can serve as an indication of
where and when attribution researchers should be concerned about the use of
atmosphere-only simulations
Recommended from our members
Forecasted attribution of the human influence on Hurricane Florence.
Changes in extreme weather, such as tropical cyclones, are one of the most serious ways society experiences the impact of climate change. Advance forecasted conditional attribution statements, using a numerical model, were made about the anthropogenic climate change influence on an individual tropical cyclone, Hurricane Florence. Mean total overland rainfall amounts associated with the forecasted storm's core were increased by 4.9 ± 4.6% with local maximum amounts experiencing increases of 3.8 ± 5.7% due to climate change. A slight increase in the forecasted storm size of 1 to 2% was also attributed. This work reviews our forecasted attribution statement with the benefit of hindsight, demonstrating credibility of advance attribution statements for tropical cyclones
Untenable nonstationarity: An assessment of the fitness for purpose of trend tests in hydrology
The detection and attribution of long-term patterns in hydrological time series have been important research topics for decades. A significant portion of the literature regards such patterns as âdeterministic componentsâ or âtrendsâ even though the complexity of hydrological systems does not allow easy deterministic explanations and attributions. Consequently, trend estimation techniques have been developed to make and justify statements about tendencies in the historical data, which are often used to predict future events. Testing trend hypothesis on observed time series is widespread in the hydro-meteorological literature mainly due to the interest in detecting consequences of human activities on the hydrological cycle. This analysis usually relies on the application of some null hypothesis significance tests (NHSTs) for slowly-varying and/or abrupt changes, such as Mann-Kendall, Pettitt, or similar, to summary statistics of hydrological time series (e.g., annual averages, maxima, minima, etc.). However, the reliability of this application has seldom been explored in detail. This paper discusses misuse, misinterpretation, and logical flaws of NHST for trends in the analysis of hydrological data from three different points of view: historic-logical, semantic-epistemological, and practical. Based on a review of NHST rationale, and basic statistical definitions of stationarity, nonstationarity, and ergodicity, we show that even if the empirical estimation of trends in hydrological time series is always feasible from a numerical point of view, it is uninformative and does not allow the inference of nonstationarity without assuming a priori additional information on the underlying stochastic process, according to deductive reasoning. This prevents the use of trend NHST outcomes to support nonstationary frequency analysis and modeling. We also show that the correlation structures characterizing hydrological time series might easily be underestimated, further compromising the attempt to draw conclusions about trends spanning the period of records. Moreover, even though adjusting procedures accounting for correlation have been developed, some of them are insufficient or are applied only to some tests, while some others are theoretically flawed but still widely applied. In particular, using 250 unimpacted stream flow time series across the conterminous United States (CONUS), we show that the test results can dramatically change if the sequences of annual values are reproduced starting from daily stream flow records, whose larger sizes enable a more reliable assessment of the correlation structures
DADA: data assimilation for the detection and attribution of weather and climate-related events
A new nudging method for data assimilation, delayâcoordinate nudging, is presented. Delayâcoordinate nudging makes explicit use of present and past observations in the formulation of the forcing driving the model evolution at each time step. Numerical experiments with a lowâorder chaotic system show that the new method systematically outperforms standard nudging in different model and observational scenarios, also when using an unoptimized formulation of the delayânudging coefficients. A connection between the optimal delay and the dominant Lyapunov exponent of the dynamics is found based on heuristic arguments and is confirmed by the numerical results, providing a guideline for the practical implementation of the algorithm. Delayâcoordinate nudging preserves the easiness of implementation, the intuitive functioning and the reduced computational cost of the standard nudging, making it a potential alternative especially in the field of seasonalâtoâdecadal predictions with large Earth system models that limit the use of more sophisticated data assimilation procedures
Breaks, trends and the attribution of climate change: a time-series analysis
Climate change detection and attribution have been the subject of intense research and debate over at least four decades. However, direct attribution of climate change to anthropogenic activities using observed climate and forcing variables through statistical methods has remained elusive, partly caused by difficulties to correctly identify the time-series properties of these variables and by the limited availability of methods to relate nonstationary variables. This paper provides strong evidence concerning the direct attribution of observed climate change to anthropogenic greenhouse gases emissions by first investigating the univariate time-series properties of observed global and hemispheric temperatures and forcing variables and then by proposing statistically adequate multivariate models. The results show that there is a clear anthropogenic fingerprint on both global and hemispheric temperatures. The signal of the well-mixed Greenhouse Gases (GHG) forcing in all temperature series is very clear and accounts for most of their secular movements since the beginning of observations. Both temperature and forcing variables are characterized by piecewise linear trends with abrupt changes in their slopes estimated to occur at different dates. Nevertheless, their long-term movements are so closely related that the observed temperature and forcing trends cancel out. The warming experimented during the last century was mainly due to the increase in GHG which was partially offset by the effect of tropospheric aerosols. Other forcing sources, such as solar, are shown to only contribute to (shorter-term) variations around the GHG forcing trend.Published versio
- âŠ