9 research outputs found
Stable machine-learning parameterization of subgrid processes for climate modeling at a range of resolutions
Global climate models represent small-scale processes such as clouds and
convection using quasi-empirical models known as parameterizations, and these
parameterizations are a leading cause of uncertainty in climate projections. A
promising alternative approach is to use machine learning to build new
parameterizations directly from high-resolution model output. However,
parameterizations learned from three-dimensional model output have not yet been
successfully used for simulations of climate. Here we use a random forest to
learn a parameterization of subgrid processes from output of a
three-dimensional high-resolution atmospheric model. Integrating this
parameterization into the atmospheric model leads to stable simulations at
coarse resolution that replicate the climate of the high-resolution simulation.
The parameterization obeys physical constraints and captures important
statistics such as precipitation extremes. The ability to learn from a fully
three-dimensional simulation presents an opportunity for learning
parameterizations from the wide range of global high-resolution simulations
that are now emerging.Comment: Main: 27 pages, 5 figures SI: 19 pages, 11 figures, 4 table
Non-local parameterization of atmospheric subgrid processes with neural networks
Subgrid processes in global climate models are represented by
parameterizations that are a major source of uncertainties in simulations of
climate. In recent years, it has been suggested that new machine-learning
parameterizations learned from high-resolution model output data could be
superior to traditional parameterizations. Currently, both traditional and
machine-learning parameterizations of subgrid processes in the atmosphere are
based on a single-column approach. Namely, the information used by these
parameterizations is taken from a single atmospheric column. However, a
single-column approach might not be ideal for the parameterization problem
since certain atmospheric phenomena, such as organized convective systems, can
cross multiple grid boxes and involve slantwise circulations that are not
purely vertical. Here we train neural networks using non-local inputs spanning
over 33 columns of inputs. We find that including the non-local inputs
substantially improves the prediction of subgrid tendencies of a range of
subgrid processes. The improvement is especially notable for cases associated
with mid-latitude fronts and convective instability. Using an explainable
artificial intelligence technique called layer-wise relevance propagation, we
find that non-local inputs from zonal and meridional winds contain information
that helps to improve the performance of the neural network parameterization.
Our results imply that use of non-local inputs has the potential to
substantially improve both traditional and machine-learning parameterizations.Comment: 31 pages, 17 figures (7 figures in the main file
Climate-Invariant Machine Learning
Data-driven algorithms, in particular neural networks, can emulate the
effects of unresolved processes in coarse-resolution climate models when
trained on high-resolution simulation data; however, they often make large
generalization errors when evaluated in conditions they were not trained on.
Here, we propose to physically rescale the inputs and outputs of machine
learning algorithms to help them generalize to unseen climates. Applied to
offline parameterizations of subgrid-scale thermodynamics in three distinct
climate models, we show that rescaled or "climate-invariant" neural networks
make accurate predictions in test climates that are 4K and 8K warmer than their
training climates. Additionally, "climate-invariant" neural nets facilitate
generalization between Aquaplanet and Earth-like simulations. Through
visualization and attribution methods, we show that compared to standard
machine learning models, "climate-invariant" algorithms learn more local and
robust relations between storm-scale convection, radiation, and their synoptic
thermodynamic environment. Overall, these results suggest that explicitly
incorporating physical knowledge into data-driven models of Earth system
processes can improve their consistency and ability to generalize across
climate regimes.Comment: 12+18 pages, 8+12 figures, 2+2 tables in the main text +
supplementary information. Submitted to PNAS on December 14th, 202
ClimSim: A large multi-scale dataset for hybrid physics-ML climate emulation
Modern climate projections lack adequate spatial and temporal resolution due to computational constraints. A consequence is inaccurate and imprecise predictions of critical processes such as storms. Hybrid methods that combine physics with machine learning (ML) have introduced a new generation of higher fidelity climate simulators that can sidestep Moore's Law by outsourcing compute-hungry, short, high-resolution simulations to ML emulators. However, this hybrid ML-physics simulation approach requires domain-specific treatment and has been inaccessible to ML experts because of lack of training data and relevant, easy-to-use workflows. We present ClimSim, the largest-ever dataset designed for hybrid ML-physics research. It comprises multi-scale climate simulations, developed by a consortium of climate scientists and ML researchers. It consists of 5.7 billion pairs of multivariate input and output vectors that isolate the influence of locally-nested, high-resolution, high-fidelity physics on a host climate simulator's macro-scale physical state.The dataset is global in coverage, spans multiple years at high sampling frequency, and is designed such that resulting emulators are compatible with downstream coupling into operational climate simulators. We implement a range of deterministic and stochastic regression baselines to highlight the ML challenges and their scoring. The data (https://huggingface.co/datasets/LEAP/ClimSim_high-res) and code (https://leap-stc.github.io/ClimSim) are released openly to support the development of hybrid ML-physics and high-fidelity climate simulations for the benefit of science and society
The intensification of winter mid-latitude storms in the Southern Hemisphere
The strength of mid-latitude storm tracks shapes weather and climate
phenomena in the extra-tropics, as these storm tracks control the daily to
multi-decadal variability of precipitation, temperature and winds. By the end
of this century, winter mid-latitude storms are projected to intensify in the
Southern Hemisphere, with large consequences over the entire extra-tropics.
Therefore, it is critical to be able to accurately assess the impacts of
anthropogenic emissions on these storms, in order to improve societal
preparedness for future changes. Here we show that current climate models
severely underestimate the intensification in mid-latitude storm-tracks in
recent decades. Specifically, the intensification obtained from reanalyses has
already reached the model-projected end of the century intensification. The
biased intensification is found to be linked to biases in the zonal flow. These
results question the ability of climate models to accurately predict the future
impacts of anthropogenic emissions in the Southern Hemisphere mid-latitudes.Comment: 3 figures in main, 8 figures in S
Use of neural networks for stable, accurate and physically consistent parameterization of subgrid atmospheric processes with good performance at reduced precision
A promising approach to improve climate-model simulations is to replace
traditional subgrid parameterizations based on simplified physical models by
machine learning algorithms that are data-driven. However, neural networks
(NNs) often lead to instabilities and climate drift when coupled to an
atmospheric model. Here we learn an NN parameterization from a high-resolution
atmospheric simulation in an idealized domain by coarse graining the model
equations and output. The NN parameterization has a structure that ensures
physical constraints are respected, and it leads to stable simulations that
replicate the climate of the high-resolution simulation with similar accuracy
to a successful random-forest parameterization while needing far less memory.
We find that the simulations are stable for a variety of NN architectures and
horizontal resolutions, and that an NN with substantially reduced numerical
precision could decrease computational costs without affecting the quality of
simulations
Response of extreme precipitation to uniform surface warming in quasi-global aquaplanet simulations at high resolution
Projections of precipitation extremes in simulations with global climate models are very uncertain in the tropics, in part because of the use of parameterizations of deep convection and model deficiencies in simulating convective organization. Here, we analyse precipitation extremes in high-resolution simulations that are run without a convective parameterization on a quasi-global aquaplanet. The frequency distributions of precipitation rates and precipitation cluster sizes in the tropics of a control simulation are similar to the observed distributions. In response to climate warming, 3 h precipitation extremes increase at rates of up to [Formula: see text] in the tropics because of a combination of positive thermodynamic and dynamic contributions. The dynamic contribution at different latitudes is connected to the vertical structure of warming using a moist static stability. When the precipitation rates are first averaged to a daily timescale and coarse-grained to a typical global climate-model resolution prior to calculating the precipitation extremes, the response of the precipitation extremes to warming becomes more similar to what was found previously in coarse-resolution aquaplanet studies. However, the simulations studied here do not exhibit the high rates of increase of tropical precipitation extremes found in projections with some global climate models. This article is part of a discussion meeting issue 'Intensification of short-duration rainfall extremes and implications for flash flood risks'