16 research outputs found
Learning to forecast diagnostic parameters using pre-trained weather embedding
Data-driven weather prediction (DDWP) models are increasingly becoming
popular for weather forecasting. However, while operational weather forecasts
predict a wide variety of weather variables, DDWPs currently forecast a
specific set of key prognostic variables. Non-prognostic ("diagnostic")
variables are sometimes modeled separately as dependent variables of the
prognostic variables (c.f. FourCastNet), or by including the diagnostic
variable as a target in the DDWP. However, the cost of training and deploying
bespoke models for each diagnostic variable can increase dramatically with more
diagnostic variables, and limit the operational use of such models. Likewise,
retraining an entire DDWP each time a new diagnostic variable is added is also
cost-prohibitive. We present an two-stage approach that allows new diagnostic
variables to be added to an end-to-end DDWP model without the expensive
retraining. In the first stage, we train an autoencoder that learns to embed
prognostic variables into a latent space. In the second stage, the autoencoder
is frozen and "downstream" models are trained to predict diagnostic variables
using only the latent representations of prognostic variables as input. Our
experiments indicate that models trained using the two-stage approach offer
accuracy comparable to training bespoke models, while leading to significant
reduction in resource utilization during training and inference. This approach
allows for new "downstream" models to be developed as needed, without affecting
existing models and thus reducing the friction in operationalizing new models.Comment: Accepted as a spotlight paper at the NeurIPS 2023 workshop on
Tackling Climate Change with Machine Learnin
Verification against in-situ observations for Data-Driven Weather Prediction
Data-driven weather prediction models (DDWPs) have made rapid strides in
recent years, demonstrating an ability to approximate Numerical Weather
Prediction (NWP) models to a high degree of accuracy. The fast, accurate, and
low-cost DDWP forecasts make their use in operational forecasting an attractive
proposition, however, there remains work to be done in rigorously evaluating
DDWPs in a true operational setting. Typically trained and evaluated using ERA5
reanalysis data, DDWPs have been tested only in a simulation, which cannot
represent the real world with complete accuracy even if it is of a very high
quality. The safe use of DDWPs in operational forecasting requires more
thorough "real-world" verification, as well as a careful examination of how
DDWPs are currently trained and evaluated. It is worth asking, for instance,
how well do the reanalysis datasets, used for training, simulate the real
world? With an eye towards climate justice and the uneven availability of
weather data: is the simulation equally good for all regions of the world, and
would DDWPs exacerbate biases present in the training data? Does a good
performance in simulation correspond to good performance in operational
settings? In addition to approximating the physics of NWP models, how can ML be
uniquely deployed to provide more accurate weather forecasts? As a first step
towards answering such questions, we present a robust dataset of in-situ
observations derived from the NOAA MADIS program to serve as a benchmark to
validate DDWPs in an operational setting. By providing a large corpus of
quality-controlled, in-situ observations, this dataset provides a meaningful
real-world task that all NWPs and DDWPs can be tested against. We hope that
this data can be used not only to rigorously and fairly compare operational
weather models but also to spur future research in new directions.Comment: 10 pages, 6 figures, under review at NeurIPS main conferenc
Recommended from our members
Acceleration of Chemical Kinetics Computation with the Learned Intelligent Tabulation (LIT) Method
In this work, a data-driven methodology for modeling combustion kinetics, Learned Intelligent Tabulation (LIT), is presented. LIT aims to accelerate the tabulation of combustion mechanisms via machine learning algorithms such as Deep Neural Networks (DNNs). The high-dimensional composition space is sampled from high-fidelity simulations covering a wide range of initial conditions to train these DNNs. The input data are clustered into subspaces, while each subspace is trained with a DNN regression model targeted to a particular part of the high-dimensional composition space. This localized approach has proven to be more tractable than having a global ANN regression model, which fails to generalize across various composition spaces. The clustering is performed using an unsupervised method, Self-Organizing Map (SOM), which automatically subdivides the space. A dense network comprised of fully connected layers is considered for the regression model, while the network hyper parameters are optimized using Bayesian optimization. A nonlinear transformation of the parameters is used to improve sensitivity to minor species and enhance the prediction of ignition delay. The LIT method is employed to model the chemistry kinetics of zero-dimensional H2-O2 and CH4-air combustion. The data-driven method achieves good agreement with the benchmark method while being cheaper in terms of computational cost. LIT is naturally extensible to different combustion models such as flamelet and PDF transport models
Improving CFD simulations by local machine-learned correction
High-fidelity computational fluid dynamics (CFD) simulations for design space
explorations can be exceedingly expensive due to the cost associated with
resolving the finer scales. This computational cost/accuracy trade-off is a
major challenge for modern CFD simulations. In the present study, we propose a
method that uses a trained machine learning model that has learned to predict
the discretization error as a function of largescale flow features to inversely
estimate the degree of lost information due to mesh coarsening. This
information is then added back to the low-resolution solution during runtime,
thereby enhancing the quality of the under-resolved coarse mesh simulation. The
use of a coarser mesh produces a non-linear benefit in speed while the cost of
inferring and correcting for the lost information has a linear cost. We
demonstrate the numerical stability of a problem of engineering interest, a 3D
turbulent channel flow. In addition to this demonstration, we further show the
potential for speedup without sacrificing solution accuracy using this method,
thereby making the cost/accuracy trade-off of CFD more favorable.Comment: 7 pages, under review at ASME IMECE 2023 conferenc