7,890 research outputs found
MeshfreeFlowNet: A Physics-Constrained Deep Continuous Space-Time Super-Resolution Framework
We propose MeshfreeFlowNet, a novel deep learning-based super-resolution
framework to generate continuous (grid-free) spatio-temporal solutions from the
low-resolution inputs. While being computationally efficient, MeshfreeFlowNet
accurately recovers the fine-scale quantities of interest. MeshfreeFlowNet
allows for: (i) the output to be sampled at all spatio-temporal resolutions,
(ii) a set of Partial Differential Equation (PDE) constraints to be imposed,
and (iii) training on fixed-size inputs on arbitrarily sized spatio-temporal
domains owing to its fully convolutional encoder. We empirically study the
performance of MeshfreeFlowNet on the task of super-resolution of turbulent
flows in the Rayleigh-Benard convection problem. Across a diverse set of
evaluation metrics, we show that MeshfreeFlowNet significantly outperforms
existing baselines. Furthermore, we provide a large scale implementation of
MeshfreeFlowNet and show that it efficiently scales across large clusters,
achieving 96.80% scaling efficiency on up to 128 GPUs and a training time of
less than 4 minutes.Comment: Supplementary Video: https://youtu.be/mjqwPch9gDo. Accepted to SC2
Sub-grid modelling for two-dimensional turbulence using neural networks
In this investigation, a data-driven turbulence closure framework is
introduced and deployed for the sub-grid modelling of Kraichnan turbulence. The
novelty of the proposed method lies in the fact that snapshots from
high-fidelity numerical data are used to inform artificial neural networks for
predicting the turbulence source term through localized grid-resolved
information. In particular, our proposed methodology successfully establishes a
map between inputs given by stencils of the vorticity and the streamfunction
along with information from two well-known eddy-viscosity kernels. Through this
we predict the sub-grid vorticity forcing in a temporally and spatially dynamic
fashion. Our study is both a-priori and a-posteriori in nature. In the former,
we present an extensive hyper-parameter optimization analysis in addition to
learning quantification through probability density function based validation
of sub-grid predictions. In the latter, we analyse the performance of our
framework for flow evolution in a classical decaying two-dimensional turbulence
test case in the presence of errors related to temporal and spatial
discretization. Statistical assessments in the form of angle-averaged kinetic
energy spectra demonstrate the promise of the proposed methodology for sub-grid
quantity inference. In addition, it is also observed that some measure of
a-posteriori error must be considered during optimal model selection for
greater accuracy. The results in this article thus represent a promising
development in the formalization of a framework for generation of
heuristic-free turbulence closures from data
Reduced order modeling of fluid flows: Machine learning, Kolmogorov barrier, closure modeling, and partitioning
In this paper, we put forth a long short-term memory (LSTM) nudging framework
for the enhancement of reduced order models (ROMs) of fluid flows utilizing
noisy measurements. We build on the fact that in a realistic application, there
are uncertainties in initial conditions, boundary conditions, model parameters,
and/or field measurements. Moreover, conventional nonlinear ROMs based on
Galerkin projection (GROMs) suffer from imperfection and solution instabilities
due to the modal truncation, especially for advection-dominated flows with slow
decay in the Kolmogorov width. In the presented LSTM-Nudge approach, we fuse
forecasts from a combination of imperfect GROM and uncertain state estimates,
with sparse Eulerian sensor measurements to provide more reliable predictions
in a dynamical data assimilation framework. We illustrate the idea with the
viscous Burgers problem, as a benchmark test bed with quadratic nonlinearity
and Laplacian dissipation. We investigate the effects of measurements noise and
state estimate uncertainty on the performance of the LSTM-Nudge behavior. We
also demonstrate that it can sufficiently handle different levels of temporal
and spatial measurement sparsity. This first step in our assessment of the
proposed model shows that the LSTM nudging could represent a viable realtime
predictive tool in emerging digital twin systems
- …