1,540 research outputs found
Model Reduction and Neural Networks for Parametric PDEs
We develop a general framework for data-driven approximation of input-output maps between infinite-dimensional spaces. The proposed approach is motivated by the recent successes of neural networks and deep learning, in combination with ideas from model reduction. This combination results in a neural network approximation which, in principle, is defined on infinite-dimensional spaces and, in practice, is robust to the dimension of finite-dimensional approximations of these spaces required for computation. For a class of input-output maps, and suitably chosen probability measures on the inputs, we prove convergence of the proposed approximation methodology. Numerically we demonstrate the effectiveness of the method on a class of parametric elliptic PDE problems, showing convergence and robustness of the approximation scheme with respect to the size of the discretization, and compare our method with existing algorithms from the literature
Deep Learning Methods for Partial Differential Equations and Related Parameter Identification Problems
Recent years have witnessed a growth in mathematics for deep learning--which
seeks a deeper understanding of the concepts of deep learning with mathematics
and explores how to make it more robust--and deep learning for mathematics,
where deep learning algorithms are used to solve problems in mathematics. The
latter has popularised the field of scientific machine learning where deep
learning is applied to problems in scientific computing. Specifically, more and
more neural network architectures have been developed to solve specific classes
of partial differential equations (PDEs). Such methods exploit properties that
are inherent to PDEs and thus solve the PDEs better than standard feed-forward
neural networks, recurrent neural networks, or convolutional neural networks.
This has had a great impact in the area of mathematical modeling where
parametric PDEs are widely used to model most natural and physical processes
arising in science and engineering. In this work, we review such methods as
well as their extensions for parametric studies and for solving the related
inverse problems. We equally proceed to show their relevance in some industrial
applications
Feature Enforcing PINN (FE-PINN): A Framework to Learn the Underlying-Physics Features Before Target Task
In this work, a new data-free framework called Feature Enforcing Physics
Informed Neural Network (FE-PINN) is introduced. This framework is capable of
learning the underlying pattern of any problem with low computational cost
before the main training loop. The loss function of vanilla PINN due to the
existence of two terms of partial differential residuals and boundary condition
mean squared error is imbalanced. FE-PINN solves this challenge with just one
minute of training instead of time-consuming hyperparameter tuning for loss
function that can take hours. The FE-PINN accomplishes this process by
performing a sequence of sub-tasks. The first sub-task learns useful features
about the underlying physics. Then, the model trains on the target task to
refine the calculations. FE-PINN is applied to three benchmarks, flow over a
cylinder, 2D heat conduction, and an inverse problem of calculating inlet
velocity. FE-PINN can solve each case with, 15x, 2x, and 5x speed up
accordingly. Another advantage of FE-PINN is that reaching lower order of value
for loss function is systematically possible. In this study, it was possible to
reach a loss value near 1e-5 which is challenging for vanilla PINN. FE-PINN
also has a smooth convergence process which allows for utilizing higher
learning rates in comparison to vanilla PINN. This framework can be used as a
fast, accurate tool for solving a wide range of Partial Differential Equations
(PDEs) across various fields.Comment: 23 pages, 8 figures, 3 table
GINNs:Graph-Informed Neural Networks for Multiscale Physics
We introduce the concept of a Graph-Informed Neural Network (GINN), a hybrid
approach combining deep learning with probabilistic graphical models (PGMs)
that acts as a surrogate for physics-based representations of multiscale and
multiphysics systems. GINNs address the twin challenges of removing intrinsic
computational bottlenecks in physics-based models and generating large data
sets for estimating probability distributions of quantities of interest (QoIs)
with a high degree of confidence. Both the selection of the complex physics
learned by the NN and its supervised learning/prediction are informed by the
PGM, which includes the formulation of structured priors for tunable control
variables (CVs) to account for their mutual correlations and ensure physically
sound CV and QoI distributions. GINNs accelerate the prediction of QoIs
essential for simulation-based decision-making where generating sufficient
sample data using physics-based models alone is often prohibitively expensive.
Using a real-world application grounded in supercapacitor-based energy storage,
we describe the construction of GINNs from a Bayesian network-embedded
homogenized model for supercapacitor dynamics, and demonstrate their ability to
produce kernel density estimates of relevant non-Gaussian, skewed QoIs with
tight confidence intervals.Comment: 20 pages, 8 figure
FO-PINNs: A First-Order formulation for Physics Informed Neural Networks
We present FO-PINNs, physics-informed neural networks that are trained using
the first-order formulation of the Partial Differential Equation (PDE) losses.
We show that FO-PINNs offer significantly higher accuracy in solving
parameterized systems compared to traditional PINNs, and reduce
time-per-iteration by removing the extra backpropagations needed to compute the
second or higher-order derivatives. Additionally, unlike standard PINNs,
FO-PINNs can be used with exact imposition of boundary conditions using
approximate distance functions, and can be trained using Automatic Mixed
Precision (AMP) to further speed up the training. Through two Helmholtz and
Navier-Stokes examples, we demonstrate the advantages of FO-PINNs over
traditional PINNs in terms of accuracy and training speedup.Comment: 6 pages, 3 figures, Selected for ML4PS workshop at NeurIPS 202
- …