674 research outputs found
Multilevel Monte Carlo for Random Degenerate Scalar Convection Diffusion Equation
We consider the numerical solution of scalar, nonlinear degenerate
convection-diffusion problems with random diffusion coefficient and with random
flux functions. Building on recent results on the existence, uniqueness and
continuous dependence of weak solutions on data in the deterministic case, we
develop a definition of random entropy solution. We establish existence,
uniqueness, measurability and integrability results for these random entropy
solutions, generalizing \cite{Mishr478,MishSch10a} to possibly degenerate
hyperbolic-parabolic problems with random data. We next address the numerical
approximation of random entropy solutions, specifically the approximation of
the deterministic first and second order statistics. To this end, we consider
explicit and implicit time discretization and Finite Difference methods in
space, and single as well as Multi-Level Monte-Carlo methods to sample the
statistics. We establish convergence rate estimates with respect to the
discretization parameters, as well as with respect to the overall work,
indicating substantial gains in efficiency are afforded under realistic
regularity assumptions by the use of the Multi-Level Monte-Carlo method.
Numerical experiments are presented which confirm the theoretical convergence
estimates.Comment: 24 Page
Sparse Finite Elements for Stochastic Elliptic Problems - Higher Order Moments
We define the higher order moments associated to the stochastic solution of an elliptic BVP in Dââ d with stochastic input data. We prove that the k-th moment solves a deterministic problem in D k ââ dk , for which we discuss well-posedness and regularity. We discretize the deterministic k-th moment problem using sparse grids and, exploiting a spline wavelet basis, we propose an efficient algorithm, of logarithmic-linear complexity, for solving the resulting syste
CORRECTION TO "QUASI-MONTE CARLO METHODS FOR HIGH-DIMENSIONAL INTEGRATION: THE STANDARD (WEIGHTED HILBERT SPACE) SETTING AND BEYONDâ
ISSN:1446-1811ISSN:1446-873
Quasi-Monte Carlo methods for high-dimensional integration: the standard (weighted Hilbert space) setting and beyond
This paper is a contemporary review of quasi-Monte Carlo (QMC) methods, that is, equal-weight rules for the approximate evaluation of high-dimensional integrals over the unit cube . It first introduces the by-now standard setting of weighted Hilbert spaces of functions with square-integrable mixed first derivatives, and then indicates alternative settings, such as non-Hilbert spaces, that can sometimes be more suitable. Original contributions include the extension of the fast component-by-component (CBC) construction of lattice rules that achieve the optimal convergence order (a rate of almost , where is the number of points, independently of dimension) to so-called ââŹĹproduct and order dependentâ⏠(POD) weights, as seen in some recent applications. Although the paper has a strong focus on lattice rules, the function space settings are applicable to all QMC methods. Furthermore, the error analysis and construction of lattice rules can be adapted to polynomial lattice rules from the family of digital nets.
doi:10.1017/S144618111200007
Electron transport through quantum wires and point contacts
We have studied quantum wires using the Green's function technique and the
density-functional theory, calculating the electronic structure and the
conductance. All the numerics are implemented using the finite-element method
with a high-order polynomial basis. For short wires, i.e. quantum point
contacts, the zero-bias conductance shows, as a function of the gate voltage
and at a finite temperature, a plateau at around 0.7G_0. (G_0 = 2e^2/h is the
quantum conductance). The behavior, which is caused in our mean-field model by
spontaneous spin polarization in the constriction, is reminiscent of the
so-called 0.7-anomaly observed in experiments. In our model the temperature and
the wire length affect the conductance-gate voltage curves in the same way as
in the measured data.Comment: 8 page
Multiresolution kernel matrix algebra
We propose a sparse arithmetic for kernel matrices, enabling efficient
scattered data analysis. The compression of kernel matrices by means of
samplets yields sparse matrices such that assembly, addition, and
multiplication of these matrices can be performed with essentially linear cost.
Since the inverse of a kernel matrix is compressible, too, we have also fast
access to the inverse kernel matrix by employing exact sparse selected
inversion techniques. As a consequence, we can rapidly evaluate series
expansions and contour integrals to access, numerically and approximately in a
data-sparse format, more complicated matrix functions such as and
. By exploiting the matrix arithmetic, also efficient Gaussian process
learning algorithms for spatial statistics can be realized. Numerical results
are presented to quantify and quality our findings
Higher-order Quasi-Monte Carlo Training of Deep Neural Networks
We present a novel algorithmic approach and an error analysis leveraging
Quasi-Monte Carlo points for training deep neural network (DNN) surrogates of
Data-to-Observable (DtO) maps in engineering design. Our analysis reveals
higher-order consistent, deterministic choices of training points in the input
data space for deep and shallow Neural Networks with holomorphic activation
functions such as tanh. These novel training points are proved to facilitate
higher-order decay (in terms of the number of training samples) of the
underlying generalization error, with consistency error bounds that are free
from the curse of dimensionality in the input data space, provided that DNN
weights in hidden layers satisfy certain summability conditions. We present
numerical experiments for DtO maps from elliptic and parabolic PDEs with
uncertain inputs that confirm the theoretical analysis
- âŚ