1,980 research outputs found
Maximum entropy properties of discrete-time first-order stable spline kernel
The first order stable spline (SS-1) kernel is used extensively in
regularized system identification. In particular, the stable spline estimator
models the impulse response as a zero-mean Gaussian process whose covariance is
given by the SS-1 kernel. In this paper, we discuss the maximum entropy
properties of this prior. In particular, we formulate the exact maximum entropy
problem solved by the SS-1 kernel without Gaussian and uniform sampling
assumptions. Under general sampling schemes, we also explicitly derive the
special structure underlying the SS-1 kernel (e.g. characterizing the
tridiagonal nature of its inverse), also giving to it a maximum entropy
covariance completion interpretation. Along the way similar maximum entropy
properties of the Wiener kernel are also given
Maximum Entropy Kernels for System Identification
A new nonparametric approach for system identification has been recently
proposed where the impulse response is modeled as the realization of a
zero-mean Gaussian process whose covariance (kernel) has to be estimated from
data. In this scheme, quality of the estimates crucially depends on the
parametrization of the covariance of the Gaussian process. A family of kernels
that have been shown to be particularly effective in the system identification
framework is the family of Diagonal/Correlated (DC) kernels. Maximum entropy
properties of a related family of kernels, the Tuned/Correlated (TC) kernels,
have been recently pointed out in the literature. In this paper we show that
maximum entropy properties indeed extend to the whole family of DC kernels. The
maximum entropy interpretation can be exploited in conjunction with results on
matrix completion problems in the graphical models literature to shed light on
the structure of the DC kernel. In particular, we prove that the DC kernel
admits a closed-form factorization, inverse and determinant. These results can
be exploited both to improve the numerical stability and to reduce the
computational complexity associated with the computation of the DC estimator.Comment: Extends results of 2014 IEEE MSC Conference Proceedings
(arXiv:1406.5706
Smoothed Particle Hydrodynamics and Magnetohydrodynamics
This paper presents an overview and introduction to Smoothed Particle
Hydrodynamics and Magnetohydrodynamics in theory and in practice. Firstly, we
give a basic grounding in the fundamentals of SPH, showing how the equations of
motion and energy can be self-consistently derived from the density estimate.
We then show how to interpret these equations using the basic SPH interpolation
formulae and highlight the subtle difference in approach between SPH and other
particle methods. In doing so, we also critique several `urban myths' regarding
SPH, in particular the idea that one can simply increase the `neighbour number'
more slowly than the total number of particles in order to obtain convergence.
We also discuss the origin of numerical instabilities such as the pairing and
tensile instabilities. Finally, we give practical advice on how to resolve
three of the main issues with SPMHD: removing the tensile instability,
formulating dissipative terms for MHD shocks and enforcing the divergence
constraint on the particles, and we give the current status of developments in
this area. Accompanying the paper is the first public release of the NDSPMHD
SPH code, a 1, 2 and 3 dimensional code designed as a testbed for SPH/SPMHD
algorithms that can be used to test many of the ideas and used to run all of
the numerical examples contained in the paper.Comment: 44 pages, 14 figures, accepted to special edition of J. Comp. Phys.
on "Computational Plasma Physics". The ndspmhd code is available for download
from http://users.monash.edu.au/~dprice/ndspmhd
Bayesian Estimation for Continuous-Time Sparse Stochastic Processes
We consider continuous-time sparse stochastic processes from which we have
only a finite number of noisy/noiseless samples. Our goal is to estimate the
noiseless samples (denoising) and the signal in-between (interpolation
problem).
By relying on tools from the theory of splines, we derive the joint a priori
distribution of the samples and show how this probability density function can
be factorized. The factorization enables us to tractably implement the maximum
a posteriori and minimum mean-square error (MMSE) criteria as two statistical
approaches for estimating the unknowns. We compare the derived statistical
methods with well-known techniques for the recovery of sparse signals, such as
the norm and Log (- relaxation) regularization
methods. The simulation results show that, under certain conditions, the
performance of the regularization techniques can be very close to that of the
MMSE estimator.Comment: To appear in IEEE TS
Regularized System Identification
This open access book provides a comprehensive treatment of recent developments in kernel-based identification that are of interest to anyone engaged in learning dynamic systems from data. The reader is led step by step into understanding of a novel paradigm that leverages the power of machine learning without losing sight of the system-theoretical principles of black-box identification. The authors’ reformulation of the identification problem in the light of regularization theory not only offers new insight on classical questions, but paves the way to new and powerful algorithms for a variety of linear and nonlinear problems. Regression methods such as regularization networks and support vector machines are the basis of techniques that extend the function-estimation problem to the estimation of dynamic models. Many examples, also from real-world applications, illustrate the comparative advantages of the new nonparametric approach with respect to classic parametric prediction error methods. The challenges it addresses lie at the intersection of several disciplines so Regularized System Identification will be of interest to a variety of researchers and practitioners in the areas of control systems, machine learning, statistics, and data science. This is an open access book
Probabilistic Numerics and Uncertainty in Computations
We deliver a call to arms for probabilistic numerical methods: algorithms for
numerical tasks, including linear algebra, integration, optimization and
solving differential equations, that return uncertainties in their
calculations. Such uncertainties, arising from the loss of precision induced by
numerical calculation with limited time or hardware, are important for much
contemporary science and industry. Within applications such as climate science
and astrophysics, the need to make decisions on the basis of computations with
large and complex data has led to a renewed focus on the management of
numerical uncertainty. We describe how several seminal classic numerical
methods can be interpreted naturally as probabilistic inference. We then show
that the probabilistic view suggests new algorithms that can flexibly be
adapted to suit application specifics, while delivering improved empirical
performance. We provide concrete illustrations of the benefits of probabilistic
numeric algorithms on real scientific problems from astrometry and astronomical
imaging, while highlighting open problems with these new algorithms. Finally,
we describe how probabilistic numerical methods provide a coherent framework
for identifying the uncertainty in calculations performed with a combination of
numerical algorithms (e.g. both numerical optimisers and differential equation
solvers), potentially allowing the diagnosis (and control) of error sources in
computations.Comment: Author Generated Postprint. 17 pages, 4 Figures, 1 Tabl
- …