17,251 research outputs found
Smoothing and filtering with a class of outer measures
Filtering and smoothing with a generalised representation of uncertainty is
considered. Here, uncertainty is represented using a class of outer measures.
It is shown how this representation of uncertainty can be propagated using
outer-measure-type versions of Markov kernels and generalised Bayesian-like
update equations. This leads to a system of generalised smoothing and filtering
equations where integrals are replaced by supremums and probability density
functions are replaced by positive functions with supremum equal to one.
Interestingly, these equations retain most of the structure found in the
classical Bayesian filtering framework. It is additionally shown that the
Kalman filter recursion can be recovered from weaker assumptions on the
available information on the corresponding hidden Markov model
Unambiguous determination of gravitational waveforms from binary black hole mergers
Gravitational radiation is properly defined only at future null infinity
(\scri), but in practice it is estimated from data calculated at a finite
radius. We have used characteristic extraction to calculate gravitational
radiation at \scri for the inspiral and merger of two equal mass non-spinning
black holes. Thus we have determined the first unambiguous merger waveforms for
this problem. The implementation is general purpose, and can be applied to
calculate the gravitational radiation, at \scri, given data at a finite
radius calculated in another computation.Comment: 4 pages, 3 figures, published versio
Recommended from our members
An Approach to Using Non Safety-Assured Programmable Components in Modest Integrity Systems
Programmable components (like personal computers or smart devices) can offer considerable benefits in terms of usability and functionality in a safety-related system. However there is a problem in justifying the use of programmable components if the components have not been safety justified to an appropriate integrity (e.g. to SIL 1 of IEC 61508). This paper outlines an approach (called LowSIL) developed in the UK CINIF nuclear industry research programme to justify the use of non safety-assured programmable components in modest integrity systems. This is a seven step approach that can be applied to new systems from an early design stage, or retrospectively to existing systems. The stages comprise: system characterisation, component suitability assessment, failure analysis, failure mitigation, identification of additional defences, identification of safety evidence requirements, and collation and evaluation of evidence. In the case of personal computers, there is supporting guidance on usage constraints, claim limits on reliability, and advice on “locking down” the component to maximise reliability. The approach is demonstrated for an example system. The approach has been applied successfully to a range of safety-related systems used in the nuclear industry
An Introduction to Wishart Matrix Moments
These lecture notes provide a comprehensive, self-contained introduction to
the analysis of Wishart matrix moments. This study may act as an introduction
to some particular aspects of random matrix theory, or as a self-contained
exposition of Wishart matrix moments. Random matrix theory plays a central role
in statistical physics, computational mathematics and engineering sciences,
including data assimilation, signal processing, combinatorial optimization,
compressed sensing, econometrics and mathematical finance, among numerous
others. The mathematical foundations of the theory of random matrices lies at
the intersection of combinatorics, non-commutative algebra, geometry,
multivariate functional and spectral analysis, and of course statistics and
probability theory. As a result, most of the classical topics in random matrix
theory are technical, and mathematically difficult to penetrate for non-experts
and regular users and practitioners. The technical aim of these notes is to
review and extend some important results in random matrix theory in the
specific context of real random Wishart matrices. This special class of
Gaussian-type sample covariance matrix plays an important role in multivariate
analysis and in statistical theory. We derive non-asymptotic formulae for the
full matrix moments of real valued Wishart random matrices. As a corollary, we
derive and extend a number of spectral and trace-type results for the case of
non-isotropic Wishart random matrices. We also derive the full matrix moment
analogues of some classic spectral and trace-type moment results. For example,
we derive semi-circle and Marchencko-Pastur-type laws in the non-isotropic and
full matrix cases. Laplace matrix transforms and matrix moment estimates are
also studied, along with new spectral and trace concentration-type
inequalities
Chemoinformatics Research at the University of Sheffield: A History and Citation Analysis
This paper reviews the work of the Chemoinformatics Research Group in the Department of Information Studies at the University of Sheffield, focusing particularly on the work carried out in the period 1985-2002. Four major research areas are discussed, these involving the development of methods for: substructure searching in databases of three-dimensional structures, including both rigid and flexible molecules; the representation and searching of the Markush structures that occur in chemical patents; similarity searching in databases of both two-dimensional and three-dimensional structures; and compound selection and the design of combinatorial libraries. An analysis of citations to 321 publications from the Group shows that it attracted a total of 3725 residual citations during the period 1980-2002. These citations appeared in 411 different journals, and involved 910 different citing organizations from 54 different countries, thus demonstrating the widespread impact of the Group's work
Generative Models For Deep Learning with Very Scarce Data
The goal of this paper is to deal with a data scarcity scenario where deep
learning techniques use to fail. We compare the use of two well established
techniques, Restricted Boltzmann Machines and Variational Auto-encoders, as
generative models in order to increase the training set in a classification
framework. Essentially, we rely on Markov Chain Monte Carlo (MCMC) algorithms
for generating new samples. We show that generalization can be improved
comparing this methodology to other state-of-the-art techniques, e.g.
semi-supervised learning with ladder networks. Furthermore, we show that RBM is
better than VAE generating new samples for training a classifier with good
generalization capabilities
- …