137,949 research outputs found
Probabilistic error estimation for non-intrusive reduced models learned from data of systems governed by linear parabolic partial differential equations
This work derives a residual-based a posteriori error estimator for reduced
models learned with non-intrusive model reduction from data of high-dimensional
systems governed by linear parabolic partial differential equations with
control inputs. It is shown that quantities that are necessary for the error
estimator can be either obtained exactly as the solutions of least-squares
problems in a non-intrusive way from data such as initial conditions, control
inputs, and high-dimensional solution trajectories or bounded in a
probabilistic sense. The computational procedure follows an offline/online
decomposition. In the offline (training) phase, the high-dimensional system is
judiciously solved in a black-box fashion to generate data and to set up the
error estimator. In the online phase, the estimator is used to bound the error
of the reduced-model predictions for new initial conditions and new control
inputs without recourse to the high-dimensional system. Numerical results
demonstrate the workflow of the proposed approach from data to reduced models
to certified predictions
Meta learning of bounds on the Bayes classifier error
Meta learning uses information from base learners (e.g. classifiers or
estimators) as well as information about the learning problem to improve upon
the performance of a single base learner. For example, the Bayes error rate of
a given feature space, if known, can be used to aid in choosing a classifier,
as well as in feature selection and model selection for the base classifiers
and the meta classifier. Recent work in the field of f-divergence functional
estimation has led to the development of simple and rapidly converging
estimators that can be used to estimate various bounds on the Bayes error. We
estimate multiple bounds on the Bayes error using an estimator that applies
meta learning to slowly converging plug-in estimators to obtain the parametric
convergence rate. We compare the estimated bounds empirically on simulated data
and then estimate the tighter bounds on features extracted from an image patch
analysis of sunspot continuum and magnetogram images.Comment: 6 pages, 3 figures, to appear in proceedings of 2015 IEEE Signal
Processing and SP Education Worksho
Influence Maximization with Bandits
We consider the problem of \emph{influence maximization}, the problem of
maximizing the number of people that become aware of a product by finding the
`best' set of `seed' users to expose the product to. Most prior work on this
topic assumes that we know the probability of each user influencing each other
user, or we have data that lets us estimate these influences. However, this
information is typically not initially available or is difficult to obtain. To
avoid this assumption, we adopt a combinatorial multi-armed bandit paradigm
that estimates the influence probabilities as we sequentially try different
seed sets. We establish bounds on the performance of this procedure under the
existing edge-level feedback as well as a novel and more realistic node-level
feedback. Beyond our theoretical results, we describe a practical
implementation and experimentally demonstrate its efficiency and effectiveness
on four real datasets.Comment: 12 page
- …