64,412 research outputs found
Penalized Orthogonal Iteration for Sparse Estimation of Generalized Eigenvalue Problem
We propose a new algorithm for sparse estimation of eigenvectors in
generalized eigenvalue problems (GEP). The GEP arises in a number of modern
data-analytic situations and statistical methods, including principal component
analysis (PCA), multiclass linear discriminant analysis (LDA), canonical
correlation analysis (CCA), sufficient dimension reduction (SDR) and invariant
co-ordinate selection. We propose to modify the standard generalized orthogonal
iteration with a sparsity-inducing penalty for the eigenvectors. To achieve
this goal, we generalize the equation-solving step of orthogonal iteration to a
penalized convex optimization problem. The resulting algorithm, called
penalized orthogonal iteration, provides accurate estimation of the true
eigenspace, when it is sparse. Also proposed is a computationally more
efficient alternative, which works well for PCA and LDA problems. Numerical
studies reveal that the proposed algorithms are competitive, and that our
tuning procedure works well. We demonstrate applications of the proposed
algorithm to obtain sparse estimates for PCA, multiclass LDA, CCA and SDR.
Supplementary materials are available online
Estimating the Expected Value of Partial Perfect Information in Health Economic Evaluations using Integrated Nested Laplace Approximation
The Expected Value of Perfect Partial Information (EVPPI) is a
decision-theoretic measure of the "cost" of parametric uncertainty in decision
making used principally in health economic decision making. Despite this
decision-theoretic grounding, the uptake of EVPPI calculations in practice has
been slow. This is in part due to the prohibitive computational time required
to estimate the EVPPI via Monte Carlo simulations. However, recent developments
have demonstrated that the EVPPI can be estimated by non-parametric regression
methods, which have significantly decreased the computation time required to
approximate the EVPPI. Under certain circumstances, high-dimensional Gaussian
Process regression is suggested, but this can still be prohibitively expensive.
Applying fast computation methods developed in spatial statistics using
Integrated Nested Laplace Approximations (INLA) and projecting from a
high-dimensional into a low-dimensional input space allows us to decrease the
computation time for fitting these high-dimensional Gaussian Processes, often
substantially. We demonstrate that the EVPPI calculated using our method for
Gaussian Process regression is in line with the standard Gaussian Process
regression method and that despite the apparent methodological complexity of
this new method, R functions are available in the package BCEA to implement it
simply and efficiently
Stable Recovery Of Sparse Vectors From Random Sinusoidal Feature Maps
Random sinusoidal features are a popular approach for speeding up
kernel-based inference in large datasets. Prior to the inference stage, the
approach suggests performing dimensionality reduction by first multiplying each
data vector by a random Gaussian matrix, and then computing an element-wise
sinusoid. Theoretical analysis shows that collecting a sufficient number of
such features can be reliably used for subsequent inference in kernel
classification and regression.
In this work, we demonstrate that with a mild increase in the dimension of
the embedding, it is also possible to reconstruct the data vector from such
random sinusoidal features, provided that the underlying data is sparse enough.
In particular, we propose a numerically stable algorithm for reconstructing the
data vector given the nonlinear features, and analyze its sample complexity.
Our algorithm can be extended to other types of structured inverse problems,
such as demixing a pair of sparse (but incoherent) vectors. We support the
efficacy of our approach via numerical experiments
Efficient hierarchical approximation of high-dimensional option pricing problems
A major challenge in computational finance is the pricing of options that depend on a large number of risk factors. Prominent examples are basket or index options where dozens or even hundreds of stocks constitute the underlying asset and determine the dimensionality of the corresponding degenerate parabolic equation. The objective of this article is to show how an efficient discretisation can be achieved by hierarchical approximation as well as asymptotic expansions of the underlying continuous problem. The relation to a number of state-of-the-art methods is highlighted
- …