11,199 research outputs found
A Variation on a Random Coordinate Minimization Method for Constrained Polynomial Optimization
In this paper, an algorithm is proposed for solving constrained and unconstrained polynomial minimization
problems. The algorithm is a variation on random coordinate descent, in which transverse steps are seldom taken.
Differently from other methods available in the literature, the proposed technique is guaranteed
to converge in probability to the global solution of the minimization problem, even when the objective polynomial is nonconvex.
The theoretical results are corroborated by a complexity
analysis and by numerical tests that validate its efficiency
Compressive sensing adaptation for polynomial chaos expansions
Basis adaptation in Homogeneous Chaos spaces rely on a suitable rotation of
the underlying Gaussian germ. Several rotations have been proposed in the
literature resulting in adaptations with different convergence properties. In
this paper we present a new adaptation mechanism that builds on compressive
sensing algorithms, resulting in a reduced polynomial chaos approximation with
optimal sparsity. The developed adaptation algorithm consists of a two-step
optimization procedure that computes the optimal coefficients and the input
projection matrix of a low dimensional chaos expansion with respect to an
optimally rotated basis. We demonstrate the attractive features of our
algorithm through several numerical examples including the application on
Large-Eddy Simulation (LES) calculations of turbulent combustion in a HIFiRE
scramjet engine.Comment: Submitted to Journal of Computational Physic
An Algorithmic Theory of Dependent Regularizers, Part 1: Submodular Structure
We present an exploration of the rich theoretical connections between several
classes of regularized models, network flows, and recent results in submodular
function theory. This work unifies key aspects of these problems under a common
theory, leading to novel methods for working with several important models of
interest in statistics, machine learning and computer vision.
In Part 1, we review the concepts of network flows and submodular function
optimization theory foundational to our results. We then examine the
connections between network flows and the minimum-norm algorithm from
submodular optimization, extending and improving several current results. This
leads to a concise representation of the structure of a large class of pairwise
regularized models important in machine learning, statistics and computer
vision.
In Part 2, we describe the full regularization path of a class of penalized
regression problems with dependent variables that includes the graph-guided
LASSO and total variation constrained models. This description also motivates a
practical algorithm. This allows us to efficiently find the regularization path
of the discretized version of TV penalized models. Ultimately, our new
algorithms scale up to high-dimensional problems with millions of variables
Macrostate Data Clustering
We develop an effective nonhierarchical data clustering method using an
analogy to the dynamic coarse graining of a stochastic system. Analyzing the
eigensystem of an interitem transition matrix identifies fuzzy clusters
corresponding to the metastable macroscopic states (macrostates) of a diffusive
system. A "minimum uncertainty criterion" determines the linear transformation
from eigenvectors to cluster-defining window functions. Eigenspectrum gap and
cluster certainty conditions identify the proper number of clusters. The
physically motivated fuzzy representation and associated uncertainty analysis
distinguishes macrostate clustering from spectral partitioning methods.
Macrostate data clustering solves a variety of test cases that challenge other
methods.Comment: keywords: cluster analysis, clustering, pattern recognition, spectral
graph theory, dynamic eigenvectors, machine learning, macrostates,
classificatio
- …