1,227 research outputs found
Supervised classification and mathematical optimization
Data Mining techniques often ask for the resolution of optimization problems. Supervised Classification, and, in particular, Support Vector Machines, can be seen as a paradigmatic instance. In this paper, some links between Mathematical Optimization methods and Supervised Classification are emphasized. It is shown that many different areas of Mathematical Optimization play a central role in off-the-shelf Supervised Classification methods. Moreover, Mathematical Optimization turns out to be extremely
useful to address important issues in Classification, such as identifying relevant variables, improving the interpretability of classifiers or dealing with vagueness/noise in the data.Ministerio de Ciencia e InnovaciónJunta de Andalucí
A D.C. Programming Approach to the Sparse Generalized Eigenvalue Problem
In this paper, we consider the sparse eigenvalue problem wherein the goal is
to obtain a sparse solution to the generalized eigenvalue problem. We achieve
this by constraining the cardinality of the solution to the generalized
eigenvalue problem and obtain sparse principal component analysis (PCA), sparse
canonical correlation analysis (CCA) and sparse Fisher discriminant analysis
(FDA) as special cases. Unlike the -norm approximation to the
cardinality constraint, which previous methods have used in the context of
sparse PCA, we propose a tighter approximation that is related to the negative
log-likelihood of a Student's t-distribution. The problem is then framed as a
d.c. (difference of convex functions) program and is solved as a sequence of
convex programs by invoking the majorization-minimization method. The resulting
algorithm is proved to exhibit \emph{global convergence} behavior, i.e., for
any random initialization, the sequence (subsequence) of iterates generated by
the algorithm converges to a stationary point of the d.c. program. The
performance of the algorithm is empirically demonstrated on both sparse PCA
(finding few relevant genes that explain as much variance as possible in a
high-dimensional gene dataset) and sparse CCA (cross-language document
retrieval and vocabulary selection for music retrieval) applications.Comment: 40 page
Supervised Classification and Mathematical Optimization
Data Mining techniques often ask for the resolution of optimization problems. Supervised Classification, and, in particular, Support Vector Machines, can be seen as a paradigmatic instance. In this paper, some links between Mathematical Optimization methods and Supervised Classification are emphasized. It is shown that many different areas of Mathematical Optimization play a central role in off-the-shelf Supervised Classification methods. Moreover, Mathematical Optimization turns out to be extremely useful to address important issues in Classification, such as identifying relevant variables, improving the interpretability of classifiers or dealing with vagueness/noise in the data
Recent Advances in Graph Partitioning
We survey recent trends in practical algorithms for balanced graph
partitioning together with applications and future research directions
Proximity Queries for Absolutely Continuous Parametric Curves
In motion planning problems for autonomous robots, such as self-driving cars,
the robot must ensure that its planned path is not in close proximity to
obstacles in the environment. However, the problem of evaluating the proximity
is generally non-convex and serves as a significant computational bottleneck
for motion planning algorithms. In this paper, we present methods for a general
class of absolutely continuous parametric curves to compute: (i) the minimum
separating distance, (ii) tolerance verification, and (iii) collision
detection. Our methods efficiently compute bounds on obstacle proximity by
bounding the curve in a convex region. This bound is based on an upper bound on
the curve arc length that can be expressed in closed form for a useful class of
parametric curves including curves with trigonometric or polynomial bases. We
demonstrate the computational efficiency and accuracy of our approach through
numerical simulations of several proximity problems.Comment: Proceedings of Robotics: Science and System
Solving the p -Median Problem with a Semi-Lagrangian Relaxation
Lagrangian relaxation is commonly used in combinatorial optimization to generate lower bounds for a minimization problem. We study a modified Lagrangian relaxation which generates an optimal integer solution. We call it semi-Lagrangian relaxation and illustrate its practical value by solving large-scale instances of the p-median proble
Ground state determination, ground state preserving fit for cluster expansion and their integration for robust CE construction
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Materials Science and Engineering, February 2018.Cataloged from PDF version of thesis.Includes bibliographical references (pages 106-111).In this thesis, we propose strategies to solve the general ground state problem for arbitrary effective cluster interactions and construct ground state preserving cluster expansions. A full mathematical definition of our problem has been formalized to illustrate its generality and clarify our discussion. We review previous methods in material science community: Monte Carlo based method, configurational polytope method, and basic ray method. Further, we investigate the connection of the ground state problem with deeper mathematical results about computational complexity and NP-hard combinatorial optimization (MAX-SAT). We have proposed a general scheme, upper bound and lower bound calculation to approach this problem. Firstly, based on the traditional configurational polytope method, we have proposed a method called cluster tree optimization method, which eliminates the necessity of introducing an exponential number of variables to counter frustration, and thus significantly improves tractability. Secondly, based on convex optimization and finite optimization without periodicity, we have introduced a beautiful MAX-MIN method to refine lower bound calculation. Finally, we present a systematic and mathematically sound method to obtain cluster expansion models that are guaranteed to preserve the ground states of the reference data.by Wenxuan Huang.Ph. D
- …