1,227 research outputs found

    Supervised classification and mathematical optimization

    Get PDF
    Data Mining techniques often ask for the resolution of optimization problems. Supervised Classification, and, in particular, Support Vector Machines, can be seen as a paradigmatic instance. In this paper, some links between Mathematical Optimization methods and Supervised Classification are emphasized. It is shown that many different areas of Mathematical Optimization play a central role in off-the-shelf Supervised Classification methods. Moreover, Mathematical Optimization turns out to be extremely useful to address important issues in Classification, such as identifying relevant variables, improving the interpretability of classifiers or dealing with vagueness/noise in the data.Ministerio de Ciencia e InnovaciónJunta de Andalucí

    A D.C. Programming Approach to the Sparse Generalized Eigenvalue Problem

    Full text link
    In this paper, we consider the sparse eigenvalue problem wherein the goal is to obtain a sparse solution to the generalized eigenvalue problem. We achieve this by constraining the cardinality of the solution to the generalized eigenvalue problem and obtain sparse principal component analysis (PCA), sparse canonical correlation analysis (CCA) and sparse Fisher discriminant analysis (FDA) as special cases. Unlike the 1\ell_1-norm approximation to the cardinality constraint, which previous methods have used in the context of sparse PCA, we propose a tighter approximation that is related to the negative log-likelihood of a Student's t-distribution. The problem is then framed as a d.c. (difference of convex functions) program and is solved as a sequence of convex programs by invoking the majorization-minimization method. The resulting algorithm is proved to exhibit \emph{global convergence} behavior, i.e., for any random initialization, the sequence (subsequence) of iterates generated by the algorithm converges to a stationary point of the d.c. program. The performance of the algorithm is empirically demonstrated on both sparse PCA (finding few relevant genes that explain as much variance as possible in a high-dimensional gene dataset) and sparse CCA (cross-language document retrieval and vocabulary selection for music retrieval) applications.Comment: 40 page

    Supervised Classification and Mathematical Optimization

    Get PDF
    Data Mining techniques often ask for the resolution of optimization problems. Supervised Classification, and, in particular, Support Vector Machines, can be seen as a paradigmatic instance. In this paper, some links between Mathematical Optimization methods and Supervised Classification are emphasized. It is shown that many different areas of Mathematical Optimization play a central role in off-the-shelf Supervised Classification methods. Moreover, Mathematical Optimization turns out to be extremely useful to address important issues in Classification, such as identifying relevant variables, improving the interpretability of classifiers or dealing with vagueness/noise in the data

    Proximity Queries for Absolutely Continuous Parametric Curves

    Full text link
    In motion planning problems for autonomous robots, such as self-driving cars, the robot must ensure that its planned path is not in close proximity to obstacles in the environment. However, the problem of evaluating the proximity is generally non-convex and serves as a significant computational bottleneck for motion planning algorithms. In this paper, we present methods for a general class of absolutely continuous parametric curves to compute: (i) the minimum separating distance, (ii) tolerance verification, and (iii) collision detection. Our methods efficiently compute bounds on obstacle proximity by bounding the curve in a convex region. This bound is based on an upper bound on the curve arc length that can be expressed in closed form for a useful class of parametric curves including curves with trigonometric or polynomial bases. We demonstrate the computational efficiency and accuracy of our approach through numerical simulations of several proximity problems.Comment: Proceedings of Robotics: Science and System

    Solving the p -Median Problem with a Semi-Lagrangian Relaxation

    Get PDF
    Lagrangian relaxation is commonly used in combinatorial optimization to generate lower bounds for a minimization problem. We study a modified Lagrangian relaxation which generates an optimal integer solution. We call it semi-Lagrangian relaxation and illustrate its practical value by solving large-scale instances of the p-median proble

    Ground state determination, ground state preserving fit for cluster expansion and their integration for robust CE construction

    Get PDF
    Thesis: Ph. D., Massachusetts Institute of Technology, Department of Materials Science and Engineering, February 2018.Cataloged from PDF version of thesis.Includes bibliographical references (pages 106-111).In this thesis, we propose strategies to solve the general ground state problem for arbitrary effective cluster interactions and construct ground state preserving cluster expansions. A full mathematical definition of our problem has been formalized to illustrate its generality and clarify our discussion. We review previous methods in material science community: Monte Carlo based method, configurational polytope method, and basic ray method. Further, we investigate the connection of the ground state problem with deeper mathematical results about computational complexity and NP-hard combinatorial optimization (MAX-SAT). We have proposed a general scheme, upper bound and lower bound calculation to approach this problem. Firstly, based on the traditional configurational polytope method, we have proposed a method called cluster tree optimization method, which eliminates the necessity of introducing an exponential number of variables to counter frustration, and thus significantly improves tractability. Secondly, based on convex optimization and finite optimization without periodicity, we have introduced a beautiful MAX-MIN method to refine lower bound calculation. Finally, we present a systematic and mathematically sound method to obtain cluster expansion models that are guaranteed to preserve the ground states of the reference data.by Wenxuan Huang.Ph. D
    corecore