28 research outputs found

    Infeasible Full-Newton-Step Interior-Point Method for the Linear Complementarity Problems

    Get PDF
    In this tesis, we present a new Infeasible Interior-Point Method (IPM) for monotone Linear Complementarity Problem (LPC). The advantage of the method is that it uses full Newton-steps, thus, avoiding the calculation of the step size at each iteration. However, by suitable choice of parameters the iterates are forced to stay in the neighborhood of the central path, hence, still guaranteeing the global convergence of the method under strict feasibility assumption. The number of iterations necessary to find -approximate solution of the problem matches the best known iteration bounds for these types of methods. The preliminary implementation of the method and numerical results indicate robustness and practical validity of the method

    An infeasible interior-point method for the P∗P_*-matrix linear complementarity‎ ‎problem based on a trigonometric kernel function with full-Newton‎ ‎step

    Get PDF
    An infeasible interior-point algorithm for solving the‎ ‎P∗P_*-matrix linear complementarity problem based on a kernel‎ ‎function with trigonometric barrier term is analyzed‎. ‎Each (main)‎ ‎iteration of the algorithm consists of a feasibility step and‎ ‎several centrality steps‎, ‎whose feasibility step is induced by a‎ ‎trigonometric kernel function‎. ‎The complexity result coincides with‎ ‎the best result for infeasible interior-point methods for‎ ‎P∗P_*-matrix linear complementarity problem

    Full-Newton-Step Interior-Point Method for the Linear Complementarity Problems

    Get PDF
    In this thesis, we present a new Interior-Point Method (IPM) for monotone Linear Complementarity Problem (LPC). The advantage of the method is that it uses full Newton-steps, thus, avoiding the calculation of the step size at each iteration. However, by suitable choice of parameters the iterates are forced to stay in the neighborhood of the central path, hence, still guaranteeing the global convergence of the method under strict feasibility assumption. The number of iterations necessary to find 2-approximate solution of the problem matches the best known iteration bounds for these types of methods. The preliminary implementation of the method and numerical results indicate robustness and practical validity of the method

    International Conference on Continuous Optimization (ICCOPT) 2019 Conference Book

    Get PDF
    The Sixth International Conference on Continuous Optimization took place on the campus of the Technical University of Berlin, August 3-8, 2019. The ICCOPT is a flagship conference of the Mathematical Optimization Society (MOS), organized every three years. ICCOPT 2019 was hosted by the Weierstrass Institute for Applied Analysis and Stochastics (WIAS) Berlin. It included a Summer School and a Conference with a series of plenary and semi-plenary talks, organized and contributed sessions, and poster sessions. This book comprises the full conference program. It contains, in particular, the scientific program in survey style as well as with all details, and information on the social program, the venue, special meetings, and more

    Conic Optimization: Optimal Partition, Parametric, and Stability Analysis

    Get PDF
    A linear conic optimization problem consists of the minimization of a linear objective function over the intersection of an affine space and a closed convex cone. In recent years, linear conic optimization has received significant attention, partly due to the fact that we can take advantage of linear conic optimization to reformulate and approximate intractable optimization problems. Steady advances in computational optimization have enabled us to approximately solve a wide variety of linear conic optimization problems in polynomial time. Nevertheless, preprocessing methods, rounding procedures and sensitivity analysis tools are still the missing parts of conic optimization solvers. Given the output of a conic optimization solver, we need methodologies to generate approximate complementary solutions or to speed up the convergence to an exact optimal solution. A preprocessing method reduces the size of a problem by finding the minimal face of the cone which contains the set of feasible solutions. However, such a preprocessing method assumes the knowledge of an exact solution. More importantly, we need robust sensitivity and post-optimal analysis tools for an optimal solution of a linear conic optimization problem. Motivated by the vital importance of linear conic optimization, we take active steps to fill this gap.This thesis is concerned with several aspects of a linear conic optimization problem, from algorithm through solution identification, to parametric analysis, which have not been fully addressed in the literature. We specifically focus on three special classes of linear conic optimization problems, namely semidefinite and second-order conic optimization, and their common generalization, symmetric conic optimization. We propose a polynomial time algorithm for symmetric conic optimization problems. We show how to approximate/identify the optimal partition of semidefinite optimization and second-order conic optimization, a concept which has its origin in linear optimization. Further, we use the optimal partition information to either generate an approximate optimal solution or to speed up the convergence of a solution identification process to the unique optimal solution of the problem. Finally, we study the parametric analysis of semidefinite and second-order conic optimization problems. We investigate the behavior of the optimal partition and the optimal set mapping under perturbation of the objective function vector

    35th Symposium on Theoretical Aspects of Computer Science: STACS 2018, February 28-March 3, 2018, Caen, France

    Get PDF
    corecore