80 research outputs found

    Optimality conditions, approximate stationarity, and applications 'a story beyond lipschitzness

    Get PDF
    Approximate necessary optimality conditions in terms of Frechet subgradients and normals for a rather general optimization problem with a potentially non-Lipschitzian objective function are established with the aid of Ekeland's variational principle, the fuzzy Frechet subdifferential sum rule, and a novel notion of lower semicontinuity relative to a set-valued mapping or set. Feasible points satisfying these optimality conditions are referred to as approximately stationary. As applications, we derive a new general version of the extremal principle. Furthermore, we study approximate stationarity conditions for an optimization problem with a composite objective function and geometric constraints, a qualification condition guaranteeing that approximately stationary points of such a problem are M-stationary, and a multiplier-penalty-method which naturally computes approximately stationary points of the underlying problem. Finally, necessary optimality conditions for an optimal control problem with a non-Lipschitzian sparsity-promoting term in the objective function are established. © The authors

    On the Finite-Time Complexity and Practical Computation of Approximate Stationarity Concepts of Lipschitz Functions

    Full text link
    We report a practical finite-time algorithmic scheme to compute approximately stationary points for nonconvex nonsmooth Lipschitz functions. In particular, we are interested in two kinds of approximate stationarity notions for nonconvex nonsmooth problems, i.e., Goldstein approximate stationarity (GAS) and near-approximate stationarity (NAS). For GAS, our scheme removes the unrealistic subgradient selection oracle assumption in (Zhang et al., 2020, Assumption 1) and computes GAS with the same finite-time complexity. For NAS, Davis & Drusvyatskiy (2019) showed that ρ\rho-weakly convex functions admit finite-time computation, while Tian & So (2021) provided the matching impossibility results of dimension-free finite-time complexity for first-order methods. Complement to these developments, in this paper, we isolate a new class of functions that could be Clarke irregular (and thus not weakly convex anymore) and show that our new algorithmic scheme can compute NAS points for functions in that class within finite time. To demonstrate the wide applicability of our new theoretical framework, we show that ρ\rho-margin SVM, 11-layer, and 22-layer ReLU neural networks, all being Clarke irregular, satisfy our new conditions.Comment: 20 pages, 3 figures, ICML 202

    Inexact proximal methods for weakly convex functions

    Full text link
    This paper proposes and develops inexact proximal methods for finding stationary points of the sum of a smooth function and a nonsmooth weakly convex one, where an error is present in the calculation of the proximal mapping of the nonsmooth term. A general framework for finding zeros of a continuous mapping is derived from our previous paper on this subject to establish convergence properties of the inexact proximal point method when the smooth term is vanished and of the inexact proximal gradient method when the smooth term satisfies a descent condition. The inexact proximal point method achieves global convergence with constructive convergence rates when the Moreau envelope of the objective function satisfies the Kurdyka-Lojasiewicz (KL) property. Meanwhile, when the smooth term is twice continuously differentiable with a Lipschitz continuous gradient and a differentiable approximation of the objective function satisfies the KL property, the inexact proximal gradient method achieves the global convergence of iterates with constructive convergence rates.Comment: 26 pages, 3 table

    A decomposition algorithm for two-stage stochastic programs with nonconvex recourse

    Full text link
    In this paper, we have studied a decomposition method for solving a class of nonconvex two-stage stochastic programs, where both the objective and constraints of the second-stage problem are nonlinearly parameterized by the first-stage variable. Due to the failure of the Clarke regularity of the resulting nonconvex recourse function, classical decomposition approaches such as Benders decomposition and (augmented) Lagrangian-based algorithms cannot be directly generalized to solve such models. By exploring an implicitly convex-concave structure of the recourse function, we introduce a novel decomposition framework based on the so-called partial Moreau envelope. The algorithm successively generates strongly convex quadratic approximations of the recourse function based on the solutions of the second-stage convex subproblems and adds them to the first-stage master problem. Convergence under both fixed scenarios and interior samplings is established. Numerical experiments are conducted to demonstrate the effectiveness of the proposed algorithm

    Nondifferentiable Optimization: Motivations and Applications

    Get PDF
    IIASA has been involved in research on nondifferentiable optimization since 1976. The Institute's research in this field has been very productive, leading to many important theoretical, algorithmic and applied results. Nondifferentiable optimization has now become a recognized and rapidly developing branch of mathematical programming. To continue this tradition and to review developments in this field IIASA held this Workshop in Sopron (Hungary) in September 1984. This volume contains selected papers presented at the Workshop. It is divided into four sections dealing with the following topics: (I) Concepts in Nonsmooth Analysis; (II) Multicriteria Optimization and Control Theory; (III) Algorithms and Optimization Methods; (IV) Stochastic Programming and Applications

    Design of Low-Order Controllers using Optimization Techniques

    Get PDF
    In many applications, especially in the process industry, low-level controllers are the workhorses of the automated production lines. The aim of this study has been to provide simple tuning procedures, either optimization-based methods or tuning rules, for design of low-order controllers. The first part of this thesis deals with PID tuning. Design methods or both SISO and MIMO PID controllers based on convex optimization are presented. The methods consist of solving a nonconvex optimization problem by deriving convex approximations of the original problem and solving these iteratively until convergence. The algorithms are fast because of the convex approximations. The controllers obtained minimize low-frequency sensitivity subject to constraints that ensure robustness to process variations and limitations of control signal effort. The second part of this thesis deals with tuning of feedforward controllers. Tuning rules that minimize the integrated-squared-error arising from measurable step disturbances are derived for a controller that can be interpreted as a filtered and possibly time-delayed PD controller. Using a controller structure that decouples the effects of the feedforward and feedback controllers, the controller is optimal both in open and closed loop settings. To improve the high-frequency noise behavior of the feedforward controller, it is proposed that the optimal controller is augmented with a second-order filter. Several aspects on the tuning of this filter are discussed. For systems with PID controllers, the response to step changes in the reference can be improved by introducing set-point weighting. This can be interpreted as feedforward from the reference signal to the control signal. It is shown how these weights can be found by solving a convex optimization problem. Proportional set-point weight that minimizes the integrated-absolute-error was obtained for a batch of over 130 different processes. From these weights, simple tuning rules were derived and the performance was evaluated on all processes in the batch using five different feedback controller tuning methods. The proposed tuning rules could improve the performance by up to 45% with a modest increase in actuation

    Variational Methods for Evolution

    Get PDF
    The meeting focused on the last advances in the applications of variational methods to evolution problems governed by partial differential equations. The talks covered a broad range of topics, including large deviation and variational principles, rate-independent evolutions and gradient flows, heat flows in metric-measure spaces, propagation of fracture, applications of optimal transport and entropy-entropy dissipation methods, phase-transitions, viscous approximation, and singular-perturbation problems
    corecore