5,912 research outputs found
Parallelization of the discrete gradient method of non-smooth optimization and its applications
We investigate parallelization and performance of the discrete gradient method of nonsmooth optimization. This derivative free method is shown to be an effective optimization tool, able to skip many shallow local minima of nonconvex nondifferentiable objective functions. Although this is a sequential iterative method, we were able to parallelize critical steps of the algorithm, and this lead to a significant improvement in performance on multiprocessor computer clusters. We applied this method to a difficult polyatomic clusters problem in computational chemistry, and found this method to outperform other algorithms. <br /
Variational image regularization with Euler's elastica using a discrete gradient scheme
This paper concerns an optimization algorithm for unconstrained non-convex
problems where the objective function has sparse connections between the
unknowns. The algorithm is based on applying a dissipation preserving numerical
integrator, the Itoh--Abe discrete gradient scheme, to the gradient flow of an
objective function, guaranteeing energy decrease regardless of step size. We
introduce the algorithm, prove a convergence rate estimate for non-convex
problems with Lipschitz continuous gradients, and show an improved convergence
rate if the objective function has sparse connections between unknowns. The
algorithm is presented in serial and parallel versions. Numerical tests show
its use in Euler's elastica regularized imaging problems and its convergence
rate and compare the execution time of the method to that of the iPiano
algorithm and the gradient descent and Heavy-ball algorithms
Non-smooth optimization methods for computation of the conditional value-at-risk and portfolio optimization
We examine numerical performance of various methods of calculation of the Conditional Value-at-risk (CVaR), and portfolio optimization with respect to this risk measure. We concentrate on the method proposed by Rockafellar and Uryasev in (Rockafellar, R.T. and Uryasev, S., 2000, Optimization of conditional value-at-risk. Journal of Risk, 2, 21-41), which converts this problem to that of convex optimization. We compare the use of linear programming techniques against a non-smooth optimization method of the discrete gradient, and establish the supremacy of the latter. We show that non-smooth optimization can be used efficiently for large portfolio optimization, and also examine parallel execution of this method on computer clusters.<br /
A Parallel Dual Fast Gradient Method for MPC Applications
We propose a parallel adaptive constraint-tightening approach to solve a
linear model predictive control problem for discrete-time systems, based on
inexact numerical optimization algorithms and operator splitting methods. The
underlying algorithm first splits the original problem in as many independent
subproblems as the length of the prediction horizon. Then, our algorithm
computes a solution for these subproblems in parallel by exploiting auxiliary
tightened subproblems in order to certify the control law in terms of
suboptimality and recursive feasibility, along with closed-loop stability of
the controlled system. Compared to prior approaches based on constraint
tightening, our algorithm computes the tightening parameter for each subproblem
to handle the propagation of errors introduced by the parallelization of the
original problem. Our simulations show the computational benefits of the
parallelization with positive impacts on performance and numerical conditioning
when compared with a recent nonparallel adaptive tightening scheme.Comment: This technical report is an extended version of the paper "A Parallel
Dual Fast Gradient Method for MPC Applications" by the same authors submitted
to the 54th IEEE Conference on Decision and Contro
- …