2,850 research outputs found
Saddle Points and Pareto Points in Multiple Objective Programming
In this paper relationships between Pareto points and saddle points in multiple objective programming are investigated. Convex and nonconvex problems are considered and the equivalence between Pareto points and saddle points is proved in both cases. The results are based on scalarizations of multiple objective programs and related linear and augmented Lagrangian functions. Partitions of the index sets of objectives and constranints are introduced to reduce the size of the problems. The relevance of the results in the context of decision making is also discussed
Approximating Pareto frontier using a hybrid line search approach
This is the post-print version of the final paper published in Information Sciences. The published article is available from the link below. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. Copyright @ 2010 Elsevier B.V.The aggregation of objectives in multiple criteria programming is one of the simplest and widely used approach. But it is well known that this technique sometimes fail in different aspects for determining the Pareto frontier. This paper proposes a new approach for multicriteria optimization, which aggregates the objective functions and uses a line search method in order to locate an approximate efficient point. Once the first Pareto solution is obtained, a simplified version of the former one is used in the context of Pareto dominance to obtain a set of efficient points, which will assure a thorough distribution of solutions on the Pareto frontier. In the current form, the proposed technique is well suitable for problems having multiple objectives (it is not limited to bi-objective problems) and require the functions to be continuous twice differentiable. In order to assess the effectiveness of this approach, some experiments were performed and compared with two recent well known population-based metaheuristics namely ParEGO and NSGA II. When compared to ParEGO and NSGA II, the proposed approach not only assures a better convergence to the Pareto frontier but also illustrates a good distribution of solutions. From a computational point of view, both stages of the line search converge within a short time (average about 150 ms for the first stage and about 20 ms for the second stage). Apart from this, the proposed technique is very simple, easy to implement and use to solve multiobjective problems.CNCSIS IDEI 2412, Romani
Meta-Modeling by Symbolic Regression and Pareto Simulated Annealing
The subject of this paper is a new approach to Symbolic Regression.Other publications on Symbolic Regression use Genetic Programming.This paper describes an alternative method based on Pareto Simulated Annealing.Our method is based on linear regression for the estimation of constants.Interval arithmetic is applied to ensure the consistency of a model.In order to prevent over-fitting, we merit a model not only on predictions in the data points, but also on the complexity of a model.For the complexity we introduce a new measure.We compare our new method with the Kriging meta-model and against a Symbolic Regression meta-model based on Genetic Programming.We conclude that Pareto Simulated Annealing based Symbolic Regression is very competitive compared to the other meta-model approachesapproximation;meta-modeling;pareto simulated annealing;symbolic regression
A Scalable Algorithm For Sparse Portfolio Selection
The sparse portfolio selection problem is one of the most famous and
frequently-studied problems in the optimization and financial economics
literatures. In a universe of risky assets, the goal is to construct a
portfolio with maximal expected return and minimum variance, subject to an
upper bound on the number of positions, linear inequalities and minimum
investment constraints. Existing certifiably optimal approaches to this problem
do not converge within a practical amount of time at real world problem sizes
with more than 400 securities. In this paper, we propose a more scalable
approach. By imposing a ridge regularization term, we reformulate the problem
as a convex binary optimization problem, which is solvable via an efficient
outer-approximation procedure. We propose various techniques for improving the
performance of the procedure, including a heuristic which supplies high-quality
warm-starts, a preprocessing technique for decreasing the gap at the root node,
and an analytic technique for strengthening our cuts. We also study the
problem's Boolean relaxation, establish that it is second-order-cone
representable, and supply a sufficient condition for its tightness. In
numerical experiments, we establish that the outer-approximation procedure
gives rise to dramatic speedups for sparse portfolio selection problems.Comment: Submitted to INFORMS Journal on Computin
Pareto-Path Multi-Task Multiple Kernel Learning
A traditional and intuitively appealing Multi-Task Multiple Kernel Learning
(MT-MKL) method is to optimize the sum (thus, the average) of objective
functions with (partially) shared kernel function, which allows information
sharing amongst tasks. We point out that the obtained solution corresponds to a
single point on the Pareto Front (PF) of a Multi-Objective Optimization (MOO)
problem, which considers the concurrent optimization of all task objectives
involved in the Multi-Task Learning (MTL) problem. Motivated by this last
observation and arguing that the former approach is heuristic, we propose a
novel Support Vector Machine (SVM) MT-MKL framework, that considers an
implicitly-defined set of conic combinations of task objectives. We show that
solving our framework produces solutions along a path on the aforementioned PF
and that it subsumes the optimization of the average of objective functions as
a special case. Using algorithms we derived, we demonstrate through a series of
experimental results that the framework is capable of achieving better
classification performance, when compared to other similar MTL approaches.Comment: Accepted by IEEE Transactions on Neural Networks and Learning System
Decision analysis: vector optimization theory
First published in Revista de la Real Academia de Ciencias Exactas, FÃsicas y Naturales in 93, 4, 1999, published by the Real Academia de Ciencias Exactas, FÃsicas y Naturales
- …