1,631 research outputs found
Preference incorporation in MOEA/D using an outranking approach with imprecise model parameters
Multi-objective Optimization Evolutionary Algorithms (MOEAs) face numerous challenges when they are used to solve Many-objective Optimization Problems (MaOPs). Decomposition-based strategies, such as MOEA/D, divide an MaOP into multiple single-optimization sub-problems, achieving better diversity and a better approximation of the Pareto front, and dealing with some of the challenges of MaOPs. However, these approaches still require one to solve a multi-criteria selection problem that will allow a Decision-Maker (DM) to choose the final solution. Incorporating preferences may provide results that are closer to the region of interest of a DM. Most of the proposals to integrate preferences in decomposition-based MOEAs prefer progressive articulation over the âa prioriâ incorporation of preferences. Progressive articulation methods can hardly work without comparable and transitive preferences, and they can significantly increase the cognitive effort required of a DM. On the other hand, the âa prioriâ strategies do not demand transitive judgements from the DM but require a direct parameter elicitation that usually is subject to imprecision. Outranking approaches have properties that allow them to suitably handle non-transitive preferences, veto conditions, and incomparability, which are typical characteristics of many real DMs. This paper explores how to incorporate DM preferences into MOEA/D using the âa prioriâ incorporation of preferences, based on interval outranking relations, to handle imprecision when preference parameters are elicited. Several experiments make it possible to analyze the proposal's performance on benchmark problems and to compare the results with the classic MOEA/D without preference incorporation and with a recent, state-of-the-art preference-based decomposition algorithm. In many instances, our results are closer to the Region of Interest, particularly when the number of objectives increases
RIGA: A Regret-Based Interactive Genetic Algorithm
In this paper, we propose an interactive genetic algorithm for solving
multi-objective combinatorial optimization problems under preference
imprecision. More precisely, we consider problems where the decision maker's
preferences over solutions can be represented by a parameterized aggregation
function (e.g., a weighted sum, an OWA operator, a Choquet integral), and we
assume that the parameters are initially not known by the recommendation
system. In order to quickly make a good recommendation, we combine elicitation
and search in the following way: 1) we use regret-based elicitation techniques
to reduce the parameter space in a efficient way, 2) genetic operators are
applied on parameter instances (instead of solutions) to better explore the
parameter space, and 3) we generate promising solutions (population) using
existing solving methods designed for the problem with known preferences. Our
algorithm, called RIGA, can be applied to any multi-objective combinatorial
optimization problem provided that the aggregation function is linear in its
parameters and that a (near-)optimal solution can be efficiently determined for
the problem with known preferences. We also study its theoretical performances:
RIGA can be implemented in such way that it runs in polynomial time while
asking no more than a polynomial number of queries. The method is tested on the
multi-objective knapsack and traveling salesman problems. For several
performance indicators (computation times, gap to optimality and number of
queries), RIGA obtains better results than state-of-the-art algorithms
A Perturbed Self-organizing Multiobjective Evolutionary Algorithm to solve Multiobjective TSP
Travelling Salesman Problem (TSP) is a very important NP-Hard problem getting focused more on these days. Having improvement on TSP, right now consider the multi-objective TSP (MOTSP), broadened occurrence of travelling salesman problem. Since TSP is NP-hard issue MOTSP is additionally a NP-hard issue. There are a lot of algorithms and methods to solve the MOTSP among which Multiobjective evolutionary algorithm based on decomposition is appropriate to solve it nowadays. This work presents a new algorithm which combines the Data Perturbation, Self-Organizing Map (SOM) and MOEA/D to solve the problem of MOTSP, named Perturbed Self-Organizing multiobjective Evolutionary Algorithm (P-SMEA). In P-SMEA Self-Organizing Map (SOM) is used extract neighborhood relationship information and with MOEA/D subproblems are generated and solved simultaneously to obtain the optimal solution. Data Perturbation is applied to avoid the local optima. So by using the P-SMEA, MOTSP can be handled efficiently. The experimental results show that P-SMEA outperforms MOEA/D and SMEA on a set of test instances
ETEA: A euclidean minimum spanning tree-Based evolutionary algorithm for multiobjective optimization
Š the Massachusetts Institute of TechnologyAbstract The Euclidean minimum spanning tree (EMST), widely used in a variety of domains, is a minimum spanning tree of a set of points in the space, where the edge weight between each pair of points is their Euclidean distance. Since the generation of an EMST is entirely determined by the Euclidean distance between solutions (points), the properties of EMSTs have a close relation with the distribution and position information of solutions. This paper explores the properties of EMSTs and proposes an EMST-based Evolutionary Algorithm (ETEA) to solve multiobjective optimization problems (MOPs). Unlike most EMO algorithms that focus on the Pareto dominance relation, the proposed algorithm mainly considers distance-based measures to evaluate and compare individuals during the evolutionary search. Specifically in ETEA, four strategies are introduced: 1) An EMST-based crowding distance (ETCD) is presented to estimate the density of individuals in the population; 2) A distance comparison approach incorporating ETCD is used to assign the fitness value for individuals; 3) A fitness adjustment technique is designed to avoid the partial overcrowding in environmental selection; 4) Three diversity indicators-the minimum edge, degree, and ETCD-with regard to EMSTs are applied to determine the survival of individuals in archive truncation. From a series of extensive experiments on 32 test instances with different characteristics, ETEA is found to be competitive against five state-of-the-art algorithms and its predecessor in providing a good balance among convergence, uniformity, and spread.Engineering and Physical Sciences Research Council (EPSRC) of the United Kingdom under
Grant EP/K001310/1, and the National Natural Science Foundation of China under Grant 61070088
Data mining in soft computing framework: a survey
The present article provides a survey of the available literature on data mining using soft computing. A categorization has been provided based on the different soft computing tools and their hybridizations used, the data mining function implemented, and the preference criterion selected by the model. The utility of the different soft computing methodologies is highlighted. Generally fuzzy sets are suitable for handling the issues related to understandability of patterns, incomplete/noisy data, mixed media information and human interaction, and can provide approximate solutions faster. Neural networks are nonparametric, robust, and exhibit good learning and generalization capabilities in data-rich environments. Genetic algorithms provide efficient search algorithms to select a model, from mixed media data, based on some preference criterion/objective function. Rough sets are suitable for handling different types of uncertainty in data. Some challenges to data mining and the application of soft computing methodologies are indicated. An extensive bibliography is also included
Data-driven aerodynamic shape design with distributionally robust optimization approaches
We formulate and solve data-driven aerodynamic shape design problems with
distributionally robust optimization (DRO) approaches. Building on the findings
of the work \cite{gotoh2018robust}, we study the connections between a class of
DRO and the Taguchi method in the context of robust design optimization. Our
preliminary computational experiments on aerodynamic shape optimization in
transonic turbulent flow show promising design results
Numerical and Evolutionary Optimization 2020
This book was established after the 8th International Workshop on Numerical and Evolutionary Optimization (NEO), representing a collection of papers on the intersection of the two research areas covered at this workshop: numerical optimization and evolutionary search techniques. While focusing on the design of fast and reliable methods lying across these two paradigms, the resulting techniques are strongly applicable to a broad class of real-world problems, such as pattern recognition, routing, energy, lines of production, prediction, and modeling, among others. This volume is intended to serve as a useful reference for mathematicians, engineers, and computer scientists to explore current issues and solutions emerging from these mathematical and computational methods and their applications
Robustness - a challenge also for the 21st century: A review of robustness phenomena in technical, biological and social systems as well as robust approaches in engineering, computer science, operations research and decision aiding
Notions on robustness exist in many facets. They come from different disciplines and reflect different worldviews. Consequently, they contradict each other very often, which makes the term less applicable in a general context. Robustness approaches are often limited to specific problems for which they have been developed. This means, notions and definitions might reveal to be wrong if put into another domain of validity, i.e. context. A definition might be correct in a specific context but need not hold in another. Therefore, in order to be able to speak of robustness we need to specify the domain of validity, i.e. system, property and uncertainty of interest. As proofed by Ho et al. in an optimization context with finite and discrete domains, without prior knowledge about the problem there exists no solution what so ever which is more robust than any other. Similar to the results of the No Free Lunch Theorems of Optimization (NLFTs) we have to exploit the problem structure in order to make a solution more robust. This optimization problem is directly linked to a robustness/fragility tradeoff which has been observed in many contexts, e.g. 'robust, yet fragile' property of HOT (Highly Optimized Tolerance) systems. Another issue is that robustness is tightly bounded to other phenomena like complexity for which themselves exist no clear definition or theoretical framework. Consequently, this review rather tries to find common aspects within many different approaches and phenomena than to build a general theorem for robustness, which anyhow might not exist because complex phenomena often need to be described from a pluralistic view to address as many aspects of a phenomenon as possible. First, many different robustness problems have been reviewed from many different disciplines. Second, different common aspects will be discussed, in particular the relationship of functional and structural properties. This paper argues that robustness phenomena are also a challenge for the 21st century. It is a useful quality of a model or system in terms of the 'maintenance of some desired system characteristics despite fluctuations in the behaviour of its component parts or its environment' (s. [Carlson and Doyle, 2002], p. 2). We define robustness phenomena as solution with balanced tradeoffs and robust design principles and robustness measures as means to balance tradeoffs. --
Recommended from our members
Computational intelligence techniques in asset risk analysis
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The problem of asset risk analysis is positioned within the computational intelligence paradigm. We suggest an algorithm for reformulating asset pricing, which involves incorporating imprecise information into the pricing factors through fuzzy variables as well as a calibration procedure for their possibility distributions. Then fuzzy mathematics is used to process the imprecise factors and obtain an asset evaluation. This evaluation is further automated using neural networks with sign restrictions on their weights. While such type of networks has been only used for up to two network inputs and hypothetical data, here we apply thirty-six inputs and empirical data. To achieve successful training, we modify the Levenberg-Marquart backpropagation algorithm. The intermediate result achieved is that the fuzzy asset evaluation inherits features of the factor imprecision and provides the basis for risk analysis. Next, we formulate a risk measure and a risk robustness measure based on the fuzzy asset evaluation under different characteristics of the pricing factors as well as different calibrations. Our database, extracted from DataStream, includes thirty-five companies traded on the London Stock Exchange. For each company, the risk and robustness measures are evaluated and an asset risk analysis is carried out through these values, indicating the implications they have on company performance. A comparative company risk analysis is also provided. Then, we employ both risk measures to formulate a two-step asset ranking method. The assets are initially rated according to the investors' risk preference. In addition, an algorithm is suggested to incorporate the asset robustness information and refine further the ranking benefiting market analysts. The rationale provided by the ranking technique serves as a point of departure in designing an asset risk classifier. We identify the fuzzy neural network structure of the classifier and develop an evolutionary training algorithm. The algorithm starts with suggesting preliminary heuristics in constructing a sufficient training set of assets with various characteristics revealed by the values of the pricing factors and the asset risk values. Then, the training algorithm works at two levels, the inner level targets weight optimization, while the outer level efficiently guides the exploration of the search space. The latter is achieved by automatically decomposing the training set into subsets of decreasing complexity and then incrementing backward the corresponding subpopulations of partially trained networks. The empirical results prove that the developed algorithm is capable of training the identified fuzzy network structure. This is a problem of such complexity that prevents single-level evolution from attaining meaningful results. The final outcome is an automatic asset classifier, based on the investorsâ perceptions of acceptable risk. All the steps described above constitute our approach to reformulating asset risk analysis within the approximate reasoning framework through the fusion of various computational intelligence techniques
- âŚ