17 research outputs found

    On the Effect of the Cooperation of Indicator-Based Multiobjective Evolutionary Algorithms

    Get PDF
    For almost 20 years, quality indicators (QIs) have promoted the design of new selection mechanisms of multiobjective evolutionary algorithms (MOEAs). Each indicator-based MOEA (IB-MOEA) has specific search preferences related to its baseline QI, producing Pareto front approximations with different properties. In consequence, an IB-MOEA based on a single QI has a limited scope of multiobjective optimization problems (MOPs) in which it is expected to have a good performance. This issue is emphasized when the associated Pareto front geometries are highly irregular. In order to overcome these issues, we propose here an island-based multiindicator algorithm (IMIA) that takes advantage of the search biases of multiple IB-MOEAs through a cooperative scheme. Our experimental results show that the cooperation of multiple IB-MOEAs allows IMIA to perform more robustly (considering several QIs) than the panmictic versions of its baseline IB-MOEAs as well as several state-of-the-art MOEAs. Additionally, IMIA shows a Pareto-front-shape invariance property, which makes it a remarkable optimizer when tackling MOPs with complex Pareto front geometries

    Quality evaluation of solution sets in multiobjective optimisation:a survey

    Get PDF

    Regularized model learning in EDAs for continuous and multi-objective optimization

    Get PDF
    Probabilistic modeling is the de�ning characteristic of estimation of distribution algorithms (EDAs) which determines their behavior and performance in optimization. Regularization is a well-known statistical technique used for obtaining an improved model by reducing the generalization error of estimation, especially in high-dimensional problems. `1-regularization is a type of this technique with the appealing variable selection property which results in sparse model estimations. In this thesis, we study the use of regularization techniques for model learning in EDAs. Several methods for regularized model estimation in continuous domains based on a Gaussian distribution assumption are presented, and analyzed from di�erent aspects when used for optimization in a high-dimensional setting, where the population size of EDA has a logarithmic scale with respect to the number of variables. The optimization results obtained for a number of continuous problems with an increasing number of variables show that the proposed EDA based on regularized model estimation performs a more robust optimization, and is able to achieve signi�cantly better results for larger dimensions than other Gaussian-based EDAs. We also propose a method for learning a marginally factorized Gaussian Markov random �eld model using regularization techniques and a clustering algorithm. The experimental results show notable optimization performance on continuous additively decomposable problems when using this model estimation method. Our study also covers multi-objective optimization and we propose joint probabilistic modeling of variables and objectives in EDAs based on Bayesian networks, speci�cally models inspired from multi-dimensional Bayesian network classi�ers. It is shown that with this approach to modeling, two new types of relationships are encoded in the estimated models in addition to the variable relationships captured in other EDAs: objectivevariable and objective-objective relationships. An extensive experimental study shows the e�ectiveness of this approach for multi- and many-objective optimization. With the proposed joint variable-objective modeling, in addition to the Pareto set approximation, the algorithm is also able to obtain an estimation of the multi-objective problem structure. Finally, the study of multi-objective optimization based on joint probabilistic modeling is extended to noisy domains, where the noise in objective values is represented by intervals. A new version of the Pareto dominance relation for ordering the solutions in these problems, namely �-degree Pareto dominance, is introduced and its properties are analyzed. We show that the ranking methods based on this dominance relation can result in competitive performance of EDAs with respect to the quality of the approximated Pareto sets. This dominance relation is then used together with a method for joint probabilistic modeling based on `1-regularization for multi-objective feature subset selection in classi�cation, where six di�erent measures of accuracy are considered as objectives with interval values. The individual assessment of the proposed joint probabilistic modeling and solution ranking methods on datasets with small-medium dimensionality, when using two di�erent Bayesian classi�ers, shows that comparable or better Pareto sets of feature subsets are approximated in comparison to standard methods

    An approach toward pareto front based dimensionality reduction of parameters in nonlinear circuits

    Get PDF
    In this thesis, we consider dimensionality reduction of parameters in a piecewise linear model of boost converter with PV input. The circuit is subject to a bi-objective optimization problem where the first and second objectives evaluate efficiency of input power and stability of periodic orbits, respectively. A trade off exists between the two objectives and is represented by a Pareto front. An approximation of the Pareto front is a criterion for the dimensionality reduction. Performing precise numerical experiment, a simple dimensionality reduction from two-dimensional parameter space into a one-dimensional parameter subspace is achieved

    Hybrid Optimisation Algorithms for Two-Objective Design of Water Distribution Systems

    Get PDF
    Multi-objective design or extended design of Water Distribution Systems (WDSs) has received more attention in recent years. It is of particular interest for obtaining the trade-offs between cost and hydraulic benefit to support the decision-making process. The design problem is usually formulated as a multi-objective optimisation problem, featuring a huge search space associated with a great number of constraints. Multi-objective evolutionary algorithms (MOEAs) are popular tools for addressing this kind of problem because they are capable of approximating the Pareto-optimal front effectively in a single run. However, these methods are often held by the “No Free Lunch” theorem (Wolpert and Macready 1997) that there is no guarantee that they can perform well on a wide range of cases. To overcome this drawback, many hybrid optimisation methods have been proposed to take advantage of multiple search mechanisms which can synergistically facilitate optimisation. In this thesis, a novel hybrid algorithm, called Genetically Adaptive Leaping Algorithm for approXimation and diversitY (GALAXY), is proposed. It is a dedicated optimiser for solving the discrete two-objective design or extended design of WDSs, minimising the total cost and maximising the network resilience, which is a surrogate indicator of hydraulic benefit. GALAXY is developed using the general framework of MOEAs with substantial improvements and modifications tailored for WDS design. It features a generational framework, a hybrid use of the traditional Pareto-dominance and the epsilon-dominance concepts, an integer coding scheme, and six search operators organised in a high-level teamwork hybrid paradigm. In addition, several important strategies are implemented within GALAXY, including the genetically adaptive strategy, the global information sharing strategy, the duplicates handling strategy and the hybrid replacement strategy. One great advantage of GALAXY over other state-of-the-art MOEAs lies in the fact that it eliminates all the individual parameters of search operators, thus providing an effective and efficient tool to researchers and practitioners alike for dealing with real-world cases. To verify the capability of GALAXY, an archive of benchmark problems of WDS design collected from the literature is first established, ranging from small to large cases. GALAXY has been applied to solve these benchmark design problems and its achievements in terms of both ultimate and dynamic performances are compared with those obtained by two state-of-the-art hybrid algorithms and two baseline MOEAs. GALAXY generally outperforms these MOEAs according to various numerical indicators and a graphical comparison tool. For the largest problem considered in this thesis, GALAXY does not perform as well as its competitors due to the limited computational budget in terms of number of function evaluations. All the algorithms have also been applied to solve the challenging Anytown rehabilitation problem, which considers both the design and operation of a system from the extended period simulation perspective. The performance of each algorithm is sensitive to the quality of the initial population and the random seed used. GALAXY and the Pareto-dominance based MOEAs are superior to the epsilon-dominance based methods; however, there is a tie between GALAXY and the Pareto-dominance based approaches. At the end, a summary of this thesis is provided and relevant conclusions are drawn. Recommendations for future research work are also made

    Handling High-Level Model Changes Using Search Based Software Engineering

    Full text link
    Model-Driven Engineering (MDE) considers models as first-class artifacts during the software lifecycle. The number of available tools, techniques, and approaches for MDE is increasing as its use gains traction in driving quality, and controlling cost in evolution of large software systems. Software models, defined as code abstractions, are iteratively refined, restructured, and evolved. This is due to many reasons such as fixing defects in design, reflecting changes in requirements, and modifying a design to enhance existing features. In this work, we focus on four main problems related to the evolution of software models: 1) the detection of applied model changes, 2) merging parallel evolved models, 3) detection of design defects in merged model, and 4) the recommendation of new changes to fix defects in software models. Regarding the first contribution, a-posteriori multi-objective change detection approach has been proposed for evolved models. The changes are expressed in terms of atomic and composite refactoring operations. The majority of existing approaches detects atomic changes but do not adequately address composite changes which mask atomic operations in intermediate models. For the second contribution, several approaches exist to construct a merged model by incorporating all non-conflicting operations of evolved models. Conflicts arise when the application of one operation disables the applicability of another one. The essence of the problem is to identify and prioritize conflicting operations based on importance and context – a gap in existing approaches. This work proposes a multi-objective formulation of model merging that aims to maximize the number of successfully applied merged operations. For the third and fourth contributions, the majority of existing works focuses on refactoring at source code level, and does not exploit the benefits of software design optimization at model level. However, refactoring at model level is inherently more challenging due to difficulty in assessing the potential impact on structural and behavioral features of the software system. This requires analysis of class and activity diagrams to appraise the overall system quality, feasibility, and inter-diagram consistency. This work focuses on designing, implementing, and evaluating a multi-objective refactoring framework for detection and fixing of design defects in software models.Ph.D.Information Systems Engineering, College of Engineering and Computer ScienceUniversity of Michigan-Dearbornhttp://deepblue.lib.umich.edu/bitstream/2027.42/136077/1/Usman Mansoor Final.pdfDescription of Usman Mansoor Final.pdf : Dissertatio

    Holistic, data-driven, service and supply chain optimisation: linked optimisation.

    Get PDF
    The intensity of competition and technological advancements in the business environment has made companies collaborate and cooperate together as a means of survival. This creates a chain of companies and business components with unified business objectives. However, managing the decision-making process (like scheduling, ordering, delivering and allocating) at the various business components and maintaining a holistic objective is a huge business challenge, as these operations are complex and dynamic. This is because the overall chain of business processes is widely distributed across all the supply chain participants; therefore, no individual collaborator has a complete overview of the processes. Increasingly, such decisions are automated and are strongly supported by optimisation algorithms - manufacturing optimisation, B2B ordering, financial trading, transportation scheduling and allocation. However, most of these algorithms do not incorporate the complexity associated with interacting decision-making systems like supply chains. It is well-known that decisions made at one point in supply chains can have significant consequences that ripple through linked production and transportation systems. Recently, global shocks to supply chains (COVID-19, climate change, blockage of the Suez Canal) have demonstrated the importance of these interdependencies, and the need to create supply chains that are more resilient and have significantly reduced impact on the environment. Such interacting decision-making systems need to be considered through an optimisation process. However, the interactions between such decision-making systems are not modelled. We therefore believe that modelling such interactions is an opportunity to provide computational extensions to current optimisation paradigms. This research study aims to develop a general framework for formulating and solving holistic, data-driven optimisation problems in service and supply chains. This research achieved this aim and contributes to scholarship by firstly considering the complexities of supply chain problems from a linked problem perspective. This leads to developing a formalism for characterising linked optimisation problems as a model for supply chains. Secondly, the research adopts a method for creating a linked optimisation problem benchmark by linking existing classical benchmark sets. This involves using a mix of classical optimisation problems, typically relating to supply chain decision problems, to describe different modes of linkages in linked optimisation problems. Thirdly, several techniques for linking supply chain fragmented data have been proposed in the literature to identify data relationships. Therefore, this thesis explores some of these techniques and combines them in specific ways to improve the data discovery process. Lastly, many state-of-the-art algorithms have been explored in the literature and these algorithms have been used to tackle problems relating to supply chain problems. This research therefore investigates the resilient state-of-the-art optimisation algorithms presented in the literature, and then designs suitable algorithmic approaches inspired by the existing algorithms and the nature of problem linkages to address different problem linkages in supply chains. Considering research findings and future perspectives, the study demonstrates the suitability of algorithms to different linked structures involving two sub-problems, which suggests further investigations on issues like the suitability of algorithms on more complex structures, benchmark methodologies, holistic goals and evaluation, processmining, game theory and dependency analysis
    corecore