1,772 research outputs found

    Efficient dynamic resampling for dominance-based multiobjective evolutionary optimization

    Get PDF
    Multi-objective optimization problems are often subject to the presence of objectives that require expensive resampling for their computation. This is the case for many robustness metrics, which are frequently used as an additional objective that accounts for the reliability of specific sections of the solution space. Typical robustness measurements use resampling, but the number of samples that constitute a precise dispersion measure has a potentially large impact on the computational cost of an algorithm. This article proposes the integration of dominance based statistical testing methods as part of the selection mechanism of evolutionary multi-objective genetic algorithms with the aim of reducing the number of fitness evaluations. The performance of the approach is tested on five classical benchmark functions integrating it into two well-known algorithms, NSGA-II and SPEA2. The experimental results show a significant reduction in the number of fitness evaluations while, at the same time, maintaining the quality of the solutions.The authors acknowledge financial support granted by the Spanish Ministry of Economy and Competitivity under grant ENE2014-56126-C2-2-R

    Elite Accumulative Sampling Strategies for Noisy Multi-Objective Optimisation

    Get PDF
    The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-319-15892-1_128th International Conference on Evolutionary Multi-Criterion Optimization 2015, Guimarães, Portugal, 29 March - 1 April 1 2015The codebase for this paper is available at https://github.com/fieldsend/EMO_2015_eliteWhen designing evolutionary algorithms one of the key concerns is the balance between expending function evaluations on exploration versus exploitation. When the optimisation problem experiences observational noise, there is also a trade-off with respect to accuracy refinement – as improving the estimate of a design’s performance typically is at the cost of additional function reevaluations. Empirically the most effective resampling approach developed so far is accumulative resampling of the elite set. In this approach elite members are regularly reevaluated, meaning they progressively accumulate reevaluations over time. This results in their approximated objective values having greater fidelity, meaning non-dominated solutions are more likely to be correctly identified. Here we examine four different approaches to accumulative resampling of elite members, embedded within a differential evolution algorithm. Comparing results on 40 variants of the unconstrained IEEE CEC’09 multi-objective test problems, we find that at low noise levels a low fixed resample rate is usually sufficient, however for larger noise magnitudes progressively raising the number of minimum resamples of elite members based on detecting estimated front oscillation tends to improve performance

    Portfolio implementation risk management using evolutionary multiobjective optimization

    Get PDF
    Portfoliomanagementbasedonmean-varianceportfoliooptimizationissubjecttodifferent sources of uncertainty. In addition to those related to the quality of parameter estimates used in the optimization process, investors face a portfolio implementation risk. The potential temporary discrepancybetweentargetandpresentportfolios,causedbytradingstrategies,mayexposeinvestors to undesired risks. This study proposes an evolutionary multiobjective optimization algorithm aiming at regions with solutions more tolerant to these deviations and, therefore, more reliable. The proposed approach incorporates a user’s preference and seeks a fine-grained approximation of the most relevant efficient region. The computational experiments performed in this study are based on a cardinality-constrained problem with investment limits for eight broad-category indexes and 15 years of data. The obtained results show the ability of the proposed approach to address the robustness issue and to support decision making by providing a preferred part of the efficient set. The results reveal that the obtained solutions also exhibit a higher tolerance to prediction errors in asset returns and variance–covariance matrix.Sandra Garcia-Rodriguez and David Quintana acknowledge financial support granted by the Spanish Ministry of Economy and Competitivity under grant ENE2014-56126-C2-2-R. Roman Denysiuk and Antonio Gaspar-Cunha were supported by the Portuguese Foundation for Science and Technology under grant PEst-C/CTM/LA0025/2013 (Projecto Estratégico-LA 25-2013-2014-Strategic Project-LA 25-2013-2014).info:eu-repo/semantics/publishedVersio

    Towards the detection and analysis of performance regression introducing code changes

    Get PDF
    In contemporary software development, developers commonly conduct regression testing to ensure that code changes do not affect software quality. Performance regression testing is an emerging research area from the regression testing domain in software engineering. Performance regression testing aims to maintain the system\u27s performance. Conducting performance regression testing is known to be expensive. It is also complex, considering the increase of committed code and developing team members working simultaneously. Many automated regression testing techniques have been proposed in prior research. However, challenges in the practice of locating and resolving performance regression still exist. Directing regression testing to the commit level provides solutions to locate the root cause, yet it hinders the development process. This thesis outlines motivations and solutions to address locating performance regression root causes. First, we challenge a deterministic state-of-art approach by expanding the testing data to find improvement areas. The deterministic approach was found to be limited in searching for the best regression-locating rule. Thus, we presented two stochastic approaches to develop models that can learn from historical commits. The goal of the first stochastic approach is to view the research problem as a search-based optimization problem seeking to reach the highest detection rate. We are applying different multi-objective evolutionary algorithms and conducting a comparison between them. This thesis also investigates whether simplifying the search space by combining objectives would achieve comparative results. The second stochastic approach addresses the severity of class imbalance any system could have since code changes introducing regression are rare but costly. We formulate the identification of problematic commits that introduce performance regression as a binary classification problem that handles class imbalance. Further, the thesis provides an exploratory study on the challenges developers face in resolving performance regression. The study is based on the questions posted on a technical form directed to performance regression. We collected around 2k questions discussing the regression of software execution time, and all were manually analyzed. The study resulted in a categorization of the challenges. We also discussed the difficulty level of performance regression issues within the development community. This study provides insights to help developers during the software design and implementation to avoid regression causes

    A multi-objective evolutionary approach to simulation-based optimisation of real-world problems.

    Get PDF
    This thesis presents a novel evolutionary optimisation algorithm that can improve the quality of solutions in simulation-based optimisation. Simulation-based optimisation is the process of finding optimal parameter settings without explicitly examining each possible configuration of settings. An optimisation algorithm generates potential configurations and sends these to the simulation, which acts as an evaluation function. The evaluation results are used to refine the optimisation such that it eventually returns a high-quality solution. The algorithm described in this thesis integrates multi-objective optimisation, parallelism, surrogate usage, and noise handling in a unique way for dealing with simulation-based optimisation problems incurred by these characteristics. In order to handle multiple, conflicting optimisation objectives, the algorithm uses a Pareto approach in which the set of best trade-off solutions is searched for and presented to the user. The algorithm supports a high degree of parallelism by adopting an asynchronous master-slave parallelisation model in combination with an incremental population refinement strategy. A surrogate evaluation function is adopted in the algorithm to quickly identify promising candidate solutions and filter out poor ones. A novel technique based on inheritance is used to compensate for the uncertainties associated with the approximative surrogate evaluations. Furthermore, a novel technique for multi-objective problems that effectively reduces noise by adopting a dynamic procedure in resampling solutions is used to tackle the problem of real-world unpredictability (noise). The proposed algorithm is evaluated on benchmark problems and two complex real-world problems of manufacturing optimisation. The first real-world problem concerns the optimisation of a production cell at Volvo Aero, while the second one concerns the optimisation of a camshaft machining line at Volvo Cars Engine. The results from the optimisations show that the algorithm finds better solutions for all the problems considered than existing, similar algorithms. The new techniques for dealing with surrogate imprecision and noise used in the algorithm are identified as key reasons for the good performance.University of Skövde Knowledge Foundation Swede

    Application of an evolutionary algorithm-based ensemble model to job-shop scheduling

    Get PDF
    In this paper, a novel evolutionary algorithm is applied to tackle job-shop scheduling tasks in manufacturing environments. Specifically, a modified micro genetic algorithm (MmGA) is used as the building block to formulate an ensemble model to undertake multi-objective optimisation problems in job-shop scheduling. The MmGA ensemble is able to approximate the optimal solution under the Pareto optimality principle. To evaluate the effectiveness of the MmGA ensemble, a case study based on real requirements is conducted. The results positively indicate the effectiveness of the MmGA ensemble in undertaking job-shop scheduling problems

    Pedestrian-Aware Supervisory Control System Interactive Optimization of Connected Hybrid Electric Vehicles via Fuzzy Adaptive Cost Map and Bees Algorithm

    Get PDF
    Electrified vehicles are increasingly being seen as a means of mitigating the pressing concerns of traffic-related pollution. Due to the nature of engine-assisted vehicle exhaust systems, pedestrians in close proximity to these vehicles may experience events where specific emission concentrations are high enough to cause health effects. To minimize pedestrians’ exposure to vehicle emissions and pollutants nearby, we present a pedestrian-aware supervisory control system for connected hybrid electric vehicles by proposing an interactive optimization methodology. This optimization methodology combines a novel fuzzy adaptive cost map and the Bees Algorithm to optimize power-split control parameters. It enables the self-regulation of inter-objective weights of fuel and exhaust emissions based on the real-time pedestrian density information during the optimization process. The evaluation of the vehicle performance by using the proposed methodology is conducted on the realistic trip map involving pedestrian density information collected from the University College Dublin campus. Moreover, two bootstrap sampling techniques and effect of communication quality are both investigated in order to examine the robustness of the improved vehicle system. The results demonstrate that 14.42% mass of exhaust emissions can be reduced for the involved pedestrians, by using the developed fuzzy adaptive cost map
    corecore