906 research outputs found

    Ant Clustering with Locally Weighting Ant Perception and Diversified Memory

    Get PDF
    Ant clustering algorithms are a robust and flexible tool for clustering data that have produced some promising results. This paper introduces two improvements that can be incorporated into any ant clustering algorithm: kernel function similarity weights and a similarity memory model replacement scheme. A kernel function weights objects within an ant’s neighborhood according to the object distance and provides an alternate interpretation of the similarity of objects in an ant’s neighborhood. Ants can hill-climb the kernel gradients as they look for a suitable place to drop a carried object. The similarity memory model equips ants with a small memory consisting of a sampling of the current clustering space. We test several kernel functions and memory replacement schemes on the Iris, Wisconsin Breast Cancer, and Lincoln Lab network intrusion datasets. Compared to a basic ant clustering algorithm, we show that kernel functions and the similarity memory model increase clustering speed and cluster quality, especially for datasets with an unbalanced class distribution, such as network intrusion

    A Guided Neighborhood Search Applied to the Split Delivery Vehicle Routing Problem

    Get PDF
    The classic vehicle routing problem considers the distribution of goods to geographically scattered customers from a central depot using a homogeneous fleet of vehicles with finite capacity. Each customer has a known demand and can be visited by exactly one vehicle. Each vehicle services the assigned customers in such a way that all customers are fully supplied and the total service does not exceed the vehicle capacity. In the split delivery vehicle routing problem, a customer can be visited by more than one vehicle, i.e., a customer demand can be split between various vehicles. Allowing split deliveries has been proven to potentially reduce the operational costs of the fleet. This study efficiently solves the split delivery vehicle routing problem using three new approaches. In the first approach, the problem is solved in two stages. During the first stage, an initial solution is found by means of a greedy approach that can produce high quality solutions comparable to those obtained with existing sophisticated approaches. The greedy approach is based on a novel concept called the route angle control measure that helps to produce spatially thin routes and avoids crossing routes. In the second stage, this constructive approach is extended to an iterative approach using adaptive memory concepts, and then a variable neighborhood descent process is added to improve the solution obtained. A new solution diversification scheme is presented in the second approach based on concentric rings centered at the depot that partitions the original problem. The resulting sub-problems are then solved using the greedy approach with route angle control measures. Different ring settings produce varied partitions and thus different solutions to the original problem are obtained and improved via a variable neighborhood descent. The third approach is a learning procedure based on a set or population of solutions. Those solutions are used to find attractive attributes and construct new solutions within a tabu search framework. As the search progresses, the existing population evolves, better solutions are included in it whereas bad solutions are removed from it. The initial set is constructed using the greedy approach with the route angle control measure whereas new solutions are created using an adaptation of the well known savings algorithm of Clarke and Wright (1964) and improved by means of an enhanced version of the variable neighborhood descent process. The proposed approaches are tested on benchmark instances and results are compared with existing implementations

    Hybrid nature-inspired computation methods for optimization

    Get PDF
    The focus of this work is on the exploration of the hybrid Nature-Inspired Computation (NIC) methods with application in optimization. In the dissertation, we first study various types of the NIC algorithms including the Clonal Selection Algorithm (CSA), Particle Swarm Optimization (PSO), Ant Colony Optimization (ACO), Simulated Annealing (SA), Harmony Search (HS), Differential Evolution (DE), and Mind Evolution Computing (MEC), and propose several new fusions of the NIC techniques, such as CSA-DE, HS-DE, and CSA-SA. Their working principles, structures, and algorithms are analyzed and discussed in details. We next investigate the performances of our hybrid NIC methods in handling nonlinear, multi-modal, and dynamical optimization problems, e.g., nonlinear function optimization, optimal LC passive power filter design, and optimization of neural networks and fuzzy classification systems. The hybridization of these NIC methods can overcome the shortcomings of standalone algorithms while still retaining all the advantages. It has been demonstrated using computer simulations that the proposed hybrid NIC approaches are capable of yielding superior optimization performances over the individual NIC methods as well as conventional methodologies with regard to the search efficiency, convergence speed, and quantity and quality of the optimal solutions achieved

    Real time tracking using nature-inspired algorithms

    Get PDF
    This thesis investigates the core difficulties in the tracking field of computer vision. The aim is to develop a suitable tuning free optimisation strategy so that a real time tracking could be achieved. The population and multi-solution based approaches have been applied first to analyse the convergence behaviours in the evolutionary test cases. The aim is to identify the core misconceptions in the manner the search characteristics of particles are defined in the literature. A general perception in the scientific community is that the particle based methods are not suitable for the real time applications. This thesis improves the convergence properties of particles by a novel scale free correlation approach. By altering the fundamental definition of a particle and by avoiding the nostalgic operations the tracking was expedited to a rate of 250 FPS. There is a reasonable amount of similarity between the tracking landscapes and the ones generated by three dimensional evolutionary test cases. Several experimental studies are conducted that compares the performances of the novel optimisation to the ones observed with the swarming methods. It is therefore concluded that the modified particle behaviour outclassed the traditional approaches by huge margins in almost every test scenario

    Incorporating Memory and Learning Mechanisms Into Meta-RaPS

    Get PDF
    Due to the rapid increase of dimensions and complexity of real life problems, it has become more difficult to find optimal solutions using only exact mathematical methods. The need to find near-optimal solutions in an acceptable amount of time is a challenge when developing more sophisticated approaches. A proper answer to this challenge can be through the implementation of metaheuristic approaches. However, a more powerful answer might be reached by incorporating intelligence into metaheuristics. Meta-RaPS (Metaheuristic for Randomized Priority Search) is a metaheuristic that creates high quality solutions for discrete optimization problems. It is proposed that incorporating memory and learning mechanisms into Meta-RaPS, which is currently classified as a memoryless metaheuristic, can help the algorithm produce higher quality results. The proposed Meta-RaPS versions were created by taking different perspectives of learning. The first approach taken is Estimation of Distribution Algorithms (EDA), a stochastic learning technique that creates a probability distribution for each decision variable to generate new solutions. The second Meta-RaPS version was developed by utilizing a machine learning algorithm, Q Learning, which has been successfully applied to optimization problems whose output is a sequence of actions. In the third Meta-RaPS version, Path Relinking (PR) was implemented as a post-optimization method in which the new algorithm learns the good attributes by memorizing best solutions, and follows them to reach better solutions. The fourth proposed version of Meta-RaPS presented another form of learning with its ability to adaptively tune parameters. The efficiency of these approaches motivated us to redesign Meta-RaPS by removing the improvement phase and adding a more sophisticated Path Relinking method. The new Meta-RaPS could solve even the largest problems in much less time while keeping up the quality of its solutions. To evaluate their performance, all introduced versions were tested using the 0-1 Multidimensional Knapsack Problem (MKP). After comparing the proposed algorithms, Meta-RaPS PR and Meta-RaPS Q Learning appeared to be the algorithms with the best and worst performance, respectively. On the other hand, they could all show superior performance than other approaches to the 0-1 MKP in the literature

    Self-aware and self-adaptive autoscaling for cloud based services

    Get PDF
    Modern Internet services are increasingly leveraging on cloud computing for flexible, elastic and on-demand provision. Typically, Quality of Service (QoS) of cloud-based services can be tuned using different underlying cloud configurations and resources, e.g., number of threads, CPU and memory etc., which are shared, leased and priced as utilities. This benefit is fundamentally grounded by autoscaling: an automatic and elastic process that adapts cloud configurations on-demand according to time-varying workloads. This thesis proposes a holistic cloud autoscaling framework to effectively and seamlessly address existing challenges related to different logical aspects of autoscaling, including architecting autoscaling system, modelling the QoS of cloudbased service, determining the granularity of control and deciding trade-off autoscaling decisions. The framework takes advantages of the principles of self-awareness and the related algorithms to adaptively handle the dynamics, uncertainties, QoS interference and trade-offs on objectives that are exhibited in the cloud. The major benefit is that, by leveraging the framework, cloud autoscaling can be effectively achieved without heavy human analysis and design time knowledge. Through conducting various experiments using RUBiS benchmark and realistic workload on real cloud setting, this thesis evaluates the effectiveness of the framework based on various quality indicators and compared with other state-of-the-art approaches

    A Comprehensive Survey on Particle Swarm Optimization Algorithm and Its Applications

    Get PDF
    Particle swarm optimization (PSO) is a heuristic global optimization method, proposed originally by Kennedy and Eberhart in 1995. It is now one of the most commonly used optimization techniques. This survey presented a comprehensive investigation of PSO. On one hand, we provided advances with PSO, including its modifications (including quantum-behaved PSO, bare-bones PSO, chaotic PSO, and fuzzy PSO), population topology (as fully connected, von Neumann, ring, star, random, etc.), hybridization (with genetic algorithm, simulated annealing, Tabu search, artificial immune system, ant colony algorithm, artificial bee colony, differential evolution, harmonic search, and biogeography-based optimization), extensions (to multiobjective, constrained, discrete, and binary optimization), theoretical analysis (parameter selection and tuning, and convergence analysis), and parallel implementation (in multicore, multiprocessor, GPU, and cloud computing forms). On the other hand, we offered a survey on applications of PSO to the following eight fields: electrical and electronic engineering, automation control systems, communication theory, operations research, mechanical engineering, fuel and energy, medicine, chemistry, and biology. It is hoped that this survey would be beneficial for the researchers studying PSO algorithms

    Applied Metaheuristic Computing

    Get PDF
    For decades, Applied Metaheuristic Computing (AMC) has been a prevailing optimization technique for tackling perplexing engineering and business problems, such as scheduling, routing, ordering, bin packing, assignment, facility layout planning, among others. This is partly because the classic exact methods are constrained with prior assumptions, and partly due to the heuristics being problem-dependent and lacking generalization. AMC, on the contrary, guides the course of low-level heuristics to search beyond the local optimality, which impairs the capability of traditional computation methods. This topic series has collected quality papers proposing cutting-edge methodology and innovative applications which drive the advances of AMC

    Modeling And Applying Biomimetic Metaheuristics To Product Life Cycle Engineering

    Get PDF
    Due to its potential for significant impact, interest continues to grow in the assessment of products from a life cycle perspective. As the nature of products shifts from mechanized and Newtonian to more adaptive and complex, the behavior of products more closely resembles biological organisms in community. The change in product nature is increasingly mirrored at the component level. The work presented in this dissertation is twofold. First, the research proposes a general, systematic and holistic classification of life cycle data to transform the design problem into an optimization problem. Second, the research proposes two new metaheuristics (bio-inspired and socio-inspired) to solve optimization problems to produce grouped solutions that are efficient, evolvable and sustainable. The bio-inspired approach is schooling genetic algorithms (SGA), while the socio-inspired approach is referred to as genetic social networks (GSN). SGA is an approach that combines fish schooling concepts with genetic algorithms (GAs) to enable a dynamic search process. The application of GA operators is subject to the perception of the immediate local environment by clusters of candidate solutions behaving as schools of fish. GSN is an approach that adds social network concepts to GAs, implementing single and dyadic social interactions of social groups (clusters of similar candidate solutions) with GA operators. SGA and GSN both use phenotypic representations of a hypothetical product or system as input. The representations are derived from the proposed life cycle engineering (LCE) data classification. The outputs of either method are the representations that are more than likely to perform better, longer, and more autonomously within their environment during their life cycle. Both methods can also be used as a decision making tool. Both approaches were tested on product design problems with differing parametric relations, underlying solution space, and problem size

    Self-Organized Specialization and Controlled Emergence in Organic Computing Systems

    Get PDF
    In this chapter we studied a first approach to generate suitable rule sets for solving classification problems on systems of autonomous, memory constrained components. It was shown that a multi agent system that uses interacting Pittsburgh-style classifier systems can evolve appropiate rule sets. The system evolves specialists for parts of the classification problem and cooperation between them. In this way the components overcome their restricted memory size and are able to solve the entire problem. It was shown that the communication topology between the components strongly influences the average number of components that a request has to pass until it is classified. It was also shown that the introduction of communication costs into the fitness function leads to a more even distribution of knowledge between the components and reduces the communication overhead without influencing the classification performance very much. If the system is used to generate rule sets to solve classification tasks on real hardware systems, communication cost in the training phase can thus lead to a better knowledge distribution and small communication cost. That is, in this way the system will be more robust against the loss of single components and longer reliable in case of limited energy resources
    • …
    corecore