429 research outputs found

    TRUSS STRUCTURE OPTIMIZATION BASED ON IMPROVED WOLF PACK ALGORITHM

    Get PDF
    Aiming at the optimization of truss structure, a wolf pack algorithm based on chaos and improved search strategy was proposed. The mathematical model of truss optimization was constructed, and the classical truss structure was optimized. The results were compared with those of other optimization algorithms. When selecting and updating the initial position of wolves, chaos idea was used to distribute the initial value evenly in the solution space; phase factor was introduced to optimize the formula of wolf detection; information interaction between wolves is increased and the number of runs is reduced. The numerical results show that the improved wolf pack algorithm has the characteristics of fewer parameters, simple programming, easy implementation, fast convergence speed, and can quickly find the optimal solution. It is suitable for the optimization design of the section size of space truss structures

    A New Enhanced Hybrid Grey Wolf Optimizer (GWO) Combined with Elephant Herding Optimization (EHO) Algorithm for Engineering Optimization

    Get PDF
    Although the exploitation of GWO advances sharply, it has limitations for continuous implementing exploration. On the other hand, the EHO algorithm easily has shown its capability to prevent local optima. For hybridization and by considering the advantages of GWO and the abilities of EHO, it would be impressive to combine these two algorithms. In this respect, the exploitation and exploration performances and the convergence speed of the GWO algorithm are improved by combining it with the EHO algorithm. Therefore, this paper proposes a new hybrid Grey Wolf Optimizer (GWO) combined with Elephant Herding Optimization (EHO) algorithm. Twenty-three benchmark mathematical optimization challenges and six constrained engineering challenges are used to validate the performance of the suggested GWOEHO compared to both the original GWO and EHO algorithms and some other well-known optimization algorithms. Wilcoxon's rank-sum test outcomes revealed that GWOEHO outperforms others in most function minimization. The results also proved that the convergence speed of GWOEHO is faster than the original algorithms

    Chaotic Sand Cat Swarm Optimization

    Get PDF
    In this study, a new hybrid metaheuristic algorithm named Chaotic Sand Cat Swarm Optimization (CSCSO) is proposed for constrained and complex optimization problems. This algorithm combines the features of the recently introduced SCSO with the concept of chaos. The basic aim of the proposed algorithm is to integrate the chaos feature of non-recurring locations into SCSO’s core search process to improve global search performance and convergence behavior. Thus, randomness in SCSO can be replaced by a chaotic map due to similar randomness features with better statistical and dynamic properties. In addition to these advantages, low search consistency, local optimum trap, inefficiency search, and low population diversity issues are also provided. In the proposed CSCSO, several chaotic maps are implemented for more efficient behavior in the exploration and exploitation phases. Experiments are conducted on a wide variety of well-known test functions to increase the reliability of the results, as well as real-world problems. In this study, the proposed algorithm was applied to a total of 39 functions and multidisciplinary problems. It found 76.3% better responses compared to a best-developed SCSO variant and other chaotic-based metaheuristics tested. This extensive experiment indicates that the CSCSO algorithm excels in providing acceptable results

    An Improved Binary Grey-Wolf Optimizer with Simulated Annealing for Feature Selection

    Get PDF
    This paper proposes improvements to the binary grey-wolf optimizer (BGWO) to solve the feature selection (FS) problem associated with high data dimensionality, irrelevant, noisy, and redundant data that will then allow machine learning algorithms to attain better classification/clustering accuracy in less training time. We propose three variants of BGWO in addition to the standard variant, applying different transfer functions to tackle the FS problem. Because BGWO generates continuous values and FS needs discrete values, a number of V-shaped, S-shaped, and U-shaped transfer functions were investigated for incorporation with BGWO to convert their continuous values to binary. After investigation, we note that the performance of BGWO is affected by the selection of the transfer function. Then, in the first variant, we look to reduce the local minima problem by integrating an exploration capability to update the position of the grey wolf randomly within the search space with a certain probability; this variant was abbreviated as IBGWO. Consequently, a novel mutation strategy is proposed to select a number of the worst grey wolves in the population which are updated toward the best solution and randomly within the search space based on a certain probability to determine if the update is either toward the best or randomly. The number of the worst grey wolf selected by this strategy is linearly increased with the iteration. Finally, this strategy is combined with IBGWO to produce the second variant of BGWO that was abbreviated as LIBGWO. In the last variant, simulated annealing (SA) was integrated with LIBGWO to search around the best-so-far solution at the end of each iteration in order to identify better solutions. The performance of the proposed variants was validated on 32 datasets taken from the UCI repository and compared with six wrapper feature selection methods. The experiments show the superiority of the proposed improved variants in producing better classification accuracy than the other selected wrapper feature selection algorithms
    corecore