3,190 research outputs found

    A hybrid swarm-based algorithm for single-objective optimization problems involving high-cost analyses

    Full text link
    In many technical fields, single-objective optimization procedures in continuous domains involve expensive numerical simulations. In this context, an improvement of the Artificial Bee Colony (ABC) algorithm, called the Artificial super-Bee enhanced Colony (AsBeC), is presented. AsBeC is designed to provide fast convergence speed, high solution accuracy and robust performance over a wide range of problems. It implements enhancements of the ABC structure and hybridizations with interpolation strategies. The latter are inspired by the quadratic trust region approach for local investigation and by an efficient global optimizer for separable problems. Each modification and their combined effects are studied with appropriate metrics on a numerical benchmark, which is also used for comparing AsBeC with some effective ABC variants and other derivative-free algorithms. In addition, the presented algorithm is validated on two recent benchmarks adopted for competitions in international conferences. Results show remarkable competitiveness and robustness for AsBeC.Comment: 19 pages, 4 figures, Springer Swarm Intelligenc

    An Evolutionary Optimization Algorithm for Automated Classical Machine Learning

    Get PDF
    Machine learning is an evolving branch of computational algorithms that allow computers to learn from experiences, make predictions, and solve different problems without being explicitly programmed. However, building a useful machine learning model is a challenging process, requiring human expertise to perform various proper tasks and ensure that the machine learning\u27s primary objective --determining the best and most predictive model-- is achieved. These tasks include pre-processing, feature selection, and model selection. Many machine learning models developed by experts are designed manually and by trial and error. In other words, even experts need the time and resources to create good predictive machine learning models. The idea of automated machine learning (AutoML) is to automate a machine learning pipeline to release the burden of substantial development costs and manual processes. The algorithms leveraged in these systems have different hyper-parameters. On the other hand, different input datasets have various features. In both cases, the final performance of the model is closely related to the final selected configuration of features and hyper-parameters. That is why they are considered as crucial tasks in the AutoML. The challenges regarding the computationally expensive nature of tuning hyper-parameters and optimally selecting features create significant opportunities for filling the research gaps in the AutoML field. This dissertation explores how to select the features and tune the hyper-parameters of conventional machine learning algorithms efficiently and automatically. To address the challenges in the AutoML area, novel algorithms for hyper-parameter tuning and feature selection are proposed. The hyper-parameter tuning algorithm aims to provide the optimal set of hyper-parameters in three conventional machine learning models (Random Forest, XGBoost and Support Vector Machine) to obtain best scores regarding performance. On the other hand, the feature selection algorithm looks for the optimal subset of features to achieve the highest performance. Afterward, a hybrid framework is designed for both hyper-parameter tuning and feature selection. The proposed framework can discover close to the optimal configuration of features and hyper-parameters. The proposed framework includes the following components: (1) an automatic feature selection component based on artificial bee colony algorithms and machine learning training, and (2) an automatic hyper-parameter tuning component based on artificial bee colony algorithms and machine learning training for faster training and convergence of the learning models. The whole framework has been evaluated using four real-world datasets in different applications. This framework is an attempt to alleviate the challenges of hyper-parameter tuning and feature selection by using efficient algorithms. However, distributed processing, distributed learning, parallel computing, and other big data solutions are not taken into consideration in this framework

    Obstacle Avoidance Scheme Based Elite Opposition Bat Algorithm for Unmanned Ground Vehicles

    Get PDF
    Unmanned Ground Vehicles (UGVs) are intelligent vehicles that operate in an obstacle environment without an onboard human operator but can be controlled autonomously using an obstacle avoidance system or by a human operator from a remote location. In this research, an obstacle avoidance scheme-based elite opposition bat algorithm (EOBA) for UGVs was developed. The obstacle avoidance system comprises a simulation map, a perception system for obstacle detection, and the implementation of EOBA for generating an optimal collision-free path that led the UGV to the goal location. Three distance thresholds of 0.1 m, 0.2 m, and 0.3 m was used in the obstacle detection stage to determine the optimal distance threshold for obstacle avoidance. The performance of the obstacle avoidance scheme was compared with that of bat algorithm (BA) and particle swarm optimization (PSO) techniques. The simulation results show that the distance threshold of 0.3 m is the optimal threshold for obstacle avoidance provided that the size of the obstacle does not exceed the size of the UGV. The EOBA based scheme when compared with BA and PSO schemes obtained an average percentage reduction of 21.82% in terms of path length and 60% in terms of time taken to reach the target destination. The uniqueness of this approach is that the UGV avoid collision with an obstacle at a distance of 0.3 m from nearby obstacles as against taking three steps backward before avoiding obstacl

    Context-Aware Clustering and the Optimized Whale Optimization Algorithm: An Effective Predictive Model for the Smart Grid

    Get PDF
    For customers to participate in key peak pricing, period-of-use fees, and individualized responsiveness to demand programmes taken from multi-dimensional data flows, energy use projection and analysis must be done well. However, it is a difficult study topic to ascertain the knowledge of use of electricity as recorded in the electricity records' Multi-Dimensional Data Streams (MDDS). Context-Aware Clustering (CAC) and the Optimized Whale Optimization Algorithm were suggested by researchers as a fresh power usage knowledge finding model from the multi-dimensional data streams (MDDS) to resolve issue (OWOA). The proposed CAC-OWOA framework first performs the data cleaning to handle the noisy and null elements. The predictive features are extracted from the novel context-aware group formation algorithm using the statistical context parameters from the pre-processed MDDS electricity logs. To perform the energy consumption prediction, researchers have proposed the novel Artificial Neural Network (ANN) predictive algorithm using the bio-inspired optimization algorithm called OWOA. The OWOA is the modified algorithm of the existing WOA to overcome the problems of slow convergence speed and easily falling into the local optimal solutions. The ANN training method is used in conjunction with the suggested bio-inspired OWOA algorithm to lower error rates and boost overall prediction accuracy. The efficiency of the CAC-OWOA framework is evaluated using the publicly available smart grid electricity consumption logs. The experimental results demonstrate the effectiveness of the CAC-OWOA framework in terms of forecasting accuracy, precision, recall, and duration when compared to underlying approaches

    A self-adaptive artificial bee colony algorithm with local search for TSK-type neuro-fuzzy system training

    Full text link
    © 2019 IEEE. In this paper, we introduce a self-adaptive artificial bee colony (ABC) algorithm for learning the parameters of a Takagi-Sugeno-Kang-type (TSK-type) neuro-fuzzy system (NFS). The proposed NFS learns fuzzy rules for the premise part of the fuzzy system using an adaptive clustering method according to the input-output data at hand for establishing the network structure. All the free parameters in the NFS, including the premise and the following TSK-type consequent parameters, are optimized by the modified ABC (MABC) algorithm. Experiments involve two parts, including numerical optimization problems and dynamic system identification problems. In the first part of investigations, the proposed MABC compares to the standard ABC on mathematical optimization problems. In the remaining experiments, the performance of the proposed method is verified with other metaheuristic methods, including differential evolution (DE), genetic algorithm (GA), particle swarm optimization (PSO) and standard ABC, to evaluate the effectiveness and feasibility of the system. The simulation results show that the proposed method provides better approximation results than those obtained by competitors methods

    Predicting Arrhythmia Based on Machine Learning Using Improved Harris Hawk Algorithm

    Get PDF
    Arrhythmia disease is widely recognized as a prominent and lethal ailment on a global scale, resulting in a significant number of fatalities annually. The timely identification of this ailment is crucial for preserving individuals' lives. Machine Learning (ML), a branch of artificial intelligence (AI), has emerged as a highly efficient and cost-effective method for illness detection. The objective of this work is to develop a machine learning (ML) model capable of accurately predicting heart illness by using the Arrhythmia disease dataset, with the purpose of achieving optimal performance. The performance of the model is greatly influenced by the selection of the machine learning method and the features in the dataset for training purposes. In order to mitigate the issue of overfitting caused by the high dimensionality of the features in the Arrhythmia dataset, a reduction of the dataset to a lower dimensional subspace was performed via the improved Harris hawk optimization algorithm (iHHO). The Harris hawk algorithm exhibits a rapid convergence rate and possesses a notable degree of adaptability in its ability to identify optimal characteristics. The performance of the models created with the feature-selected dataset using various machine learning techniques was evaluated and compared. In this work, total seven classifiers like SVM, GB, GNB, RF, LR, DT, and KNN are used to classify the data produced by the iHHO algorithm. The results clearly show the improvement of 3%, 4%, 4%, 9%, 8%, 3%, and 9% with the classifiers KNN, RF, GB, SVM, LR, DT, and GNB respectively
    • …
    corecore