25,247 research outputs found
A Review of the Family of Artificial Fish Swarm Algorithms: Recent Advances and Applications
The Artificial Fish Swarm Algorithm (AFSA) is inspired by the ecological
behaviors of fish schooling in nature, viz., the preying, swarming, following
and random behaviors. Owing to a number of salient properties, which include
flexibility, fast convergence, and insensitivity to the initial parameter
settings, the family of AFSA has emerged as an effective Swarm Intelligence
(SI) methodology that has been widely applied to solve real-world optimization
problems. Since its introduction in 2002, many improved and hybrid AFSA models
have been developed to tackle continuous, binary, and combinatorial
optimization problems. This paper aims to present a concise review of the
family of AFSA, encompassing the original ASFA and its improvements,
continuous, binary, discrete, and hybrid models, as well as the associated
applications. A comprehensive survey on the AFSA from its introduction to 2012
can be found in [1]. As such, we focus on a total of {\color{blue}123} articles
published in high-quality journals since 2013. We also discuss possible AFSA
enhancements and highlight future research directions for the family of
AFSA-based models.Comment: 37 pages, 3 figure
Discrete Particle Swarm Optimization for the minimum labelling Steiner tree problem
Particle Swarm Optimization is an evolutionary method inspired by the
social behaviour of individuals inside swarms in nature. Solutions of the problem are
modelled as members of the swarm which fly in the solution space. The evolution is
obtained from the continuous movement of the particles that constitute the swarm
submitted to the effect of the inertia and the attraction of the members who lead the
swarm. This work focuses on a recent Discrete Particle Swarm Optimization for combinatorial optimization, called Jumping Particle Swarm Optimization. Its effectiveness is
illustrated on the minimum labelling Steiner tree problem: given an undirected labelled
connected graph, the aim is to find a spanning tree covering a given subset of nodes,
whose edges have the smallest number of distinct labels
Bat Algorithm: Literature Review and Applications
Bat algorithm (BA) is a bio-inspired algorithm developed by Yang in 2010 and
BA has been found to be very efficient. As a result, the literature has
expanded significantly in the last 3 years. This paper provides a timely review
of the bat algorithm and its new variants. A wide range of diverse applications
and case studies are also reviewed and summarized briefly here. Further
research topics are also discussed.Comment: 10 page
k-Nearest Neighbour Classifiers: 2nd Edition (with Python examples)
Perhaps the most straightforward classifier in the arsenal or machine
learning techniques is the Nearest Neighbour Classifier -- classification is
achieved by identifying the nearest neighbours to a query example and using
those neighbours to determine the class of the query. This approach to
classification is of particular importance because issues of poor run-time
performance is not such a problem these days with the computational power that
is available. This paper presents an overview of techniques for Nearest
Neighbour classification focusing on; mechanisms for assessing similarity
(distance), computational issues in identifying nearest neighbours and
mechanisms for reducing the dimension of the data.
This paper is the second edition of a paper previously published as a
technical report. Sections on similarity measures for time-series, retrieval
speed-up and intrinsic dimensionality have been added. An Appendix is included
providing access to Python code for the key methods.Comment: 22 pages, 15 figures: An updated edition of an older tutorial on kN
Applications of Nature-Inspired Algorithms for Dimension Reduction: Enabling Efficient Data Analytics
In [1], we have explored the theoretical aspects of feature selection and evolutionary algorithms. In this chapter, we focus on optimization algorithms for enhancing data analytic process, i.e., we propose to explore applications of nature-inspired algorithms in data science. Feature selection optimization is a hybrid approach leveraging feature selection techniques and evolutionary algorithms process to optimize the selected features. Prior works solve this problem iteratively to converge to an optimal feature subset. Feature selection optimization is a non-specific domain approach. Data scientists mainly attempt to find an advanced way to analyze data n with high computational efficiency and low time complexity, leading to efficient data analytics. Thus, by increasing generated/measured/sensed data from various sources, analysis, manipulation and illustration of data grow exponentially. Due to the large scale data sets, Curse of dimensionality (CoD) is one of the NP-hard problems in data science. Hence, several efforts have been focused on leveraging evolutionary algorithms (EAs) to address the complex issues in large scale data analytics problems. Dimension reduction, together with EAs, lends itself to solve CoD and solve complex problems, in terms of time complexity, efficiently. In this chapter, we first provide a brief overview of previous studies that focused on solving CoD using feature extraction optimization process. We then discuss practical examples of research studies are successfully tackled some application domains, such as image processing, sentiment analysis, network traffics / anomalies analysis, credit score analysis and other benchmark functions/data sets analysis
- …