46 research outputs found

    Adaptive binary artificial bee colony algorithm

    Get PDF
    Metaheuristics and swarm intelligence algorithms are bio-inspired algorithms, which have long standing track record of success in problem solving. Due to the nature and the complexity of the problems, problem solving approaches may not achieve the same success level in every type of problems. Artificial bee colony (ABC) algorithm is a swarm intelligence algorithm and has originally been developed to solve numerical optimisation problems. It has a sound track record in numerical problems, but has not yet been tested sufficiently for combinatorial and binary problems. This paper proposes an adaptive hybrid approach to devise ABC algorithms with multiple and complementary binary operators for higher efficiency in solving binary problems.} Three prominent operator selection schemes have been comparatively investigated for the best configuration in this regard. The proposed approach has been applied to uncapacitated facility location problems, a renown NP-Hard combinatorial problem type modelled with 0-1 programming, and successfully solved the well-known benchmarks outperforming state-of-art algorithms

    A heuristic approach for multi-product capacitated single-allocation hub location problems

    Get PDF
    Tese de mestrado, Estatística e Investigação Operacional, Universidade de Lisboa, Faculdade de Ciências, 2015Em redes onde o fluxo entre nodos é muito elevado (como pode ser o caso do transporte de pessoas e mercadorias ou até mesmo fluxo de dados numa rede), torna-se menos dispendioso criar pontos onde se concentram os fluxos provenientes das diferentes origens para depois serem consolidados e redistribuídos até aos destinos. A esses pontos dá-se o nome de hubs. O problema de localização de hubs consiste na localização de hubs numa rede e na alocação de todos os nodos da rede a esses hubs, de modo a que se possa encaminhar os fluxos entre os pares origem-destino a menos que sejam hubs. A rede constituída pelos hubs é normalmente definida como completa e não se permitem ligações diretas entre os pares origem-destino. Para além disso, assume-se que existe um factor de desconto para o fluxo que circula entre hubs. Neste tipo de redes (hub-and-spoke networks) podem aparecer duas variantes, no que diz respeito à alocação dos nodos aos hubs: single-allocation e multiple-allocation. No primeiro caso, permite-se apenas uma ligação de cada nodo não hub a um hub de modo a que todo o fluxo com origem e destino a cada nodo saia e chegue a esse nodo através de apenas um hub. No caso em que se tem multiple-allocation, cada nodo poderá ser afecto a mais do que um hub e o fluxo que chega e sai desse nodo poderá usar mais do que um hub. Algumas variantes que se poderão considerar para este problema incluem restrições de capacidade nos hubs (restrições que limitam a capacidade de um hub processar uma certa quantidade de fluxo de origem, limitações na capacidade total, limitações no processamento de fluxo que sai do hub, etc.), restrições de capacidade nos arcos, problemas multi-periódicos, presença de incerteza, o número de hubs ser fixo, o tipo de objectivo (minimizar custos, minimizar distâncias entre hubs, etc.) entre outras. A necessidade de aproximar este tipo de problemas aos casos que se observam no mundo real leva à inclusão de cada vez mais restrições dando origem a mais variantes do problema. Neste trabalho, será abordado o problema de localização de hubs na variante single-Allocation, com restrições de capacidade em relação ao fluxo que cada hub é capaz de processar. Para além disso, considera-se fluxos relativos a mais do que um tipo de produto. Este problema é designado por Problema Multi-produto de Localização de Hubs com Capacidade1. Cada hub poderá ser dedicado a processar apenas um tipo de produto, poderá processar mais do que um, ou mesmo todos. A rede de hubs é completa para cada produto mas, no entanto, se se considerar a rede de hubs para todos os produtos, esta poderá não ser completa. Como constatado em Correia et al. [17], no caso em que cada hub processa todos os tipos de produto, resolver o problema multi-produto ao invés de se resolver vários problemas, um para cada produto em separado, dá origem a melhores resultados. A complexidade inerente a este tipo de problemas leva a que sejam classificados como problemas NP-Hard pois não existem algoritmos que sejam capazes de os resolver em tempo polinomial. Por esta razão faz sentido desenvolver algoritmos heurísticos de modo a se conseguir obter, em tempo útil, soluções para instâncias maiores do problema . Como referido em Meyer et al. [51], em problemas de localização de hubs, duas soluções com valores objectivo muito semelhantes poderão ser estruturalmente muito diferentes, e portanto, através um mecanismo de pesquisa local poderá ser muito difícil a passagem de uma boa solução para outra melhor. Por esta razão, neste trabalho opta-se por uma heurística que se baseia num método em que se constroem soluções repetidamente. Para a construção das soluções, considerando que um processo de construção do tipo Greedy poderia dar origem a um número limitado de soluções e que as componentes da solução que são escolhidas por último são as piores, optou-se pelo desenvolvimento de um algoritmo de Ant Colony Optimization (ACO). Esta meta-heurística baseia-se no comportamento apresentado pelas formigas quando estas procuram alimento. Quando uma formiga deixa a colónia em busca de alimento, no seu trajeto, deposita um químico (feromona) que pode ser detectado por outras formigas. Quanto maior a concentração de feromona, maior a atração de cada formiga por esse trajeto e, portanto, os trajetos com maiores concentrações de feromonas serão percorridos por mais formigas. Por outro lado, se o caminho de ida e volta até ao alimento for mais curto, mais vezes será percorrido e maior será a concentração de feromona nesse caminho. O resultado destes dois tipos de reforço positivo nas concentrações de feromona nos trajetos percorridos pelas formigas, aliados ao facto de que existe evaporação do químico (a concentração de feromona diminui nos caminhos menos percorridos ao longo do tempo) dá origem aos \carreirinhos" de formigas que se podem observar na natureza e que normalmente representam o caminho mais curto entre o alimento e a Colónia de formigas. Considere-se o problema em questão em que se tem n nodos e p produtos. Para a representação das soluções, em vez de se considerar uma matriz binária n χ n χ p, onde o valor 1 representa uma afetação, considerou-se uma matriz n χ p, em que cada entrada representa, para cada produto, o hub ao qual o nodo foi afecto. O caso em que um nodo é afecto a si mesmo indica que esse nodo é hub para o produto correspondente. Este tipo de representação permite reduzir o tamanho da matriz e diminuir o uso da memória computacional. Antes da construção de uma solução, é aplicado um pré-processamento que vai evitar, com base nas restrições do problema, que certas componentes da solução sejam consideradas durante o processo de construção da solução. Deste modo, reduz-se o espaço de procura de soluções e algum esforço computacional. Para a construção de uma solução, escolhe-se o tamanho da colonia (o número de formigas que pertencem à colónia) e cada formiga vai escolhendo, sucessivamente, componentes da solução através de uma regra pseudo-aleatória onde algumas componentes da solução são escolhidas de um modo greedy e outras são escolhidas através de roulette wheel selection. A cada componente da solução é atribuído um valor inicial de feromona e, à medida que cada formiga vai adicionando componentes à solução, o valor da feromona associado à componente adicionada vai decrescendo, o que resulta na diminuição da probabilidade de que essa componente seja escolhida pela próxima formiga, dando origem à diversificação do conjunto de soluções construído por cada colónia. No fim, depois de todas as formigas terem construído uma solução, escolhe-se a melhor solução e reforça-se a concentração de feromona na melhor solução construída pela colónia. Se, por acaso, uma formiga der origem a uma solução não admissível, a solução construída por essa formiga não é considerada. Para mais detalhe em relação a este processo consultar Dorigo et al. [20]. Este tipo de algoritmo permite a inclusão de métodos de pesquisa local de modo a que a solução obtida por cada colónia seja melhorada. Com o objectivo de obter um algoritmo mais eficiente, escolheu-se incluir esta possibilidade e procedeu-se ao reforço da concentração de feromona após feita uma pesquisa local. Na pesquisa local efectuada, usaram-se três tipos de vizinhança. Um deles fecha os hubs dedicados que só servem a si próprios e realoca-os a outros já abertos para esse mesmo produto. Outro, escolhe aleatoriamente um nodo alocado a um hub dedicado para um dado produto e realoca-o a outro hub dedicado ao mesmo produto. Um terceiro, escolhe um hub aleatoriamente e transforma-o num nodo, realocando-o a outro hub dedicado ao mesmo tipo de produto. De modo a obter soluções iniciais melhores, explora-se a possibilidade de atribuir valores iniciais de feromona mais altos às componentes de solução pertencentes à solução da relaxação linear, na proporção do valor correspondente no caso das variáveis 0-1. Uma outra variação explorada consiste em fazer o reforço do valor de feromona às componentes da solução, apenas quando esta é a melhor de todas encontrada até ao momento, permitindo que haja evaporação de certas componentes de solução que poderão estar a ser escolhidas consecutivamente e permitindo que se escape mais facilmente de óptimos locais. Após implementação do algoritmo procede-se à fase dos testes computacionais em instâncias do problema com 10, 20, 25 e 40 nodos, 1, 2 e 3 produtos e hubs que processam 1, 2 e 3 produtos. As instâncias usadas nos testes computacionais pertencem ao Australian Post data set e foram adaptados por Correia et al. [17] de modo a que se tivesse dados para mais do que um tipo de produto.In this thesis, an heuristic procedure is proposed for the the multi-product capacitated single-allocation hub location problem. When addressing a problem in which it is necessary to determine the transportation of large commodity flows between many origin-destination (O-D) pairs, instead of using direct links, it becomes more efficient to design the networks in such a way that some of the nodes become consolidation centers or hubs. The Multi-Product Capacitated Single-Allocation Hub Location Problem (MP-CSAHLP according to Correia et al. [17]), is a NP-Hard problem in which several types of ow are considered, making it possible to consider the case when multiple types of products are to be shipped between each O-D pair. It can be seen as an extension of the classical Capacitated Single-Allocation Hub Location Problem. In the problem investigated in this work, no more than one hub can be located in each node and the hubs can be either dedicated (each hub can only handle one type of product) or non-dedicated (one hub can handle more than one type product). The hubs have capacity limitations regarding the incoming flow. Furthermore, the hub network is complete for each product but, when considering the hub network as a whole, it does not necessarily have to be complete. The goal is to locate the hubs in the network, allocate the non-hub nodes to the opened hubs and route the flow between each O-D pair. The objective is to minimize the total ow routing cost plus the setup costs of the hubs and costs of preparing the hubs to handle the different types of products. In order to obtain feasible solutions to the above problem, an Ant Colony Optimization procedure is proposed, which is a constructive, population-based meta-heuristic based in the foraging behavior of ants. Indirect communication between the ants through pheromones reflects the colony search experience. High-quality solutions are found as an outcome of the global cooperation among all the ants of the colony. A preprocessing procedure is also proposed in which some solution components are forbidden based on the problems restrictions. Such preprocessing reduces the search space and thus may reduce the computational effort. The proposed heuristic uses a single ant colony, which simultaneously chooses the hubs and allocates the nodes to the hubs. Once these solutions are found, the routing of the flow is computed in a short amount of time, using the optimization models for the MP-CSAHLP in which some variables (location and allocation) are fixed. The results show that the proposed heuristic has the potential to find good quality solutions for the MP-CSAHLP and that its performance can be improved with finer parameter tuning, longer runs and more intense local search

    Generation and optimisation of real-world static and dynamic location-allocation problems with application to the telecommunications industry.

    Get PDF
    The location-allocation (LA) problem concerns the location of facilities and the allocation of demand, to minimise or maximise a particular function such as cost, profit or a measure of distance. Many formulations of LA problems have been presented in the literature to capture and study the unique aspects of real-world problems. However, some real-world aspects, such as resilience, are still lacking in the literature. Resilience ensures uninterrupted supply of demand and enhances the quality of service. Due to changes in population shift, market size, and the economic and labour markets - which often cause demand to be stochastic - a reasonable LA problem formulation should consider some aspect of future uncertainties. Almost all LA problem formulations in the literature that capture some aspect of future uncertainties fall in the domain of dynamic optimisation problems, where new facilities are located every time the environment changes. However, considering the substantial cost associated with locating a new facility, it becomes infeasible to locate facilities each time the environment changes. In this study, we propose and investigate variations of LA problem formulations. Firstly, we develop and study new LA formulations, which extend the location of facilities and the allocation of demand to add a layer of resilience. We apply the population-based incremental learning algorithm for the first time in the literature to solve the new novel LA formulations. Secondly, we propose and study a new dynamic formulation of the LA problem where facilities are opened once at the start of a defined period and are expected to be satisfactory in servicing customers' demands irrespective of changes in customer distribution. The problem is based on the idea that customers will change locations over a defined period and that these changes have to be taken into account when establishing facilities to service changing customers' distributions. Thirdly, we employ a simulation-based optimisation approach to tackle the new dynamic formulation. Owing to the high computational costs associated with simulation-based optimisation, we investigate the concept of Racing, an approach used in model selection, to reduce the high computational cost by employing the minimum number of simulations for solution selection

    Problemas de localização-distribuição de serviços semiobnóxios: aproximações e apoio à decisão

    Get PDF
    Doutoramento em Gestão IndustrialA presente tese resulta de um trabalho de investigação cujo objectivo se centrou no problema de localização-distribuição (PLD) que pretende abordar, de forma integrada, duas actividades logísticas intimamente relacionadas: a localização de equipamentos e a distribuição de produtos. O PLD, nomeadamente a sua modelação matemática, tem sido estudado na literatura, dando origem a diversas aproximações que resultam de diferentes cenários reais. Importa portanto agrupar as diferentes variantes por forma a facilitar e potenciar a sua investigação. Após fazer uma revisão e propor uma taxonomia dos modelos de localização-distribuição, este trabalho foca-se na resolução de alguns modelos considerados como mais representativos. É feita assim a análise de dois dos PLDs mais básicos (os problema capacitados com procura nos nós e nos arcos), sendo apresentadas, para ambos, propostas de resolução. Posteriormente, é abordada a localização-distribuição de serviços semiobnóxios. Este tipo de serviços, ainda que seja necessário e indispensável para o público em geral, dada a sua natureza, exerce um efeito desagradável sobre as comunidades contíguas. Assim, aos critérios tipicamente utilizados na tomada de decisão sobre a localização destes serviços (habitualmente a minimização de custo) é necessário adicionar preocupações que reflectem a manutenção da qualidade de vida das regiões que sofrem o impacto do resultado da referida decisão. A abordagem da localização-distribuição de serviços semiobnóxios requer portanto uma análise multi-objectivo. Esta análise pode ser feita com recurso a dois métodos distintos: não interactivos e interactivos. Ambos são abordados nesta tese, com novas propostas, sendo o método interactivo proposto aplicável a outros problemas de programação inteira mista multi-objectivo. Por último, é desenvolvida uma ferramenta de apoio à decisão para os problemas abordados nesta tese, sendo apresentada a metodologia adoptada e as suas principais funcionalidades. A ferramenta desenvolvida tem grandes preocupações com a interface de utilizador, visto ser direccionada para decisores que tipicamente não têm conhecimentos sobre os modelos matemáticos subjacentes a este tipo de problemas.This thesis main objective is to address the location-routing problem (LRP) which intends to tackle, using an integrated approach, two highly related logistics activities: the location of facilities and the distribution of materials. The LRP, namely its mathematical formulation, has been studied in the literature, and several approaches have emerged, corresponding to different real-world scenarios. Therefore, it is important to identify and group the different LRP variants, in order to segment current research and foster future studies. After presenting a review and a taxonomy of location-routing models, the following research focuses on solving some of its variants. Thus, a study of two of the most basic LRPs (capacitated problems with demand either on the nodes or on the arcs) is performed, and new approaches are presented. Afterwards, the location-routing of semi-obnoxious facilities is addressed. These are facilities that, although providing useful and indispensible services, given their nature, bring about an undesirable effect to adjacent communities. Consequently, to the usual objectives when considering their location (cost minimization), new ones must be added that are able to reflect concerns regarding the quality of life of the communities impacted by the outcome of these decisions. The location-routing of semi-obnoxious facilities therefore requires to be analysed using multi-objective approaches, which can be of two types: noninteractive or interactive. Both are discussed and new methods proposed in this thesis; the proposed interactive method is suitable to other multi-objective mixed integer programming problems. Finally, a newly developed decision-support tool to address the LRP is presented (being the adopted methodology discussed, and its main functionalities shown). This tool has great concerns regarding the user interface, as it is directed at decision makers who typically don’t have specific knowledge of the underlying models of this type of problems

    Graph-based Algorithms for Smart Mobility Planning and Large-scale Network Discovery

    Get PDF
    Graph theory has become a hot topic in the past two decades as evidenced by the increasing number of citations in research. Its applications are found in many fields, e.g. database, clustering, routing, etc. In this thesis, two novel graph-based algorithms are presented. The first algorithm finds itself in the thriving carsharing service, while the second algorithm is about large graph discovery to unearth the unknown graph before any analyses can be performed. In the first scenario, the automatisation of the fleet planning process in carsharing is proposed. The proposed work enhances the accuracy of the planning to the next level by taking an advantage of the open data movement such as street networks, building footprints, and demographic data. By using the street network (based on graph), it solves the questionable aspect in many previous works, feasibility as they tended to use rasterisation to simplify the map, but that comes with the price of accuracy and feasibility. A benchmark suite for further research in this problem is also provided. Along with it, two optimisation models with different sets of objectives and contexts are proposed. Through a series of experiment, a novel hybrid metaheuristic algorithm is proposed. The algorithm is called NGAP, which is based on Reference Point based Non-dominated Sorting genetic Algorithm (NSGA-III) and Pareto Local Search (PLS) and a novel problem specific local search operator designed for the fleet placement problem in carsharing called Extensible Neighbourhood Search (ENS). The designed local search operator exploits the graph structure of the street network and utilises the local knowledge to improve the exploration capability. The results show that the proposed hybrid algorithm outperforms the original NSGA-III in convergence under the same execution time. The work in smart mobility is done on city scale graphs which are considered to be medium size. However, the scale of the graphs in other fields in the real-world can be much larger than that which is why the large graph discovery algorithm is proposed as the second algorithm. To elaborate on the definition of large, some examples are required. The internet graph has over 30 billion nodes. Another one is a human brain network contains around 1011 nodes. Apart of the size, there is another aspect in real-world graph and that is the unknown. With the dynamic nature of the real-world graphs, it is almost impossible to have a complete knowledge of the graph to perform an analysis that is why graph traversal is crucial as the preparation process. I propose a novel memoryless chaos-based graph traversal algorithm called Chaotic Traversal (CHAT). CHAT is the first graph traversal algorithm that utilises the chaotic attractor directly. An experiment with two well-known chaotic attractors, Lozi map and Rössler system is conducted. The proposed algorithm is compared against the memoryless state-of-the-art algorithm, Random Walk. The results demonstrate the superior performance in coverage rate over Random Walk on five tested topologies; ring, small world, random, grid and power-law. In summary, the contribution of this research is twofold. Firstly, it contributes to the research society by introducing new study problems and novel approaches to propel the advance of the current state-of-the-art. And Secondly, it demonstrates a strong case for the conversion of research to the industrial sector to solve a real-world problem

    Holistic, data-driven, service and supply chain optimisation: linked optimisation.

    Get PDF
    The intensity of competition and technological advancements in the business environment has made companies collaborate and cooperate together as a means of survival. This creates a chain of companies and business components with unified business objectives. However, managing the decision-making process (like scheduling, ordering, delivering and allocating) at the various business components and maintaining a holistic objective is a huge business challenge, as these operations are complex and dynamic. This is because the overall chain of business processes is widely distributed across all the supply chain participants; therefore, no individual collaborator has a complete overview of the processes. Increasingly, such decisions are automated and are strongly supported by optimisation algorithms - manufacturing optimisation, B2B ordering, financial trading, transportation scheduling and allocation. However, most of these algorithms do not incorporate the complexity associated with interacting decision-making systems like supply chains. It is well-known that decisions made at one point in supply chains can have significant consequences that ripple through linked production and transportation systems. Recently, global shocks to supply chains (COVID-19, climate change, blockage of the Suez Canal) have demonstrated the importance of these interdependencies, and the need to create supply chains that are more resilient and have significantly reduced impact on the environment. Such interacting decision-making systems need to be considered through an optimisation process. However, the interactions between such decision-making systems are not modelled. We therefore believe that modelling such interactions is an opportunity to provide computational extensions to current optimisation paradigms. This research study aims to develop a general framework for formulating and solving holistic, data-driven optimisation problems in service and supply chains. This research achieved this aim and contributes to scholarship by firstly considering the complexities of supply chain problems from a linked problem perspective. This leads to developing a formalism for characterising linked optimisation problems as a model for supply chains. Secondly, the research adopts a method for creating a linked optimisation problem benchmark by linking existing classical benchmark sets. This involves using a mix of classical optimisation problems, typically relating to supply chain decision problems, to describe different modes of linkages in linked optimisation problems. Thirdly, several techniques for linking supply chain fragmented data have been proposed in the literature to identify data relationships. Therefore, this thesis explores some of these techniques and combines them in specific ways to improve the data discovery process. Lastly, many state-of-the-art algorithms have been explored in the literature and these algorithms have been used to tackle problems relating to supply chain problems. This research therefore investigates the resilient state-of-the-art optimisation algorithms presented in the literature, and then designs suitable algorithmic approaches inspired by the existing algorithms and the nature of problem linkages to address different problem linkages in supply chains. Considering research findings and future perspectives, the study demonstrates the suitability of algorithms to different linked structures involving two sub-problems, which suggests further investigations on issues like the suitability of algorithms on more complex structures, benchmark methodologies, holistic goals and evaluation, processmining, game theory and dependency analysis

    La métaheuristique CAT pour le design de réseaux logistiques déterministes et stochastiques

    Get PDF
    De nos jours, les entreprises d’ici et d’ailleurs sont confrontées à une concurrence mondiale sans cesse plus féroce. Afin de survivre et de développer des avantages concurrentiels, elles doivent s’approvisionner et vendre leurs produits sur les marchés mondiaux. Elles doivent aussi offrir simultanément à leurs clients des produits d’excellente qualité à prix concurrentiels et assortis d’un service impeccable. Ainsi, les activités d’approvisionnement, de production et de marketing ne peuvent plus être planifiées et gérées indépendamment. Dans ce contexte, les grandes entreprises manufacturières se doivent de réorganiser et reconfigurer sans cesse leur réseau logistique pour faire face aux pressions financières et environnementales ainsi qu’aux exigences de leurs clients. Tout doit être révisé et planifié de façon intégrée : sélection des fournisseurs, choix d’investissements, planification du transport et préparation d’une proposition de valeur incluant souvent produits et services au fournisseur. Au niveau stratégique, ce problème est fréquemment désigné par le vocable « design de réseau logistique ». Une approche intéressante pour résoudre ces problématiques décisionnelles complexes consiste à formuler et résoudre un modèle mathématique en nombres entiers représentant la problématique. Plusieurs modèles ont ainsi été récemment proposés pour traiter différentes catégories de décision en matière de design de réseau logistique. Cependant, ces modèles sont très complexes et difficiles à résoudre, et même les solveurs les plus performants échouent parfois à fournir une solution de qualité. Les travaux développés dans cette thèse proposent plusieurs contributions. Tout d’abord, un modèle de design de réseau logistique incorporant plusieurs innovations proposées récemment dans la littérature a été développé; celui-ci intègre les dimensions du choix des fournisseurs, la localisation, la configuration et l’assignation de mission aux installations (usines, entrepôts, etc.) de l’entreprise, la planification stratégique du transport et la sélection de politiques de marketing et d’offre de valeur au consommateur. Des innovations sont proposées au niveau de la modélisation des inventaires ainsi que de la sélection des options de transport. En deuxième lieu, une méthode de résolution distribuée inspirée du paradigme des systèmes multi-agents a été développée afin de résoudre des problèmes d’optimisation de grande taille incorporant plusieurs catégories de décisions. Cette approche, appelée CAT (pour collaborative agent teams), consiste à diviser le problème en un ensemble de sous-problèmes, et assigner chacun de ces sous-problèmes à un agent qui devra le résoudre. Par la suite, les solutions à chacun de ces sous-problèmes sont combinées par d’autres agents afin d’obtenir une solution de qualité au problème initial. Des mécanismes efficaces sont conçus pour la division du problème, pour la résolution des sous-problèmes et pour l’intégration des solutions. L’approche CAT ainsi développée est utilisée pour résoudre le problème de design de réseaux logistiques en univers certain (déterministe). Finalement, des adaptations sont proposées à CAT permettant de résoudre des problèmes de design de réseaux logistiques en univers incertain (stochastique)

    Evolving machine learning and deep learning models using evolutionary algorithms

    Get PDF
    Despite the great success in data mining, machine learning and deep learning models are yet subject to material obstacles when tackling real-life challenges, such as feature selection, initialization sensitivity, as well as hyperparameter optimization. The prevalence of these obstacles has severely constrained conventional machine learning and deep learning methods from fulfilling their potentials. In this research, three evolving machine learning and one evolving deep learning models are proposed to eliminate above bottlenecks, i.e. improving model initialization, enhancing feature representation, as well as optimizing model configuration, respectively, through hybridization between the advanced evolutionary algorithms and the conventional ML and DL methods. Specifically, two Firefly Algorithm based evolutionary clustering models are proposed to optimize cluster centroids in K-means and overcome initialization sensitivity as well as local stagnation. Secondly, a Particle Swarm Optimization based evolving feature selection model is developed for automatic identification of the most effective feature subset and reduction of feature dimensionality for tackling classification problems. Lastly, a Grey Wolf Optimizer based evolving Convolutional Neural Network-Long Short-Term Memory method is devised for automatic generation of the optimal topological and learning configurations for Convolutional Neural Network-Long Short-Term Memory networks to undertake multivariate time series prediction problems. Moreover, a variety of tailored search strategies are proposed to eliminate the intrinsic limitations embedded in the search mechanisms of the three employed evolutionary algorithms, i.e. the dictation of the global best signal in Particle Swarm Optimization, the constraint of the diagonal movement in Firefly Algorithm, as well as the acute contraction of search territory in Grey Wolf Optimizer, respectively. The remedy strategies include the diversification of guiding signals, the adaptive nonlinear search parameters, the hybrid position updating mechanisms, as well as the enhancement of population leaders. As such, the enhanced Particle Swarm Optimization, Firefly Algorithm, and Grey Wolf Optimizer variants are more likely to attain global optimality on complex search landscapes embedded in data mining problems, owing to the elevated search diversity as well as the achievement of advanced trade-offs between exploration and exploitation
    corecore