569 research outputs found

    Reinforcement Learning-assisted Evolutionary Algorithm: A Survey and Research Opportunities

    Full text link
    Evolutionary algorithms (EA), a class of stochastic search methods based on the principles of natural evolution, have received widespread acclaim for their exceptional performance in various real-world optimization problems. While researchers worldwide have proposed a wide variety of EAs, certain limitations remain, such as slow convergence speed and poor generalization capabilities. Consequently, numerous scholars actively explore improvements to algorithmic structures, operators, search patterns, etc., to enhance their optimization performance. Reinforcement learning (RL) integrated as a component in the EA framework has demonstrated superior performance in recent years. This paper presents a comprehensive survey on integrating reinforcement learning into the evolutionary algorithm, referred to as reinforcement learning-assisted evolutionary algorithm (RL-EA). We begin with the conceptual outlines of reinforcement learning and the evolutionary algorithm. We then provide a taxonomy of RL-EA. Subsequently, we discuss the RL-EA integration method, the RL-assisted strategy adopted by RL-EA, and its applications according to the existing literature. The RL-assisted procedure is divided according to the implemented functions including solution generation, learnable objective function, algorithm/operator/sub-population selection, parameter adaptation, and other strategies. Finally, we analyze potential directions for future research. This survey serves as a rich resource for researchers interested in RL-EA as it overviews the current state-of-the-art and highlights the associated challenges. By leveraging this survey, readers can swiftly gain insights into RL-EA to develop efficient algorithms, thereby fostering further advancements in this emerging field.Comment: 26 pages, 16 figure

    A Review on Computational Intelligence Techniques in Cloud and Edge Computing

    Get PDF
    Cloud computing (CC) is a centralized computing paradigm that accumulates resources centrally and provides these resources to users through Internet. Although CC holds a large number of resources, it may not be acceptable by real-time mobile applications, as it is usually far away from users geographically. On the other hand, edge computing (EC), which distributes resources to the network edge, enjoys increasing popularity in the applications with low-latency and high-reliability requirements. EC provides resources in a decentralized manner, which can respond to users’ requirements faster than the normal CC, but with limited computing capacities. As both CC and EC are resource-sensitive, several big issues arise, such as how to conduct job scheduling, resource allocation, and task offloading, which significantly influence the performance of the whole system. To tackle these issues, many optimization problems have been formulated. These optimization problems usually have complex properties, such as non-convexity and NP-hardness, which may not be addressed by the traditional convex optimization-based solutions. Computational intelligence (CI), consisting of a set of nature-inspired computational approaches, recently exhibits great potential in addressing these optimization problems in CC and EC. This article provides an overview of research problems in CC and EC and recent progresses in addressing them with the help of CI techniques. Informative discussions and future research trends are also presented, with the aim of offering insights to the readers and motivating new research directions

    Advances in De Novo Drug Design : From Conventional to Machine Learning Methods

    Get PDF
    De novo drug design is a computational approach that generates novel molecular structures from atomic building blocks with no a priori relationships. Conventional methods include structure-based and ligand-based design, which depend on the properties of the active site of a biological target or its known active binders, respectively. Artificial intelligence, including ma-chine learning, is an emerging field that has positively impacted the drug discovery process. Deep reinforcement learning is a subdivision of machine learning that combines artificial neural networks with reinforcement-learning architectures. This method has successfully been em-ployed to develop novel de novo drug design approaches using a variety of artificial networks including recurrent neural networks, convolutional neural networks, generative adversarial networks, and autoencoders. This review article summarizes advances in de novo drug design, from conventional growth algorithms to advanced machine-learning methodologies and high-lights hot topics for further development.Peer reviewe

    Learning a Swarm Foraging Behavior with Microscopic Fuzzy Controllers Using Deep Reinforcement Learning

    Get PDF
    This article presents a macroscopic swarm foraging behavior obtained using deep reinforcement learning. The selected behavior is a complex task in which a group of simple agents must be directed towards an object to move it to a target position without the use of special gripping mechanisms, using only their own bodies. Our system has been designed to use and combine basic fuzzy behaviors to control obstacle avoidance and the low-level rendezvous processes needed for the foraging task. We use a realistically modeled swarm based on differential robots equipped with light detection and ranging (LiDAR) sensors. It is important to highlight that the obtained macroscopic behavior, in contrast to that of end-to-end systems, combines existing microscopic tasks, which allows us to apply these learning techniques even with the dimensionality and complexity of the problem in a realistic robotic swarm system. The presented behavior is capable of correctly developing the macroscopic foraging task in a robust and scalable way, even in situations that have not been seen in the training phase. An exhaustive analysis of the obtained behavior is carried out, where both the movement of the swarm while performing the task and the swarm scalability are analyzed.This work was supported by the Ministerio de Ciencia, InnovaciĂłn y Universidades (Spain), project RTI2018-096219-B-I00. Project co-financed with FEDER funds

    Particle Swarm Optimisation of Spoken Dialogue System Strategies

    No full text
    International audienceDialogue management optimisation has been cast into a plan- ning under uncertainty problem for long. Some methods such as Reinforcement Learning (RL) are now part of the state of the art. Whatever the solving method, strong assumptions are made about the dialogue system properties. For instance, RL assumes that the dialogue state space is Markovian. Such con- straints may involve important engineering work. This paper introduces a more general approach, based on fewer modelling assumptions. A Black Box Optimisation (BBO) method and more precisely a Particle Swarm Optimisation (PSO) is used to solve the control problem. In addition, PSO allows taking ad- vantage of the parallel aspect of the problem of optimising a system online with many users calling at the same time. Some preliminary results are presented

    A review of key planning and scheduling in the rail industry in Europe and UK

    Get PDF
    Planning and scheduling activities within the rail industry have benefited from developments in computer-based simulation and modelling techniques over the last 25 years. Increasingly, the use of computational intelligence in such tasks is featuring more heavily in research publications. This paper examines a number of common rail-based planning and scheduling activities and how they benefit from five broad technology approaches. Summary tables of papers are provided relating to rail planning and scheduling activities and to the use of expert and decision systems in the rail industry.EPSR

    Comparison of Genetic and Reinforcement Learning Algorithms for Energy Cogeneration Optimization

    Get PDF
    Large process plants generally require energy in different forms: mechanical, electrical, or thermal (in the form of steam or hot water). A commonly used source of energy is cogeneration, also defined as Combined Heat and Power (CHP). Cogeneration can offer substantial economic as well as energy savings; however, its real-time operation scheduling is still a challenge today. Multiple algorithms have been proposed for the CHP control problem in the literature, such as genetic algorithms (GAs), particle swarm optimization algorithms, artificial neural networks, fuzzy decision making systems and, most recently, reinforcement learning (RL) algorithms.This paper presents the comparison of a RL approach and a GA for the control of a cogenerator, using as a case study a thermal power plant serving a factory during the year 2021. The two methods were compared based on an earnings before interest, taxes, depreciation, and amortization (EBITDA) metric. The EBITDA that could be obtained using the RL algorithm, exceeds both the EBITDA that could be generated using a per-week genetic algorithm and the one from the manual scheduling of the CHP. Thus, the RL algorithm proves to be the most cost-effective strategy for the control of a CHP

    An optimal scheduling method in iot-fog-cloud network using combination of aquila optimizer and african vultures optimization

    Get PDF
    Today, fog and cloud computing environments can be used to further develop the Internet of Things (IoT). In such environments, task scheduling is very efficient for executing user requests, and the optimal scheduling of IoT task requests increases the productivity of the IoT-fog-cloud system. In this paper, a hybrid meta-heuristic (MH) algorithm is developed to schedule the IoT requests in IoT-fog-cloud networks using the Aquila Optimizer (AO) and African Vultures Optimization Algorithm (AVOA) called AO_AVOA. In AO_AVOA, the exploration phase of AVOA is improved by using AO operators to obtain the best solution during the process of finding the optimal scheduling solution. A comparison between AO_AVOA and methods of AVOA, AO, Firefly Algorithm (FA), particle swarm optimization (PSO), and Harris Hawks Optimization (HHO) according to performance metrics such as makespan and throughput shows the high ability of AO_AVOA to solve the scheduling problem in IoT-fog-cloud networks. © 2023 by the authors

    Trajectory Generation for a Multibody Robotic System: Modern Methods Based on Product of Exponentials

    Get PDF
    This work presents several trajectory generation algorithms for multibody robotic systems based on the Product of Exponentials (PoE) formulation, also known as screw theory. A PoE formulation is first developed to model the kinematics and dynamics of a multibody robotic manipulator (Sawyer Robot) with 7 revolute joints and an end-effector. In the first method, an Inverse Kinematics (IK) algorithm based on the Newton-Raphson iterative method is applied to generate constrained joint-space trajectories corresponding to straight-line and curvilinear motions of the end effector in Cartesian space with finite jerk. The second approach describes Constant Screw Axis (CSA) trajectories which are generated using Machine Learning (ML) and Artificial Neural Networks (ANNs) techniques. The CSA method smooths the trajectory in the Special Euclidean (SE(3)) space. In the third approach, a multi-objective Swarm Intelligence (SI) trajectory generation algorithm is developed, where the IK problem is tackled using a combined SI-PoE ML technique resulting in a joint trajectory that avoids obstacles in the workspace, and satisfies the finite jerk constraint on end-effector while minimizing the torque profiles. The final method is a different approach to solving the IK problem using the Deep Q-Learning (DQN) Reinforcement Learning (RL) algorithm which can generate different joint space trajectories given the Cartesian end-effector path. For all methods above, the Newton-Euler recursive algorithm is implemented to compute the inverse dynamics, which generates the joint torques profiles. The simulated torque profiles are experimentally validated by feeding the generated joint trajectories to the Sawyer robotic arm through the developed Robot Operating System (ROS) - Python environment in the Software Development Kit (SDK) mode. The developed algorithms can be used to generate various trajectories for robotic arms (e.g. spacecraft servicing missions)
    • …
    corecore