5,168 research outputs found

    UltraSwarm: A Further Step Towards a Flock of Miniature Helicopters

    Get PDF
    We describe further progress towards the development of a MAV (micro aerial vehicle) designed as an enabling tool to investigate aerial flocking. Our research focuses on the use of low cost off the shelf vehicles and sensors to enable fast prototyping and to reduce development costs. Details on the design of the embedded electronics and the modification of the chosen toy helicopter are presented, and the technique used for state estimation is described. The fusion of inertial data through an unscented Kalman filter is used to estimate the helicopter’s state, and this forms the main input to the control system. Since no detailed dynamic model of the helicopter in use is available, a method is proposed for automated system identification, and for subsequent controller design based on artificial evolution. Preliminary results obtained with a dynamic simulator of a helicopter are reported, along with some encouraging results for tackling the problem of flocking

    Neural network controller against environment: A coevolutive approach to generalize robot navigation behavior

    Get PDF
    In this paper, a new coevolutive method, called Uniform Coevolution, is introduced to learn weights of a neural network controller in autonomous robots. An evolutionary strategy is used to learn high-performance reactive behavior for navigation and collisions avoidance. The introduction of coevolutive over evolutionary strategies allows evolving the environment, to learn a general behavior able to solve the problem in different environments. Using a traditional evolutionary strategy method, without coevolution, the learning process obtains a specialized behavior. All the behaviors obtained, with/without coevolution have been tested in a set of environments and the capability of generalization is shown for each learned behavior. A simulator based on a mini-robot Khepera has been used to learn each behavior. The results show that Uniform Coevolution obtains better generalized solutions to examples-based problems.Publicad

    Gradient-free Policy Architecture Search and Adaptation

    Full text link
    We develop a method for policy architecture search and adaptation via gradient-free optimization which can learn to perform autonomous driving tasks. By learning from both demonstration and environmental reward we develop a model that can learn with relatively few early catastrophic failures. We first learn an architecture of appropriate complexity to perceive aspects of world state relevant to the expert demonstration, and then mitigate the effect of domain-shift during deployment by adapting a policy demonstrated in a source domain to rewards obtained in a target environment. We show that our approach allows safer learning than baseline methods, offering a reduced cumulative crash metric over the agent's lifetime as it learns to drive in a realistic simulated environment.Comment: Accepted in Conference on Robot Learning, 201

    Investigation on Mlp Artificial Neural Network Using FPGA For Autonomous Cart Follower System

    Get PDF
    Dengan kos alat pengesan yang semakin rendah, masa depan sistem pedati pengikut autonomi akan dilengkapi dengan lebih banyak alat pengesan. Ini menjadi cabaran rekabentuk dalam mengendalikan data besar dan kerumitan perkukuhan. Kebanyakan sistem yang sedia ada menggunakan papan mikropengawal yang mempunyai prestasi yang terhad dan pengembangan tidak mungkin tanpa penggantian yang lebih baru. Projek ini mencadangkan perlaksanaan alternatif sistem pedati pengikut autonomi dengan model rangkaian neural MLP menggunakan FPGA. Sistem pedati pengikut autonomi yang mengguakan papan mikropengawal telah diubah suai untuk menggunakan papan FPGA dan dilaksanakan melalui Sistem pada Chip (SOC). System rangkaian neural dilatih dalam simulasi dengan vektor latihan yang dikumpul daripada sistem pedati pengikut autonomi yang sedia ada. System rangkaian neural kemudian dilaksanakan sebagai perkukuhan dalam SOC itu. Dalam pemerhatian, jejak perkukuhan model rangkaian neural kekal saiz kecil tanpa mengira saiz rangkaian neural. Hasil kajian menunjukkan bahawa dengan penggunaan sumber tambahan sebanyak 40%, penambahbaikan sistem secara keseluruhan sebanyak 27 kali dicapai dengan penggunaan blok pecutan perkakasan di SOC, berbanding dengan SOC tanpa penggunaan blok pecutan perkakasan. ________________________________________________________________________________________________________________________ The future of the autonomous cart follower system will equipped with lots of sensory data, due to the ever lower cost of sensory device. This provides design challenge on handling large data and firmware complexity. Most of the existing systems are implemented via usage of microcontroller board, which has limited performance and expansion is not possible without replacement of newer board. The project proposes an alternative approach of running the autonomous cart follower systems on neural network model using Field Programmable Gates Array (FPGA). A microcontroller based autonomous cart follower systems is modified to use the FPGA board and implemented via the System on Chip (SOC) approach. The neural network is trained offline in simulation tools with training vector collected from running the existing autonomous cart follower systems. The trained neural network model then implemented as software code in the SOC. By observation the firmware footprint of the neural network model remains small size regardless of the neural network size. The result shows that with 40% more additional resource utilization, the overall system improvement of 27 times is achieved with the usage of hardware acceleration block in SOC compared to SOC without hardware acceleration

    Autonomous Drone Landings on an Unmanned Marine Vehicle using Deep Reinforcement Learning

    Get PDF
    This thesis describes with the integration of an Unmanned Surface Vehicle (USV) and an Unmanned Aerial Vehicle (UAV, also commonly known as drone) in a single Multi-Agent System (MAS). In marine robotics, the advantage offered by a MAS consists of exploiting the key features of a single robot to compensate for the shortcomings in the other. In this way, a USV can serve as the landing platform to alleviate the need for a UAV to be airborne for long periods time, whilst the latter can increase the overall environmental awareness thanks to the possibility to cover large portions of the prevailing environment with a camera (or more than one) mounted on it. There are numerous potential applications in which this system can be used, such as deployment in search and rescue missions, water and coastal monitoring, and reconnaissance and force protection, to name but a few. The theory developed is of a general nature. The landing manoeuvre has been accomplished mainly identifying, through artificial vision techniques, a fiducial marker placed on a flat surface serving as a landing platform. The raison d'etre for the thesis was to propose a new solution for autonomous landing that relies solely on onboard sensors and with minimum or no communications between the vehicles. To this end, initial work solved the problem while using only data from the cameras mounted on the in-flight drone. In the situation in which the tracking of the marker is interrupted, the current position of the USV is estimated and integrated into the control commands. The limitations of classic control theory used in this approached suggested the need for a new solution that empowered the flexibility of intelligent methods, such as fuzzy logic or artificial neural networks. The recent achievements obtained by deep reinforcement learning (DRL) techniques in end-to-end control in playing the Atari video-games suite represented a fascinating while challenging new way to see and address the landing problem. Therefore, novel architectures were designed for approximating the action-value function of a Q-learning algorithm and used to map raw input observation to high-level navigation actions. In this way, the UAV learnt how to land from high latitude without any human supervision, using only low-resolution grey-scale images and with a level of accuracy and robustness. Both the approaches have been implemented on a simulated test-bed based on Gazebo simulator and the model of the Parrot AR-Drone. The solution based on DRL was further verified experimentally using the Parrot Bebop 2 in a series of trials. The outcomes demonstrate that both these innovative methods are both feasible and practicable, not only in an outdoor marine scenario but also in indoor ones as well

    Robustness analysis of evolutionary controller tuning using real systems

    Get PDF
    A genetic algorithm (GA) presents an excellent method for controller parameter tuning. In our work, we evolved the heading as well as the altitude controller for a small lightweight helicopter. We use the real flying robot to evaluate the GA's individuals rather than an artificially consistent simulator. By doing so we avoid the ldquoreality gaprdquo, taking the controller from the simulator to the real world. In this paper we analyze the evolutionary aspects of this technique and discuss the issues that need to be considered for it to perform well and result in robust controllers
    corecore