960 research outputs found

    UAV Optimal Cooperative Obstacle Avoidance and Target Tracking in Dynamic Stochastic Environments

    Get PDF
    Cette thèse propose une stratégie de contrôle avancée pour guider une flotte d'aéronefs sans pilote (UAV) dans un environnement à la fois stochastique et dynamique. Pour ce faire, un simulateur de vol 3D a été développé avec MATLAB® pour tester les algorithmes de la stratégie de guidage en fonctions de différents scénarios. L'objectif des missions simulées est de s'assurer que chaque UAV intercepte une cible ellipsoïdale mobile tout en évitant une panoplie d'obstacles ellipsoïdaux mobiles détectés en route. Les UAVs situés à l'intérieur des limites de communication peuvent coopérer afin d'améliorer leurs performances au cours de la mission. Le simulateur a été conçu de façon à ce que les UAV soient dotés de capteurs et d'appareils de communication de portée limitée. De plus, chaque UAV possède un pilote automatique qui stabilise l'aéronef en vol et un planificateur de trajectoires qui génère les commandes à envoyer au pilote automatique. Au coeur du planificateur de trajectoires se trouve un contrôleur prédictif à horizon fuyant qui détermine les commandes à envoyer à l'UAV. Ces commandes optimisent un critère de performance assujetti à des contraintes. Le critère de performance est conçu de sorte que les UAV atteignent les objectifs de la mission, alors que les contraintes assurent que les commandes générées adhèrent aux limites de manoeuvrabilité de l'aéronef. La planification de trajectoires pour UAV opérant dans un environnement dynamique et stochastique dépend fortement des déplacements anticipés des objets (obstacle, cible). Un filtre de Kalman étendu est donc utilisé pour prédire les trajectoires les plus probables des objets à partir de leurs états estimés. Des stratégies de poursuite et d'évitement ont aussi été développées en fonction des trajectoires prédites des objets détectés. Pour des raisons de sécurité, la conception de stratégies d'évitement de collision à la fois efficaces et robustes est primordiale au guidage d'UAV. Une nouvelle stratégie d'évitement d'obstacles par approche probabiliste a donc été développée. La méthode cherche à minimiser la probabilité de collision entre l'UAV et tous ses obstacles détectés sur l'horizon de prédiction, tout en s'assurant que, à chaque pas de prédiction, la probabilité de collision entre l'UAV et chacun de ses obstacles détectés ne surpasse pas un seuil prescrit. Des simulations sont présentées au cours de cette thèse pour démontrer l'efficacité des algorithmes proposés

    Multi-target detection and recognition by UAVs using online POMDPs

    Get PDF
    This paper tackles high-level decision-making techniques for robotic missions, which involve both active sensing and symbolic goal reaching, under uncertain probabilistic environments and strong time constraints. Our case study is a POMDP model of an online multi-target detection and recognition mission by an autonomous UAV.The POMDP model of the multi-target detection and recognition problem is generated online from a list of areas of interest, which are automatically extracted at the beginning of the flight from a coarse-grained high altitude observation of the scene. The POMDP observation model relies on a statistical abstraction of an image processing algorithm's output used to detect targets. As the POMDP problem cannot be known and thus optimized before the beginning of the flight, our main contribution is an ``optimize-while-execute'' algorithmic framework: it drives a POMDP sub-planner to optimize and execute the POMDP policy in parallel under action duration constraints. We present new results from real outdoor flights and SAIL simulations, which highlight both the benefits of using POMDPs in multi-target detection and recognition missions, and of our`optimize-while-execute'' paradigm

    Optimal control approaches for consensus and path planning in multi-agent systems

    Get PDF
    Optimal control is one of the most powerful, important and advantageous topics in control engineering. The two challenges in every optimal control problem are defining the proper cost function and obtaining the best method to minimize it. In this study, innovative optimal control approaches are developed to solve the two problems of consensus and path planning in multi-agent systems (MASs). The consensus problem for general Linear-Time Invariant systems is solved by implementing an inverse optimal control approach which enables us to start by deriving a control law based on the stability and optimality condition and then according to the derived control define the cost function. We will see that this method in which the cost function is not specified a priori as the conventional optimal control design has the benefit that the resulting control law is guaranteed to be both stabilizing and optimal. Three new theorems in related linear algebra are developed to enable us to use the algorithm for all the general LTI systems. The designed optimal control is distributed and only needs local neighbor-to-neighbor information based on the communication topology to make the agents achieve consensus and track a desired trajectory. Path planning problem is solved for a group are Unmanned Aerial Vehicles (UAVs) that are assigned to track the fronts of a fires in a process of wildfire management. We use Partially Observable Markov Decision Process (POMDP) in order to minimize the cost function that is defined according to the tracking error. Here the challenge is designing the algorithm such that (1) the UAVs are able to make decisions autonomously on which fire front to track and (2) they are able to track the fire fronts which evolve over time in random directions. We will see that by defining proper models, the designed algorithms provides real-time calculation of control variables which enables the UAVs to track the fronts and find their way autonomously. Furthermore, by implementing Nominal Belief-state Optimization (NBO) method, the dynamic constraints of the UAVs is considered and challenges such as collision avoidance is addressed completely in the context of POMDP

    Adaptive and learning-based formation control of swarm robots

    Get PDF
    Autonomous aerial and wheeled mobile robots play a major role in tasks such as search and rescue, transportation, monitoring, and inspection. However, these operations are faced with a few open challenges including robust autonomy, and adaptive coordination based on the environment and operating conditions, particularly in swarm robots with limited communication and perception capabilities. Furthermore, the computational complexity increases exponentially with the number of robots in the swarm. This thesis examines two different aspects of the formation control problem. On the one hand, we investigate how formation could be performed by swarm robots with limited communication and perception (e.g., Crazyflie nano quadrotor). On the other hand, we explore human-swarm interaction (HSI) and different shared-control mechanisms between human and swarm robots (e.g., BristleBot) for artistic creation. In particular, we combine bio-inspired (i.e., flocking, foraging) techniques with learning-based control strategies (using artificial neural networks) for adaptive control of multi- robots. We first review how learning-based control and networked dynamical systems can be used to assign distributed and decentralized policies to individual robots such that the desired formation emerges from their collective behavior. We proceed by presenting a novel flocking control for UAV swarm using deep reinforcement learning. We formulate the flocking formation problem as a partially observable Markov decision process (POMDP), and consider a leader-follower configuration, where consensus among all UAVs is used to train a shared control policy, and each UAV performs actions based on the local information it collects. In addition, to avoid collision among UAVs and guarantee flocking and navigation, a reward function is added with the global flocking maintenance, mutual reward, and a collision penalty. We adapt deep deterministic policy gradient (DDPG) with centralized training and decentralized execution to obtain the flocking control policy using actor-critic networks and a global state space matrix. In the context of swarm robotics in arts, we investigate how the formation paradigm can serve as an interaction modality for artists to aesthetically utilize swarms. In particular, we explore particle swarm optimization (PSO) and random walk to control the communication between a team of robots with swarming behavior for musical creation
    corecore