14 research outputs found

    Contributions to improve the technologies supporting unmanned aircraft operations

    Get PDF
    Mención Internacional en el título de doctorUnmanned Aerial Vehicles (UAVs), in their smaller versions known as drones, are becoming increasingly important in today's societies. The systems that make them up present a multitude of challenges, of which error can be considered the common denominator. The perception of the environment is measured by sensors that have errors, the models that interpret the information and/or define behaviors are approximations of the world and therefore also have errors. Explaining error allows extending the limits of deterministic models to address real-world problems. The performance of the technologies embedded in drones depends on our ability to understand, model, and control the error of the systems that integrate them, as well as new technologies that may emerge. Flight controllers integrate various subsystems that are generally dependent on other systems. One example is the guidance systems. These systems provide the engine's propulsion controller with the necessary information to accomplish a desired mission. For this purpose, the flight controller is made up of a control law for the guidance system that reacts to the information perceived by the perception and navigation systems. The error of any of the subsystems propagates through the ecosystem of the controller, so the study of each of them is essential. On the other hand, among the strategies for error control are state-space estimators, where the Kalman filter has been a great ally of engineers since its appearance in the 1960s. Kalman filters are at the heart of information fusion systems, minimizing the error covariance of the system and allowing the measured states to be filtered and estimated in the absence of observations. State Space Models (SSM) are developed based on a set of hypotheses for modeling the world. Among the assumptions are that the models of the world must be linear, Markovian, and that the error of their models must be Gaussian. In general, systems are not linear, so linearization are performed on models that are already approximations of the world. In other cases, the noise to be controlled is not Gaussian, but it is approximated to that distribution in order to be able to deal with it. On the other hand, many systems are not Markovian, i.e., their states do not depend only on the previous state, but there are other dependencies that state space models cannot handle. This thesis deals a collection of studies in which error is formulated and reduced. First, the error in a computer vision-based precision landing system is studied, then estimation and filtering problems from the deep learning approach are addressed. Finally, classification concepts with deep learning over trajectories are studied. The first case of the collection xviiistudies the consequences of error propagation in a machine vision-based precision landing system. This paper proposes a set of strategies to reduce the impact on the guidance system, and ultimately reduce the error. The next two studies approach the estimation and filtering problem from the deep learning approach, where error is a function to be minimized by learning. The last case of the collection deals with a trajectory classification problem with real data. This work completes the two main fields in deep learning, regression and classification, where the error is considered as a probability function of class membership.Los vehículos aéreos no tripulados (UAV) en sus versiones de pequeño tamaño conocidos como drones, van tomando protagonismo en las sociedades actuales. Los sistemas que los componen presentan multitud de retos entre los cuales el error se puede considerar como el denominador común. La percepción del entorno se mide mediante sensores que tienen error, los modelos que interpretan la información y/o definen comportamientos son aproximaciones del mundo y por consiguiente también presentan error. Explicar el error permite extender los límites de los modelos deterministas para abordar problemas del mundo real. El rendimiento de las tecnologías embarcadas en los drones, dependen de nuestra capacidad de comprender, modelar y controlar el error de los sistemas que los integran, así como de las nuevas tecnologías que puedan surgir. Los controladores de vuelo integran diferentes subsistemas los cuales generalmente son dependientes de otros sistemas. Un caso de esta situación son los sistemas de guiado. Estos sistemas son los encargados de proporcionar al controlador de los motores información necesaria para cumplir con una misión deseada. Para ello se componen de una ley de control de guiado que reacciona a la información percibida por los sistemas de percepción y navegación. El error de cualquiera de estos sistemas se propaga por el ecosistema del controlador siendo vital su estudio. Por otro lado, entre las estrategias para abordar el control del error se encuentran los estimadores en espacios de estados, donde el filtro de Kalman desde su aparición en los años 60, ha sido y continúa siendo un gran aliado para los ingenieros. Los filtros de Kalman son el corazón de los sistemas de fusión de información, los cuales minimizan la covarianza del error del sistema, permitiendo filtrar los estados medidos y estimarlos cuando no se tienen observaciones. Los modelos de espacios de estados se desarrollan en base a un conjunto de hipótesis para modelar el mundo. Entre las hipótesis se encuentra que los modelos del mundo han de ser lineales, markovianos y que el error de sus modelos ha de ser gaussiano. Generalmente los sistemas no son lineales por lo que se realizan linealizaciones sobre modelos que a su vez ya son aproximaciones del mundo. En otros casos el ruido que se desea controlar no es gaussiano, pero se aproxima a esta distribución para poder abordarlo. Por otro lado, multitud de sistemas no son markovianos, es decir, sus estados no solo dependen del estado anterior, sino que existen otras dependencias que los modelos de espacio de estados no son capaces de abordar. Esta tesis aborda un compendio de estudios sobre los que se formula y reduce el error. En primer lugar, se estudia el error en un sistema de aterrizaje de precisión basado en visión por computador. Después se plantean problemas de estimación y filtrado desde la aproximación del aprendizaje profundo. Por último, se estudian los conceptos de clasificación con aprendizaje profundo sobre trayectorias. El primer caso del compendio estudia las consecuencias de la propagación del error de un sistema de aterrizaje de precisión basado en visión artificial. En este trabajo se propone un conjunto de estrategias para reducir el impacto sobre el sistema de guiado, y en última instancia reducir el error. Los siguientes dos estudios abordan el problema de estimación y filtrado desde la perspectiva del aprendizaje profundo, donde el error es una función que minimizar mediante aprendizaje. El último caso del compendio aborda un problema de clasificación de trayectorias con datos reales. Con este trabajo se completan los dos campos principales en aprendizaje profundo, regresión y clasificación, donde se plantea el error como una función de probabilidad de pertenencia a una clase.I would like to thank the Ministry of Science and Innovation for granting me the funding with reference PRE2018-086793, associated to the project TEC2017-88048-C2-2-R, which provide me the opportunity to carry out all my PhD. activities, including completing an international research internship.Programa de Doctorado en Ciencia y Tecnología Informática por la Universidad Carlos III de MadridPresidente: Antonio Berlanga de Jesús.- Secretario: Daniel Arias Medina.- Vocal: Alejandro Martínez Cav

    Proceedings of the NASA Conference on Space Telerobotics, volume 3

    Get PDF
    The theme of the Conference was man-machine collaboration in space. The Conference provided a forum for researchers and engineers to exchange ideas on the research and development required for application of telerobotics technology to the space systems planned for the 1990s and beyond. The Conference: (1) provided a view of current NASA telerobotic research and development; (2) stimulated technical exchange on man-machine systems, manipulator control, machine sensing, machine intelligence, concurrent computation, and system architectures; and (3) identified important unsolved problems of current interest which can be dealt with by future research

    Advances in Robot Kinematics : Proceedings of the 15th international conference on Advances in Robot Kinematics

    Get PDF
    International audienceThe motion of mechanisms, kinematics, is one of the most fundamental aspect of robot design, analysis and control but is also relevant to other scientific domains such as biome- chanics, molecular biology, . . . . The series of books on Advances in Robot Kinematics (ARK) report the latest achievement in this field. ARK has a long history as the first book was published in 1991 and since then new issues have been published every 2 years. Each book is the follow-up of a single-track symposium in which the participants exchange their results and opinions in a meeting that bring together the best of world’s researchers and scientists together with young students. Since 1992 the ARK symposia have come under the patronage of the International Federation for the Promotion of Machine Science-IFToMM.This book is the 13th in the series and is the result of peer-review process intended to select the newest and most original achievements in this field. For the first time the articles of this symposium will be published in a green open-access archive to favor free dissemination of the results. However the book will also be o↵ered as a on-demand printed book.The papers proposed in this book show that robot kinematics is an exciting domain with an immense number of research challenges that go well beyond the field of robotics.The last symposium related with this book was organized by the French National Re- search Institute in Computer Science and Control Theory (INRIA) in Grasse, France

    Proceedings of the International Micro Air Vehicles Conference and Flight Competition 2017 (IMAV 2017)

    Get PDF
    The IMAV 2017 conference has been held at ISAE-SUPAERO, Toulouse, France from Sept. 18 to Sept. 21, 2017. More than 250 participants coming from 30 different countries worldwide have presented their latest research activities in the field of drones. 38 papers have been presented during the conference including various topics such as Aerodynamics, Aeroacoustics, Propulsion, Autopilots, Sensors, Communication systems, Mission planning techniques, Artificial Intelligence, Human-machine cooperation as applied to drones

    Multimodal Navigation for Accurate Space Rendezvous Missions

    Get PDF
    © Cranfield University 2021. All rights reserved. No part of this publication may be reproduced without the written permission of the copyright ownerRelative navigation is paramount in space missions that involve rendezvousing between two spacecraft. It demands accurate and continuous estimation of the six degree-of-freedom relative pose, as this stage involves close-proximity-fast-reaction operations that can last up to five orbits. This has been routinely achieved thanks to active sensors such as lidar, but their large size, cost, power and limited operational range remain a stumbling block for en masse on-board integration. With the onset of faster processing units, lighter and cheaper passive optical sensors are emerging as the suitable alternative for autonomous rendezvous in combination with computer vision algorithms. Current vision-based solutions, however, are limited by adverse illumination conditions such as solar glare, shadowing, and eclipse. These effects are exacerbated when the target does not hold cooperative markers to accommodate the estimation process and is incapable of controlling its rotational state. This thesis explores novel model-based methods that exploit sequences of monoc ular images acquired by an on-board camera to accurately carry out spacecraft relative pose estimation for non-cooperative close-range rendezvous with a known artificial target. The proposed solutions tackle the current challenges of imaging in the visible spectrum and investigate the contribution of the long wavelength infrared (or “thermal”) band towards a combined multimodal approach. As part of the research, a visible-thermal synthetic dataset of a rendezvous approach with the defunct satellite Envisat is generated from the ground up using a realistic orbital camera simulator. From the rendered trajectories, the performance of several state-of-the-art feature detectors and descriptors is first evaluated for both modalities in a tailored scenario for short and wide baseline image processing transforms. Multiple combinations, including the pairing of algorithms with their non-native counterparts, are tested. Computational runtimes are assessed in an embedded hardware board. From the insight gained, a method to estimate the pose on the visible band is derived from minimising geometric constraints between online local point and edge contour features matched to keyframes generated offline from a 3D model of the target. The combination of both feature types is demonstrated to achieve a pose solution for a tumbling target using a sparse set of training images, bypassing the need for hardware-accelerated real-time renderings of the model. The proposed algorithm is then augmented with an extended Kalman filter which processes each feature-induced minimisation output as individual pseudo measurements, fusing them to estimate the relative pose and velocity states at each time-step. Both the minimisation and filtering are established using Lie group formalisms, allowing for the covariance of the solution computed by the former to be automatically incorporated as measurement noise in the latter, providing an automatic weighing of each feature type directly related to the quality of the matches. The predicted states are then used to search for new feature matches in the subsequent time-step. Furthermore, a method to derive a coarse viewpoint estimate to initialise the nominal algorithm is developed based on probabilistic modelling of the target’s shape. The robustness of the complete approach is demonstrated for several synthetic and laboratory test cases involving two types of target undergoing extreme illumination conditions. Lastly, an innovative deep learning-based framework is developed by processing the features extracted by a convolutional front-end with long short-term memory cells, thus proposing the first deep recurrent convolutional neural network for spacecraft pose estimation. The framework is used to compare the performance achieved by visible-only and multimodal input sequences, where the addition of the thermal band is shown to greatly improve the performance during sunlit sequences. Potential limitations of this modality are also identified, such as when the target’s thermal signature is comparable to Earth’s during eclipse.PH

    Third International Symposium on Artificial Intelligence, Robotics, and Automation for Space 1994

    Get PDF
    The Third International Symposium on Artificial Intelligence, Robotics, and Automation for Space (i-SAIRAS 94), held October 18-20, 1994, in Pasadena, California, was jointly sponsored by NASA, ESA, and Japan's National Space Development Agency, and was hosted by the Jet Propulsion Laboratory (JPL) of the California Institute of Technology. i-SAIRAS 94 featured presentations covering a variety of technical and programmatic topics, ranging from underlying basic technology to specific applications of artificial intelligence and robotics to space missions. i-SAIRAS 94 featured a special workshop on planning and scheduling and provided scientists, engineers, and managers with the opportunity to exchange theoretical ideas, practical results, and program plans in such areas as space mission control, space vehicle processing, data analysis, autonomous spacecraft, space robots and rovers, satellite servicing, and intelligent instruments

    Optimisation of docking locations for remotely operated vehicles.

    Get PDF
    This thesis describes work aimed at developing practical methods for determining the best docking locations for an underwater remotely operated vehicle (ROV) when inspecting an offshore platform. ROVs are used extensively in the offshore oil and gas industry to conduct a large variety of intervention tasks such as visual inspection, operational monitoring, equipment installation and operation, debris recovery, and so on. However, they have found only limited use in the more difficult tasks such as the detailed inspection of complex weld geometries. These complex welds are, however, found extensively in the construction of the majority of offshore structures and platforms ('oil rigs'). Furthermore, there is a safety requirement to have them inspected regularly since failure of these welds can potentially lead to catastrophic failure of the structures, the majority of which are manned. A number of specialist ROV systems have been developed that are able to attach onto platform structures and use their manipulators to conduct inspection. However, due to the short reach of the manipulators and the complex geometry of the welds (often encumbered with protruding pipes and other fittings) the success of any inspection is crucially dependent on a good initial choice of ROV docking position. This thesis will describe the problems and current manual planning methods, and then detail the development of two new methods for automated optimisation of docking positions - firstly using neural networks, and secondly using more conventional numerical processing. This thesis will also review related work in the field, such as the development of neural networks and their applications in the general offshore environment and in the control of ROVs and robot manipulator arms, and other approaches to ROV docking. It will further describe the use of the system developed here for planning docking positions on example commercial ROV inspection work programmes

    Revisiting the determination of the singularity cases in the visual servoing of image points through the concept of hidden robot

    Get PDF
    International audienceThe determination of the singularity cases in visual servoing is a tricky problem, which is unsolved for most of the image-based approaches. In order to avoid singularities, redundant measurements may be used. However, they lead to the presence of local minima. Moreover, they do not always ensure that singularities can be avoided. Here, we show that a concept named the “hidden robot,” which was formerly used for understanding the singularities of a vision-based controller dedicated to parallel robots, can be used for interpreting the singularities in the visual servoing of image points. These singularity cases were already found in the case in which three points are observed, but we show that the hidden robot concept considerably simplifies the analysis by using geometric interpretations of the mapping degeneracy and tools provided by the mechanical engineering community. Moreover, to the best of our knowledge, for the first time, we provide the singularity conditions when more than three points are observed. We also discuss how these tools could be extended in order to find the singularity cases of other visual servoing techniques (e.g., when lines are observed)

    Revisiting the determination of the singularity cases in the visual servoing of image points through the concept of hidden robot

    No full text
    International audienceThe determination of the singularity cases in visual servoing is a tricky problem which is unsolved for most image- based approaches. In order to avoid singularities, redundant measurements may be used. However, they lead to the presence of local minima. Moreover, they do not always ensure that singularities can be avoided. Here, we show that a concept named the “hidden robot”, which was formerly used for understanding the singularities of a vision-based controller dedicated to parallel robots, can be used for interpreting the singularities in the visual servoing of image points. These singularity cases were already found in the case where three points are observed but we show that the hidden robot concept considerably simplifies the analysis by using geometric interpretations of the mapping degeneracy and tools provided by the mechanical engineering community. Moreover, to the best of our knowledge, we provide for the first time the singularity conditions when more than three points are observed. We also discuss how these tools could be extended in order to find the singularity cases of other visual servoing techniques (e.g. when lines are observed)
    corecore