12 research outputs found

    Categorization of indoor places using the Kinect sensor

    Get PDF
    The categorization of places in indoor environments is an important capability for service robots working and interacting with humans. In this paper we present a method to categorize different areas in indoor environments using a mobile robot equipped with a Kinect camera. Our approach transforms depth and grey scale images taken at each place into histograms of local binary patterns (LBPs) whose dimensionality is further reduced following a uniform criterion. The histograms are then combined into a single feature vector which is categorized using a supervised method. In this work we compare the performance of support vector machines and random forests as supervised classifiers. Finally, we apply our technique to distinguish five different place categories: corridors, laboratories, offices, kitchens, and study rooms. Experimental results show that we can categorize these places with high accuracy using our approach

    Towards a Probabilistic Roadmap for Multi-robot Coordination

    Get PDF
    International audienceIn this paper, we discuss the problem of multi-robot coordination and propose an approach for coordinated multi-robot motion planning by using a probabilistic roadmap (PRM) based on adaptive cross sampling (ACS). The proposed approach, called ACS-PRM, is a sampling-based method and consists of three steps including C-space sampling, roadmap building and motion planning. In contrast to previous approaches, our approach is designed to plan separate kinematic paths for multiple robots to minimize the problem of congestion and collision in an effective way so as to improve the system efficiency. Our approach has been implemented and evaluated in simulation. The experimental results demonstrate the total planning time can be obviously reduced by our ACS-PRM approach compared with previous approaches

    Toward understanding natural language directions

    Get PDF
    Speaking using unconstrained natural language is an intuitive and flexible way for humans to interact with robots. Understanding this kind of linguistic input is challenging because diverse words and phrases must be mapped into structures that the robot can understand, and elements in those structures must be grounded in an uncertain environment. We present a system that follows natural language directions by extracting a sequence of spatial description clauses from the linguistic input and then infers the most probable path through the environment given only information about the environmental geometry and detected visible objects. We use a probabilistic graphical model that factors into three key components. The first component grounds landmark phrases such as "the computers" in the perceptual frame of the robot by exploiting co-occurrence statistics from a database of tagged images such as Flickr. Second, a spatial reasoning component judges how well spatial relations such as "past the computers" describe a path. Finally, verb phrases such as "turn right" are modeled according to the amount of change in orientation in the path. Our system follows 60% of the directions in our corpus to within 15 meters of the true destination, significantly outperforming other approaches.United States. Office of Naval Research (MURI N00014-07-1-0749

    Influence of complex environments on LiDAR-Based robot navigation

    Get PDF
    La navigation sécuritaire et efficace des robots mobiles repose grandement sur l’utilisation des capteurs embarqués. L’un des capteurs qui est de plus en plus utilisé pour cette tâche est le Light Detection And Ranging (LiDAR). Bien que les recherches récentes montrent une amélioration des performances de navigation basée sur les LiDARs, faire face à des environnements non structurés complexes ou des conditions météorologiques difficiles reste problématique. Dans ce mémoire, nous présentons une analyse de l’influence de telles conditions sur la navigation basée sur les LiDARs. Notre première contribution est d’évaluer comment les LiDARs sont affectés par les flocons de neige durant les tempêtes de neige. Pour ce faire, nous créons un nouvel ensemble de données en faisant l’acquisition de données durant six précipitations de neige. Une analyse statistique de ces ensembles de données, nous caractérisons la sensibilité de chaque capteur et montrons que les mesures de capteurs peuvent être modélisées de manière probabilistique. Nous montrons aussi que les précipitations de neige ont peu d’influence au-delà de 10 m. Notre seconde contribution est d’évaluer l’impact de structures tridimensionnelles complexes présentes en forêt sur les performances d’un algorithme de reconnaissance d’endroits. Nous avons acquis des données dans un environnement extérieur structuré et en forêt, ce qui permet d’évaluer l’influence de ces derniers sur les performances de reconnaissance d’endroits. Notre hypothèse est que, plus deux balayages laser sont proches l’un de l’autre, plus la croyance que ceux-ci proviennent du même endroit sera élevée, mais modulé par le niveau de complexité de l’environnement. Nos expériences confirment que la forêt, avec ses réseaux de branches compliqués et son feuillage, produit plus de données aberrantes et induit une chute plus rapide des performances de reconnaissance en fonction de la distance. Notre conclusion finale est que, les environnements complexes étudiés influencent négativement les performances de navigation basée sur les LiDARs, ce qui devrait être considéré pour développer des algorithmes de navigation robustes.To ensure safe and efficient navigation, mobile robots heavily rely on their ability to use on-board sensors. One such sensor, increasingly used for robot navigation, is the Light Detection And Ranging (LiDAR). Although recent research showed improvement in LiDAR-based navigation, dealing with complex unstructured environments or difficult weather conditions remains problematic. In this thesis, we present an analysis of the influence of such challenging conditions on LiDAR-based navigation. Our first contribution is to evaluate how LiDARs are affected by snowflakes during snowstorms. To this end, we create a novel dataset by acquiring data during six snowfalls using four sensors simultaneously. Based on statistical analysis of this dataset, we characterized the sensitivity of each device and showed that sensor measurements can be modelled in a probabilistic manner. We also showed that falling snow has little impact beyond a range of 10 m. Our second contribution is to evaluate the impact of complex of three-dimensional structures, present in forests, on the performance of a LiDAR-based place recognition algorithm. We acquired data in structured outdoor environment and in forest, which allowed evaluating the impact of the environment on the place recognition performance. Our hypothesis was that the closer two scans are acquired from each other, the higher the belief that the scans originate from the same place will be, but modulated by the level of complexity of the environments. Our experiments confirmed that forests, with their intricate network of branches and foliage, produce more outliers and induce recognition performance to decrease more quickly with distance when compared with structured outdoor environment. Our conclusion is that falling snow conditions and forest environments negatively impact LiDAR-based navigation performance, which should be considered to develop robust navigation algorithms

    The More you Learn, the Less you Store: Memory-controlled Incremental SVM for Visual Place Recognition

    Get PDF
    The capability to learn from experience is a key property for autonomous cognitive systems working in realistic settings. To this end, this paper presents an SVM-based algorithm, capable of learning model representations incrementally while keeping under control memory requirements. We combine an incremental extension of SVMs with a method reducing the number of support vectors needed to build the decision function without any loss in performance introducing a parameter which permits a user-set trade-off between performance and memory. The resulting algorithm is able to achieve the same recognition results as the original incremental method while reducing the memory growth. Our method is especially suited to work for autonomous systems in realistic settings. We present experiments on two common scenarios in this domain: adaptation in presence of dynamic changes and transfer of knowledge between two different autonomous agents, focusing in both cases on the problem of visual place recognition applied to mobile robot topological localization. Experiments in both scenarios clearly show the power of our approach

    Topological Mapping and Navigation in Real-World Environments

    Full text link
    We introduce the Hierarchical Hybrid Spatial Semantic Hierarchy (H2SSH), a hybrid topological-metric map representation. The H2SSH provides a more scalable representation of both small and large structures in the world than existing topological map representations, providing natural descriptions of a hallway lined with offices as well as a cluster of buildings on a college campus. By considering the affordances in the environment, we identify a division of space into three distinct classes: path segments afford travel between places at their ends, decision points present a choice amongst incident path segments, and destinations typically exist at the start and end of routes. Constructing an H2SSH map of the environment requires understanding both its local and global structure. We present a place detection and classification algorithm to create a semantic map representation that parses the free space in the local environment into a set of discrete areas representing features like corridors, intersections, and offices. Using these areas, we introduce a new probabilistic topological simultaneous localization and mapping algorithm based on lazy evaluation to estimate a probability distribution over possible topological maps of the global environment. After construction, an H2SSH map provides the necessary representations for navigation through large-scale environments. The local semantic map provides a high-fidelity metric map suitable for motion planning in dynamic environments, while the global topological map is a graph-like map that allows for route planning using simple graph search algorithms. For navigation, we have integrated the H2SSH with Model Predictive Equilibrium Point Control (MPEPC) to provide safe and efficient motion planning for our robotic wheelchair, Vulcan. However, navigation in human environments entails more than safety and efficiency, as human behavior is further influenced by complex cultural and social norms. We show how social norms for moving along corridors and through intersections can be learned by observing how pedestrians around the robot behave. We then integrate these learned norms with MPEPC to create a socially-aware navigation algorithm, SA-MPEPC. Through real-world experiments, we show how SA-MPEPC improves not only Vulcan’s adherence to social norms, but the adherence of pedestrians interacting with Vulcan as well.PHDComputer Science & EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttps://deepblue.lib.umich.edu/bitstream/2027.42/144014/1/collinej_1.pd

    Multi-robot deployment planning in communication-constrained environments

    Get PDF
    A lo largo de los últimos años se ha podido observar el aumento del uso de equipos de robots en tareas en las cuales es imposible o poco eficiente la intervención de los humanos, e incluso que implica un cierto grado de riesgo para una persona. Por ejemplo, monitorización de entornos de difícil acceso, como podrían ser túneles, minas, etc. Éste es el tema en el que se ha enfocado el trabajo realizado durante esta tesis: la planificación del despliegue de un equipo de agentes para la monitorización de entornos.La misión de los agentes es alcanzar unas localizaciones de interés y transmitirle la información observada a una estación base estática. Ante la ausencia de una infraestructura de comunicaciones, una transmisión directa a la base es imposible. Por tanto, los agentes se deben coordinar de manera autónoma, de modo que algunos de ellos alcancen los objetivos y otros realicen la función de repetidor para retransmitir la información.Nos hemos centrado en dos líneas de investigación principales, relacionadas con dos maneras del envío de la información a la estación base. En el primer enfoque, los agentes deben mantener un enlace de comunicación con la base en el momento de alcanzar los objetivos. Con el fin de, por ejemplo, poder interactuar desde la base con un robot que ha alcanzado el objetivo. Para ello hemos desarrollado un método que obtiene las posiciones óptimas para los agentes utilizados a modo de repetidor. A continuación, hemos implementado un método de planificación de caminos de modo que los agentes pudiesen navegar el máximo tiempo posible dentro de zonas con señal. Empleando conjuntamente ambos métodos, los agentes extienden el área de cobertura de la estación base, estableciendo un enlace de comunicación desde la misma hasta los objetivos marcados.Utilizando este método, el equipo es capaz de lidiar con variaciones del entorno si la comunicación entre los agentes no se pierde. Sin embargo, los eventos tan comunes e irrelevantes para los seres humanos, como el simple cierre de una puerta, pueden llegar a ser críticos para el equipo de robots. Ya que esto podría interrumpir la comunicación entre el equipo. Por ello, hemos propuesto un método distribuido para que el equipo sea capaz de reconectarse, formando una cadena hacia un objetivo, en escenarios donde haya variaciones con respecto al mapa inicial que poseían los robots.La segunda parte de la presente tesis se ha centrado en misiones de recopilación de datos de un entorno. Aquí la comunicación con la estación base, en el instante de alcanzar un objetivo, no es necesaria y a menudo imposible. Por tanto, en este tipo de escenarios, es más eficiente que algunos agentes, llamados trabajadores, recopilen datos del entorno, y otros, denominados colectores, reúnan la información de los que trabajan para periódicamente retransmitirla a la base. De este modo tan solo los colectores realizan largos viajes a la estación base, mientras que los trabajadores emplean la mayor parte de su tiempo exclusivamente a la recopilación de datos.Primero, hemos desarrollado dos métodos para la planificación de caminos para la sincronización entre los trabajadores y colectores. El primero, muestrea el espacio de manera aleatoria, para obtener una solución lo más rápido posible. El segundo, usando FMM, es más lento, pero obtiene soluciones óptimas.Finalmente, hemos propuesto una técnica global para la misión de recopilación de datos. Este método consiste en: encontrar el mejor balance entre la cantidad de trabajadores y colectores, la mejor división del escenario en áreas de trabajo para los trabajadores, la asociación de los trabajadores para transmitir los datos recopilados a los colectores o directamente a la estación base, así como los caminos de los colectores. El método propuesto trata de encontrar la mejor solución con el fin de entregar la mayor cantidad de datos y que el tiempo de "refresco" de los mismos sea el menor posible.<br /

    Vision-based guidance and control of a hovering vehicle in unknown environments

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Mechanical Engineering, 2008.Includes bibliographical references (leaves 115-122).This thesis presents a methodology, architecture, hardware implementation, and results of a system capable of controlling and guiding a hovering vehicle in unknown environments, emphasizing cluttered indoor spaces. Six-axis inertial data and a low-resolution onboard camera yield sufficient information for image processing, Kalman filtering, and novel mapping algorithms to generate a, high-performance estimate of vehicle motion, as well as an accurate three-dimensional map of the environment. This combination of mapping and localization enables a quadrotor vehicle to autonomously navigate cluttered, unknown environments safely. Communication limitations are considered, and a hybrid control architecture is presented to demonstrate the feasibility of combining separated proactive offboard and reactive onboard planners simultaneously, including a detailed presentation of a novel reactive obstacle avoidance algorithm and preliminary results integrating the MIT Darpa Urban Challenge planner for high-level control. The RAVEN testbed is successfully employed as a prototyping facility for rapid development of these algorithms using emulated inertial data and offboard processing as a precursor to embedded development. An analysis of computational demand and a comparison of the emulated inertial system to an embedded sensor package demonstrates the feasibility of porting the onboard algorithms to an embedded autopilot. Finally, flight results using only the single camera and emulated inertial data for closed-loop trajectory following, environment mapping, and obstacle avoidance are presented and discussed.by Spencer Greg Ahrens.S.M
    corecore