6,182 research outputs found

    Fast, Autonomous Flight in GPS-Denied and Cluttered Environments

    Full text link
    One of the most challenging tasks for a flying robot is to autonomously navigate between target locations quickly and reliably while avoiding obstacles in its path, and with little to no a-priori knowledge of the operating environment. This challenge is addressed in the present paper. We describe the system design and software architecture of our proposed solution, and showcase how all the distinct components can be integrated to enable smooth robot operation. We provide critical insight on hardware and software component selection and development, and present results from extensive experimental testing in real-world warehouse environments. Experimental testing reveals that our proposed solution can deliver fast and robust aerial robot autonomous navigation in cluttered, GPS-denied environments.Comment: Pre-peer reviewed version of the article accepted in Journal of Field Robotic

    A Survey of Positioning Systems Using Visible LED Lights

    Get PDF
    © 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.As Global Positioning System (GPS) cannot provide satisfying performance in indoor environments, indoor positioning technology, which utilizes indoor wireless signals instead of GPS signals, has grown rapidly in recent years. Meanwhile, visible light communication (VLC) using light devices such as light emitting diodes (LEDs) has been deemed to be a promising candidate in the heterogeneous wireless networks that may collaborate with radio frequencies (RF) wireless networks. In particular, light-fidelity has a great potential for deployment in future indoor environments because of its high throughput and security advantages. This paper provides a comprehensive study of a novel positioning technology based on visible white LED lights, which has attracted much attention from both academia and industry. The essential characteristics and principles of this system are deeply discussed, and relevant positioning algorithms and designs are classified and elaborated. This paper undertakes a thorough investigation into current LED-based indoor positioning systems and compares their performance through many aspects, such as test environment, accuracy, and cost. It presents indoor hybrid positioning systems among VLC and other systems (e.g., inertial sensors and RF systems). We also review and classify outdoor VLC positioning applications for the first time. Finally, this paper surveys major advances as well as open issues, challenges, and future research directions in VLC positioning systems.Peer reviewe

    Distributed implementations of the particle filter with performance bounds

    Get PDF
    The focus of the thesis is on developing distributed estimation algorithms for systems with nonlinear dynamics. Of particular interest are the agent or sensor networks (AN/SN) consisting of a large number of local processing and observation agents/nodes, which can communicate and cooperate with each other to perform a predefined task. Examples of such AN/SNs are distributed camera networks, acoustic sensor networks, networks of unmanned aerial vehicles, social networks, and robotic networks. Signal processing in the AN/SNs is traditionally centralized and developed for systems with linear dynamics. In the centralized architecture, the participating nodes communicate their observations (either directly or indirectly via a multi-hop relay) to a central processing unit, referred to as the fusion centre, which is responsible for performing the predefined task. For centralized systems with linear dynamics, the Kalman filter provides the optimal approach but suffers from several drawbacks, e.g., it is generally unscalable and also susceptible to failure in case the fusion centre breaks down. In general, no analytic solution can be determined for systems with nonlinear dynamics. Consequently, the conventional Kalman filter cannot be used and one has to rely on numerical approaches. In such cases, the sequential Monte Carlo approaches, also known as the particle filters, are widely used as approximates to the Bayesian estimators but mostly in the centralized configuration. Recently there has been a growing interest in distributed signal processing algorithms where: (i) There is no fusion centre; (ii) The local nodes do not have (require) global knowledge of the network topology, and; (iii) Each node exchanges data only within its local neighborhood. Distributed estimation have been widely explored for estimation/tracking problems in linear systems. Distributed particle filter implementations for nonlinear systems are still in their infancy and are the focus of this thesis. In the first part of this thesis, four different consensus-based distributed particle filter implementations are proposed. First, a constrained sufficient statistic based distributed implementation of the particle filter (CSS/DPF) is proposed for bearing-only tracking (BOT) and joint bearing/range tracking problems encountered in a number of applications including radar target tracking and robot localization. Although the number of parallel consensus runs in the CSS/DPF is lower compared to the existing distributed implementations of the particle filter, the CSS/DPF still requires a large number of iterations for the consensus runs to converge. To further reduce the consensus overhead, the CSS/DPF is extended to distributed implementation of the unscented particle filter, referred to as the CSS/DUPF, which require a limited number of consensus iterations. Both CSS/DPF and CSS/DUPF are specific to BOT and joint bearing/range tracking problems. Next, the unscented, consensus-based, distributed implementation of the particle filter (UCD /DPF) is proposed which is generalizable to systems with any dynamics. In terms of contributions, the UCD /DPF makes two important improvements to the existing distributed particle filter framework: (i) Unlike existing distributed implementations of the particle filter, the UCD /DPF uses all available global observations including the most recent ones in deriving the proposal distribution based on the distributed UKF, and; (ii) Computation of the global estimates from local estimates during the consensus step is based on an optimal fusion rule. Finally, a multi-rate consensus/fusion based framework for distributed implementation of the particle filter, referred to as the CF /DPF, is proposed. Separate fusion filters are designed to consistently assimilate the local filtering distributions into the global posterior by compensating for the common past information between neighbouring nodes. The CF /DPF offers two distinct advantages over its counterparts. First, the CF /DPF framework is suitable for scenarios where network connectivity is intermittent and consensus can not be reached between two consecutive observations. Second, the CF /DPF is not limited to the Gaussian approximation for the global posterior density. Numerical simulations verify the near-optimal performance of the proposed distributed particle filter implementations. The second half of the thesis focuses on the distributed computation of the posterior Cramer-Rao lower bounds (PCRLB). The current PCRLB approaches assume a centralized or hierarchical architecture. The exact expression for distributed computation of the PCRLB is not yet available and only an approximate expression has recently been derived. Motivated by the distributed adaptive resource management problems with the objective of dynamically activating a time-variant subset of observation nodes to optimize the network's performance, the thesis derives the exact expression, referred to as the dPCRLB, for computing the PCRLB for any AN/SN configured in a distributed fashion. The dPCRLB computational algorithms are derived for both the off-line conventional (non-conditional) PCRLB determined primarily from the state model, observation model, and prior knowledge of the initial state of the system, and the online conditional PCRLB expressed as a function of past history of the observations. Compared to the non-conditional dPCRLB, its conditional counterpart provides a more accurate representation of the estimator's performance and, consequently, a better criteria for sensor selection. The thesis then extends the dPCRLB algorithms to quantized observations. Particle filter realizations are used to compute these bounds numerically and quantify their performance for data fusion problems through Monte-Carlo simulations

    Trajectory Poisson multi-Bernoulli mixture filter for traffic monitoring using a drone

    Full text link
    This paper proposes a multi-object tracking (MOT) algorithm for traffic monitoring using a drone equipped with optical and thermal cameras. Object detections on the images are obtained using a neural network for each type of camera. The cameras are modelled as direction-of-arrival (DOA) sensors. Each DOA detection follows a von-Mises Fisher distribution, whose mean direction is obtain by projecting a vehicle position on the ground to the camera. We then use the trajectory Poisson multi-Bernoulli mixture filter (TPMBM), which is a Bayesian MOT algorithm, to optimally estimate the set of vehicle trajectories. We have also developed a parameter estimation algorithm for the measurement model. We have tested the accuracy of the resulting TPMBM filter in synthetic and experimental data sets

    Multisensor-based human detection and tracking for mobile service robots

    Get PDF
    The one of fundamental issues for service robots is human-robot interaction. In order to perform such a task and provide the desired services, these robots need to detect and track people in the surroundings. In the present paper, we propose a solution for human tracking with a mobile robot that implements multisensor data fusion techniques. The system utilizes a new algorithm for laser-based legs detection using the on-board LRF. The approach is based on the recognition of typical leg patterns extracted from laser scans, which are shown to be very discriminative also in cluttered environments. These patterns can be used to localize both static and walking persons, even when the robot moves. Furthermore, faces are detected using the robot's camera and the information is fused to the legs position using a sequential implementation of Unscented Kalman Filter. The proposed solution is feasible for service robots with a similar device configuration and has been successfully implemented on two different mobile platforms. Several experiments illustrate the effectiveness of our approach, showing that robust human tracking can be performed within complex indoor environments

    Multi-object tracking using sensor fusion

    Get PDF

    Vision Science and Technology at NASA: Results of a Workshop

    Get PDF
    A broad review is given of vision science and technology within NASA. The subject is defined and its applications in both NASA and the nation at large are noted. A survey of current NASA efforts is given, noting strengths and weaknesses of the NASA program

    Data fusion in ubiquitous networked robot systems for urban services

    Get PDF
    There is a clear trend in the use of robots to accomplish services that can help humans. In this paper, robots acting in urban environments are considered for the task of person guiding. Nowadays, it is common to have ubiquitous sensors integrated within the buildings, such as camera networks, and wireless communications like 3G or WiFi. Such infrastructure can be directly used by robotic platforms. The paper shows how combining the information from the robots and the sensors allows tracking failures to be overcome, by being more robust under occlusion, clutter, and lighting changes. The paper describes the algorithms for tracking with a set of fixed surveillance cameras and the algorithms for position tracking using the signal strength received by a wireless sensor network (WSN). Moreover, an algorithm to obtain estimations on the positions of people from cameras on board robots is described. The estimate from all these sources are then combined using a decentralized data fusion algorithm to provide an increase in performance. This scheme is scalable and can handle communication latencies and failures. We present results of the system operating in real time on a large outdoor environment, including 22 nonoverlapping cameras, WSN, and several robots.Universidad Pablo de Olavide. Departamento de Deporte e InformáticaPostprin

    Exploring space situational awareness using neuromorphic event-based cameras

    Get PDF
    The orbits around earth are a limited natural resource and one that hosts a vast range of vital space-based systems that support international systems use by both commercial industries, civil organisations, and national defence. The availability of this space resource is rapidly depleting due to the ever-growing presence of space debris and rampant overcrowding, especially in the limited and highly desirable slots in geosynchronous orbit. The field of Space Situational Awareness encompasses tasks aimed at mitigating these hazards to on-orbit systems through the monitoring of satellite traffic. Essential to this task is the collection of accurate and timely observation data. This thesis explores the use of a novel sensor paradigm to optically collect and process sensor data to enhance and improve space situational awareness tasks. Solving this issue is critical to ensure that we can continue to utilise the space environment in a sustainable way. However, these tasks pose significant engineering challenges that involve the detection and characterisation of faint, highly distant, and high-speed targets. Recent advances in neuromorphic engineering have led to the availability of high-quality neuromorphic event-based cameras that provide a promising alternative to the conventional cameras used in space imaging. These cameras offer the potential to improve the capabilities of existing space tracking systems and have been shown to detect and track satellites or ‘Resident Space Objects’ at low data rates, high temporal resolutions, and in conditions typically unsuitable for conventional optical cameras. This thesis presents a thorough exploration of neuromorphic event-based cameras for space situational awareness tasks and establishes a rigorous foundation for event-based space imaging. The work conducted in this project demonstrates how to enable event-based space imaging systems that serve the goals of space situational awareness by providing accurate and timely information on the space domain. By developing and implementing event-based processing techniques, the asynchronous operation, high temporal resolution, and dynamic range of these novel sensors are leveraged to provide low latency target acquisition and rapid reaction to challenging satellite tracking scenarios. The algorithms and experiments developed in this thesis successfully study the properties and trade-offs of event-based space imaging and provide comparisons with traditional observing methods and conventional frame-based sensors. The outcomes of this thesis demonstrate the viability of event-based cameras for use in tracking and space imaging tasks and therefore contribute to the growing efforts of the international space situational awareness community and the development of the event-based technology in astronomy and space science applications
    corecore