9 research outputs found

    Information Acquisition with Sensing Robots: Algorithms and Error Bounds

    Full text link
    Utilizing the capabilities of configurable sensing systems requires addressing difficult information gathering problems. Near-optimal approaches exist for sensing systems without internal states. However, when it comes to optimizing the trajectories of mobile sensors the solutions are often greedy and rarely provide performance guarantees. Notably, under linear Gaussian assumptions, the problem becomes deterministic and can be solved off-line. Approaches based on submodularity have been applied by ignoring the sensor dynamics and greedily selecting informative locations in the environment. This paper presents a non-greedy algorithm with suboptimality guarantees, which does not rely on submodularity and takes the sensor dynamics into account. Our method performs provably better than the widely used greedy one. Coupled with linearization and model predictive control, it can be used to generate adaptive policies for mobile sensors with non-linear sensing models. Applications in gas concentration mapping and target tracking are presented.Comment: 9 pages (two-column); 2 figures; Manuscript submitted to the 2014 IEEE International Conference on Robotics and Automatio

    Localization from semantic observations via the matrix permanent

    Get PDF
    Most approaches to robot localization rely on low-level geometric features such as points, lines, and planes. In this paper, we use object recognition to obtain semantic information from the robot’s sensors and consider the task of localizing the robot within a prior map of landmarks, which are annotated with semantic labels. As object recognition algorithms miss detections and produce false alarms, correct data association between the detections and the landmarks on the map is central to the semantic localization problem. Instead of the traditional vector-based representation, we propose a sensor model, which encodes the semantic observations via random finite sets and enables a unified treatment of missed detections, false alarms, and data association. Our second contribution is to reduce the problem of computing the likelihood of a set-valued observation to the problem of computing a matrix permanent. It is this crucial transformation that allows us to solve the semantic localization problem with a polynomial-time approximation to the set-based Bayes filter. Finally, we address the active semantic localization problem, in which the observer’s trajectory is planned in order to improve the accuracy and efficiency of the localization process. The performance of our approach is demonstrated in simulation and in real environments using deformable-part-model-based object detectors. Robust global localization from semantic observations is demonstrated for a mobile robot, for the Project Tango phone, and on the KITTI visual odometry dataset. Comparisons are made with the traditional lidar-based geometric Monte Carlo localization

    Recognizing Fine-Grained Object Instances for Robotics Applications

    Get PDF
    State-of-the-art object recognition systems, and computer vision methods in general, are getting better and better at a variety of tasks. Many of these techniques focus on general object categories, like person or bottle, and use many hundreds or thousands of training examples per category. Some domains, such as robotics, require recognition of fine-grained object instances, such as a 16oz. bottle of Coke, and only have access to limited training data. In this work we look at how to build data and adapt successful computer vision techniques for fine-grained object recognition tasks. Our Active Vision Dataset is the first deep-learning scale dataset for instance detection and motion simulation using real images. Our Target Driven Instance Detection method uses a siamese convolutional neural network to focus on specific fine-grained object instances, and has the capacity to generalize to instances outside its training set. Our SymGan method used the Generative Adversarial Network paradigm to train a 3D object orientation predictor that is robust to object symmetries.Doctor of Philosoph

    Controlled Recognition Bounds for Visual Learning and Exploration

    No full text
    We describe the tradeoff between the performance in a visual recognition problem and the control authority that the agent can exercise on the sensing process. We focus on the problem of “visual search ” of an object in an otherwise known and static scene, propose a measure of control authority, and relate it to the expected risk and its proxy (conditional entropy of the posterior density). We show this analytically, as well as empirically by simulation using the simplest known model that captures the phenomenology of image formation, including scaling and occlusions. We show that a “passive ” agent given a training set can provide no guarantees on performance beyond what is afforded by the priors, and that an “omnipotent ” agent, capable of infinite control authority, can achieve arbitrarily good performance (asymptotically). In between these limiting cases, the tradeoff can be characterized empirically.
    corecore