178 research outputs found

    Building and Navigating Maps Of Road Scenes Using An Active Sensor

    No full text
    The author presents algorithms for building maps of road scenes using an active range and reflectance sensor and for using the maps to traverse a portion of the world already explored. He describes some advantages of an active sensor, namely, that it is independent of the illumination conditions, does not require complex calibration in order to transform observed features to the vehicle's reflectance frame, and provides 3-D terrain models as well as road models. Using this map built from sensor data facilitates navigation in two respects: the vehicle may navigate faster, since less perception processing is necessary, and the vehicle may follow a more accurate path, since the navigation system does not rely entirely on inaccurate visual data. The author presents a complete system that includes road-following, map-building, and map-based navigation using the ERIM laser rangefinder. Experimental results are presente

    Active and Passive Range Sensing for Robotics

    No full text
    In this paper, we present a brief survey of the technologies currently available for range sensing, of their use in robotics applications, and of emerging technologies for future systems. The paper is organized by type of sensing: laser range finders, triangulation range finders, and passive stereo. A separate section focuses on current development in the area of nonscanning sensors, a critical area to achieve range sensing performance comparable to that of conventional cameras. The presentation of the different technologies is based on many recent examples from robotics research

    Omni-Directional Structure From Motion

    No full text
    In this paper, we describe our work on a robot navigation system using an omnidirectional camera as the primary sensor. An omnidirectional structure from motion algorithm is presented with its uncertainty analyzed. We evaluate the omnidirectional SFM on both synthetic data and on real image sequences. Comparison with the conventional camera is made and we show that in certain situations the omnidirectional SFM gives better results than the conventional one by taking advantage of its larger field of view

    Incorporating Background Invariance into Feature-Based Object Recognition

    No full text
    Current feature-based object recognition methods use information derived from local image patches. For robustness, features are engineered for invariance to various transformations, such as rotation, scaling, or affine warping. When patches overlap object boundaries, however, errors in both detection and matching will almost certainly occur due to inclusion of unwanted background pixels. This is common in real images, which often contain significant background clutter, objects which are not heavily textured, or objects which occupy a relatively small portion of the image. We suggest improvements to the popular Scale Invariant Feature Transform (SIFT) which incorporate local object boundary information. The resulting feature detection and descriptor creation processes are invariant to changes in background.We call this method the Background and Scale Invariant Feature Transform (BSIFT).We demonstrate BSIFT’s superior performance in feature detection and matching on synthetic and natural imag

    A Hierarchical Field Framework for Unified Context-Based Classification

    No full text
    We present a two-layer hierarchical formulation to exploit different levels of contextual information in images for robust classification. Each layer is modeled as a conditional field that allows one to capture arbitrary observationdependent label interactions. The proposed framework has two main advantages. First, it encodes both the short-range interactions (e.g., pixelwise label smoothing) as well as the long-range interactions (e.g., relative configurations of objects or regions) in a tractable manner. Second, the formulation is general enough to be applied to different domains ranging from pixelwise image labeling to contextual object detection. The parameters of the model are learned using a sequential maximum-likelihood approximation. The benefits of the proposed framework are demonstrated on four different datasets and comparison results are presented

    Pairwise Grouping Using Color

    No full text
    Grouping was recognized in computer vision early on as having the potential of improving both matching and recognition. Most papers consider grouping as a segmentation problem and a hard decision is made about which pixels in the image belong to the same object. In this paper we instead focus on soft pairwise grouping, that is computing affinities between pairs of pixels that reflect how likely that pair is to belong to the same object. This fits perfectly with our recognition approach, where we consider pairwise relationships between features/pixels. Some other papers also considered soft pairwise grouping between features, but they focused more on geometry than appearance. In this paper we take a different approach and show how color could also be used for pairwise grouping. We present a simple but effective method to group pixels based on color statistics. By using only color information and no prior higher level knowledge about objects and scenes we develop an efficient classifier that can separate the pixels that belong to the same object from those that do not. In the context of segmentation where color is also used only nearby pixels are generally considered, and very simple color information is taken into account. We use global color information instead and develop an efficient algorithm that can successfully classify even pairs of pixels that are far apar

    Multi-Scale Classification of 3D Objects

    No full text
    We describe an approach to the classification of 3-D objects using a multi-scale representation. This approach starts with a smoothing algorithm for representing objects at different scales. In a way similar to the classical scale space representations, larger amount of smoothing removes more details from the surfaces. Smoothing is applied in curvature space directly, thus avoiding the usual shrinkage problems and allowing for efficient implementations. A 3-D similarity measure that integrates the representations of the objects at multiple scales is introduced. This similarity measure is designed to give higher weight to the coarse scale representations, while ignoring the finer scale details of the surfaces. Given a library of models, objects that are similar based on this multi-scale measure are grouped together into classes. We show how shapes in a given class can be combined into a single prototype object. This is achieved by using a powerful property, introduced earlier, of inverse mapping from representation to shape. Finally, the prototypes are used for hierarchical recognition by first comparing the scene representation to the prototypes and then matching it only to the objects in the most likely class rather than to the entire library of models. Beyond its application to object recognition, this approach provides an attractive implementation of the intuitive notions of scale and approximate similarity for 3-D shapes

    Experimental Comparison of Techniques for Localization and Mapping Using A Bearing-Only Sensor

    No full text
    We present a comparison of an extended Kalman filter and an adaptation of bundle adjustment from computer vision for mobile robot localization and mapping using a bearing-only sensor. We show results on synthetic and real examples and discuss some advantages and disadvantages of the techniques. The comparison leads to a novel combination of the two techniques which results in computational complexity near Kalman filters and performance near bundle adjustment on the examples shown

    Evaluation of Image-Based Landmark Recognition Techniques

    No full text
    Recognizing landmarks in sequences of images is a challenging problem for a number of reasons. First of all, the appearance of any given landmark varies substantially from one observation to the next. In addition to variations due to different aspects, an illumination change, external clutter, and changing geometry of the imaging devices are other factors affecting the variability of the observed landmarks. Finally, it is typically difficult to make use of accurate 3D information in landmark recognition applications. For those reasons, it is not possible to use many of the object recognition techniques based on strong geometric models. The alternative is to use image-based techniques in which landmarks are represented by collections of images which capture the “typical” appearance of the object. The information most relevant to recognition is extracted from the collection of raw images and used as the model for recognition. This process is often referred to as “visual learning.” Models of landmarks are acquired from image sequences and later recognized for vehicle localization in urban environments. In the acquisition phase, a vehicle drives and collects images of an unknown area. The algorithm organizes these images into groups with similar image features. The feature distribution for each group describes a landmark. In the recognition phase, while navigating through the same general area, the vehicle collects new images. The algorithm classifies these images into one of the learned groups, thus recognizing a landmark. Unlike computationally intensive model-based approaches that build models from known objects observed in isolation, our image-based approach automatically learns the most salient landmarks in complex environments. It delivers a robust performance under a wide range of lighting and imaging angle variations

    Where and When to Look: How to Extend the Myopic Planning Horizon

    No full text
    In this paper we describe an approach towards integrating mid-range sensing data into a dynamic path planning algorithm. The key problem, sensing for planning is addressed in the context of outdoor navigation. An algorithmic approach is described towards solving these problems and both simulation results and initial experimental results for outdoor navigation using wide baseline stereo data are presented
    • …
    corecore