37 research outputs found

    Visual identification by signature tracking

    Get PDF
    We propose a new camera-based biometric: visual signature identification. We discuss the importance of the parameterization of the signatures in order to achieve good classification results, independently of variations in the position of the camera with respect to the writing surface. We show that affine arc-length parameterization performs better than conventional time and Euclidean arc-length ones. We find that the system verification performance is better than 4 percent error on skilled forgeries and 1 percent error on random forgeries, and that its recognition performance is better than 1 percent error rate, comparable to the best camera-based biometrics

    Automatic discovery of image families: Global vs. local features

    Get PDF
    Gathering a large collection of images has been made quite easy by social and image sharing websites, e.g. flickr.com. However, using such collections faces the problem that they contain a large number of duplicates and highly similar images. This work tackles the problem of how to automatically organize image collections into sets of similar images, called image families hereinafter. We thoroughly compare the performance of two approaches to measure image similarity: global descriptors vs. a set of local descriptors. We assess the performance of these approaches as the problem scales up to thousands of images and hundreds of families. We present our results on a new dataset of CD/DVD game covers

    Continuous dynamic time warping for translation-invariant curve alignment with applications to signature verification

    Get PDF
    The problem of establishing correspondence and measuring the similarity of a pair of planar curves arises in many applications in computer vision and pattern recognition. This paper presents a new method for comparing planar curves and for performing matching at sub-sampling resolution. The analysis of the algorithm as well as its structural properties are described. The performance of the new technique applied to the problem of signature verification is shown and compared with the performance of the well-known Dynamic Time Warping algorithm

    Visual input for pen-based computers

    Get PDF
    The design and implementation of a camera-based, human-computer interface for acquisition of handwriting is presented. The camera focuses on a standard sheet of paper and images a common pen; the trajectory of the tip of the pen is tracked and the contact with the paper is detected. The recovered trajectory is shown to have sufficient spatio-temporal resolution and accuracy to enable handwritten character recognition. More than 100 subjects have used the system and have provided a large and heterogeneous set of examples showing that the system is both convenient and accurate

    Visual signature verification using affine arc-length

    Get PDF
    Signatures can be acquired with a camera-based system with enough resolution to perform verification. This paper presents the performance of a visual-acquisition signature verification system, emphasizing on the importance of the parameterisation of the signature in order to achieve good classification results. A technique to overcome the lack of examples in order to estimate the generalization error of the algorithm is also described

    Distributed Kd-Trees for Ultra Large Scale Object Recognition

    Get PDF
    Distributed Kd-Trees is a method for building image retrieval systems that can handle hundreds of millions of images. It is based on dividing the Kd-Tree into a “root subtree” that resides on a root machine, and several “leaf subtrees”, each residing on a leaf machine. The root machine handles incoming queries and farms out feature matching to an appropriate small subset of the leaf machines. Our implementation employs the MapReduce architecture to efficiently build and distribute the Kd-Tree for millions of images. It can run on thousands of machines, and provides orders of magnitude more throughput than the state-of-the-art, with better recognition performance. We show experiments with up to 100 million images running on 2048 machines, with run time of a fraction of a second for each query image

    A Constant-Time Algorithm for Vector Field SLAM using an Exactly Sparse Extended Information Filter

    Full text link
    Abstract — Designing a localization system for a low-cost robotic consumer product poses a major challenge. In previous work, we introduced Vector Field SLAM [5], a system for simultaneously estimating robot pose and a vector field induced by stationary signal sources present in the environment. In this paper we show how this method can be realized on a low-cost embedded processing unit by applying the concepts of the Exactly Sparse Extended Information Filter [15]. By restricting the set of active features to the 4 nodes of the current cell, the size of the map becomes linear in the area explored by the robot while the time for updating the state can be held constant under certain approximations. We report results from running our method on an ARM 7 embedded board with 64 kByte RAM controlling a Roomba 510 vacuum cleaner in a standard test environment. NS spot1 X (sensor units) Spot1 X readings Node X1 estimate

    Distributed Kd-Trees for Ultra Large Scale Object Recognition

    Get PDF
    Distributed Kd-Trees is a method for building image retrieval systems that can handle hundreds of millions of images. It is based on dividing the Kd-Tree into a “root subtree” that resides on a root machine, and several “leaf subtrees”, each residing on a leaf machine. The root machine handles incoming queries and farms out feature matching to an appropriate small subset of the leaf machines. Our implementation employs the MapReduce architecture to efficiently build and distribute the Kd-Tree for millions of images. It can run on thousands of machines, and provides orders of magnitude more throughput than the state-of-the-art, with better recognition performance. We show experiments with up to 100 million images running on 2048 machines, with run time of a fraction of a second for each query image

    Monocular graph SLAM with complexity reduction

    Get PDF
    Abstract-We present a graph-based SLAM approach, using monocular vision and odometry, designed to operate on computationally constrained platforms. When computation and memory are limited, visual tracking becomes difficult or impossible, and map representation and update costs must remain low. Our system constructs a map of structured views using only weak temporal assumptions, and performs recognition and relative pose estimation over the set of views. Visual observations are fused with differential sensors in an incrementally optimized graph representation. Using variable elimination and constraint pruning, the graph complexity and storage is kept linear in explored space rather than in time. We evaluate performance on sequences with ground truth, and also compare to a standard graph SLAM approach

    View management for lifelong visual maps

    Full text link
    The time complexity of making observations and loop closures in a graph-based visual SLAM system is a function of the number of views stored. Clever algorithms, such as approximate nearest neighbor search, can make this function sub-linear. Despite this, over time the number of views can still grow to a point at which the speed and/or accuracy of the system becomes unacceptable, especially in computation- and memory-constrained SLAM systems. However, not all views are created equal. Some views are rarely observed, because they have been created in an unusual lighting condition, or from low quality images, or in a location whose appearance has changed. These views can be removed to improve the overall performance of a SLAM system. In this paper, we propose a method for pruning views in a visual SLAM system to maintain its speed and accuracy for long term use.Comment: IEEE International Conference on Intelligent Robots and Systems (IROS), 201
    corecore