31 research outputs found

    The brightness clustering transform and locally contrasting keypoints

    No full text
    In recent years a new wave of feature descriptors has been presented to the computer vision community, ORB, BRISK and FREAK amongst others. These new descriptors allow reduced time and memory consumption on the processing and storage stages of tasks such as image matching or visual odometry, enabling real time applications. The problem is now the lack of fast interest point detectors with good repeatability to use with these new descriptors. We present a new blob- detector which can be implemented in real time and is faster than most of the currently used feature-detectors. The detection is achieved with an innovative non-deterministic low-level operator called the Brightness Clustering Transform (BCT). The BCT can be thought as a coarse-to- fine search through scale spaces for the true derivative of the image; it also mimics trans-saccadic perception of human vision. We call the new algorithm Locally Contrasting Keypoints detector or LOCKY. Showing good repeatability and robustness to image transformations included in the Oxford dataset, LOCKY is amongst the fastest affine-covariant feature detectors

    GeoDesc: Learning Local Descriptors by Integrating Geometry Constraints

    Full text link
    Learned local descriptors based on Convolutional Neural Networks (CNNs) have achieved significant improvements on patch-based benchmarks, whereas not having demonstrated strong generalization ability on recent benchmarks of image-based 3D reconstruction. In this paper, we mitigate this limitation by proposing a novel local descriptor learning approach that integrates geometry constraints from multi-view reconstructions, which benefits the learning process in terms of data generation, data sampling and loss computation. We refer to the proposed descriptor as GeoDesc, and demonstrate its superior performance on various large-scale benchmarks, and in particular show its great success on challenging reconstruction tasks. Moreover, we provide guidelines towards practical integration of learned descriptors in Structure-from-Motion (SfM) pipelines, showing the good trade-off that GeoDesc delivers to 3D reconstruction tasks between accuracy and efficiency.Comment: Accepted to ECCV'1

    Progressive Structure from Motion

    Full text link
    Structure from Motion or the sparse 3D reconstruction out of individual photos is a long studied topic in computer vision. Yet none of the existing reconstruction pipelines fully addresses a progressive scenario where images are only getting available during the reconstruction process and intermediate results are delivered to the user. Incremental pipelines are capable of growing a 3D model but often get stuck in local minima due to wrong (binding) decisions taken based on incomplete information. Global pipelines on the other hand need the access to the complete viewgraph and are not capable of delivering intermediate results. In this paper we propose a new reconstruction pipeline working in a progressive manner rather than in a batch processing scheme. The pipeline is able to recover from failed reconstructions in early stages, avoids to take binding decisions, delivers a progressive output and yet maintains the capabilities of existing pipelines. We demonstrate and evaluate our method on diverse challenging public and dedicated datasets including those with highly symmetric structures and compare to the state of the art.Comment: Accepted to ECCV 201

    Unbiased evaluation of keypoint detectors with respect to rotation invariance

    No full text
    The authors present the results of a comparative performance study of algorithms for detecting keypoints in digital images. The Harris, good features to track (GFTT), SIFT, SURF, FAST, ORB, BRISK, and the MSER keypoint detectors were tested using two types of images: POV‐Ray simulated images and photographs from the Caltech 256 image dataset. They tested the repeatability of detection of the image keypoints for the evaluated detectors for a series of images with one degree rotations from 0 to 180° (3982 images in total). In the evaluation scenario they adopted an original approach in which they did not hold back a single image to be the reference image. They conclude that the most computationally complex detector, i.e. the SIFT performs best under rotation transformation of images. However, the FAST and ORB detectors, while being less computationally demanding, perform almost equally well. Hence, they can be viable choices in image processing tasks for mobile applications

    Semi-independent Stereo Visual Odometry for Different Field of View Cameras

    No full text
    International audienceThis paper presents a pipeline for stereo visual odometry using cameras with different fields of view. It gives a proof of concept about how a constraint on the respective field of view of each camera can lead to both an accurate 3D reconstruction and a robust pose estimation. Indeed, when considering a fixed resolution, a narrow field of view has a higher angular resolution and can preserve image texture details. On the other hand, a wide field of view allows to track features over longer periods since the overlap between two successive frames is more substantial. We propose a semi-independent stereo system where each camera performs individually temporal multi-view optimization but their initial parameters are still jointly optimized in an iterative framework. Furthermore , the concept of lead and follow camera is introduced to adaptively propagate information between the cameras. We evaluate the method qualitatively on two indoor datasets, and quantitatively on a synthetic dataset to allow the comparison across different fields of view

    CD7 and CD28 Are Required for Murine CD4 +

    No full text

    Extremely Fast Unsupervised Codebook Learning for Landmark Recognition

    No full text
    corecore