175 research outputs found
Interest point detectors for visual SLAM
In this paper we present several interest points detectors and we analyze their suitability when used as landmark extractors for vision-based simultaneous localization and mapping (vSLAM). For this purpose, we evaluate the detectors according to their repeatability under changes in viewpoint and scale. These are the desired requirements for visual landmarks. Several experiments were carried out using sequence of images captured with high precision. The sequences represent planar objects as well as 3D scenes
MapSnapper: Engineering an Efficient Algorithm for Matching Images of Maps from Mobile Phones
The MapSnapper project aimed to develop a system for robust matching of low-quality images of a paper map taken from a mobile phone against a high quality digital raster representation of the same map. The paper presents a novel methodology for performing content-based image retrieval and object recognition from query images that have been degraded by noise and subjected to transformations through the imaging system. In addition the paper also provides an insight into the evaluation-driven development process that was used to incrementally improve the matching performance until the design specifications were met
Video Acceleration Magnification
The ability to amplify or reduce subtle image changes over time is useful in
contexts such as video editing, medical video analysis, product quality control
and sports. In these contexts there is often large motion present which
severely distorts current video amplification methods that magnify change
linearly. In this work we propose a method to cope with large motions while
still magnifying small changes. We make the following two observations: i)
large motions are linear on the temporal scale of the small changes; ii) small
changes deviate from this linearity. We ignore linear motion and propose to
magnify acceleration. Our method is pure Eulerian and does not require any
optical flow, temporal alignment or region annotations. We link temporal
second-order derivative filtering to spatial acceleration magnification. We
apply our method to moving objects where we show motion magnification and color
magnification. We provide quantitative as well as qualitative evidence for our
method while comparing to the state-of-the-art.Comment: Accepted paper at CVPR 2017. Project webpage:
http://acceleration-magnification.github.io
Large scale evaluation of local image feature detectors on homography datasets
We present a large scale benchmark for the evaluation of local feature
detectors. Our key innovation is the introduction of a new evaluation protocol
which extends and improves the standard detection repeatability measure. The
new protocol is better for assessment on a large number of images and reduces
the dependency of the results on unwanted distractors such as the number of
detected features and the feature magnification factor. Additionally, our
protocol provides a comprehensive assessment of the expected performance of
detectors under several practical scenarios. Using images from the
recently-introduced HPatches dataset, we evaluate a range of state-of-the-art
local feature detectors on two main tasks: viewpoint and illumination invariant
detection. Contrary to previous detector evaluations, our study contains an
order of magnitude more image sequences, resulting in a quantitative evaluation
significantly more robust to over-fitting. We also show that traditional
detectors are still very competitive when compared to recent deep-learning
alternatives.Comment: Accepted to BMVC 201
Learning Object Categories From Internet Image Searches
In this paper, we describe a simple approach to learning models of visual object categories from images gathered from Internet image search engines. The images for a given keyword are typically highly variable, with a large fraction being unrelated to the query term, and thus pose a challenging environment from which to learn. By training our models directly from Internet images, we remove the need to laboriously compile training data sets, required by most other recognition approaches-this opens up the possibility of learning object category models “on-the-fly.” We describe two simple approaches, derived from the probabilistic latent semantic analysis (pLSA) technique for text document analysis, that can be used to automatically learn object models from these data. We show two applications of the learned model: first, to rerank the images returned by the search engine, thus improving the quality of the search engine; and second, to recognize objects in other image data sets
Neuroanatomic-based detection algorithm for automatic labeling of brain structures in brain injury
The number and grade of injured neuroanatomic structures and the type of injury determine the degree of impairment after a brain injury event and the recovery options of the patient. However, the body of knowledge and clinical intervention guides are basically focused on functional disorder and they still do not take into account the location of injuries. The prognostic value of location information is not known in detail either. This paper proposes a feature-based detection algorithm, named Neuroanatomic-Based Detection Algorithm (NBDA), based on SURF (Speeded Up Robust Feature) to label anatomical brain structures on cortical and sub-cortical areas. Themain goal is to register injured neuroanatomic structures to generate a database containing patient?s structural impairment profile. This kind of information permits to establish a relation with functional disorders and the prognostic evolution during neurorehabilitation procedures
Recommended from our members
An On-Board Visual-Based Attitude Estimation System For Unmanned Aerial Vehicle Mapping
This paper evaluates the performances of several salient feature detectors, namely; Harris detector, Minimum Eigenvalue (MinEig), Scale Invariant Feature Transform (SIFT), Maximally Stable Extremal Region (MSER), Speeded Up Robust Feature (SURF), Features from Accelerated Segment Test (FAST), and Binary Robust Scale Invariant Keypoint (BRISK), in order to assess the suitability in the application of the proposed visual-based attitude estimation system. Throughout the experiment, three main requirements have been investigated which include Time-to-Complete (TTC), detection rate, and matching rate. It was found that SURF fulfills each of the system’s requirements. Moreover, it was also found that keypoints detection capabilities affect the processing time, and the clustering patterns in the results may assist in automated inspection of correct and false matching
- …