22,931 research outputs found
Link anchors in images: is there truth?
While automatic linking in text collections is well understood, little is known about links in images. In this work, we investigate two aspects of anchors, the origin of a link, in images: 1) the requirements of users for such anchors, e.g. the things users would like more information on, and 2) possible evaluation methods assessing anchor selection al- gorithms. To investigate these aspects, we perform a study with 102 users. We find that 59% of the required anchors are image segments, as opposed to the whole image, and most users require information on displayed persons. The agreement of users on the required anchors is too low (often below 30%) for a ground truth-based evaluation, which is the standard IR evaluation method. As an alternative, we propose a novel evaluation method based on improved search performance and user experience
Mapping, Localization and Path Planning for Image-based Navigation using Visual Features and Map
Building on progress in feature representations for image retrieval,
image-based localization has seen a surge of research interest. Image-based
localization has the advantage of being inexpensive and efficient, often
avoiding the use of 3D metric maps altogether. That said, the need to maintain
a large number of reference images as an effective support of localization in a
scene, nonetheless calls for them to be organized in a map structure of some
kind.
The problem of localization often arises as part of a navigation process. We
are, therefore, interested in summarizing the reference images as a set of
landmarks, which meet the requirements for image-based navigation. A
contribution of this paper is to formulate such a set of requirements for the
two sub-tasks involved: map construction and self-localization. These
requirements are then exploited for compact map representation and accurate
self-localization, using the framework of a network flow problem. During this
process, we formulate the map construction and self-localization problems as
convex quadratic and second-order cone programs, respectively. We evaluate our
methods on publicly available indoor and outdoor datasets, where they
outperform existing methods significantly.Comment: CVPR 2019, for implementation see https://github.com/janinethom
Embedding based on function approximation for large scale image search
The objective of this paper is to design an embedding method that maps local
features describing an image (e.g. SIFT) to a higher dimensional representation
useful for the image retrieval problem. First, motivated by the relationship
between the linear approximation of a nonlinear function in high dimensional
space and the stateof-the-art feature representation used in image retrieval,
i.e., VLAD, we propose a new approach for the approximation. The embedded
vectors resulted by the function approximation process are then aggregated to
form a single representation for image retrieval. Second, in order to make the
proposed embedding method applicable to large scale problem, we further derive
its fast version in which the embedded vectors can be efficiently computed,
i.e., in the closed-form. We compare the proposed embedding methods with the
state of the art in the context of image search under various settings: when
the images are represented by medium length vectors, short vectors, or binary
vectors. The experimental results show that the proposed embedding methods
outperform existing the state of the art on the standard public image retrieval
benchmarks.Comment: Accepted to TPAMI 2017. The implementation and precomputed features
of the proposed F-FAemb are released at the following link:
http://tinyurl.com/F-FAem
- …