2,250 research outputs found
Scale and Orientation-invariant Scene Similarity Metrics for Image Queries
In this paper we extend our previous work on shape-based queries to support queries on configurations of image objects. Here we consider spatial reasoning, especially directional and metric object relationships. Existing models for spatial reasoning tend to rely on pre-identified cardinal directions and minimal scale variations, assumption that cannot be considered as given in our image applications, where orientations and scale may vary substantially, and are often unknown. Accordingly, we have developed the method of varying baselines to identify similarities in direction and distance relations. Our method allows us to evaluate directional similarities without a priori knowledge of cardinal directions, and to compare distance relations even when query scene and database content differ in scale by unknown amounts. We use our method to evaluate similarity between a user-defined query scene and object configurations. Here we present this new method, and discuss its role within a broader image retrieval framework
Review of Person Re-identification Techniques
Person re-identification across different surveillance cameras with disjoint
fields of view has become one of the most interesting and challenging subjects
in the area of intelligent video surveillance. Although several methods have
been developed and proposed, certain limitations and unresolved issues remain.
In all of the existing re-identification approaches, feature vectors are
extracted from segmented still images or video frames. Different similarity or
dissimilarity measures have been applied to these vectors. Some methods have
used simple constant metrics, whereas others have utilised models to obtain
optimised metrics. Some have created models based on local colour or texture
information, and others have built models based on the gait of people. In
general, the main objective of all these approaches is to achieve a
higher-accuracy rate and lowercomputational costs. This study summarises
several developments in recent literature and discusses the various available
methods used in person re-identification. Specifically, their advantages and
disadvantages are mentioned and compared.Comment: Published 201
Defining the Pose of any 3D Rigid Object and an Associated Distance
The pose of a rigid object is usually regarded as a rigid transformation,
described by a translation and a rotation. However, equating the pose space
with the space of rigid transformations is in general abusive, as it does not
account for objects with proper symmetries -- which are common among man-made
objects.In this article, we define pose as a distinguishable static state of an
object, and equate a pose with a set of rigid transformations. Based solely on
geometric considerations, we propose a frame-invariant metric on the space of
possible poses, valid for any physical rigid object, and requiring no arbitrary
tuning. This distance can be evaluated efficiently using a representation of
poses within an Euclidean space of at most 12 dimensions depending on the
object's symmetries. This makes it possible to efficiently perform neighborhood
queries such as radius searches or k-nearest neighbor searches within a large
set of poses using off-the-shelf methods. Pose averaging considering this
metric can similarly be performed easily, using a projection function from the
Euclidean space onto the pose space. The practical value of those theoretical
developments is illustrated with an application of pose estimation of instances
of a 3D rigid object given an input depth map, via a Mean Shift procedure
Video Registration in Egocentric Vision under Day and Night Illumination Changes
With the spread of wearable devices and head mounted cameras, a wide range of
application requiring precise user localization is now possible. In this paper
we propose to treat the problem of obtaining the user position with respect to
a known environment as a video registration problem. Video registration, i.e.
the task of aligning an input video sequence to a pre-built 3D model, relies on
a matching process of local keypoints extracted on the query sequence to a 3D
point cloud. The overall registration performance is strictly tied to the
actual quality of this 2D-3D matching, and can degrade if environmental
conditions such as steep changes in lighting like the ones between day and
night occur. To effectively register an egocentric video sequence under these
conditions, we propose to tackle the source of the problem: the matching
process. To overcome the shortcomings of standard matching techniques, we
introduce a novel embedding space that allows us to obtain robust matches by
jointly taking into account local descriptors, their spatial arrangement and
their temporal robustness. The proposal is evaluated using unconstrained
egocentric video sequences both in terms of matching quality and resulting
registration performance using different 3D models of historical landmarks. The
results show that the proposed method can outperform state of the art
registration algorithms, in particular when dealing with the challenges of
night and day sequences
- …