998 research outputs found
IMAGE RETRIEVAL BASED ON COMPLEX DESCRIPTIVE QUERIES
The amount of visual data such as images and videos available over web has increased exponentially over the last few years. In order to efficiently organize and exploit these massive collections, a system, apart from being able to answer simple classification based questions such as whether a specific object is present (or absent) in an image, should also be capable of searching images and videos based on more complex descriptive questions. There is also a considerable amount of structure present in the visual world which, if effectively utilized, can help achieve this goal. To this end, we first present an approach for image ranking and retrieval based on queries consisting of multiple semantic attributes. We further show that there are significant correlations present between these attributes and accounting for them can lead to superior performance. Next, we extend this by proposing an image retrieval framework for descriptive queries composed of object categories, semantic attributes and spatial relationships. The proposed framework also includes a unique multi-view hashing technique, which enables query specification in three different modalities - image, sketch and text.
We also demonstrate the effectiveness of leveraging contextual information to reduce the supervision requirements for learning object and scene recognition models. We present an active learning framework to simultaneously learn appearance and contextual models for scene understanding. Within this framework we introduce new kinds of labeling questions that are designed to collect appearance as well as contextual information and which mimic the way in which humans actively learn about their environment. Furthermore we explicitly model the contextual interactions between the regions within an image and select the question which leads to the maximum reduction in the combined entropy of all the regions in the image (image entropy)
A Novel Medical Freehand Sketch 3D Model Retrieval Method by Dimensionality Reduction and Feature Vector Transformation
To assist physicians to quickly find the required 3D model from the mass medical model, we propose a novel retrieval method, called DRFVT, which combines the characteristics of dimensionality reduction (DR) and feature vector transformation (FVT) method. The DR method reduces the dimensionality of feature vector; only the top M low frequency Discrete Fourier Transform coefficients are retained. The FVT method does the transformation of the original feature vector and generates a new feature vector to solve the problem of noise sensitivity. The experiment results demonstrate that the DRFVT method achieves more effective and efficient retrieval results than other proposed methods
Exploring Fine-Grained Representation and Recomposition for Cloth-Changing Person Re-Identification
Cloth-changing person Re-IDentification (Re-ID) is a particularly challenging
task, suffering from two limitations of inferior identity-relevant features and
limited training samples. Existing methods mainly leverage auxiliary
information to facilitate discriminative feature learning, including
soft-biometrics features of shapes and gaits, and additional labels of
clothing. However, these information may be unavailable in real-world
applications. In this paper, we propose a novel FIne-grained Representation and
Recomposition (FIRe) framework to tackle both limitations without any
auxiliary information. Specifically, we first design a Fine-grained Feature
Mining (FFM) module to separately cluster images of each person. Images with
similar so-called fine-grained attributes (e.g., clothes and viewpoints) are
encouraged to cluster together. An attribute-aware classification loss is
introduced to perform fine-grained learning based on cluster labels, which are
not shared among different people, promoting the model to learn
identity-relevant features. Furthermore, by taking full advantage of the
clustered fine-grained attributes, we present a Fine-grained Attribute
Recomposition (FAR) module to recompose image features with different
attributes in the latent space. It can significantly enhance representations
for robust feature learning. Extensive experiments demonstrate that FIRe
can achieve state-of-the-art performance on five widely-used cloth-changing
person Re-ID benchmarks
Data-Driven Shape Analysis and Processing
Data-driven methods play an increasingly important role in discovering
geometric, structural, and semantic relationships between 3D shapes in
collections, and applying this analysis to support intelligent modeling,
editing, and visualization of geometric data. In contrast to traditional
approaches, a key feature of data-driven approaches is that they aggregate
information from a collection of shapes to improve the analysis and processing
of individual shapes. In addition, they are able to learn models that reason
about properties and relationships of shapes without relying on hard-coded
rules or explicitly programmed instructions. We provide an overview of the main
concepts and components of these techniques, and discuss their application to
shape classification, segmentation, matching, reconstruction, modeling and
exploration, as well as scene analysis and synthesis, through reviewing the
literature and relating the existing works with both qualitative and numerical
comparisons. We conclude our report with ideas that can inspire future research
in data-driven shape analysis and processing.Comment: 10 pages, 19 figure
- …