10 research outputs found
Deep learning of appearance affinity for multi-object tracking and re-identification: a comparative view
Recognizing the identity of a query individual in a surveillance sequence is the core of Multi-Object Tracking (MOT) and Re-Identification (Re-Id) algorithms. Both tasks can be addressed by measuring the appearance affinity between people observations with a deep neural model. Nevertheless, the differences in their specifications and, consequently, in the characteristics and constraints of the available training data for each one of these tasks, arise from the necessity of employing different learning approaches to attain each one of them. This article offers a comparative view of the Double-Margin-Contrastive and the Triplet loss function, and analyzes the benefits and drawbacks of applying each one of them to learn an Appearance Affinity model for Tracking and Re-Identification. A batch of experiments have been conducted, and their results support the hypothesis concluded from the presented study: Triplet loss function is more effective than the Contrastive one when an Re-Id model is learnt, and, conversely, in the MOT domain, the Contrastive loss can better discriminate between pairs of images rendering the same person or not.This research was funded by the Spanish Government through the CICYT projects
(TRA2016-78886-C3-1-R and RTI2018-096036-B-C21), Universidad Carlos III of Madrid through (PEAVAUTO-CM-UC3M), the Comunidad de Madrid through SEGVAUTO-4.0-CM (P2018/EMT-4362), and the Ministerio de Educación, Cultura y Deporte para la Formación de Profesorado Universitario (FPU14/02143)
Minimising Human Annotation for Scalable Person Re-Identification
PhDAmong the diverse tasks performed by an intelligent distributed multi-camera surveillance system,
person re-identification (re-id) is one of the most essential. Re-id refers to associating an
individual or a group of people across non-overlapping cameras at different times and locations,
and forms the foundation of a variety of applications ranging from security and forensic search
to quotidian retail and health care. Though attracted rapidly increasing academic interests over
the past decade, it still remains a non-trivial and unsolved problem for launching a practical reid
system in real-world environments, due to the ambiguous and noisy feature of surveillance
data and the potentially dramatic visual appearance changes caused by uncontrolled variations in
human poses and divergent viewing conditions across distributed camera views.
To mitigate such visual ambiguity and appearance variations, most existing re-id approaches
rely on constructing fully supervised machine learning models with extensively labelled training
datasets which is unscalable for practical applications in the real-world. Particularly, human annotators
must exhaustively search over a vast quantity of offline collected data, manually label
cross-view matched images of a large population between every possible camera pair. Nonetheless,
having the prohibitively expensive human efforts dissipated, a trained re-id model is often
not easily generalisable and transferable, due to the elastic and dynamic operating conditions
of a surveillance system. With such motivations, this thesis proposes several scalable re-id approaches
with significantly reduced human supervision, readily applied to practical applications.
More specifically, this thesis has developed and investigated four new approaches for reducing
human labelling effort in real-world re-id as follows:
Chapter 3 The first approach is affinity mining from unlabelled data. Different from most
existing supervised approaches, this work aims to model the discriminative information for reid
without exploiting human annotations, but from the vast amount of unlabelled person image
data, thus applicable to both semi-supervised and unsupervised re-id. It is non-trivial since the
human annotated identity matching correspondence is often the key to discriminative re-id modelling.
In this chapter, an alternative strategy is explored by specifically mining two types of
affinity relationships among unlabelled data: (1) inter-view data affinity and (2) intra-view data
affinity. In particular, with such affinity information encoded as constraints, a Regularised Kernel
Subspace Learning model is developed to explicitly reduce inter-view appearance variations
and meanwhile enhance intra-view appearance disparity for more discriminative re-id matching.
Consequently, annotation costs can be immensely alleviated and a scalable re-id model is readily
to be leveraged to plenty of unlabelled data which is inexpensive to collect.
Chapter 4 The second approach is saliency discovery from unlabelled data. This chapter
continues to investigate the problem of what can be learned in unlabelled images without identity
labels annotated by human. Other than affinity mining as proposed by Chapter 3, a different solution
is proposed. That is, to discover localised visual appearance saliency of person appearances.
Intuitively, salient and atypical appearances of human are able to uniquely and representatively
describe and identify an individual, whilst also often robust to view changes and detection variances.
Motivated by this, an unsupervised Generative Topic Saliency model is proposed to jointly
perform foreground extraction, saliency detection, as well as discriminative re-id matching. This
approach completely avoids the exhaustive annotation effort for model training, and thus better
scales to real-world applications. Moreover, its automatically discovered re-id saliency representations
are shown to be semantically interpretable, suitable for generating useful visual analysis
for deployable user-oriented software tools.
Chapter 5 The third approach is incremental learning from actively labelled data. Since
learning from unlabelled data alone yields less discriminative matching results, and in some cases
there will be limited human labelling resources available for re-id modelling, this chapter thus
investigate the problem of how to maximise a model’s discriminative capability with minimised
labelling efforts. The challenges are to (1) automatically select the most representative data from
a vast number of noisy/ambiguous unlabelled data in order to maximise model discrimination
capacity; and (2) incrementally update the model parameters to accelerate machine responses
and reduce human waiting time. To that end, this thesis proposes a regression based re-id model,
characterised by its very fast and efficient incremental model updates. Furthermore, an effective
active data sampling algorithm with three novel joint exploration-exploitation criteria is designed,
to make automatic data selection feasible with notably reduced human labelling costs. Such an
approach ensures annotations to be spent only on very few data samples which are most critical
to model’s generalisation capability, instead of being exhausted by blindly labelling many noisy
and redundant training samples.
Chapter 6 The last technical area of this thesis is human-in-the-loop learning from relevance
feedback. Whilst former chapters mainly investigate techniques to reduce human supervision for
model training, this chapter motivates a novel research area to further minimise human efforts
spent in the re-id deployment stage. In real-world applications where camera network and potential
gallery size increases dramatically, even the state-of-the-art re-id models generate much
inferior re-id performances and human involvements at deployment stage is inevitable. To minimise
such human efforts and maximise re-id performance, this thesis explores an alternative
approach to re-id by formulating a hybrid human-computer learning paradigm with humans in
the model matching loop. Specifically, a Human Verification Incremental Learning model is formulated
which does not require any pre-labelled training data, therefore scalable to new camera
pairs; Moreover, the proposed model learns cumulatively from human feedback to provide an instant
improvement to re-id ranking of each probe on-the-fly, thus scalable to large gallery sizes. It
has been demonstrated that the proposed re-id model achieves significantly superior re-id results
whilst only consumes much less human supervision effort.
For facilitating a holistic understanding about this thesis, the main studies are summarised
and framed into a graphical abstract as shown in Figur
Recommended from our members
Three-Dimensional Object Search, Understanding, and Pose Estimation with Low-Cost Sensors
With the recent development of low-cost depth sensors, an entirely new type of 3D data is being generated rapidly by regular consumers. Traditionally, 3D data is produced by a small number of professional designers (i.e., the Computer Aided Design (CAD) model); however, 3D data from massive consumer-level sensors has the potential of introducing many new applications, such as user-captured 3D warehouse and search engines, robots with 3D sensing capability, and customized 3D printing. Nevertheless, the low-cost sensors used by general consumers also pose new technological challenges. First, they have relatively high levels of sensor noise. Second, the use of such consumer devices is often in uncontrolled settings, resulting in challenging conditions, such as poor lighting, cluttered scenes, and object occlusion. To address such emerging opportunities and associated challenges, this dissertation is dedicated to the development of novel algorithms and systems for 3D data understanding and processing, using input from a consumer-level 3D sensor.
In particular, the key problems of 3D shape retrieval, scene understanding, and pose recognition are explored in order to present a comprehensive coverage of the key aspects of content-based 3D shape analysis. To resolve the aforementioned challenges, we propose a flexible Markov Random Field (MRF) framework that uses local information to allow partial matching, and thus address the model incompleteness problem; the framework also uses higher-order correlation to provide additional robustness against sensor noise. With the MRF framework, these 3D analysis problems can be transformed into a unified potential energy minimization problem, while preserving the flexibility to adapt to different settings and resolve the unique challenges of each problem. The contributions of the dissertation include:
a. Cross-Domain 3D Retrieval: First we tackle the problem of searching 3D noise- free models using noisy data captured by low-cost 3D sensors – a unique cross-domain setting. To manage the challenges of sensor noise and model incompleteness from consumer-level sensors, we propose a novel MRF formulation for the retrieval problem. The potential function of the random field is designed to capture both the local shape and global spatial consistency in order to preserve the local matching capability, while offering robustness against the sensor noise. The specific form of the potential functions is determined efficiently by a series of weak classifiers, thus forming a variant of the Regression Tree Field (RTF). We achieve better retrieval precision and recall in the cross-domain settings with a consumer-level depth sensor compared with state-of-the-art approaches.
b. 3D Scene Understanding: We develop a scene understanding system based on input from consumer-level depth sensors. To resolve the key challenge of the lack of annotated 3D training data, we construct an MRF that connects the input 3D point cloud and the associated 2D reference images, based on which the 3D point cloud is stitched. A series of weak classifiers are trained to obtain an approximate semantic segmentation result from the reference images. The potential function of the field is designed to integrate the results from the classifiers, while taking advantage of the 3D spatial consistency in order to output a comprehensive scene understanding result. We achieve comparable accuracy and much faster speed compared with state-of-the-art 3D scene understanding systems, with the difference that we do not require annotated 3D training data.
c. Pose Recognition of Deformable Objects: We develop a method for supporting a robotics system to recognize pose and manipulate deformable objects. More specifically, garment pose is recognized with the help of an offline simulated database and the proposed retrieval approach. We use a novel binary feature representation extracted from the reconstructed 3D surfaces in order to allow efficient matching, thus achieving real-time performance. A spatial weight is further learned in order to integrate the local matching result. The system shows superior recognition accuracy and faster speed than the state-of-the-art approaches.
d. Application with 2D Data: In addition to the traditional 3D applications, we explore the possibility of extending MRF formulation to 2D data, especially those used in classical low-level 2D vision problems, such as image deblurring and denoising. One well-known technique that uses image prior, the probabilistic patched-based prior, is known to have bottlenecks in finding the most similar model from a model set, which can be posed as a retrieval problem. Therefore, we apply the MRF formulation originally developed for 3D shape retrieval, and extend it to this 2D problem by introducing a grid-like random field structure. We can achieve 40x acceleration compared with the state-of-the-art algorithm, while preserving quality.
We organize the dissertation as follows. First, the core problems of 3D shape retrieval, scene understanding, and pose recognition, and with the proposed solutions that use MRF and RTF are explored in Part I. In Part II, the extension to 2D data is discussed. Extensive evaluation is performed in each specific task in order to compare the proposed approaches with state-of-the-art algorithms and systems, and also to justify the components of the proposed methods. Finally, in Part III, we include the conclusion remarks and discussion of open issues and future work
IMAGE RETRIEVAL BASED ON COMPLEX DESCRIPTIVE QUERIES
The amount of visual data such as images and videos available over web has increased exponentially over the last few years. In order to efficiently organize and exploit these massive collections, a system, apart from being able to answer simple classification based questions such as whether a specific object is present (or absent) in an image, should also be capable of searching images and videos based on more complex descriptive questions. There is also a considerable amount of structure present in the visual world which, if effectively utilized, can help achieve this goal. To this end, we first present an approach for image ranking and retrieval based on queries consisting of multiple semantic attributes. We further show that there are significant correlations present between these attributes and accounting for them can lead to superior performance. Next, we extend this by proposing an image retrieval framework for descriptive queries composed of object categories, semantic attributes and spatial relationships. The proposed framework also includes a unique multi-view hashing technique, which enables query specification in three different modalities - image, sketch and text.
We also demonstrate the effectiveness of leveraging contextual information to reduce the supervision requirements for learning object and scene recognition models. We present an active learning framework to simultaneously learn appearance and contextual models for scene understanding. Within this framework we introduce new kinds of labeling questions that are designed to collect appearance as well as contextual information and which mimic the way in which humans actively learn about their environment. Furthermore we explicitly model the contextual interactions between the regions within an image and select the question which leads to the maximum reduction in the combined entropy of all the regions in the image (image entropy)
Ranking and Retrieval under Semantic Relevance
This thesis presents a series of conceptual and empirical developments on the ranking and retrieval of candidates under semantic relevance. Part I of the thesis introduces the concept of uncertainty in various semantic tasks (such as recognizing textual entailment) in natural language processing, and the machine learning techniques commonly employed to model these semantic phenomena. A unified view of ranking and retrieval will be presented, and the trade-off between model expressiveness, performance, and scalability in model design will be discussed.
Part II of the thesis focuses on applying these ranking and retrieval techniques to text: Chapter 3 examines the feasibility of ranking hypotheses given a premise with respect to a human's subjective probability of the hypothesis happening, effectively extending the traditional categorical task of natural language inference. Chapter 4 focuses on detecting situation frames for documents using ranking methods. Then we extend the ranking notion to retrieval, and develop both sparse (Chapter 5) and dense (Chapter 6) vector-based methods to facilitate scalable retrieval for potential answer paragraphs in question answering.
Part III turns the focus to mentions and entities in text, while continuing the theme on ranking and retrieval: Chapter 7 discusses the ranking of fine-grained types that an entity mention could belong to, leading to state-of-the-art performance on hierarchical multi-label fine-grained entity typing. Chapter 8 extends the semantic relation of coreference to a cross-document setting, enabling models to retrieve from a large corpus, instead of in a single document, when resolving coreferent entity mentions
Semantic Systems. The Power of AI and Knowledge Graphs
This open access book constitutes the refereed proceedings of the 15th International Conference on Semantic Systems, SEMANTiCS 2019, held in Karlsruhe, Germany, in September 2019. The 20 full papers and 8 short papers presented in this volume were carefully reviewed and selected from 88 submissions. They cover topics such as: web semantics and linked (open) data; machine learning and deep learning techniques; semantic information management and knowledge integration; terminology, thesaurus and ontology management; data mining and knowledge discovery; semantics in blockchain and distributed ledger technologies