4,058 research outputs found
Bags of Affine Subspaces for Robust Object Tracking
We propose an adaptive tracking algorithm where the object is modelled as a
continuously updated bag of affine subspaces, with each subspace constructed
from the object's appearance over several consecutive frames. In contrast to
linear subspaces, affine subspaces explicitly model the origin of subspaces.
Furthermore, instead of using a brittle point-to-subspace distance during the
search for the object in a new frame, we propose to use a subspace-to-subspace
distance by representing candidate image areas also as affine subspaces.
Distances between subspaces are then obtained by exploiting the non-Euclidean
geometry of Grassmann manifolds. Experiments on challenging videos (containing
object occlusions, deformations, as well as variations in pose and
illumination) indicate that the proposed method achieves higher tracking
accuracy than several recent discriminative trackers.Comment: in International Conference on Digital Image Computing: Techniques
and Applications, 201
Simultaneous Inference of User Representations and Trust
Inferring trust relations between social media users is critical for a number
of applications wherein users seek credible information. The fact that
available trust relations are scarce and skewed makes trust prediction a
challenging task. To the best of our knowledge, this is the first work on
exploring representation learning for trust prediction. We propose an approach
that uses only a small amount of binary user-user trust relations to
simultaneously learn user embeddings and a model to predict trust between user
pairs. We empirically demonstrate that for trust prediction, our approach
outperforms classifier-based approaches which use state-of-the-art
representation learning methods like DeepWalk and LINE as features. We also
conduct experiments which use embeddings pre-trained with DeepWalk and LINE
each as an input to our model, resulting in further performance improvement.
Experiments with a dataset of 356K user pairs show that the proposed
method can obtain an high F-score of 92.65%.Comment: To appear in the proceedings of ASONAM'17. Please cite that versio
Recommended from our members
Semantic Concept Co-Occurrence Patterns for Image Annotation and Retrieval.
Describing visual image contents by semantic concepts is an effective and straightforward way to facilitate various high level applications. Inferring semantic concepts from low-level pictorial feature analysis is challenging due to the semantic gap problem, while manually labeling concepts is unwise because of a large number of images in both online and offline collections. In this paper, we present a novel approach to automatically generate intermediate image descriptors by exploiting concept co-occurrence patterns in the pre-labeled training set that renders it possible to depict complex scene images semantically. Our work is motivated by the fact that multiple concepts that frequently co-occur across images form patterns which could provide contextual cues for individual concept inference. We discover the co-occurrence patterns as hierarchical communities by graph modularity maximization in a network with nodes and edges representing concepts and co-occurrence relationships separately. A random walk process working on the inferred concept probabilities with the discovered co-occurrence patterns is applied to acquire the refined concept signature representation. Through experiments in automatic image annotation and semantic image retrieval on several challenging datasets, we demonstrate the effectiveness of the proposed concept co-occurrence patterns as well as the concept signature representation in comparison with state-of-the-art approaches
Improving Texture Categorization with Biologically Inspired Filtering
Within the domain of texture classification, a lot of effort has been spent
on local descriptors, leading to many powerful algorithms. However,
preprocessing techniques have received much less attention despite their
important potential for improving the overall classification performance. We
address this question by proposing a novel, simple, yet very powerful
biologically-inspired filtering (BF) which simulates the performance of human
retina. In the proposed approach, given a texture image, after applying a DoG
filter to detect the "edges", we first split the filtered image into two "maps"
alongside the sides of its edges. The feature extraction step is then carried
out on the two "maps" instead of the input image. Our algorithm has several
advantages such as simplicity, robustness to illumination and noise, and
discriminative power. Experimental results on three large texture databases
show that with an extremely low computational cost, the proposed method
improves significantly the performance of many texture classification systems,
notably in noisy environments. The source codes of the proposed algorithm can
be downloaded from https://sites.google.com/site/nsonvu/code.Comment: 11 page
Deep Self-Taught Learning for Handwritten Character Recognition
Recent theoretical and empirical work in statistical machine learning has
demonstrated the importance of learning algorithms for deep architectures,
i.e., function classes obtained by composing multiple non-linear
transformations. Self-taught learning (exploiting unlabeled examples or
examples from other distributions) has already been applied to deep learners,
but mostly to show the advantage of unlabeled examples. Here we explore the
advantage brought by {\em out-of-distribution examples}. For this purpose we
developed a powerful generator of stochastic variations and noise processes for
character images, including not only affine transformations but also slant,
local elastic deformations, changes in thickness, background images, grey level
changes, contrast, occlusion, and various types of noise. The
out-of-distribution examples are obtained from these highly distorted images or
by including examples of object classes different from those in the target test
set. We show that {\em deep learners benefit more from out-of-distribution
examples than a corresponding shallow learner}, at least in the area of
handwritten character recognition. In fact, we show that they beat previously
published results and reach human-level performance on both handwritten digit
classification and 62-class handwritten character recognition
- …