40,745 research outputs found
From Cluster Assumption to Graph Convolution: Graph-based Semi-Supervised Learning Revisited
Graph-based semi-supervised learning (GSSL) has long been a hot research
topic. Traditional methods are generally shallow learners, based on the cluster
assumption. Recently, graph convolutional networks (GCNs) have become the
predominant techniques for their promising performance. In this paper, we
theoretically discuss the relationship between these two types of methods in a
unified optimization framework. One of the most intriguing findings is that,
unlike traditional ones, typical GCNs may not jointly consider the graph
structure and label information at each layer. Motivated by this, we further
propose three simple but powerful graph convolution methods. The first is a
supervised method OGC which guides the graph convolution process with labels.
The others are two unsupervised methods: GGC and its multi-scale version GGCM,
both aiming to preserve the graph structure information during the convolution
process. Finally, we conduct extensive experiments to show the effectiveness of
our methods
Recommended from our members
Uncertainty quantification for semi-supervised multi-class classification in image processing and ego-motion analysis of body-worn videos
Semi-supervised learning uses underlying relationships in data with a scarcity of ground-truth labels. In this paper, we introduce an uncertainty quantification (UQ) method for graph-based semi-supervised multi-class classification problems. We not only predict the class label for each data point, but also provide a confidence score for the prediction. We adopt a Bayesian approach and propose a graphical multi-class probit model together with an effective Gibbs sampling procedure. Furthermore, we propose a confidence measure for each data point that correlates with the classification performance. We use the empirical properties of the proposed confidence measure to guide the design of a human-in-the-loop system. The uncertainty quantification algorithm and the human-in-the-loop system are successfully applied to classification
problems in image processing and ego-motion analysis of
body-worn videos
Visual Understanding via Multi-Feature Shared Learning with Global Consistency
Image/video data is usually represented with multiple visual features. Fusion
of multi-source information for establishing the attributes has been widely
recognized. Multi-feature visual recognition has recently received much
attention in multimedia applications. This paper studies visual understanding
via a newly proposed l_2-norm based multi-feature shared learning framework,
which can simultaneously learn a global label matrix and multiple
sub-classifiers with the labeled multi-feature data. Additionally, a group
graph manifold regularizer composed of the Laplacian and Hessian graph is
proposed for better preserving the manifold structure of each feature, such
that the label prediction power is much improved through the semi-supervised
learning with global label consistency. For convenience, we call the proposed
approach Global-Label-Consistent Classifier (GLCC). The merits of the proposed
method include: 1) the manifold structure information of each feature is
exploited in learning, resulting in a more faithful classification owing to the
global label consistency; 2) a group graph manifold regularizer based on the
Laplacian and Hessian regularization is constructed; 3) an efficient
alternative optimization method is introduced as a fast solver owing to the
convex sub-problems. Experiments on several benchmark visual datasets for
multimedia understanding, such as the 17-category Oxford Flower dataset, the
challenging 101-category Caltech dataset, the YouTube & Consumer Videos dataset
and the large-scale NUS-WIDE dataset, demonstrate that the proposed approach
compares favorably with the state-of-the-art algorithms. An extensive
experiment on the deep convolutional activation features also show the
effectiveness of the proposed approach. The code is available on
http://www.escience.cn/people/lei/index.htmlComment: 13 pages,6 figures, this paper is accepted for publication in IEEE
Transactions on Multimedi
Graph-based Poisson learning for image co-segmentation
Faculty advisor: Jeff CalderThis paper focuses on applying the state-of-the-art graph-based semi-supervised learning algorithm Poisson Learning to the task of image co-segmentation. The weighted matrix built based on the image texture is used as the feature to do the classification by Poisson learning. The results presented in the paper show that Poisson learning does a good job in segmentation and has high accuracy at very low label rates. Moreover, Poisson learning is tolerant with multi-weight metrics and simple to implement.This research was supported by the Undergraduate Research Opportunities Program (UROP)
Graph-based Multi-View Fusion and Local Adaptation: Mitigating Within-Household Confusability for Speaker Identification
Speaker identification (SID) in the household scenario (e.g., for smart
speakers) is an important but challenging problem due to limited number of
labeled (enrollment) utterances, confusable voices, and demographic imbalances.
Conventional speaker recognition systems generalize from a large random sample
of speakers, causing the recognition to underperform for households drawn from
specific cohorts or otherwise exhibiting high confusability. In this work, we
propose a graph-based semi-supervised learning approach to improve
household-level SID accuracy and robustness with locally adapted graph
normalization and multi-signal fusion with multi-view graphs. Unlike other work
on household SID, fairness, and signal fusion, this work focuses on speaker
label inference (scoring) and provides a simple solution to realize
household-specific adaptation and multi-signal fusion without tuning the
embeddings or training a fusion network. Experiments on the VoxCeleb dataset
demonstrate that our approach consistently improves the performance across
households with different customer cohorts and degrees of confusability.Comment: To appear in Interspeech 2022. arXiv admin note: text overlap with
arXiv:2106.0820
- …