1,385 research outputs found
Representation Learning: A Review and New Perspectives
The success of machine learning algorithms generally depends on data
representation, and we hypothesize that this is because different
representations can entangle and hide more or less the different explanatory
factors of variation behind the data. Although specific domain knowledge can be
used to help design representations, learning with generic priors can also be
used, and the quest for AI is motivating the design of more powerful
representation-learning algorithms implementing such priors. This paper reviews
recent work in the area of unsupervised feature learning and deep learning,
covering advances in probabilistic models, auto-encoders, manifold learning,
and deep networks. This motivates longer-term unanswered questions about the
appropriate objectives for learning good representations, for computing
representations (i.e., inference), and the geometrical connections between
representation learning, density estimation and manifold learning
Structure-Aware Classification using Supervised Dictionary Learning
In this paper, we propose a supervised dictionary learning algorithm that
aims to preserve the local geometry in both dimensions of the data. A
graph-based regularization explicitly takes into account the local manifold
structure of the data points. A second graph regularization gives similar
treatment to the feature domain and helps in learning a more robust dictionary.
Both graphs can be constructed from the training data or learned and adapted
along the dictionary learning process. The combination of these two terms
promotes the discriminative power of the learned sparse representations and
leads to improved classification accuracy. The proposed method was evaluated on
several different datasets, representing both single-label and multi-label
classification problems, and demonstrated better performance compared with
other dictionary based approaches
- …