744 research outputs found
Combined Group and Exclusive Sparsity for Deep Neural Networks
Department of Computer Science and EngineeringThe number of parameters in a deep neural network is usually very large, which helps with its learning capacity but also hinders its scalability and practicality due to memory/time inefficiency and overfitting. To resolve this issue, we propose a sparsity regularization method that exploits both positive and negative correlations among the features to enforce the network to be sparse, and at the same time remove any redundancies among the features to fully utilize the capacity of the network. Specifically, we propose to use an exclusive sparsity regularization based on (1,2)-norm, which promotes competition for features between different weights, thus enforcing them to fit to disjoint sets of features. We further combine the exclusive sparsity with the group sparsity based on (2,1)-norm, to promote both sharing and competition for features in training of a deep neural network. We validate our method on multiple public datasets, and the results show that our method can obtain more compact and efficient networks while also improving the performance over the base networks with full weights, as opposed to existing sparsity regularizations that often obtain efficiency at the expense of prediction accuracy.ope
Non-Rigid Puzzles
Shape correspondence is a fundamental problem in computer graphics and vision, with applications in various problems including animation, texture mapping, robotic vision, medical imaging, archaeology and many more. In settings where the shapes are allowed to undergo non-rigid deformations and only partial views are available, the problem becomes very challenging. To this end, we present a non-rigid multi-part shape matching algorithm. We assume to be given a reference shape and its multiple parts undergoing a non-rigid deformation. Each of these query parts can be additionally contaminated by clutter, may overlap with other parts, and there might be missing parts or redundant ones. Our method simultaneously solves for the segmentation of the reference model, and for a dense correspondence to (subsets of) the parts. Experimental results on synthetic as well as real scans demonstrate the effectiveness of our method in dealing with this challenging matching scenario
OOGAN: Disentangling GAN with One-Hot Sampling and Orthogonal Regularization
Exploring the potential of GANs for unsupervised disentanglement learning,
this paper proposes a novel GAN-based disentanglement framework with One-Hot
Sampling and Orthogonal Regularization (OOGAN). While previous works mostly
attempt to tackle disentanglement learning through VAE and seek to implicitly
minimize the Total Correlation (TC) objective with various sorts of
approximation methods, we show that GANs have a natural advantage in
disentangling with an alternating latent variable (noise) sampling method that
is straightforward and robust. Furthermore, we provide a brand-new perspective
on designing the structure of the generator and discriminator, demonstrating
that a minor structural change and an orthogonal regularization on model
weights entails an improved disentanglement. Instead of experimenting on simple
toy datasets, we conduct experiments on higher-resolution images and show that
OOGAN greatly pushes the boundary of unsupervised disentanglement.Comment: AAAI 202
- …