14 research outputs found
Compressive Sensing with Tensorized Autoencoder
Deep networks can be trained to map images into a low-dimensional latent
space. In many cases, different images in a collection are articulated versions
of one another; for example, same object with different lighting, background,
or pose. Furthermore, in many cases, parts of images can be corrupted by noise
or missing entries. In this paper, our goal is to recover images without access
to the ground-truth (clean) images using the articulations as structural prior
of the data. Such recovery problems fall under the domain of compressive
sensing. We propose to learn autoencoder with tensor ring factorization on the
the embedding space to impose structural constraints on the data. In
particular, we use a tensor ring structure in the bottleneck layer of the
autoencoder that utilizes the soft labels of the structured dataset. We
empirically demonstrate the effectiveness of the proposed approach for
inpainting and denoising applications. The resulting method achieves better
reconstruction quality compared to other generative prior-based self-supervised
recovery approaches for compressive sensing
Recommended from our members
Generative Models for Low-Dimensional Video Representation and Reconstruction
Finding compact representation of videos is an essential component in almost
every problem related to video processing or understanding. In this paper, we
propose a generative model to learn compact latent codes that can efficiently
represent and reconstruct a video sequence from its missing or under-sampled
measurements. We use a generative network that is trained to map a compact code
into an image. We first demonstrate that if a video sequence belongs to the
range of the pretrained generative network, then we can recover it by
estimating the underlying compact latent codes. Then we demonstrate that even
if the video sequence does not belong to the range of a pretrained network, we
can still recover the true video sequence by jointly updating the latent codes
and the weights of the generative network. To avoid overfitting in our model,
we regularize the recovery problem by imposing low-rank and similarity
constraints on the latent codes of the neighboring frames in the video
sequence. We use our methods to recover a variety of videos from compressive
measurements at different compression rates. We also demonstrate that we can
generate missing frames in a video sequence by interpolating the latent codes
of the observed frames in the low-dimensional space
Generative Models for Low-Rank Video Representation and Reconstruction
Finding compact representation of videos is an essential component in almost
every problem related to video processing or understanding. In this paper, we
propose a generative model to learn compact latent codes that can efficiently
represent and reconstruct a video sequence from its missing or under-sampled
measurements. We use a generative network that is trained to map a compact code
into an image. We first demonstrate that if a video sequence belongs to the
range of the pretrained generative network, then we can recover it by
estimating the underlying compact latent codes. Then we demonstrate that even
if the video sequence does not belong to the range of a pretrained network, we
can still recover the true video sequence by jointly updating the latent codes
and the weights of the generative network. To avoid overfitting in our model,
we regularize the recovery problem by imposing low-rank and similarity
constraints on the latent codes of the neighboring frames in the video
sequence. We use our methods to recover a variety of videos from compressive
measurements at different compression rates. We also demonstrate that we can
generate missing frames in a video sequence by interpolating the latent codes
of the observed frames in the low-dimensional space
Factorized Tensor Networks for Multi-Task and Multi-Domain Learning
Multi-task and multi-domain learning methods seek to learn multiple
tasks/domains, jointly or one after another, using a single unified network.
The key challenge and opportunity is to exploit shared information across tasks
and domains to improve the efficiency of the unified network. The efficiency
can be in terms of accuracy, storage cost, computation, or sample complexity.
In this paper, we propose a factorized tensor network (FTN) that can achieve
accuracy comparable to independent single-task/domain networks with a small
number of additional parameters. FTN uses a frozen backbone network from a
source model and incrementally adds task/domain-specific low-rank tensor
factors to the shared frozen network. This approach can adapt to a large number
of target domains and tasks without catastrophic forgetting. Furthermore, FTN
requires a significantly smaller number of task-specific parameters compared to
existing methods. We performed experiments on widely used multi-domain and
multi-task datasets. We show the experiments on convolutional-based
architecture with different backbones and on transformer-based architecture. We
observed that FTN achieves similar accuracy as single-task/domain methods while
using only a fraction of additional parameters per task
Incremental Task Learning with Incremental Rank Updates
Incremental Task learning (ITL) is a category of continual learning that
seeks to train a single network for multiple tasks (one after another), where
training data for each task is only available during the training of that task.
Neural networks tend to forget older tasks when they are trained for the newer
tasks; this property is often known as catastrophic forgetting. To address this
issue, ITL methods use episodic memory, parameter regularization, masking and
pruning, or extensible network structures. In this paper, we propose a new
incremental task learning framework based on low-rank factorization. In
particular, we represent the network weights for each layer as a linear
combination of several rank-1 matrices. To update the network for a new task,
we learn a rank-1 (or low-rank) matrix and add that to the weights of every
layer. We also introduce an additional selector vector that assigns different
weights to the low-rank matrices learned for the previous tasks. We show that
our approach performs better than the current state-of-the-art methods in terms
of accuracy and forgetting. Our method also offers better memory efficiency
compared to episodic memory- and mask-based approaches. Our code will be
available at https://github.com/CSIPlab/task-increment-rank-update.gitComment: Code will be available at
https://github.com/CSIPlab/task-increment-rank-update.gi
Non-Adversarial Video Synthesis with Learned Priors
Most of the existing works in video synthesis focus on generating videos
using adversarial learning. Despite their success, these methods often require
input reference frame or fail to generate diverse videos from the given data
distribution, with little to no uniformity in the quality of videos that can be
generated. Different from these methods, we focus on the problem of generating
videos from latent noise vectors, without any reference input frames. To this
end, we develop a novel approach that jointly optimizes the input latent space,
the weights of a recurrent neural network and a generator through
non-adversarial learning. Optimizing for the input latent space along with the
network weights allows us to generate videos in a controlled environment, i.e.,
we can faithfully generate all videos the model has seen during the learning
process as well as new unseen videos. Extensive experiments on three
challenging and diverse datasets well demonstrate that our approach generates
superior quality videos compared to the existing state-of-the-art methods.Comment: Accepted to CVPR 202
Recommended from our members
Simple Structures in Deep Networks
Deep networks have received considerable attention in recent years due to their applications in different problems of science and engineering. This dissertation explores the application of deep networks in continual learning and inverse problems. In this work, we enforce some simple structures on the networks to achieve better solution in terms of performance, memory and computational cost.Continual Learning with Low-rank Increment: Continual learning is a process of training a single neural network on multiple tasks one after another, where training data for each task is often available only during the training of that task. Neural networks tend to forget older tasks when they are trained for the newer tasks; this property is often known as catastrophic forgetting. To address this issue, continual learning methods use episodic memory, parameter regularization, masking and pruning, or extensible network structures. This work proposes a continual learning framework based on low-rank factorization of the network weights. To update the network for a new task, a rank-1 (or low-rank) matrix is learned and added to the weights of every layer. An additional selector vector is also introduced that assigns different weights to the low-rank matrices learned for the previous tasks. Our proposed approach demonstrates superior performance compared to the current state-of-the-art methods with much lower number of network parameters.Inverse Problems with Deep Networks: Inverse problems form a family of problems where we try to recover the true signal given the modified version of the signal. Since inverse problems are often ill-posed in nature, we often need to impose some constraints on the solution set. This dissertation mainly focuses on deep generative networks as a prior for solving inverse problems. Low-rank matrix and tensor structures have been used in this work as constraints on the input latent vectors of the deep generative networks to improve quality of the reconstruction with reduced parameters. This dissertation also explores unrolled networks where classical iterative solution approaches are structured as fixed layer networks with each iteration forming a layer of the network. We use such unrolled network structures to design sensing parameters for nonlinear inverse problems that led to achieving good reconstruction quality with a fixed number of layers (or iterations)