32 research outputs found
The Neural Tangent Link Between CNN Denoisers and Non-Local Filters
Convolutional Neural Networks (CNNs) are now a well-established tool for
solving computational imaging problems. Modern CNN-based algorithms obtain
state-of-the-art performance in diverse image restoration problems.
Furthermore, it has been recently shown that, despite being highly
overparameterized, networks trained with a single corrupted image can still
perform as well as fully trained networks. We introduce a formal link between
such networks through their neural tangent kernel (NTK), and well-known
non-local filtering techniques, such as non-local means or BM3D. The filtering
function associated with a given network architecture can be obtained in closed
form without need to train the network, being fully characterized by the random
initialization of the network weights. While the NTK theory accurately predicts
the filter associated with networks trained using standard gradient descent,
our analysis shows that it falls short to explain the behaviour of networks
trained using the popular Adam optimizer. The latter achieves a larger change
of weights in hidden layers, adapting the non-local filtering function during
training. We evaluate our findings via extensive image denoising experiments
Sampling Theorems for Unsupervised Learning in Linear Inverse Problems
Solving a linear inverse problem requires knowledge about the underlying
signal model. In many applications, this model is a priori unknown and has to
be learned from data. However, it is impossible to learn the model using
observations obtained via a single incomplete measurement operator, as there is
no information outside the range of the inverse operator, resulting in a
chicken-and-egg problem: to learn the model we need reconstructed signals, but
to reconstruct the signals we need to know the model. Two ways to overcome this
limitation are using multiple measurement operators or assuming that the signal
model is invariant to a certain group action. In this paper, we present
necessary and sufficient sampling conditions for learning the signal model from
partial measurements which only depend on the dimension of the model, and the
number of operators or properties of the group action that the model is
invariant to. As our results are agnostic of the learning algorithm, they shed
light into the fundamental limitations of learning from incomplete data and
have implications in a wide range set of practical algorithms, such as
dictionary learning, matrix completion and deep neural networks.Comment: arXiv admin note: substantial text overlap with arXiv:2201.1215
Equivariant imaging: Learning beyond the range space
In various imaging problems, we only have access to compressed measurements
of the underlying signals, hindering most learning-based strategies which
usually require pairs of signals and associated measurements for training.
Learning only from compressed measurements is impossible in general, as the
compressed observations do not contain information outside the range of the
forward sensing operator. We propose a new end-to-end self-supervised framework
that overcomes this limitation by exploiting the equivariances present in
natural signals. Our proposed learning strategy performs as well as fully
supervised methods. Experiments demonstrate the potential of this framework on
inverse problems including sparse-view X-ray computed tomography on real
clinical data and image inpainting on natural images. Code will be released.Comment: Technical repor
Sensing Theorems for Unsupervised Learning in Linear Inverse Problems
International audienceSolving an ill-posed linear inverse problem requires knowledge about the underlying signal model. In many applications, this model is a priori unknown and has to be learned from data. However, it is impossible to learn the model using observations obtained via a single incomplete measurement operator, as there is no information about the signal model in the nullspace of the operator, resulting in a chicken-and-egg problem: to learn the model we need reconstructed signals, but to reconstruct the signals we need to know the model. Two ways to overcome this limitation are using multiple measurement operators or assuming that the signal model is invariant to a certain group action. In this paper, we present necessary and sufficient sensing conditions for learning the signal model from measurement data alone which only depend on the dimension of the model and the number of operators or properties of the group action that the model is invariant to. As our results are agnostic of the learning algorithm, they shed light into the fundamental limitations of learning from incomplete data and have implications in a wide range set of practical algorithms, such as dictionary learning, matrix completion and deep neural networks
A Sketching Framework for Reduced Data Transfer in Photon Counting Lidar
Single-photon lidar has become a prominent tool for depth imaging in recent
years. At the core of the technique, the depth of a target is measured by
constructing a histogram of time delays between emitted light pulses and
detected photon arrivals. A major data processing bottleneck arises on the
device when either the number of photons per pixel is large or the resolution
of the time stamp is fine, as both the space requirement and the complexity of
the image reconstruction algorithms scale with these parameters. We solve this
limiting bottleneck of existing lidar techniques by sampling the characteristic
function of the time of flight (ToF) model to build a compressive statistic, a
so-called sketch of the time delay distribution, which is sufficient to infer
the spatial distance and intensity of the object. The size of the sketch scales
with the degrees of freedom of the ToF model (number of objects) and not,
fundamentally, with the number of photons or the time stamp resolution.
Moreover, the sketch is highly amenable for on-chip online processing. We show
theoretically that the loss of information for compression is controlled and
the mean squared error of the inference quickly converges towards the optimal
Cram\'er-Rao bound (i.e. no loss of information) for modest sketch sizes. The
proposed compressed single-photon lidar framework is tested and evaluated on
real life datasets of complex scenes where it is shown that a compression rate
of up-to 150 is achievable in practice without sacrificing the overall
resolution of the reconstructed image.Comment: 16 pages, 20 figure
Fast surface detection in single-photon Lidar waveforms
International audienc
3D reconstruction using single-photon Lidar data exploiting the widths of the returns
International audienc
Equivariant Bootstrapping for Uncertainty Quantification in Imaging Inverse Problems
Scientific imaging problems are often severely ill-posed and hence have significant intrinsic uncertainty. Accurately quantifying the uncertainty in the solutions to such problems is therefore critical for the rigorous interpretation of experimental results as well as for reliably using the reconstructed images as scientific evidence. Unfortunately, existing imaging methods are unable to quantify the uncertainty in the reconstructed images in a way that is robust to experiment replications. This paper presents a new uncertainty quantification methodology based on an equivariant formulation of the parametric bootstrap algorithm that leverages symmetries and invariance properties commonly encountered in imaging problems. Additionally, the proposed methodology is general and can be easily applied with any image reconstruction technique, including unsupervised training strategies that can be trained from observed data alone, thus enabling uncertainty quantification in situations where there is no ground truth data available. We demonstrate the proposed approach with a series of experiments and comparisons with alternative state-of-the-art uncertainty quantification strategies. In all our experiments, the proposed equivariant bootstrap delivers remarkably accurate high-dimensional confidence regions and outperforms the competing approaches in terms of estimation accuracy, uncertainty quantification accuracy, and computing time. These empirical findings are supported by a detailed theoretical analysis of equivariant bootstrap for linear estimators