69,024 research outputs found
Learning Optimal Deep Projection of F-FDG PET Imaging for Early Differential Diagnosis of Parkinsonian Syndromes
Several diseases of parkinsonian syndromes present similar symptoms at early
stage and no objective widely used diagnostic methods have been approved until
now. Positron emission tomography (PET) with F-FDG was shown to be able
to assess early neuronal dysfunction of synucleinopathies and tauopathies.
Tensor factorization (TF) based approaches have been applied to identify
characteristic metabolic patterns for differential diagnosis. However, these
conventional dimension-reduction strategies assume linear or multi-linear
relationships inside data, and are therefore insufficient to distinguish
nonlinear metabolic differences between various parkinsonian syndromes. In this
paper, we propose a Deep Projection Neural Network (DPNN) to identify
characteristic metabolic pattern for early differential diagnosis of
parkinsonian syndromes. We draw our inspiration from the existing TF methods.
The network consists of a (i) compression part: which uses a deep network to
learn optimal 2D projections of 3D scans, and a (ii) classification part: which
maps the 2D projections to labels. The compression part can be pre-trained
using surplus unlabelled datasets. Also, as the classification part operates on
these 2D projections, it can be trained end-to-end effectively with limited
labelled data, in contrast to 3D approaches. We show that DPNN is more
effective in comparison to existing state-of-the-art and plausible baselines.Comment: 8 pages, 3 figures, conference, MICCAI DLMIA, 201
Semantic Perceptual Image Compression using Deep Convolution Networks
It has long been considered a significant problem to improve the visual
quality of lossy image and video compression. Recent advances in computing
power together with the availability of large training data sets has increased
interest in the application of deep learning cnns to address image recognition
and image processing tasks. Here, we present a powerful cnn tailored to the
specific task of semantic image understanding to achieve higher visual quality
in lossy compression. A modest increase in complexity is incorporated to the
encoder which allows a standard, off-the-shelf jpeg decoder to be used. While
jpeg encoding may be optimized for generic images, the process is ultimately
unaware of the specific content of the image to be compressed. Our technique
makes jpeg content-aware by designing and training a model to identify multiple
semantic regions in a given image. Unlike object detection techniques, our
model does not require labeling of object positions and is able to identify
objects in a single pass. We present a new cnn architecture directed
specifically to image compression, which generates a map that highlights
semantically-salient regions so that they can be encoded at higher quality as
compared to background regions. By adding a complete set of features for every
class, and then taking a threshold over the sum of all feature activations, we
generate a map that highlights semantically-salient regions so that they can be
encoded at a better quality compared to background regions. Experiments are
presented on the Kodak PhotoCD dataset and the MIT Saliency Benchmark dataset,
in which our algorithm achieves higher visual quality for the same compressed
size.Comment: Accepted to Data Compression Conference, 11 pages, 5 figure
Generative Compression
Traditional image and video compression algorithms rely on hand-crafted
encoder/decoder pairs (codecs) that lack adaptability and are agnostic to the
data being compressed. Here we describe the concept of generative compression,
the compression of data using generative models, and suggest that it is a
direction worth pursuing to produce more accurate and visually pleasing
reconstructions at much deeper compression levels for both image and video
data. We also demonstrate that generative compression is orders-of-magnitude
more resilient to bit error rates (e.g. from noisy wireless channels) than
traditional variable-length coding schemes
Linear and Geometric Mixtures - Analysis
Linear and geometric mixtures are two methods to combine arbitrary models in
data compression. Geometric mixtures generalize the empirically well-performing
PAQ7 mixture. Both mixture schemes rely on weight vectors, which heavily
determine their performance. Typically weight vectors are identified via Online
Gradient Descent. In this work we show that one can obtain strong code length
bounds for such a weight estimation scheme. These bounds hold for arbitrary
input sequences. For this purpose we introduce the class of nice mixtures and
analyze how Online Gradient Descent with a fixed step size combined with a nice
mixture performs. These results translate to linear and geometric mixtures,
which are nice, as we show. The results hold for PAQ7 mixtures as well, thus we
provide the first theoretical analysis of PAQ7.Comment: Data Compression Conference (DCC) 201
- …