4,213 research outputs found
Efficient LiDAR data compression for embedded V2I or V2V data handling
LiDAR are increasingly being used in intelligent vehicles (IV) or intelligent
transportation systems (ITS). Storage and transmission of data generated by
LiDAR sensors are one of the most challenging aspects of their deployment. In
this paper we present a method that can be used to efficiently compress LiDAR
data in order to facilitate storage and transmission in V2V or V2I
applications. This method can be used to perform lossless or lossy compression
and is specifically designed for embedded applications with low processing
power. This method is also designed to be easily applicable to existing
processing chains by keeping the structure of the data stream intact. We
benchmarked our method using several publicly available datasets and compared
it with state-of-the-art LiDAR data compression methods from the literature
Robust Spatial-spread Deep Neural Image Watermarking
Watermarking is an operation of embedding an information into an image in a
way that allows to identify ownership of the image despite applying some
distortions on it. In this paper, we presented a novel end-to-end solution for
embedding and recovering the watermark in the digital image using convolutional
neural networks. The method is based on spreading the message over the spatial
domain of the image, hence reducing the "local bits per pixel" capacity. To
obtain the model we used adversarial training and applied noiser layers between
the encoder and the decoder. Moreover, we broadened the spectrum of typically
considered attacks on the watermark and by grouping the attacks according to
their scope, we achieved high general robustness, most notably against JPEG
compression, Gaussian blurring, subsampling or resizing. To help us in the
models training we also proposed a precise differentiable approximation of
JPEG.Comment: The article was accepted on TrustCom 2020: The 19th IEEE
International Conference on Trust, Security and Privacy in Computing and
Communication
Image compression overview
Compression plays a significant role in a data storage and a transmission. If
we speak about a generall data compression, it has to be a lossless one. It
means, we are able to recover the original data 1:1 from the compressed file.
Multimedia data (images, video, sound...), are a special case. In this area, we
can use something called a lossy compression. Our main goal is not to recover
data 1:1, but only keep them visually similar. This article is about an image
compression, so we will be interested only in image compression. For a human
eye, it is not a huge difference, if we recover RGB color with values
[150,140,138] instead of original [151,140,137]. The magnitude of a difference
determines the loss rate of the compression. The bigger difference usually
means a smaller file, but also worse image quality and noticable differences
from the original image. We want to cover compression techniques mainly from
the last decade. Many of them are variations of existing ones, only some of
them uses new principes
Exploiting Errors for Efficiency: A Survey from Circuits to Algorithms
When a computational task tolerates a relaxation of its specification or when
an algorithm tolerates the effects of noise in its execution, hardware,
programming languages, and system software can trade deviations from correct
behavior for lower resource usage. We present, for the first time, a synthesis
of research results on computing systems that only make as many errors as their
users can tolerate, from across the disciplines of computer aided design of
circuits, digital system design, computer architecture, programming languages,
operating systems, and information theory.
Rather than over-provisioning resources at each layer to avoid errors, it can
be more efficient to exploit the masking of errors occurring at one layer which
can prevent them from propagating to a higher layer. We survey tradeoffs for
individual layers of computing systems from the circuit level to the operating
system level and illustrate the potential benefits of end-to-end approaches
using two illustrative examples. To tie together the survey, we present a
consistent formalization of terminology, across the layers, which does not
significantly deviate from the terminology traditionally used by research
communities in their layer of focus.Comment: 35 page
DSSLIC: Deep Semantic Segmentation-based Layered Image Compression
Deep learning has revolutionized many computer vision fields in the last few
years, including learning-based image compression. In this paper, we propose a
deep semantic segmentation-based layered image compression (DSSLIC) framework
in which the semantic segmentation map of the input image is obtained and
encoded as the base layer of the bit-stream. A compact representation of the
input image is also generated and encoded as the first enhancement layer. The
segmentation map and the compact version of the image are then employed to
obtain a coarse reconstruction of the image. The residual between the input and
the coarse reconstruction is additionally encoded as another enhancement layer.
Experimental results show that the proposed framework outperforms the
H.265/HEVC-based BPG and other codecs in both PSNR and MS-SSIM metrics across a
wide range of bit rates in RGB domain. Besides, since semantic segmentation map
is included in the bit-stream, the proposed scheme can facilitate many other
tasks such as image search and object-based adaptive image compression.Comment: - More Experimental results adde
Learning based Facial Image Compression with Semantic Fidelity Metric
Surveillance and security scenarios usually require high efficient facial
image compression scheme for face recognition and identification. While either
traditional general image codecs or special facial image compression schemes
only heuristically refine codec separately according to face verification
accuracy metric. We propose a Learning based Facial Image Compression (LFIC)
framework with a novel Regionally Adaptive Pooling (RAP) module whose
parameters can be automatically optimized according to gradient feedback from
an integrated hybrid semantic fidelity metric, including a successfully
exploration to apply Generative Adversarial Network (GAN) as metric directly in
image compression scheme. The experimental results verify the framework's
efficiency by demonstrating performance improvement of 71.41%, 48.28% and
52.67% bitrate saving separately over JPEG2000, WebP and neural network-based
codecs under the same face verification accuracy distortion metric. We also
evaluate LFIC's superior performance gain compared with latest specific facial
image codecs. Visual experiments also show some interesting insight on how LFIC
can automatically capture the information in critical areas based on semantic
distortion metrics for optimized compression, which is quite different from the
heuristic way of optimization in traditional image compression algorithms.Comment: Accepted by Neurocomputin
Implicit Dual-domain Convolutional Network for Robust Color Image Compression Artifact Reduction
Several dual-domain convolutional neural network-based methods show
outstanding performance in reducing image compression artifacts. However, they
suffer from handling color images because the compression processes for
gray-scale and color images are completely different. Moreover, these methods
train a specific model for each compression quality and require multiple models
to achieve different compression qualities. To address these problems, we
proposed an implicit dual-domain convolutional network (IDCN) with the pixel
position labeling map and the quantization tables as inputs. Specifically, we
proposed an extractor-corrector framework-based dual-domain correction unit
(DCU) as the basic component to formulate the IDCN. A dense block was
introduced to improve the performance of extractor in DRU. The implicit
dual-domain translation allows the IDCN to handle color images with the
discrete cosine transform (DCT)-domain priors. A flexible version of IDCN
(IDCN-f) was developed to handle a wide range of compression qualities.
Experiments for both objective and subjective evaluations on benchmark datasets
show that IDCN is superior to the state-of-the-art methods and IDCN-f exhibits
excellent abilities to handle a wide range of compression qualities with little
performance sacrifice and demonstrates great potential for practical
applications.Comment: accepted by IEEE Transactions on Circuits and Systems for Video
Technology(T-CSVT
Deep Generative Models for Distribution-Preserving Lossy Compression
We propose and study the problem of distribution-preserving lossy
compression. Motivated by recent advances in extreme image compression which
allow to maintain artifact-free reconstructions even at very low bitrates, we
propose to optimize the rate-distortion tradeoff under the constraint that the
reconstructed samples follow the distribution of the training data. The
resulting compression system recovers both ends of the spectrum: On one hand,
at zero bitrate it learns a generative model of the data, and at high enough
bitrates it achieves perfect reconstruction. Furthermore, for intermediate
bitrates it smoothly interpolates between learning a generative model of the
training data and perfectly reconstructing the training samples. We study
several methods to approximately solve the proposed optimization problem,
including a novel combination of Wasserstein GAN and Wasserstein Autoencoder,
and present an extensive theoretical and empirical characterization of the
proposed compression systems.Comment: NIPS 2018. Code: https://github.com/mitscha/dplc . Changes w.r.t. v1:
Some clarifications in the text and additional numerical result
To Compress or Not To Compress: Processing vs Transmission Tradeoffs for Energy Constrained Sensor Networking
In the past few years, lossy compression has been widely applied in the field
of wireless sensor networks (WSN), where energy efficiency is a crucial concern
due to the constrained nature of the transmission devices. Often, the common
thinking among researchers and implementers is that compression is always a
good choice, because the major source of energy consumption in a sensor node
comes from the transmission of the data. Lossy compression is deemed a viable
solution as the imperfect reconstruction of the signal is often acceptable in
WSN. In this paper, we thoroughly review a number of lossy compression methods
from the literature, and analyze their performance in terms of compression
efficiency, computational complexity and energy consumption. We consider two
different scenarios, namely, wireless and underwater communications, and show
that signal compression may or may not help in the reduction of the overall
energy consumption, depending on factors such as the compression algorithm, the
signal statistics and the hardware characteristics, i.e., micro-controller and
transmission technology. The lesson that we have learned, is that signal
compression may in fact provide some energy savings. However, its usage should
be carefully evaluated, as in quite a few cases processing and transmission
costs are of the same order of magnitude, whereas, in some other cases, the
former may even dominate the latter. In this paper, we show quantitative
comparisons to assess these tradeoffs in the above mentioned scenarios.
Finally, we provide formulas, obtained through numerical fittings, to gauge
computational complexity, overall energy consumption and signal representation
accuracy for the best performing algorithms as a function of the most relevant
system parameters
Five Modulus Method For Image Compression
Data is compressed by reducing its redundancy, but this also makes the data
less reliable, more prone to errors. In this paper a novel approach of image
compression based on a new method that has been created for image compression
which is called Five Modulus Method (FMM). The new method consists of
converting each pixel value in an 8-by-8 block into a multiple of 5 for each of
the R, G and B arrays. After that, the new values could be divided by 5 to get
new values which are 6-bit length for each pixel and it is less in storage
space than the original value which is 8-bits. Also, a new protocol for
compression of the new values as a stream of bits has been presented that gives
the opportunity to store and transfer the new compressed image easily.Comment: 10 pages, 2 figures, 9 table
- …