1,554 research outputs found
Statistical Atmospheric Parameter Retrieval Largely Benefits from Spatial-Spectral Image Compression
The Infrared Atmospheric Sounding Interferometer
(IASI) is flying on board of the Metop satellite series, which is
part of the EUMETSAT Polar System (EPS). Products obtained
from IASI data represent a significant improvement in the
accuracy and quality of the measurements used for meteorological models. Notably, IASI collects rich spectral information to
derive temperature and moisture profiles –among other relevant
trace gases–, essential for atmospheric forecasts and for the
understanding of weather. Here, we investigate the impact of
near-lossless and lossy compression on IASI L1C data when
statistical retrieval algorithms are later applied. We search for
those compression ratios that yield a positive impact on the
accuracy of the statistical retrievals. The compression techniques
help reduce certain amount of noise on the original data and,
at the same time, incorporate spatial-spectral feature relations in
an indirect way without increasing the computational complexity.
We observed that compressing images, at relatively low bitrates, improves results in predicting temperature and dew point
temperature, and we advocate that some amount of compression
prior to model inversion is beneficial. This research can benefit
the development of current and upcoming retrieval chains in
infrared sounding and hyperspectral sensors
Further results on dissimilarity spaces for hyperspectral images RF-CBIR
Content-Based Image Retrieval (CBIR) systems are powerful search tools in
image databases that have been little applied to hyperspectral images.
Relevance feedback (RF) is an iterative process that uses machine learning
techniques and user's feedback to improve the CBIR systems performance. We
pursued to expand previous research in hyperspectral CBIR systems built on
dissimilarity functions defined either on spectral and spatial features
extracted by spectral unmixing techniques, or on dictionaries extracted by
dictionary-based compressors. These dissimilarity functions were not suitable
for direct application in common machine learning techniques. We propose to use
a RF general approach based on dissimilarity spaces which is more appropriate
for the application of machine learning algorithms to the hyperspectral
RF-CBIR. We validate the proposed RF method for hyperspectral CBIR systems over
a real hyperspectral dataset.Comment: In Pattern Recognition Letters (2013
Practical Full Resolution Learned Lossless Image Compression
We propose the first practical learned lossless image compression system,
L3C, and show that it outperforms the popular engineered codecs, PNG, WebP and
JPEG 2000. At the core of our method is a fully parallelizable hierarchical
probabilistic model for adaptive entropy coding which is optimized end-to-end
for the compression task. In contrast to recent autoregressive discrete
probabilistic models such as PixelCNN, our method i) models the image
distribution jointly with learned auxiliary representations instead of
exclusively modeling the image distribution in RGB space, and ii) only requires
three forward-passes to predict all pixel probabilities instead of one for each
pixel. As a result, L3C obtains over two orders of magnitude speedups when
sampling compared to the fastest PixelCNN variant (Multiscale-PixelCNN).
Furthermore, we find that learning the auxiliary representation is crucial and
outperforms predefined auxiliary representations such as an RGB pyramid
significantly.Comment: Updated preprocessing and Table 1, see A.1 in supplementary. Code and
models: https://github.com/fab-jul/L3C-PyTorc
Statistical atmospheric parameter retrieval largely benefits from spatial-spectral image compression
The infrared atmospheric sounding interferometer (IASI) is flying on board of the Metop satellite series, which is part of the EUMETSAT Polar System. Products obtained from IASI data represent a significant improvement in the accuracy and quality of the measurements used for meteorological models. Notably, the IASI collects rich spectral information to derive temperature and moisture profiles, among other relevant trace gases, essential for atmospheric forecasts and for the understanding of weather. Here, we investigate the impact of near-lossless and lossy compression on IASI L1C data when statistical retrieval algorithms are later applied. We search for those compression ratios that yield a positive impact on the accuracy of the statistical retrievals. The compression techniques help reduce certain amount of noise on the original data and, at the same time, incorporate spatial-spectral feature relations in an indirect way without increasing the computational complexity. We observed that compressing images, at relatively low bit rates, improves results in predicting temperature and dew point temperature, and we advocate that some amount of compression prior to model inversion is beneficial. This research can benefit the development of current and upcoming retrieval chains in infrared sounding and hyperspectral sensors
Ultrafast processing of pixel detector data with machine learning frameworks
Modern photon science performed at high repetition rate free-electron laser
(FEL) facilities and beyond relies on 2D pixel detectors operating at
increasing frequencies (towards 100 kHz at LCLS-II) and producing rapidly
increasing amounts of data (towards TB/s). This data must be rapidly stored for
offline analysis and summarized in real time. While at LCLS all raw data has
been stored, at LCLS-II this would lead to a prohibitive cost; instead,
enabling real time processing of pixel detector raw data allows reducing the
size and cost of online processing, offline processing and storage by orders of
magnitude while preserving full photon information, by taking advantage of the
compressibility of sparse data typical for LCLS-II applications. We
investigated if recent developments in machine learning are useful in data
processing for high speed pixel detectors and found that typical deep learning
models and autoencoder architectures failed to yield useful noise reduction
while preserving full photon information, presumably because of the very
different statistics and feature sets between computer vision and radiation
imaging. However, we redesigned in Tensorflow mathematically equivalent
versions of the state-of-the-art, "classical" algorithms used at LCLS. The
novel Tensorflow models resulted in elegant, compact and hardware agnostic
code, gaining 1 to 2 orders of magnitude faster processing on an inexpensive
consumer GPU, reducing by 3 orders of magnitude the projected cost of online
analysis at LCLS-II. Computer vision a decade ago was dominated by hand-crafted
filters; their structure inspired the deep learning revolution resulting in
modern deep convolutional networks; similarly, our novel Tensorflow filters
provide inspiration for designing future deep learning architectures for
ultrafast and efficient processing and classification of pixel detector images
at FEL facilities.Comment: 9 pages, 9 figure
Healing failures and improving generalization in deep generative modelling
Deep generative modeling is a crucial and rapidly developing area of machine learning, with numerous potential applications, including data generation, anomaly detection, data compression, and more. Despite the significant empirical success of many generative models, some limitations still need to be addressed to improve their performance in certain cases. This thesis focuses on understanding the limitations of generative modeling in common scenarios and proposes corresponding techniques to alleviate these limitations and improve performance in practical generative modeling applications. Specifically, the thesis is divided into two sub-topics: one focusing on the training and the other on the generalization of generative models. A brief introduction to each sub-topic is provided below.
Generative models are typically trained by optimizing their fit to the data distribution. This is achieved by minimizing a statistical divergence between the model and data distributions. However, there are cases where these divergences fail to accurately capture the differences between the model and data distributions, resulting in poor performance of the trained model. In the first part of the thesis, we discuss the two situations where the classic divergences are ineffective for training the models:
1. KL divergence fails to train implicit models for manifold modeling tasks.
2. Fisher divergence cannot distinguish the mixture proportions for modeling target multi-modality distribution.
For both failure modes, we investigate the theoretical reasons underlying the failures of KL and Fisher divergences in modelling certain types of data distributions. We propose techniques that address the limitations of these divergences, enabling more reliable estimation of the underlying data distributions.
While the generalization of classification or regression models has been extensively studied in machine learning, the generalization of generative models is a relatively under-explored area. In the second part of this thesis, we aim to address this gap by investigating the generalization properties of generative models.
Specifically, we investigate two generalization scenarios:
1. In-distribution (ID) generalization of probabilistic models, where the test data and the training data are from the same distribution.
2. Out-of-distribution (OOD) generalization of probabilistic models, where the test data and the training data can come from different distributions.
In the context of ID generalization, our emphasis rests on the Variational Auto-Encoder (VAE) model, and for OOD generalization, we primarily explore autoregressive models.
By studying the generalization properties of the models, we demonstrate how to design new models or training criteria that improve the performance of practical applications, such as lossless compression and OOD detection.
The findings of this thesis shed light on the intricate challenges faced by generative models in both training and generalization scenarios. Our investigations into the inefficacies of classic divergences like KL and Fisher highlight the importance of tailoring modeling techniques to the specific characteristics of data distributions. Additionally, by delving into the generalization aspects of generative models, this work pioneers insights into the ID and OOD scenarios, a domain not extensively covered in current literature. Collectively, the insights and techniques presented in this thesis provide valuable contributions to the community, fostering an environment for the development of more robust and reliable generative models. It's our hope that these take-home messages will serve as a foundation for future research and applications in the realm of deep generative modeling
- …