24 research outputs found
Deep Generative Model for Simultaneous Range Error Mitigation and Environment Identification
Received waveforms contain rich information for both range information and
environment semantics. However, its full potential is hard to exploit under
multipath and non-line-of-sight conditions. This paper proposes a deep
generative model (DGM) for simultaneous range error mitigation and environment
identification. In particular, we present a Bayesian model for the generative
process of the received waveform composed by latent variables for both
range-related features and environment semantics. The simultaneous range error
mitigation and environment identification is interpreted as an inference
problem based on the DGM, and implemented in a unique end-to-end learning
scheme. Comprehensive experiments on a general Ultra-wideband dataset
demonstrate the superior performance on range error mitigation, scalability to
different environments, and novel capability on simultaneous environment
identification.Comment: 6 pages, 5 figures, Published in: 2021 IEEE Global Communications
Conference (GLOBECOM
A Semi-Supervised Learning Approach for Ranging Error Mitigation Based on UWB Waveform
Localization systems based on ultra-wide band (UWB) measurements can have
unsatisfactory performance in harsh environments due to the presence of
non-line-of-sight (NLOS) errors. Learning-based methods for error mitigation
have shown great performance improvement via directly exploiting the wideband
waveform instead of handcrafted features. However, these methods require data
samples fully labeled with actual measurement errors for training, which leads
to time-consuming data collection. In this paper, we propose a semi-supervised
learning method based on variational Bayes for UWB ranging error mitigation.
Combining deep learning techniques and statistic tools, our method can
efficiently accumulate knowledge from both labeled and unlabeled data samples.
Extensive experiments illustrate the effectiveness of the proposed method under
different supervision rates, and the superiority compared to other fully
supervised methods even at a low supervision rate.Comment: 5 pages, 3 figures, Published in: MILCOM 2021 - 2021 IEEE Military
Communications Conference (MILCOM
Minimax risk classifiers with 0-1 loss
Supervised classification techniques use training samples to learn a
classification rule with small expected 0-1 loss (error probability).
Conventional methods enable tractable learning and provide out-of-sample
generalization by using surrogate losses instead of the 0-1 loss and
considering specific families of rules (hypothesis classes). This paper
presents minimax risk classifiers (MRCs) that minimize the worst-case 0-1 loss
over general classification rules and provide tight performance guarantees at
learning. We show that MRCs are strongly universally consistent using feature
mappings given by characteristic kernels. The paper also proposes efficient
optimization techniques for MRC learning and shows that the methods presented
can provide accurate classification together with tight performance guarantees
in practice
Variational Bayesian Framework for Advanced Image Generation with Domain-Related Variables
Deep generative models (DGMs) and their conditional counterparts provide a
powerful ability for general-purpose generative modeling of data distributions.
However, it remains challenging for existing methods to address advanced
conditional generative problems without annotations, which can enable multiple
applications like image-to-image translation and image editing. We present a
unified Bayesian framework for such problems, which introduces an inference
stage on latent variables within the learning process. In particular, we
propose a variational Bayesian image translation network (VBITN) that enables
multiple image translation and editing tasks. Comprehensive experiments show
the effectiveness of our method on unsupervised image-to-image translation, and
demonstrate the novel advanced capabilities for semantic editing and mixed
domain translation.Comment: 5 pages, 2 figures
Minimax Forward and Backward Learning of Evolving Tasks with Performance Guarantees
For a sequence of classification tasks that arrive over time, it is common
that tasks are evolving in the sense that consecutive tasks often have a higher
similarity. The incremental learning of a growing sequence of tasks holds
promise to enable accurate classification even with few samples per task by
leveraging information from all the tasks in the sequence (forward and backward
learning). However, existing techniques developed for continual learning and
concept drift adaptation are either designed for tasks with time-independent
similarities or only aim to learn the last task in the sequence. This paper
presents incremental minimax risk classifiers (IMRCs) that effectively exploit
forward and backward learning and account for evolving tasks. In addition, we
analytically characterize the performance improvement provided by forward and
backward learning in terms of the tasks' expected quadratic change and the
number of tasks. The experimental evaluation shows that IMRCs can result in a
significant performance improvement, especially for reduced sample sizes
Double-Weighting for Covariate Shift Adaptation
Supervised learning is often affected by a covariate shift in which the
marginal distributions of instances (covariates ) of training and testing
samples and are different
but the label conditionals coincide. Existing approaches address such covariate
shift by either using the ratio
to weight training samples
(reweighted methods) or using the ratio
to weight testing samples
(robust methods). However, the performance of such approaches can be poor under
support mismatch or when the above ratios take large values. We propose a
minimax risk classification (MRC) approach for covariate shift adaptation that
avoids such limitations by weighting both training and testing samples. In
addition, we develop effective techniques that obtain both sets of weights and
generalize the conventional kernel mean matching method. We provide novel
generalization bounds for our method that show a significant increase in the
effective sample size compared with reweighted methods. The proposed method
also achieves enhanced classification performance in both synthetic and
empirical experiments
On the Performance Limits of Map-Aware Localization
Establishing bounds on the accuracy achievable by localization techniques represents a fundamental technical issue. Bounds on localization accuracy have been derived for cases in which the position of an agent is estimated on the basis of a set of observations and, possibly, of some a priori information related to them (e.g., information about anchor positions and properties of the communication channel). In this paper, new bounds are derived under the assumption that the localization system is map-aware, i.e., it can benefit not only from the availability of observations, but also from the a priori knowledge provided by the map of the environment where it operates. Our results show that: a) map-aware estimation accuracy can be related to some features of the map (e.g., its shape and area) even though, in general, the relation is complicated; b) maps are really useful in the presence of some combination of low SNRs and specific geometrical features of the map (e.g., the size of obstructions); c) in most cases, there is no need of refined maps since additional details do not improve estimation accuracy.United States. Air Force Office of Scientific Research (Grant FA9550-12-0287)United States. Office of Naval Research (Grant N00014-11-1-0397)Massachusetts Institute of Technology. Institute for Soldier Nanotechnologie