294 research outputs found
Learning How to Demodulate from Few Pilots via Meta-Learning
Consider an Internet-of-Things (IoT) scenario in which devices transmit
sporadically using short packets with few pilot symbols. Each device transmits
over a fading channel and is characterized by an amplifier with a unique
non-linear transfer function. The number of pilots is generally insufficient to
obtain an accurate estimate of the end-to-end channel, which includes the
effects of fading and of the amplifier's distortion. This paper proposes to
tackle this problem using meta-learning. Accordingly, pilots from previous IoT
transmissions are used as meta-training in order to learn a demodulator that is
able to quickly adapt to new end-to-end channel conditions from few pilots.
Numerical results validate the advantages of the approach as compared to
training schemes that either do not leverage prior transmissions or apply a
standard learning algorithm on previously received data
Hypernetwork approach to Bayesian MAML
The main goal of Few-Shot learning algorithms is to enable learning from
small amounts of data. One of the most popular and elegant Few-Shot learning
approaches is Model-Agnostic Meta-Learning (MAML). The main idea behind this
method is to learn the shared universal weights of a meta-model, which are then
adapted for specific tasks. However, the method suffers from over-fitting and
poorly quantifies uncertainty due to limited data size. Bayesian approaches
could, in principle, alleviate these shortcomings by learning weight
distributions in place of point-wise weights. Unfortunately, previous
modifications of MAML are limited due to the simplicity of Gaussian posteriors,
MAML-like gradient-based weight updates, or by the same structure enforced for
universal and adapted weights.
In this paper, we propose a novel framework for Bayesian MAML called
BayesianHMAML, which employs Hypernetworks for weight updates. It learns the
universal weights point-wise, but a probabilistic structure is added when
adapted for specific tasks. In such a framework, we can use simple Gaussian
distributions or more complicated posteriors induced by Continuous Normalizing
Flows.Comment: arXiv admin note: text overlap with arXiv:2205.1574
- …