294 research outputs found

    Learning How to Demodulate from Few Pilots via Meta-Learning

    Full text link
    Consider an Internet-of-Things (IoT) scenario in which devices transmit sporadically using short packets with few pilot symbols. Each device transmits over a fading channel and is characterized by an amplifier with a unique non-linear transfer function. The number of pilots is generally insufficient to obtain an accurate estimate of the end-to-end channel, which includes the effects of fading and of the amplifier's distortion. This paper proposes to tackle this problem using meta-learning. Accordingly, pilots from previous IoT transmissions are used as meta-training in order to learn a demodulator that is able to quickly adapt to new end-to-end channel conditions from few pilots. Numerical results validate the advantages of the approach as compared to training schemes that either do not leverage prior transmissions or apply a standard learning algorithm on previously received data

    Hypernetwork approach to Bayesian MAML

    Full text link
    The main goal of Few-Shot learning algorithms is to enable learning from small amounts of data. One of the most popular and elegant Few-Shot learning approaches is Model-Agnostic Meta-Learning (MAML). The main idea behind this method is to learn the shared universal weights of a meta-model, which are then adapted for specific tasks. However, the method suffers from over-fitting and poorly quantifies uncertainty due to limited data size. Bayesian approaches could, in principle, alleviate these shortcomings by learning weight distributions in place of point-wise weights. Unfortunately, previous modifications of MAML are limited due to the simplicity of Gaussian posteriors, MAML-like gradient-based weight updates, or by the same structure enforced for universal and adapted weights. In this paper, we propose a novel framework for Bayesian MAML called BayesianHMAML, which employs Hypernetworks for weight updates. It learns the universal weights point-wise, but a probabilistic structure is added when adapted for specific tasks. In such a framework, we can use simple Gaussian distributions or more complicated posteriors induced by Continuous Normalizing Flows.Comment: arXiv admin note: text overlap with arXiv:2205.1574
    corecore