18 research outputs found

    Strong Coordination over Noisy Channels: Is Separation Sufficient?

    Full text link
    We study the problem of strong coordination of actions of two agents XX and YY that communicate over a noisy communication channel such that the actions follow a given joint probability distribution. We propose two novel schemes for this noisy strong coordination problem, and derive inner bounds for the underlying strong coordination capacity region. The first scheme is a joint coordination-channel coding scheme that utilizes the randomness provided by the communication channel to reduce the local randomness required in generating the action sequence at agent YY. The second scheme exploits separate coordination and channel coding where local randomness is extracted from the channel after decoding. Finally, we present an example in which the joint scheme is able to outperform the separate scheme in terms of coordination rate.Comment: 9 pages, 4 figures. An extended version of a paper accepted for the IEEE International Symposium on Information Theory (ISIT), 201

    Strong Coordination over Noisy Channels: Is Separation Sufficient?

    Full text link
    We study the problem of strong coordination of actions of two agents XX and YY that communicate over a noisy communication channel such that the actions follow a given joint probability distribution. We propose two novel schemes for this noisy strong coordination problem, and derive inner bounds for the underlying strong coordination capacity region. The first scheme is a joint coordination-channel coding scheme that utilizes the randomness provided by the communication channel to reduce the local randomness required in generating the action sequence at agent YY. The second scheme exploits separate coordination and channel coding where local randomness is extracted from the channel after decoding. Finally, we present an example in which the joint scheme is able to outperform the separate scheme in terms of coordination rate.Comment: 9 pages, 4 figures. An extended version of a paper accepted for the IEEE International Symposium on Information Theory (ISIT), 201

    Data Processing Bounds for Scalar Lossy Source Codes with Side Information at the Decoder

    Full text link
    In this paper, we introduce new lower bounds on the distortion of scalar fixed-rate codes for lossy compression with side information available at the receiver. These bounds are derived by presenting the relevant random variables as a Markov chain and applying generalized data processing inequalities a la Ziv and Zakai. We show that by replacing the logarithmic function with other functions, in the data processing theorem we formulate, we obtain new lower bounds on the distortion of scalar coding with side information at the decoder. The usefulness of these results is demonstrated for uniform sources and the convex function Q(t)=t1αQ(t)=t^{1-\alpha}, α>1\alpha>1. The bounds in this case are shown to be better than one can obtain from the Wyner-Ziv rate-distortion function.Comment: 35 pages, 9 figure

    Simulation of a Channel with Another Channel

    Full text link
    In this paper, we study the problem of simulating a DMC channel from another DMC channel under an average-case and an exact model. We present several achievability and infeasibility results, with tight characterizations in special cases. In particular for the exact model, we fully characterize when a BSC channel can be simulated from a BEC channel when there is no shared randomness. We also provide infeasibility and achievability results for simulation of a binary channel from another binary channel in the case of no shared randomness. To do this, we use properties of R\'enyi capacity of a given order. We also introduce a notion of "channel diameter" which is shown to be additive and satisfy a data processing inequality.Comment: 31 pages, 10 figures, and some parts of this work were published at ITW 201

    Publicness, Privacy and Confidentiality in the Single-Serving Quantum Broadcast Channel

    Full text link
    The 2-receiver broadcast channel is studied: a network with three parties where the transmitter and one of the receivers are the primarily involved parties and the other receiver considered as third party. The messages that are determined to be communicated are classified into public, private and confidential based on the information they convey. The public message contains information intended for both parties and is required to be decoded correctly by both of them, the private message is intended for the primary party only, however, there is no secrecy requirement imposed upon it meaning that it can possibly be exposed to the third party and finally the confidential message containing information intended exclusively for the primary party such that this information must be kept completely secret from the other receiver. A trade-off arises between the rates of the three messages, when one of the rates is high, the other rates may need to be reduced to guarantee the reliable transmission of all three messages. The encoder performs the necessary equivocation by virtue of dummy random numbers whose rate is assumed to be limited and should be considered in the trade-off as well. We study this trade-off in the one-shot regime of a quantum broadcast channel by providing achievability and (weak) converse regions. In the achievability, we prove and use a conditional version of the convex-split lemma as well as position-based decoding. By studying the asymptotic behaviour of our bounds, we will recover several well-known asymptotic results in the literature.Comment: 23 pages, 1 figure, journa

    Distributed Channel Synthesis

    Full text link
    Two familiar notions of correlation are rediscovered as the extreme operating points for distributed synthesis of a discrete memoryless channel, in which a stochastic channel output is generated based on a compressed description of the channel input. Wyner's common information is the minimum description rate needed. However, when common randomness independent of the input is available, the necessary description rate reduces to Shannon's mutual information. This work characterizes the optimal trade-off between the amount of common randomness used and the required rate of description. We also include a number of related derivations, including the effect of limited local randomness, rate requirements for secrecy, applications to game theory, and new insights into common information duality. Our proof makes use of a soft covering lemma, known in the literature for its role in quantifying the resolvability of a channel. The direct proof (achievability) constructs a feasible joint distribution over all parts of the system using a soft covering, from which the behavior of the encoder and decoder is inferred, with no explicit reference to joint typicality or binning. Of auxiliary interest, this work also generalizes and strengthens this soft covering tool.Comment: To appear in IEEE Trans. on Information Theory (submitted Aug., 2012, accepted July, 2013), 26 pages, using IEEEtran.cl
    corecore