111,816 research outputs found

    Error Exponents for Variable-length Block Codes with Feedback and Cost Constraints

    Get PDF
    Variable-length block-coding schemes are investigated for discrete memoryless channels with ideal feedback under cost constraints. Upper and lower bounds are found for the minimum achievable probability of decoding error Pe,minP_{e,\min} as a function of constraints R, \AV, and τˉ\bar \tau on the transmission rate, average cost, and average block length respectively. For given RR and \AV, the lower and upper bounds to the exponent (lnPe,min)/τˉ-(\ln P_{e,\min})/\bar \tau are asymptotically equal as τˉ\bar \tau \to \infty. The resulting reliability function, limτˉ(lnPe,min)/τˉ\lim_{\bar \tau\to \infty} (-\ln P_{e,\min})/\bar \tau, as a function of RR and \AV, is concave in the pair (R, \AV) and generalizes the linear reliability function of Burnashev to include cost constraints. The results are generalized to a class of discrete-time memoryless channels with arbitrary alphabets, including additive Gaussian noise channels with amplitude and power constraints

    Variable block length coding for channels with feedback and cost constraints

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2005.This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.Includes bibliographical references (p. 95-96).Variable-decoding-time/generalized block-coding schemes are investigated for discrete memoryless channels (DMC) with perfect feedback (error free, delay free, infinite capacity) under cost constraints. For a given number of messages and average error probability, upper and lower bounds are found for expected decoding time. These coincide with each other up to a proportionality constant which approaches one in a certain asymptotic sense. A resulting reliability function is found for variable decoding time DMC's with perfect feedback under a cost constraint. The results in this work generalize Burnashev's results, to the cost constrained case.by Bariş Nakiboḡlu.S.M

    A Unified Approach for Network Information Theory

    Full text link
    In this paper, we take a unified approach for network information theory and prove a coding theorem, which can recover most of the achievability results in network information theory that are based on random coding. The final single-letter expression has a very simple form, which was made possible by many novel elements such as a unified framework that represents various network problems in a simple and unified way, a unified coding strategy that consists of a few basic ingredients but can emulate many known coding techniques if needed, and new proof techniques beyond the use of standard covering and packing lemmas. For example, in our framework, sources, channels, states and side information are treated in a unified way and various constraints such as cost and distortion constraints are unified as a single joint-typicality constraint. Our theorem can be useful in proving many new achievability results easily and in some cases gives simpler rate expressions than those obtained using conventional approaches. Furthermore, our unified coding can strictly outperform existing schemes. For example, we obtain a generalized decode-compress-amplify-and-forward bound as a simple corollary of our main theorem and show it strictly outperforms previously known coding schemes. Using our unified framework, we formally define and characterize three types of network duality based on channel input-output reversal and network flow reversal combined with packing-covering duality.Comment: 52 pages, 7 figures, submitted to IEEE Transactions on Information theory, a shorter version will appear in Proc. IEEE ISIT 201

    Writing on the Facade of RWTH ICT Cubes: Cost Constrained Geometric Huffman Coding

    Get PDF
    In this work, a coding technique called cost constrained Geometric Huffman coding (ccGhc) is developed. ccGhc minimizes the Kullback-Leibler distance between a dyadic probability mass function (pmf) and a target pmf subject to an affine inequality constraint. An analytical proof is given that when ccGhc is applied to blocks of symbols, the optimum is asymptotically achieved when the blocklength goes to infinity. The derivation of ccGhc is motivated by the problem of encoding a text to a sequence of slats subject to architectural design criteria. For the considered architectural problem, for a blocklength of 3, the codes found by ccGhc match the design criteria. For communications channels with average cost constraints, ccGhc can be used to efficiently find prefix-free modulation codes that are provably capacity achieving.Comment: 5 pages, submitted to ISWCS 201

    Finite Blocklength and Dispersion Bounds for the Arbitrarily-Varying Channel

    Full text link
    Finite blocklength and second-order (dispersion) results are presented for the arbitrarily-varying channel (AVC), a classical model wherein an adversary can transmit arbitrary signals into the channel. A novel finite blocklength achievability bound is presented, roughly analogous to the random coding union bound for non-adversarial channels. This finite blocklength bound, along with a known converse bound, is used to derive bounds on the dispersion of discrete memoryless AVCs without shared randomness, and with cost constraints on the input and the state. These bounds are tight for many channels of interest, including the binary symmetric AVC. However, the bounds are not tight if the deterministic and random code capacities differ.Comment: 7 pages, full version of paper submitted to the 2018 IEEE International Symposium on Information Theor

    Distortion Minimization in Gaussian Layered Broadcast Coding with Successive Refinement

    Full text link
    A transmitter without channel state information (CSI) wishes to send a delay-limited Gaussian source over a slowly fading channel. The source is coded in superimposed layers, with each layer successively refining the description in the previous one. The receiver decodes the layers that are supported by the channel realization and reconstructs the source up to a distortion. The expected distortion is minimized by optimally allocating the transmit power among the source layers. For two source layers, the allocation is optimal when power is first assigned to the higher layer up to a power ceiling that depends only on the channel fading distribution; all remaining power, if any, is allocated to the lower layer. For convex distortion cost functions with convex constraints, the minimization is formulated as a convex optimization problem. In the limit of a continuum of infinite layers, the minimum expected distortion is given by the solution to a set of linear differential equations in terms of the density of the fading distribution. As the bandwidth ratio b (channel uses per source symbol) tends to zero, the power distribution that minimizes expected distortion converges to the one that maximizes expected capacity. While expected distortion can be improved by acquiring CSI at the transmitter (CSIT) or by increasing diversity from the realization of independent fading paths, at high SNR the performance benefit from diversity exceeds that from CSIT, especially when b is large.Comment: Accepted for publication in IEEE Transactions on Information Theor
    corecore