5,033 research outputs found
Capacity Scaling in MIMO Systems with General Unitarily Invariant Random Matrices
We investigate the capacity scaling of MIMO systems with the system
dimensions. To that end, we quantify how the mutual information varies when the
number of antennas (at either the receiver or transmitter side) is altered. For
a system comprising receive and transmit antennas with , we find
the following: By removing as many receive antennas as needed to obtain a
square system (provided the channel matrices before and after the removal have
full rank) the maximum resulting loss of mutual information over all
signal-to-noise ratios (SNRs) depends only on , and the matrix of
left-singular vectors of the initial channel matrix, but not on its singular
values. In particular, if the latter matrix is Haar distributed the ergodic
rate loss is given by nats. Under
the same assumption, if with the ratio
fixed, the rate loss normalized by converges almost surely to
bits with denoting the binary entropy function. We also quantify and
study how the mutual information as a function of the system dimensions
deviates from the traditionally assumed linear growth in the minimum of the
system dimensions at high SNR.Comment: Accepted for publication in the IEEE Transactions on Information
Theor
Cores of Cooperative Games in Information Theory
Cores of cooperative games are ubiquitous in information theory, and arise
most frequently in the characterization of fundamental limits in various
scenarios involving multiple users. Examples include classical settings in
network information theory such as Slepian-Wolf source coding and multiple
access channels, classical settings in statistics such as robust hypothesis
testing, and new settings at the intersection of networking and statistics such
as distributed estimation problems for sensor networks. Cooperative game theory
allows one to understand aspects of all of these problems from a fresh and
unifying perspective that treats users as players in a game, sometimes leading
to new insights. At the heart of these analyses are fundamental dualities that
have been long studied in the context of cooperative games; for information
theoretic purposes, these are dualities between information inequalities on the
one hand and properties of rate, capacity or other resource allocation regions
on the other.Comment: 12 pages, published at
http://www.hindawi.com/GetArticle.aspx?doi=10.1155/2008/318704 in EURASIP
Journal on Wireless Communications and Networking, Special Issue on "Theory
and Applications in Multiuser/Multiterminal Communications", April 200
Capacity per Unit Energy of Fading Channels with a Peak Constraint
A discrete-time single-user scalar channel with temporally correlated
Rayleigh fading is analyzed. There is no side information at the transmitter or
the receiver. A simple expression is given for the capacity per unit energy, in
the presence of a peak constraint. The simple formula of Verdu for capacity per
unit cost is adapted to a channel with memory, and is used in the proof. In
addition to bounding the capacity of a channel with correlated fading, the
result gives some insight into the relationship between the correlation in the
fading process and the channel capacity. The results are extended to a channel
with side information, showing that the capacity per unit energy is one nat per
Joule, independently of the peak power constraint.
A continuous-time version of the model is also considered. The capacity per
unit energy subject to a peak constraint (but no bandwidth constraint) is given
by an expression similar to that for discrete time, and is evaluated for
Gauss-Markov and Clarke fading channels.Comment: Journal version of paper presented in ISIT 2003 - now accepted for
publication in IEEE Transactions on Information Theor
Elias Bound for General Distances and Stable Sets in Edge-Weighted Graphs
This paper presents an extension of the Elias bound on the minimum distance
of codes for discrete alphabets with general, possibly infinite-valued,
distances. The bound is obtained by combining a previous extension of the Elias
bound, introduced by Blahut, with an extension of a bound previously introduced
by the author which builds upon ideas of Gallager, Lov\'asz and Marton. The
result can in fact be interpreted as a unification of the Elias bound and of
Lov\'asz's bound on graph (or zero-error) capacity, both being recovered as
particular cases of the one presented here. Previous extensions of the Elias
bound by Berlekamp, Blahut and Piret are shown to be included as particular
cases of our bound. Applications to the reliability function are then
discussed.Comment: Accepted, IEEE Transaction on Information Theor
- …