3,892 research outputs found
Fundamental Limits of Caching
Caching is a technique to reduce peak traffic rates by prefetching popular
content into memories at the end users. Conventionally, these memories are used
to deliver requested content in part from a locally cached copy rather than
through the network. The gain offered by this approach, which we term local
caching gain, depends on the local cache size (i.e, the memory available at
each individual user). In this paper, we introduce and exploit a second,
global, caching gain not utilized by conventional caching schemes. This gain
depends on the aggregate global cache size (i.e., the cumulative memory
available at all users), even though there is no cooperation among the users.
To evaluate and isolate these two gains, we introduce an
information-theoretic formulation of the caching problem focusing on its basic
structure. For this setting, we propose a novel coded caching scheme that
exploits both local and global caching gains, leading to a multiplicative
improvement in the peak rate compared to previously known schemes. In
particular, the improvement can be on the order of the number of users in the
network. Moreover, we argue that the performance of the proposed scheme is
within a constant factor of the information-theoretic optimum for all values of
the problem parameters.Comment: To appear in IEEE Transactions on Information Theor
Synchronization in wireless communications
The last decade has witnessed an immense increase of wireless communications services in order to keep pace with the ever increasing demand for higher data rates combined with higher mobility. To satisfy this demand for higher data rates, the throughput over the existing transmission media had to be increased. Several techniques were proposed to boost up the data rate: multicarrier systems to combat selective fading, ultra wide band (UWB) communications systems to share the spectrum with other users, MIMO transmissions to increase the capacity of wireless links, iteratively decodable codes (e.g., turbo codes and LDPC codes) to improve the quality of the link, cognitive radios, and so forth
Information Switching Processor (ISP) contention analysis and control
Future satellite communications, as a viable means of communications and an alternative to terrestrial networks, demand flexibility and low end-user cost. On-board switching/processing satellites potentially provide these features, allowing flexible interconnection among multiple spot beams, direct to the user communications services using very small aperture terminals (VSAT's), independent uplink and downlink access/transmission system designs optimized to user's traffic requirements, efficient TDM downlink transmission, and better link performance. A flexible switching system on the satellite in conjunction with low-cost user terminals will likely benefit future satellite network users
From Theory to Practice: Sub-Nyquist Sampling of Sparse Wideband Analog Signals
Conventional sub-Nyquist sampling methods for analog signals exploit prior
information about the spectral support. In this paper, we consider the
challenging problem of blind sub-Nyquist sampling of multiband signals, whose
unknown frequency support occupies only a small portion of a wide spectrum. Our
primary design goals are efficient hardware implementation and low
computational load on the supporting digital processing. We propose a system,
named the modulated wideband converter, which first multiplies the analog
signal by a bank of periodic waveforms. The product is then lowpass filtered
and sampled uniformly at a low rate, which is orders of magnitude smaller than
Nyquist. Perfect recovery from the proposed samples is achieved under certain
necessary and sufficient conditions. We also develop a digital architecture,
which allows either reconstruction of the analog input, or processing of any
band of interest at a low rate, that is, without interpolating to the high
Nyquist rate. Numerical simulations demonstrate many engineering aspects:
robustness to noise and mismodeling, potential hardware simplifications,
realtime performance for signals with time-varying support and stability to
quantization effects. We compare our system with two previous approaches:
periodic nonuniform sampling, which is bandwidth limited by existing hardware
devices, and the random demodulator, which is restricted to discrete multitone
signals and has a high computational load. In the broader context of Nyquist
sampling, our scheme has the potential to break through the bandwidth barrier
of state-of-the-art analog conversion technologies such as interleaved
converters.Comment: 17 pages, 12 figures, to appear in IEEE Journal of Selected Topics in
Signal Processing, the special issue on Compressed Sensin
High-resolution distributed sampling of bandlimited fields with low-precision sensors
The problem of sampling a discrete-time sequence of spatially bandlimited
fields with a bounded dynamic range, in a distributed,
communication-constrained, processing environment is addressed. A central unit,
having access to the data gathered by a dense network of fixed-precision
sensors, operating under stringent inter-node communication constraints, is
required to reconstruct the field snapshots to maximum accuracy. Both
deterministic and stochastic field models are considered. For stochastic
fields, results are established in the almost-sure sense. The feasibility of
having a flexible tradeoff between the oversampling rate (sensor density) and
the analog-to-digital converter (ADC) precision, while achieving an exponential
accuracy in the number of bits per Nyquist-interval per snapshot is
demonstrated. This exposes an underlying ``conservation of bits'' principle:
the bit-budget per Nyquist-interval per snapshot (the rate) can be distributed
along the amplitude axis (sensor-precision) and space (sensor density) in an
almost arbitrary discrete-valued manner, while retaining the same (exponential)
distortion-rate characteristics. Achievable information scaling laws for field
reconstruction over a bounded region are also derived: With N one-bit sensors
per Nyquist-interval, Nyquist-intervals, and total network
bitrate (per-sensor bitrate ), the maximum pointwise distortion goes to zero as
or . This is shown to be possible
with only nearest-neighbor communication, distributed coding, and appropriate
interpolation algorithms. For a fixed, nonzero target distortion, the number of
fixed-precision sensors and the network rate needed is always finite.Comment: 17 pages, 6 figures; paper withdrawn from IEEE Transactions on Signal
Processing and re-submitted to the IEEE Transactions on Information Theor
- …