1,727 research outputs found
On the Performance of MRC Receiver with Unknown Timing Mismatch-A Large Scale Analysis
There has been extensive research on large scale multi-user multiple-input
multiple-output (MU-MIMO) systems recently. Researchers have shown that there
are great opportunities in this area, however, there are many obstacles in the
way to achieve full potential of using large number of receive antennas. One of
the main issues, which will be investigated thoroughly in this paper, is timing
asynchrony among signals of different users. Most of the works in the
literature, assume that received signals are perfectly aligned which is not
practical. We show that, neglecting the asynchrony can significantly degrade
the performance of existing designs, particularly maximum ratio combining
(MRC). We quantify the uplink achievable rates obtained by MRC receiver with
perfect channel state information (CSI) and imperfect CSI while the system is
impaired by unknown time delays among received signals. We then use these
results to design new algorithms in order to alleviate the effects of timing
mismatch. We also analyze the performance of introduced receiver design, which
is called MRC-ZF, with perfect and imperfect CSI. For performing MRC-ZF, the
only required information is the distribution of timing mismatch which
circumvents the necessity of time delay acquisition or synchronization. To
verify our analytical results, we present extensive simulation results which
thoroughly investigate the performance of the traditional MRC receiver and the
introduced MRC-ZF receiver
A stochastic approximation algorithm for stochastic semidefinite programming
Motivated by applications to multi-antenna wireless networks, we propose a
distributed and asynchronous algorithm for stochastic semidefinite programming.
This algorithm is a stochastic approximation of a continous- time matrix
exponential scheme regularized by the addition of an entropy-like term to the
problem's objective function. We show that the resulting algorithm converges
almost surely to an -approximation of the optimal solution
requiring only an unbiased estimate of the gradient of the problem's stochastic
objective. When applied to throughput maximization in wireless multiple-input
and multiple-output (MIMO) systems, the proposed algorithm retains its
convergence properties under a wide array of mobility impediments such as user
update asynchronicities, random delays and/or ergodically changing channels.
Our theoretical analysis is complemented by extensive numerical simulations
which illustrate the robustness and scalability of the proposed method in
realistic network conditions.Comment: 25 pages, 4 figure
Sparse Signal Processing Concepts for Efficient 5G System Design
As it becomes increasingly apparent that 4G will not be able to meet the
emerging demands of future mobile communication systems, the question what
could make up a 5G system, what are the crucial challenges and what are the key
drivers is part of intensive, ongoing discussions. Partly due to the advent of
compressive sensing, methods that can optimally exploit sparsity in signals
have received tremendous attention in recent years. In this paper we will
describe a variety of scenarios in which signal sparsity arises naturally in 5G
wireless systems. Signal sparsity and the associated rich collection of tools
and algorithms will thus be a viable source for innovation in 5G wireless
system design. We will discribe applications of this sparse signal processing
paradigm in MIMO random access, cloud radio access networks, compressive
channel-source network coding, and embedded security. We will also emphasize
important open problem that may arise in 5G system design, for which sparsity
will potentially play a key role in their solution.Comment: 18 pages, 5 figures, accepted for publication in IEEE Acces
Performance of direct-oversampling correlator-type receivers in chaos-based DS-CDMA systems over frequency non-selective fading channels
In this paper, we present a study on the performance of direct-oversampling correlator-type receivers in chaos-based direct-sequence code division multiple access systems over frequency non-selective fading channels. At the input, the received signal is sampled at a sampling rate higher than the chip rate. This oversampling step is used to precisely determine the delayed-signal components from multipath fading channels, which can be combined together by a correlator for the sake of increasing the SNR at its output. The main advantage of using direct-oversampling correlator-type receivers is not only their low energy consumption due to their simple structure, but also their ability to exploit the non-selective fading characteristic of multipath channels to improve the overall system performance in scenarios with limited data speeds and low energy requirements, such as low-rate wireless personal area networks. Mathematical models in discrete-time domain for the conventional transmitting side with multiple access operation, the generalized non-selective Rayleigh fading channel, and the proposed receiver are provided and described. A rough theoretical bit-error-rate (BER) expression is first derived by means of Gaussian approximation. We then define the main component in the expression and build its probability mass function through numerical computation. The final BER estimation is carried out by integrating the rough expression over possible discrete values of the PFM. In order to validate our findings, PC simulation is performed and simulated performance is compared with the corresponding estimated one. Obtained results show that the system performance get better with the increment of the number of paths in the channel.Peer ReviewedPostprint (author's final draft
Distributed stochastic optimization via matrix exponential learning
In this paper, we investigate a distributed learning scheme for a broad class
of stochastic optimization problems and games that arise in signal processing
and wireless communications. The proposed algorithm relies on the method of
matrix exponential learning (MXL) and only requires locally computable gradient
observations that are possibly imperfect and/or obsolete. To analyze it, we
introduce the notion of a stable Nash equilibrium and we show that the
algorithm is globally convergent to such equilibria - or locally convergent
when an equilibrium is only locally stable. We also derive an explicit linear
bound for the algorithm's convergence speed, which remains valid under
measurement errors and uncertainty of arbitrarily high variance. To validate
our theoretical analysis, we test the algorithm in realistic
multi-carrier/multiple-antenna wireless scenarios where several users seek to
maximize their energy efficiency. Our results show that learning allows users
to attain a net increase between 100% and 500% in energy efficiency, even under
very high uncertainty.Comment: 31 pages, 3 figure
- …