38,106 research outputs found
An Average-case Analysis of the Gaussian Algorithm for Lattice Reduction
The Gaussian algorithm for lattice reduction in dimension 2 is analysed under its standard version. It is found that, when applied to random inputs in a continuous model, the complexity is constant on average, the probability distribution decays geometrically, and the dynamics is characterized by a conditional invariant measure. The proofs make use of connections between lattice reduction, continued fractions, continuants, and functional operators. Analysis in the discrete model and detailed numerical data are also presented
Decoding by Sampling: A Randomized Lattice Algorithm for Bounded Distance Decoding
Despite its reduced complexity, lattice reduction-aided decoding exhibits a
widening gap to maximum-likelihood (ML) performance as the dimension increases.
To improve its performance, this paper presents randomized lattice decoding
based on Klein's sampling technique, which is a randomized version of Babai's
nearest plane algorithm (i.e., successive interference cancelation (SIC)). To
find the closest lattice point, Klein's algorithm is used to sample some
lattice points and the closest among those samples is chosen. Lattice reduction
increases the probability of finding the closest lattice point, and only needs
to be run once during pre-processing. Further, the sampling can operate very
efficiently in parallel. The technical contribution of this paper is two-fold:
we analyze and optimize the decoding radius of sampling decoding resulting in
better error performance than Klein's original algorithm, and propose a very
efficient implementation of random rounding. Of particular interest is that a
fixed gain in the decoding radius compared to Babai's decoding can be achieved
at polynomial complexity. The proposed decoder is useful for moderate
dimensions where sphere decoding becomes computationally intensive, while
lattice reduction-aided decoding starts to suffer considerable loss. Simulation
results demonstrate near-ML performance is achieved by a moderate number of
samples, even if the dimension is as high as 32
Decoding by Embedding: Correct Decoding Radius and DMT Optimality
The closest vector problem (CVP) and shortest (nonzero) vector problem (SVP)
are the core algorithmic problems on Euclidean lattices. They are central to
the applications of lattices in many problems of communications and
cryptography. Kannan's \emph{embedding technique} is a powerful technique for
solving the approximate CVP, yet its remarkable practical performance is not
well understood. In this paper, the embedding technique is analyzed from a
\emph{bounded distance decoding} (BDD) viewpoint. We present two complementary
analyses of the embedding technique: We establish a reduction from BDD to
Hermite SVP (via unique SVP), which can be used along with any Hermite SVP
solver (including, among others, the Lenstra, Lenstra and Lov\'asz (LLL)
algorithm), and show that, in the special case of LLL, it performs at least as
well as Babai's nearest plane algorithm (LLL-aided SIC). The former analysis
helps to explain the folklore practical observation that unique SVP is easier
than standard approximate SVP. It is proven that when the LLL algorithm is
employed, the embedding technique can solve the CVP provided that the noise
norm is smaller than a decoding radius , where
is the minimum distance of the lattice, and . This
substantially improves the previously best known correct decoding bound . Focusing on the applications of BDD to decoding of
multiple-input multiple-output (MIMO) systems, we also prove that BDD of the
regularized lattice is optimal in terms of the diversity-multiplexing gain
tradeoff (DMT), and propose practical variants of embedding decoding which
require no knowledge of the minimum distance of the lattice and/or further
improve the error performance.Comment: To appear in IEEE Transactions on Information Theor
Solving the Closest Vector Problem in Time--- The Discrete Gaussian Strikes Again!
We give a -time and space randomized algorithm for solving the
exact Closest Vector Problem (CVP) on -dimensional Euclidean lattices. This
improves on the previous fastest algorithm, the deterministic
-time and -space algorithm of
Micciancio and Voulgaris.
We achieve our main result in three steps. First, we show how to modify the
sampling algorithm from [ADRS15] to solve the problem of discrete Gaussian
sampling over lattice shifts, , with very low parameters. While the
actual algorithm is a natural generalization of [ADRS15], the analysis uses
substantial new ideas. This yields a -time algorithm for
approximate CVP for any approximation factor .
Second, we show that the approximate closest vectors to a target vector can
be grouped into "lower-dimensional clusters," and we use this to obtain a
recursive reduction from exact CVP to a variant of approximate CVP that
"behaves well with these clusters." Third, we show that our discrete Gaussian
sampling algorithm can be used to solve this variant of approximate CVP.
The analysis depends crucially on some new properties of the discrete
Gaussian distribution and approximate closest vectors, which might be of
independent interest
Efficient Integer Coefficient Search for Compute-and-Forward
Integer coefficient selection is an important decoding step in the
implementation of compute-and-forward (C-F) relaying scheme. Choosing the
optimal integer coefficients in C-F has been shown to be a shortest vector
problem (SVP) which is known to be NP hard in its general form. Exhaustive
search of the integer coefficients is only feasible in complexity for small
number of users while approximation algorithms such as Lenstra-Lenstra-Lovasz
(LLL) lattice reduction algorithm only find a vector within an exponential
factor of the shortest vector. An optimal deterministic algorithm was proposed
for C-F by Sahraei and Gastpar specifically for the real valued channel case.
In this paper, we adapt their idea to the complex valued channel and propose an
efficient search algorithm to find the optimal integer coefficient vectors over
the ring of Gaussian integers and the ring of Eisenstein integers. A second
algorithm is then proposed that generalises our search algorithm to the
Integer-Forcing MIMO C-F receiver. Performance and efficiency of the proposed
algorithms are evaluated through simulations and theoretical analysis.Comment: IEEE Transactions on Wireless Communications, to appear.12 pages, 8
figure
Lattice Gaussian Sampling by Markov Chain Monte Carlo: Bounded Distance Decoding and Trapdoor Sampling
Sampling from the lattice Gaussian distribution plays an important role in
various research fields. In this paper, the Markov chain Monte Carlo
(MCMC)-based sampling technique is advanced in several fronts. Firstly, the
spectral gap for the independent Metropolis-Hastings-Klein (MHK) algorithm is
derived, which is then extended to Peikert's algorithm and rejection sampling;
we show that independent MHK exhibits faster convergence. Then, the performance
of bounded distance decoding using MCMC is analyzed, revealing a flexible
trade-off between the decoding radius and complexity. MCMC is further applied
to trapdoor sampling, again offering a trade-off between security and
complexity. Finally, the independent multiple-try Metropolis-Klein (MTMK)
algorithm is proposed to enhance the convergence rate. The proposed algorithms
allow parallel implementation, which is beneficial for practical applications.Comment: submitted to Transaction on Information Theor
- âŠ