243 research outputs found
Lattice sparsification and the Approximate Closest Vector Problem
We give a deterministic algorithm for solving the
(1+\eps)-approximate Closest Vector Problem (CVP) on any
-dimensional lattice and in any near-symmetric norm in
2^{O(n)}(1+1/\eps)^n time and 2^n\poly(n) space. Our algorithm
builds on the lattice point enumeration techniques of Micciancio and
Voulgaris (STOC 2010, SICOMP 2013) and Dadush, Peikert and Vempala
(FOCS 2011), and gives an elegant, deterministic alternative to the
"AKS Sieve"-based algorithms for (1+\eps)-CVP (Ajtai, Kumar, and
Sivakumar; STOC 2001 and CCC 2002). Furthermore, assuming the
existence of a \poly(n)-space and -time algorithm for
exact CVP in the norm, the space complexity of our algorithm
can be reduced to polynomial.
Our main technical contribution is a method for "sparsifying" any
input lattice while approximately maintaining its metric structure. To
this end, we employ the idea of random sublattice restrictions, which
was first employed by Khot (FOCS 2003, J. Comp. Syst. Sci. 2006) for
the purpose of proving hardness for the Shortest Vector Problem (SVP)
under norms.
A preliminary version of this paper appeared in the Proc. 24th Annual
ACM-SIAM Symp. on Discrete Algorithms (SODA'13)
(http://dx.doi.org/10.1137/1.9781611973105.78)
Lattice Sparsification and the approximate closest vector problem
We give a deterministic algorithm for solving the
(1+\eps)-approximate Closest Vector Problem (CVP) on any
-dimensional lattice and in any near-symmetric norm in
2^{O(n)}(1+1/\eps)^n time and 2^n\poly(n) space. Our algorithm
builds on the lattice point enumeration techniques of Micciancio and
Voulgaris (STOC 2010, SICOMP 2013) and Dadush, Peikert and Vempala
(FOCS 2011), and gives an elegant, deterministic alternative to the
"AKS Sieve"-based algorithms for (1+\eps)-CVP (Ajtai, Kumar, and
Sivakumar; STOC 2001 and CCC 2002). Furthermore, assuming the
existence of a \poly(n)-space and -time algorithm for
exact CVP in the norm, the space complexity of our algorithm
can be reduced to polynomial.
Our main technical contribution is a method for "sparsifying" any
input lattice while approximately maintaining its metric structure. To
this end, we employ the idea of random sublattice restrictions, which
was first employed by Khot (FOCS 2003, J. Comp. Syst. Sci. 2006) for
the purpose of proving hardness for the Shortest Vector Problem (SVP)
under norms.
A preliminary version of this paper appeared in the Proc. 24th Annual
ACM-SIAM Symp. on Discrete Algorithms (SODA'13)
(http://dx.doi.org/10.1137/1.9781611973105.78)
Search-to-Decision Reductions for Lattice Problems with Approximation Factors (Slightly) Greater Than One
We show the first dimension-preserving search-to-decision reductions for
approximate SVP and CVP. In particular, for any ,
we obtain an efficient dimension-preserving reduction from -SVP to -GapSVP and an efficient dimension-preserving reduction
from -CVP to -GapCVP. These results generalize the known
equivalences of the search and decision versions of these problems in the exact
case when . For SVP, we actually obtain something slightly stronger
than a search-to-decision reduction---we reduce -SVP to
-unique SVP, a potentially easier problem than -GapSVP.Comment: Updated to acknowledge additional prior wor
On the Closest Vector Problem with a Distance Guarantee
We present a substantially more efficient variant, both in terms of running
time and size of preprocessing advice, of the algorithm by Liu, Lyubashevsky,
and Micciancio for solving CVPP (the preprocessing version of the Closest
Vector Problem, CVP) with a distance guarantee. For instance, for any , our algorithm finds the (unique) closest lattice point for any target
point whose distance from the lattice is at most times the length of
the shortest nonzero lattice vector, requires as preprocessing advice only vectors, and runs in
time .
As our second main contribution, we present reductions showing that it
suffices to solve CVP, both in its plain and preprocessing versions, when the
input target point is within some bounded distance of the lattice. The
reductions are based on ideas due to Kannan and a recent sparsification
technique due to Dadush and Kun. Combining our reductions with the LLM
algorithm gives an approximation factor of for search
CVPP, improving on the previous best of due to Lagarias, Lenstra,
and Schnorr. When combined with our improved algorithm we obtain, somewhat
surprisingly, that only O(n) vectors of preprocessing advice are sufficient to
solve CVPP with (the only slightly worse) approximation factor of O(n).Comment: An early version of the paper was titled "On Bounded Distance
Decoding and the Closest Vector Problem with Preprocessing". Conference on
Computational Complexity (2014
On the Quantitative Hardness of CVP
For odd
integers (and ), we show that the Closest Vector Problem
in the norm (\CVP_p) over rank lattices cannot be solved in
2^{(1-\eps) n} time for any constant \eps > 0 unless the Strong Exponential
Time Hypothesis (SETH) fails. We then extend this result to "almost all" values
of , not including the even integers. This comes tantalizingly close
to settling the quantitative time complexity of the important special case of
\CVP_2 (i.e., \CVP in the Euclidean norm), for which a -time
algorithm is known. In particular, our result applies for any
that approaches as .
We also show a similar SETH-hardness result for \SVP_\infty; hardness of
approximating \CVP_p to within some constant factor under the so-called
Gap-ETH assumption; and other quantitative hardness results for \CVP_p and
\CVPP_p for any under different assumptions
Improved Reduction from the Bounded Distance Decoding Problem to the Unique Shortest Vector Problem in Lattices
We present a probabilistic polynomial-time reduction from the lattice Bounded Distance Decoding (BDD) problem with parameter 1/( sqrt(2) * gamma) to the unique Shortest Vector Problem (uSVP) with parameter gamma for any gamma > 1 that is polynomial in the lattice dimension n. It improves the BDD to uSVP reductions of [Lyubashevsky and Micciancio, CRYPTO, 2009] and [Liu, Wang, Xu and Zheng, Inf. Process. Lett., 2014], which rely on Kannan\u27s embedding technique. The main ingredient to the improvement is the use of Khot\u27s lattice sparsification [Khot, FOCS, 2003] before resorting to Kannan\u27s embedding, in order to boost the uSVP parameter
Improved Hardness of BDD and SVP Under Gap-(S)ETH
We show improved fine-grained hardness of two key lattice problems in the
norm: Bounded Distance Decoding to within an factor of the
minimum distance () and the (decisional)
-approximate Shortest Vector Problem (),
assuming variants of the Gap (Strong) Exponential Time Hypothesis (Gap-(S)ETH).
Specifically, we show:
1. For all , there is no -time algorithm for
for any constant ,
where and
is the kissing-number constant, unless non-uniform Gap-ETH is false.
2. For all , there is no -time algorithm for
for any constant , where
is explicit and satisfies for , , and as , unless randomized Gap-ETH is false.
3. For all and all , there
is no -time algorithm for for any constant
, where is explicit and
satisfies as for any fixed , unless non-uniform Gap-SETH is false.
4. For all , , and all , there is no -time algorithm for for
some constant , where is explicit and satisfies as , unless randomized Gap-SETH is false.Comment: ITCS 202
A Size-Free CLT for Poisson Multinomials and its Applications
An -Poisson Multinomial Distribution (PMD) is the distribution of the
sum of independent random vectors supported on the set of standard basis vectors in . We show
that any -PMD is -close in total
variation distance to the (appropriately discretized) multi-dimensional
Gaussian with the same first two moments, removing the dependence on from
the Central Limit Theorem of Valiant and Valiant. Interestingly, our CLT is
obtained by bootstrapping the Valiant-Valiant CLT itself through the structural
characterization of PMDs shown in recent work by Daskalakis, Kamath, and
Tzamos. In turn, our stronger CLT can be leveraged to obtain an efficient PTAS
for approximate Nash equilibria in anonymous games, significantly improving the
state of the art, and matching qualitatively the running time dependence on
and of the best known algorithm for two-strategy anonymous
games. Our new CLT also enables the construction of covers for the set of
-PMDs, which are proper and whose size is shown to be essentially
optimal. Our cover construction combines our CLT with the Shapley-Folkman
theorem and recent sparsification results for Laplacian matrices by Batson,
Spielman, and Srivastava. Our cover size lower bound is based on an algebraic
geometric construction. Finally, leveraging the structural properties of the
Fourier spectrum of PMDs we show that these distributions can be learned from
samples in -time, removing
the quasi-polynomial dependence of the running time on from the
algorithm of Daskalakis, Kamath, and Tzamos.Comment: To appear in STOC 201
- …