188 research outputs found

    Weighted Polynomial Approximations: Limits for Learning and Pseudorandomness

    Get PDF
    Polynomial approximations to boolean functions have led to many positive results in computer science. In particular, polynomial approximations to the sign function underly algorithms for agnostically learning halfspaces, as well as pseudorandom generators for halfspaces. In this work, we investigate the limits of these techniques by proving inapproximability results for the sign function. Firstly, the polynomial regression algorithm of Kalai et al. (SIAM J. Comput. 2008) shows that halfspaces can be learned with respect to log-concave distributions on Rn\mathbb{R}^n in the challenging agnostic learning model. The power of this algorithm relies on the fact that under log-concave distributions, halfspaces can be approximated arbitrarily well by low-degree polynomials. We ask whether this technique can be extended beyond log-concave distributions, and establish a negative result. We show that polynomials of any degree cannot approximate the sign function to within arbitrarily low error for a large class of non-log-concave distributions on the real line, including those with densities proportional to expā”(āˆ’āˆ£xāˆ£0.99)\exp(-|x|^{0.99}). Secondly, we investigate the derandomization of Chernoff-type concentration inequalities. Chernoff-type tail bounds on sums of independent random variables have pervasive applications in theoretical computer science. Schmidt et al. (SIAM J. Discrete Math. 1995) showed that these inequalities can be established for sums of random variables with only O(logā”(1/Ī“))O(\log(1/\delta))-wise independence, for a tail probability of Ī“\delta. We show that their results are tight up to constant factors. These results rely on techniques from weighted approximation theory, which studies how well functions on the real line can be approximated by polynomials under various distributions. We believe that these techniques will have further applications in other areas of computer science.Comment: 22 page

    Pseudorandom Generators for Width-3 Branching Programs

    Full text link
    We construct pseudorandom generators of seed length O~(logā”(n)ā‹…logā”(1/Ļµ))\tilde{O}(\log(n)\cdot \log(1/\epsilon)) that Ļµ\epsilon-fool ordered read-once branching programs (ROBPs) of width 33 and length nn. For unordered ROBPs, we construct pseudorandom generators with seed length O~(logā”(n)ā‹…poly(1/Ļµ))\tilde{O}(\log(n) \cdot \mathrm{poly}(1/\epsilon)). This is the first improvement for pseudorandom generators fooling width 33 ROBPs since the work of Nisan [Combinatorica, 1992]. Our constructions are based on the `iterated milder restrictions' approach of Gopalan et al. [FOCS, 2012] (which further extends the Ajtai-Wigderson framework [FOCS, 1985]), combined with the INW-generator [STOC, 1994] at the last step (as analyzed by Braverman et al. [SICOMP, 2014]). For the unordered case, we combine iterated milder restrictions with the generator of Chattopadhyay et al. [CCC, 2018]. Two conceptual ideas that play an important role in our analysis are: (1) A relabeling technique allowing us to analyze a relabeled version of the given branching program, which turns out to be much easier. (2) Treating the number of colliding layers in a branching program as a progress measure and showing that it reduces significantly under pseudorandom restrictions. In addition, we achieve nearly optimal seed-length O~(logā”(n/Ļµ))\tilde{O}(\log(n/\epsilon)) for the classes of: (1) read-once polynomials on nn variables, (2) locally-monotone ROBPs of length nn and width 33 (generalizing read-once CNFs and DNFs), and (3) constant-width ROBPs of length nn having a layer of width 22 in every consecutive polylogā”(n)\mathrm{poly}\log(n) layers.Comment: 51 page

    Algorithms and lower bounds for de Morgan formulas of low-communication leaf gates

    Get PDF
    The class FORMULA[s]āˆ˜GFORMULA[s] \circ \mathcal{G} consists of Boolean functions computable by size-ss de Morgan formulas whose leaves are any Boolean functions from a class G\mathcal{G}. We give lower bounds and (SAT, Learning, and PRG) algorithms for FORMULA[n1.99]āˆ˜GFORMULA[n^{1.99}]\circ \mathcal{G}, for classes G\mathcal{G} of functions with low communication complexity. Let R(k)(G)R^{(k)}(\mathcal{G}) be the maximum kk-party NOF randomized communication complexity of G\mathcal{G}. We show: (1) The Generalized Inner Product function GIPnkGIP^k_n cannot be computed in FORMULA[s]āˆ˜GFORMULA[s]\circ \mathcal{G} on more than 1/2+Īµ1/2+\varepsilon fraction of inputs for s=oā€‰ā£(n2(kā‹…4kā‹…R(k)(G)ā‹…logā”(n/Īµ)ā‹…logā”(1/Īµ))2). s = o \! \left ( \frac{n^2}{ \left(k \cdot 4^k \cdot {R}^{(k)}(\mathcal{G}) \cdot \log (n/\varepsilon) \cdot \log(1/\varepsilon) \right)^{2}} \right). As a corollary, we get an average-case lower bound for GIPnkGIP^k_n against FORMULA[n1.99]āˆ˜PTFkāˆ’1FORMULA[n^{1.99}]\circ PTF^{k-1}. (2) There is a PRG of seed length n/2+O(sā‹…R(2)(G)ā‹…logā”(s/Īµ)ā‹…logā”(1/Īµ))n/2 + O\left(\sqrt{s} \cdot R^{(2)}(\mathcal{G}) \cdot\log(s/\varepsilon) \cdot \log (1/\varepsilon) \right) that Īµ\varepsilon-fools FORMULA[s]āˆ˜GFORMULA[s] \circ \mathcal{G}. For FORMULA[s]āˆ˜LTFFORMULA[s] \circ LTF, we get the better seed length O(n1/2ā‹…s1/4ā‹…logā”(n)ā‹…logā”(n/Īµ))O\left(n^{1/2}\cdot s^{1/4}\cdot \log(n)\cdot \log(n/\varepsilon)\right). This gives the first non-trivial PRG (with seed length o(n)o(n)) for intersections of nn half-spaces in the regime where Īµā‰¤1/n\varepsilon \leq 1/n. (3) There is a randomized 2nāˆ’t2^{n-t}-time #\#SAT algorithm for FORMULA[s]āˆ˜GFORMULA[s] \circ \mathcal{G}, where t=Ī©(nsā‹…logā”2(s)ā‹…R(2)(G))1/2.t=\Omega\left(\frac{n}{\sqrt{s}\cdot\log^2(s)\cdot R^{(2)}(\mathcal{G})}\right)^{1/2}. In particular, this implies a nontrivial #SAT algorithm for FORMULA[n1.99]āˆ˜LTFFORMULA[n^{1.99}]\circ LTF. (4) The Minimum Circuit Size Problem is not in FORMULA[n1.99]āˆ˜XORFORMULA[n^{1.99}]\circ XOR. On the algorithmic side, we show that FORMULA[n1.99]āˆ˜XORFORMULA[n^{1.99}] \circ XOR can be PAC-learned in time 2O(n/logā”n)2^{O(n/\log n)}
    • ā€¦
    corecore