30,425 research outputs found

    Testing k-Monotonicity

    Get PDF
    A Boolean k-monotone function defined over a finite poset domain D alternates between the values 0 and 1 at most k times on any ascending chain in D. Therefore, k-monotone functions are natural generalizations of the classical monotone functions, which are the 1-monotone functions. Motivated by the recent interest in k-monotone functions in the context of circuit complexity and learning theory, and by the central role that monotonicity testing plays in the context of property testing, we initiate a systematic study of k-monotone functions, in the property testing model. In this model, the goal is to distinguish functions that are k-monotone (or are close to being k-monotone) from functions that are far from being k-monotone. Our results include the following: 1. We demonstrate a separation between testing k-monotonicity and testing monotonicity, on the hypercube domain {0,1}^d, for k >= 3; 2. We demonstrate a separation between testing and learning on {0,1}^d, for k=omega(log d): testing k-monotonicity can be performed with 2^{O(sqrt d . log d . log{1/eps})} queries, while learning k-monotone functions requires 2^{Omega(k . sqrt d .{1/eps})} queries (Blais et al. (RANDOM 2015)). 3. We present a tolerant test for functions fcolon[n]^dto {0,1}$with complexity independent of n, which makes progress on a problem left open by Berman et al. (STOC 2014). Our techniques exploit the testing-by-learning paradigm, use novel applications of Fourier analysis on the grid [n]^d, and draw connections to distribution testing techniques. Our techniques exploit the testing-by-learning paradigm, use novel applications of Fourier analysis on the grid [n]^d, and draw connections to distribution testing techniques

    On Optimality of Stepdown and Stepup Multiple Test Procedures

    Full text link
    Consider the multiple testing problem of testing k null hypotheses, where the unknown family of distributions is assumed to satisfy a certain monotonicity assumption. Attention is restricted to procedures that control the familywise error rate in the strong sense and which satisfy a monotonicity condition. Under these assumptions, we prove certain maximin optimality results for some well-known stepdown and stepup procedures.Comment: Published at http://dx.doi.org/10.1214/009053605000000066 in the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Testing k-monotonicity of a discrete distribution. Application to the estimation of the number of classes in a population

    Full text link
    We develop here several goodness-of-fit tests for testing the k-monotonicity of a discrete density, based on the empirical distribution of the observations. Our tests are non-parametric, easy to implement and are proved to be asymptotically of the desired level and consistent. We propose an estimator of the degree of k-monotonicity of the distribution based on the non-parametric goodness-of-fit tests. We apply our work to the estimation of the total number of classes in a population. A large simulation study allows to assess the performances of our procedures.Comment: 32 pages, 8 figure

    Nearly Optimal Bounds for Sample-Based Testing and Learning of kk-Monotone Functions

    Full text link
    We study monotonicity testing of functions f ⁣:{0,1}d{0,1}f \colon \{0,1\}^d \to \{0,1\} using sample-based algorithms, which are only allowed to observe the value of ff on points drawn independently from the uniform distribution. A classic result by Bshouty-Tamon (J. ACM 1996) proved that monotone functions can be learned with exp(O(min{1εd,d}))\exp(O(\min\{\frac{1}{\varepsilon}\sqrt{d},d\})) samples and it is not hard to show that this bound extends to testing. Prior to our work the only lower bound for this problem was Ω(exp(d)/ε)\Omega(\sqrt{\exp(d)/\varepsilon}) in the small ε\varepsilon parameter regime, when ε=O(d3/2)\varepsilon = O(d^{-3/2}), due to Goldreich-Goldwasser-Lehman-Ron-Samorodnitsky (Combinatorica 2000). Thus, the sample complexity of monotonicity testing was wide open for εd3/2\varepsilon \gg d^{-3/2}. We resolve this question, obtaining a tight lower bound of exp(Ω(min{1εd,d}))\exp(\Omega(\min\{\frac{1}{\varepsilon}\sqrt{d},d\})) for all ε\varepsilon at most a sufficiently small constant. In fact, we prove a much more general result, showing that the sample complexity of kk-monotonicity testing and learning for functions f ⁣:{0,1}d[r]f \colon \{0,1\}^d \to [r] is exp(Θ(min{rkεd,d}))\exp(\Theta(\min\{\frac{rk}{\varepsilon}\sqrt{d},d\})). For testing with one-sided error we show that the sample complexity is exp(Θ(d))\exp(\Theta(d)). Beyond the hypercube, we prove nearly tight bounds (up to polylog factors of d,k,r,1/εd,k,r,1/\varepsilon in the exponent) of exp(Θ~(min{rkεd,d}))\exp(\widetilde{\Theta}(\min\{\frac{rk}{\varepsilon}\sqrt{d},d\})) on the sample complexity of testing and learning measurable kk-monotone functions f ⁣:Rd[r]f \colon \mathbb{R}^d \to [r] under product distributions. Our upper bound improves upon the previous bound of exp(O~(min{kε2d,d}))\exp(\widetilde{O}(\min\{\frac{k}{\varepsilon^2}\sqrt{d},d\})) by Harms-Yoshida (ICALP 2022) for Boolean functions (r=2r=2)

    Mildly Exponential Lower Bounds on Tolerant Testers for Monotonicity, Unateness, and Juntas

    Full text link
    We give the first super-polynomial (in fact, mildly exponential) lower bounds for tolerant testing (equivalently, distance estimation) of monotonicity, unateness, and juntas with a constant separation between the "yes" and "no" cases. Specifically, we give \bullet A 2Ω(n1/4/ε)2^{\Omega(n^{1/4}/\sqrt{\varepsilon})}-query lower bound for non-adaptive, two-sided tolerant monotonicity testers and unateness testers when the "gap" parameter ε2ε1\varepsilon_2-\varepsilon_1 is equal to ε\varepsilon, for any ε1/n\varepsilon \geq 1/\sqrt{n}; \bullet A 2Ω(k1/2)2^{\Omega(k^{1/2})}-query lower bound for non-adaptive, two-sided tolerant junta testers when the gap parameter is an absolute constant. In the constant-gap regime no non-trivial prior lower bound was known for monotonicity, the best prior lower bound known for unateness was Ω~(n3/2)\tilde{\Omega}(n^{3/2}) queries, and the best prior lower bound known for juntas was poly(k)\mathrm{poly}(k) queries.Comment: 20 pages, 1 figur

    Exact and approximate stepdown methods for multiple hypothesis testing

    Get PDF
    Consider the problem of testing k hypotheses simultaneously. In this paper, we discuss finite and large sample theory of stepdown methods that provide control of the familywise error rate (FWE). In order to improve upon the Bonferroni method or Holm's (1979) stepdown method, Westfall and Young (1993) make eective use of resampling to construct stepdown methods that implicitly estimate the dependence structure of the test statistics. However, their methods depend on an assumption called subset pivotality. The goal of this paper is to construct general stepdown methods that do not require such an assumption. In order to accomplish this, we take a close look at what makes stepdown procedures work, and a key component is a monotonicity requirement of critical values. By imposing such monotonicity on estimated critical values (which is not an assumption on the model but an assumption on the method), it is demonstrated that the problem of constructing a valid multiple test procedure which controls the FWE can be reduced to the problem of contructing a single test which controls the usual probability of a Type 1 error. This reduction allows us to draw upon an enormous resampling literature as a general means of test contruction.Bootstrap, familywise error rate, multiple testing, permutation test, randomization test, stepdown procedure, subsampling

    Bayesian Inference for kk-Monotone Densities with Applications to Multiple Testing

    Full text link
    Shape restriction, like monotonicity or convexity, imposed on a function of interest, such as a regression or density function, allows for its estimation without smoothness assumptions. The concept of kk-monotonicity encompasses a family of shape restrictions, including decreasing and convex decreasing as special cases corresponding to k=1k=1 and k=2k=2. We consider Bayesian approaches to estimate a kk-monotone density. By utilizing a kernel mixture representation and putting a Dirichlet process or a finite mixture prior on the mixing distribution, we show that the posterior contraction rate in the Hellinger distance is (n/logn)k/(2k+1)(n/\log n)^{- k/(2k + 1)} for a kk-monotone density, which is minimax optimal up to a polylogarithmic factor. When the true kk-monotone density is a finite J0J_0-component mixture of the kernel, the contraction rate improves to the nearly parametric rate (J0logn)/n\sqrt{(J_0 \log n)/n}. Moreover, by putting a prior on kk, we show that the same rates hold even when the best value of kk is unknown. A specific application in modeling the density of pp-values in a large-scale multiple testing problem is considered. Simulation studies are conducted to evaluate the performance of the proposed method

    Finding Monotone Patterns in Sublinear Time

    Get PDF
    We study the problem of finding monotone subsequences in an array from the viewpoint of sublinear algorithms. For fixed k ϵ N and ε > 0, we show that the non-adaptive query complexity of finding a length-k monotone subsequence of f : [n] → R, assuming that f is ε-far from free of such subsequences, is Θ((log n) ^{[log_2k]}). Prior to our work, the best algorithm for this problem, due to Newman, Rabinovich, Rajendraprasad, and Sohler (2017), made (log n) ^{O(k2)} non-adaptive queries; and the only lower bound known, of Ω(log n) queries for the case k = 2, followed from that on testing monotonicity due to Ergün, Kannan, Kumar, Rubinfeld, and Viswanathan (2000) and Fischer (2004)
    corecore