30,425 research outputs found
Testing k-Monotonicity
A Boolean k-monotone function defined over a finite poset domain D alternates between the values 0 and 1 at most k times on any ascending chain in D. Therefore, k-monotone functions are natural generalizations of the classical monotone functions, which are the 1-monotone functions.
Motivated by the recent interest in k-monotone functions in the context of circuit complexity and learning theory, and by the central role that monotonicity testing plays in the context of property testing, we initiate a systematic study of k-monotone functions, in the property testing model. In this model, the goal is to distinguish functions that are k-monotone (or are close to being k-monotone) from functions that are far from being k-monotone.
Our results include the following:
1. We demonstrate a separation between testing k-monotonicity and testing monotonicity, on the hypercube domain {0,1}^d, for k >= 3;
2. We demonstrate a separation between testing and learning on {0,1}^d, for k=omega(log d): testing k-monotonicity can be performed with 2^{O(sqrt d . log d . log{1/eps})} queries, while learning k-monotone functions requires 2^{Omega(k . sqrt d .{1/eps})} queries (Blais et al. (RANDOM 2015)).
3. We present a tolerant test for functions fcolon[n]^dto {0,1}$with complexity independent of n, which makes progress on a problem left open by Berman et al. (STOC 2014).
Our techniques exploit the testing-by-learning paradigm, use novel applications of Fourier analysis on the grid [n]^d, and draw connections to distribution testing techniques.
Our techniques exploit the testing-by-learning paradigm, use novel applications of Fourier analysis on the grid [n]^d, and draw connections to distribution testing techniques
On Optimality of Stepdown and Stepup Multiple Test Procedures
Consider the multiple testing problem of testing k null hypotheses, where the
unknown family of distributions is assumed to satisfy a certain monotonicity
assumption. Attention is restricted to procedures that control the familywise
error rate in the strong sense and which satisfy a monotonicity condition.
Under these assumptions, we prove certain maximin optimality results for some
well-known stepdown and stepup procedures.Comment: Published at http://dx.doi.org/10.1214/009053605000000066 in the
Annals of Statistics (http://www.imstat.org/aos/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Testing k-monotonicity of a discrete distribution. Application to the estimation of the number of classes in a population
We develop here several goodness-of-fit tests for testing the k-monotonicity
of a discrete density, based on the empirical distribution of the observations.
Our tests are non-parametric, easy to implement and are proved to be
asymptotically of the desired level and consistent. We propose an estimator of
the degree of k-monotonicity of the distribution based on the non-parametric
goodness-of-fit tests. We apply our work to the estimation of the total number
of classes in a population. A large simulation study allows to assess the
performances of our procedures.Comment: 32 pages, 8 figure
Nearly Optimal Bounds for Sample-Based Testing and Learning of -Monotone Functions
We study monotonicity testing of functions
using sample-based algorithms, which are only allowed to observe the value of
on points drawn independently from the uniform distribution. A classic
result by Bshouty-Tamon (J. ACM 1996) proved that monotone functions can be
learned with samples and it
is not hard to show that this bound extends to testing. Prior to our work the
only lower bound for this problem was in
the small parameter regime, when , due
to Goldreich-Goldwasser-Lehman-Ron-Samorodnitsky (Combinatorica 2000). Thus,
the sample complexity of monotonicity testing was wide open for . We resolve this question, obtaining a tight lower bound of
for all
at most a sufficiently small constant. In fact, we prove a much more general
result, showing that the sample complexity of -monotonicity testing and
learning for functions is
. For testing with
one-sided error we show that the sample complexity is .
Beyond the hypercube, we prove nearly tight bounds (up to polylog factors of
in the exponent) of
on the
sample complexity of testing and learning measurable -monotone functions under product distributions. Our upper bound
improves upon the previous bound of
by
Harms-Yoshida (ICALP 2022) for Boolean functions ()
Mildly Exponential Lower Bounds on Tolerant Testers for Monotonicity, Unateness, and Juntas
We give the first super-polynomial (in fact, mildly exponential) lower bounds
for tolerant testing (equivalently, distance estimation) of monotonicity,
unateness, and juntas with a constant separation between the "yes" and "no"
cases. Specifically, we give
A -query lower bound for
non-adaptive, two-sided tolerant monotonicity testers and unateness testers
when the "gap" parameter is equal to
, for any ;
A -query lower bound for non-adaptive,
two-sided tolerant junta testers when the gap parameter is an absolute
constant.
In the constant-gap regime no non-trivial prior lower bound was known for
monotonicity, the best prior lower bound known for unateness was
queries, and the best prior lower bound known for
juntas was queries.Comment: 20 pages, 1 figur
Exact and approximate stepdown methods for multiple hypothesis testing
Consider the problem of testing k hypotheses simultaneously. In this paper, we discuss finite and large sample theory of stepdown methods that provide control of the familywise error rate (FWE). In order to improve upon the Bonferroni method or Holm's (1979) stepdown method, Westfall and Young (1993) make eective use of resampling to construct stepdown methods that implicitly estimate the dependence structure of the test statistics. However, their methods depend on an assumption called subset pivotality. The goal of this paper is to construct general stepdown methods that do not require such an assumption. In order to accomplish this, we take a close look at what makes stepdown procedures work, and a key component is a monotonicity requirement of critical values. By imposing such monotonicity on estimated critical values (which is not an assumption on the model but an assumption on the method), it is demonstrated that the problem of constructing a valid multiple test procedure which controls the FWE can be reduced to the problem of contructing a single test which controls the usual probability of a Type 1 error. This reduction allows us to draw upon an enormous resampling literature as a general means of test contruction.Bootstrap, familywise error rate, multiple testing, permutation test, randomization test, stepdown procedure, subsampling
Bayesian Inference for -Monotone Densities with Applications to Multiple Testing
Shape restriction, like monotonicity or convexity, imposed on a function of
interest, such as a regression or density function, allows for its estimation
without smoothness assumptions. The concept of -monotonicity encompasses a
family of shape restrictions, including decreasing and convex decreasing as
special cases corresponding to and . We consider Bayesian approaches
to estimate a -monotone density. By utilizing a kernel mixture
representation and putting a Dirichlet process or a finite mixture prior on the
mixing distribution, we show that the posterior contraction rate in the
Hellinger distance is for a -monotone density,
which is minimax optimal up to a polylogarithmic factor. When the true
-monotone density is a finite -component mixture of the kernel, the
contraction rate improves to the nearly parametric rate . Moreover, by putting a prior on , we show that the same rates hold
even when the best value of is unknown. A specific application in modeling
the density of -values in a large-scale multiple testing problem is
considered. Simulation studies are conducted to evaluate the performance of the
proposed method
Finding Monotone Patterns in Sublinear Time
We study the problem of finding monotone subsequences in an array from the viewpoint of sublinear algorithms. For fixed k ϵ N and ε > 0, we show that the non-adaptive query complexity of finding a length-k monotone subsequence of f : [n] → R, assuming that f is ε-far from free of such subsequences, is Θ((log n) ^{[log_2k]}). Prior to our work, the best algorithm for this problem, due to Newman, Rabinovich, Rajendraprasad, and Sohler (2017), made (log n) ^{O(k2)} non-adaptive queries; and the only lower bound known, of Ω(log n) queries for the case k = 2, followed from that on testing monotonicity due to Ergün, Kannan, Kumar, Rubinfeld, and Viswanathan (2000) and Fischer (2004)
- …