2 research outputs found

    Geometric group testing

    Full text link
    Group testing is concerned with identifying tt defective items in a set of mm items, where each test reports whether a specific subset of items contains at least one defective. In non-adaptive group testing, the subsets to be tested are fixed in advance. By testing multiple items at once, the required number of tests can be made much smaller than mm. In fact, for tO(1)t \in \mathcal{O}(1), the optimal number of (non-adaptive) tests is known to be Θ(logm)\Theta(\log{m}). In this paper, we consider the problem of non-adaptive group testing in a geometric setting, where the items are points in dd-dimensional Euclidean space and the tests are axis-parallel boxes (hyperrectangles). We present upper and lower bounds on the required number of tests under this geometric constraint. In contrast to the general, combinatorial case, the bounds in our geometric setting are polynomial in mm. For instance, our results imply that identifying a defective pair in a set of mm points in the plane always requires Ω(m3/5)\Omega(m^{3/5}) tests, and there exist configurations of mm points for which O(m2/3)\mathcal{O}(m^{2/3}) tests are sufficient, whereas to identify a single defective point in the plane, Θ(m1/2)\Theta(m^{1/2}) tests are always necessary and sometimes sufficient

    On Adaptive Distance Estimation

    Full text link
    We provide a static data structure for distance estimation which supports {\it adaptive} queries. Concretely, given a dataset X={xi}i=1nX = \{x_i\}_{i = 1}^n of nn points in Rd\mathbb{R}^d and 0<p20 < p \leq 2, we construct a randomized data structure with low memory consumption and query time which, when later given any query point qRdq \in \mathbb{R}^d, outputs a (1+ϵ)(1+\epsilon)-approximation of qxip\lVert q - x_i \rVert_p with high probability for all i[n]i\in[n]. The main novelty is our data structure's correctness guarantee holds even when the sequence of queries can be chosen adaptively: an adversary is allowed to choose the jjth query point qjq_j in a way that depends on the answers reported by the data structure for q1,,qj1q_1,\ldots,q_{j-1}. Previous randomized Monte Carlo methods do not provide error guarantees in the setting of adaptively chosen queries. Our memory consumption is O~((n+d)d/ϵ2)\tilde O((n+d)d/\epsilon^2), slightly more than the O(nd)O(nd) required to store XX in memory explicitly, but with the benefit that our time to answer queries is only O~(ϵ2(n+d))\tilde O(\epsilon^{-2}(n + d)), much faster than the naive Θ(nd)\Theta(nd) time obtained from a linear scan in the case of nn and dd very large. Here O~\tilde O hides log(nd/ϵ)\log(nd/\epsilon) factors. We discuss applications to nearest neighbor search and nonparametric estimation. Our method is simple and likely to be applicable to other domains: we describe a generic approach for transforming randomized Monte Carlo data structures which do not support adaptive queries to ones that do, and show that for the problem at hand, it can be applied to standard nonadaptive solutions to p\ell_p norm estimation with negligible overhead in query time and a factor dd overhead in memory.Comment: Minor correction in proof of Lemma B.
    corecore