12 research outputs found
Estimation of Sparsity via Simple Measurements
We consider several related problems of estimating the 'sparsity' or number
of nonzero elements in a length vector by observing only
, where is a predesigned test matrix
independent of , and the operation varies between problems.
We aim to provide a -approximation of sparsity for some constant
with a minimal number of measurements (rows of ). This framework
generalizes multiple problems, such as estimation of sparsity in group testing
and compressed sensing. We use techniques from coding theory as well as
probabilistic methods to show that rows are sufficient
when the operation is logical OR (i.e., group testing), and nearly this
many are necessary, where is a known upper bound on . When instead the
operation is multiplication over or a finite field
, we show that respectively and measurements are necessary and sufficient.Comment: 13 pages; shortened version presented at ISIT 201
Engineering Competitive and Query-Optimal Minimal-Adaptive Randomized Group Testing Strategies
Suppose that given is a collection of elements where of them are \emph{defective}. We can query an arbitrarily chosen subset of elements which returns Yes if the subset contains at least one defective and No if the subset is free of defectives. The problem of group testing is to identify the defectives with a minimum number of such queries. By the information-theoretic lower bound at least queries are needed. Using adaptive group testing, i.e., asking one query at a time, the lower bound can be easily achieved. However, strategies are preferred that work in a fixed small number of stages, where queries in a stage are asked in parallel. A group testing strategy is called \emph{competitive} if it works for completely unknown and requires only queries. Usually competitive group testing is based on sequential queries. We have shown that actually competitive group testing with expected queries is possible in only or stages. Then we have focused on minimizing the hidden constant factor in the query number and proposed a systematic approach for this purpose. Another main result is related to the design of query-optimal and minimal-adaptive strategies. We have shown that a -stage randomized strategy with prescribed success probability can asymptotically achieve the information-theoretic lower bound for and growing much slower than . Similarly, we can approach the entropy lower bound in stages when
New Constructions for Competitive and Minimal-Adaptive Group Testing
Group testing (GT) was originally proposed during the World War II in an attempt to minimize the \emph{cost} and \emph{waiting time} in performing identical blood tests of the soldiers for a low-prevalence disease.
Formally, the GT problem asks to find \emph{defective} elements out of elements by querying subsets (pools) for the presence of defectives.
By the information-theoretic lower bound, essentially queries are needed in the worst-case.
An \emph{adaptive} strategy proceeds sequentially by performing one query at a time, and it can achieve the lower bound. In various applications, nothing is known about beforehand and a strategy for this scenario is called \emph{competitive}. Such strategies are usually adaptive and achieve query optimality within a constant factor called the \emph{competitive ratio}.
In many applications, queries are time-consuming. Therefore, \emph{minimal-adaptive} strategies which run in a small number of stages of parallel
queries are favorable.
This work is mainly devoted to the design of minimal-adaptive strategies combined with other demands of both theoretical and practical interest. First we target unknown and show that actually competitive GT is possible in as few as stages only.
The main ingredient is our randomized estimate of a previously unknown using nonadaptive queries.
In addition, we have developed a systematic approach to obtain optimal competitive ratios for our strategies.
When is a known upper bound,
we propose randomized GT strategies which asymptotically achieve query optimality in just , or stages depending upon the growth of versus .
Inspired by application settings, such as at American Red Cross, where in most cases GT is applied to small instances, \textit{e.g.}, . We extended our study of query-optimal GT strategies to solve a given problem instance with fixed values , and . We also considered the situation when
elements to test cannot be divided physically (electronic devices), thus the pools must be disjoint. For GT with \emph{disjoint} simultaneous pools, we show that tests are sufficient, and also necessary for certain ranges of the parameters
Bounds for nonadaptive group tests to estimate the amount of defectives
Group testing is the problem of finding d defectives in a set of n elements, by asking carefully chosen subsets (pools) whether they contain defectives. Strategies are preferred that use both a small number of tests close to the information-theoretic lower bound d log n, and a small
constant number of stages, where tests in every stage are done in parallel, in order to save time. They should even work if d is not known in advance. In fact, one can succeed with O(d log n) queries in two stages, if certain tests are randomized and a constant failure probability is allowed. An essential ingredient of such strategies is to get an estimate of d within a constant factor. This problem is also interesting in its own right. It can be solved with O(log n) randomized group tests of a certain type. We prove
that O(log n) tests are also necessary, if elements for the pools are chosen independently. The proof builds upon an analysis of the influence of tests on the searcher's ability to distinguish between any two candidate numbers with a constant ratio. The next challenge is to get optimal constant factors in the O(log n) test number, depending on the prescribed error probability and the accuracy of d. We give practical methods to derive upper bound tradeoffs and conjecture that they are already close to optimal. One of them uses a linear programming formulation
Bounds for nonadaptive group tests to estimate the amount of defectives
The classical and well-studied group testing problem is to find d defectives in a set of n elements by group tests, which tell us for any chosen subset whether it contains defectives or not. Strategies are preferred that use both
a small number of tests close to the information-theoretic lower bound d log n, and a small constant number of stages, where tests in every stage are done in parallel, in order to save time. They should even work if d is completely unknown in advance. An essential ingredient of such competitive and minimal-adaptive group testing strategies is an estimate of d within a constant factor. More precisely, d shall be underestimated only with some
given error probability, and overestimated only by a constant factor, called the competitive ratio. The latter problem is also interesting in its own right. It can be solved with O(log n) randomized group tests of a certain type. In this paper we prove that O(log n) tests are really needed. The proof is based on an analysis of the influence of tests on the searcher's ability to distinguish between any two candidate numbers with a constant ratio. Once we know this lower bound, the next challenge is to get optimal constant factors in the O(log n) test number, depending on the desired error probability and competitive ratio. We give a method to derive upper bounds and conjecture that our particular strategy is already optimal