94,048 research outputs found

    On the complexity of range searching among curves

    Full text link
    Modern tracking technology has made the collection of large numbers of densely sampled trajectories of moving objects widely available. We consider a fundamental problem encountered when analysing such data: Given nn polygonal curves SS in Rd\mathbb{R}^d, preprocess SS into a data structure that answers queries with a query curve qq and radius ρ\rho for the curves of SS that have \Frechet distance at most ρ\rho to qq. We initiate a comprehensive analysis of the space/query-time trade-off for this data structuring problem. Our lower bounds imply that any data structure in the pointer model model that achieves Q(n)+O(k)Q(n) + O(k) query time, where kk is the output size, has to use roughly Ω((n/Q(n))2)\Omega\left((n/Q(n))^2\right) space in the worst case, even if queries are mere points (for the discrete \Frechet distance) or line segments (for the continuous \Frechet distance). More importantly, we show that more complex queries and input curves lead to additional logarithmic factors in the lower bound. Roughly speaking, the number of logarithmic factors added is linear in the number of edges added to the query and input curve complexity. This means that the space/query time trade-off worsens by an exponential factor of input and query complexity. This behaviour addresses an open question in the range searching literature: whether it is possible to avoid the additional logarithmic factors in the space and query time of a multilevel partition tree. We answer this question negatively. On the positive side, we show we can build data structures for the \Frechet distance by using semialgebraic range searching. Our solution for the discrete \Frechet distance is in line with the lower bound, as the number of levels in the data structure is O(t)O(t), where tt denotes the maximal number of vertices of a curve. For the continuous \Frechet distance, the number of levels increases to O(t2)O(t^2)

    A New Lower Bound for Semigroup Orthogonal Range Searching

    Get PDF
    We report the first improvement in the space-time trade-off of lower bounds for the orthogonal range searching problem in the semigroup model, since Chazelle's result from 1990. This is one of the very fundamental problems in range searching with a long history. Previously, Andrew Yao's influential result had shown that the problem is already non-trivial in one dimension~\cite{Yao-1Dlb}: using mm units of space, the query time Q(n)Q(n) must be Ω(α(m,n)+nmn+1)\Omega( \alpha(m,n) + \frac{n}{m-n+1}) where α(,)\alpha(\cdot,\cdot) is the inverse Ackermann's function, a very slowly growing function. In dd dimensions, Bernard Chazelle~\cite{Chazelle.LB.II} proved that the query time must be Q(n)=Ω((logβn)d1)Q(n) = \Omega( (\log_\beta n)^{d-1}) where β=2m/n\beta = 2m/n. Chazelle's lower bound is known to be tight for when space consumption is `high' i.e., m=Ω(nlogd+εn)m = \Omega(n \log^{d+\varepsilon}n). We have two main results. The first is a lower bound that shows Chazelle's lower bound was not tight for `low space': we prove that we must have m(n)=Ω(n(lognloglogn)d1)m (n) = \Omega(n (\log n \log\log n)^{d-1}). Our lower bound does not close the gap to the existing data structures, however, our second result is that our analysis is tight. Thus, we believe the gap is in fact natural since lower bounds are proven for idempotent semigroups while the data structures are built for general semigroups and thus they cannot assume (and use) the properties of an idempotent semigroup. As a result, we believe to close the gap one must study lower bounds for non-idempotent semigroups or building data structures for idempotent semigroups. We develope significantly new ideas for both of our results that could be useful in pursuing either of these directions

    Online Searching with an Autonomous Robot

    Full text link
    We discuss online strategies for visibility-based searching for an object hidden behind a corner, using Kurt3D, a real autonomous mobile robot. This task is closely related to a number of well-studied problems. Our robot uses a three-dimensional laser scanner in a stop, scan, plan, go fashion for building a virtual three-dimensional environment. Besides planning trajectories and avoiding obstacles, Kurt3D is capable of identifying objects like a chair. We derive a practically useful and asymptotically optimal strategy that guarantees a competitive ratio of 2, which differs remarkably from the well-studied scenario without the need of stopping for surveying the environment. Our strategy is used by Kurt3D, documented in a separate video.Comment: 16 pages, 8 figures, 12 photographs, 1 table, Latex, submitted for publicatio

    On Range Searching with Semialgebraic Sets II

    Full text link
    Let PP be a set of nn points in Rd\R^d. We present a linear-size data structure for answering range queries on PP with constant-complexity semialgebraic sets as ranges, in time close to O(n11/d)O(n^{1-1/d}). It essentially matches the performance of similar structures for simplex range searching, and, for d5d\ge 5, significantly improves earlier solutions by the first two authors obtained in~1994. This almost settles a long-standing open problem in range searching. The data structure is based on the polynomial-partitioning technique of Guth and Katz [arXiv:1011.4105], which shows that for a parameter rr, 1<rn1 < r \le n, there exists a dd-variate polynomial ff of degree O(r1/d)O(r^{1/d}) such that each connected component of RdZ(f)\R^d\setminus Z(f) contains at most n/rn/r points of PP, where Z(f)Z(f) is the zero set of ff. We present an efficient randomized algorithm for computing such a polynomial partition, which is of independent interest and is likely to have additional applications

    Next decade of sterile neutrino studies

    Get PDF
    We review the status of sterile neutrino dark matter and discuss astrophysical and cosmological bounds on its properties as well as future prospects for its experimental searches. We argue that if sterile neutrinos are the dominant fraction of dark matter, detecting an astrophysical signal from their decay (the so-called 'indirect detection') may be the only way to identify these particles experimentally. However, it may be possible to check the dark matter origin of the observed signal unambiguously using its characteristic properties and/or using synergy with accelerator experiments, searching for other sterile neutrinos, responsible for neutrino flavor oscillations. We argue that to fully explore this possibility a dedicated cosmic mission - an X-ray spectrometer - is needed.Comment: 23 pages, 6 figure
    corecore