3 research outputs found

    Fast algorithms for collision and proximity problems involving moving geometric objects

    No full text
    Consider a set of geometric objects, such as points, line segments, or axes-parallel hyperrectangles in \IR^d, that move with constant but possibly different velocities along linear trajectories. Efficient algorithms are presented for several problems defined on such objects, such as determining whether any two objects ever collide and computing the minimum inter-point separation or minimum diameter that ever occurs. The strategy used involves reducing the given problem on moving objects to a different problem on a set of static objects, and then solving the latter problem using techniques based on sweeping, orthogonal range searching, simplex composition, and parametric search

    On Geometric Range Searching, Approximate Counting and Depth Problems

    Get PDF
    In this thesis we deal with problems connected to range searching, which is one of the central areas of computational geometry. The dominant problems in this area are halfspace range searching, simplex range searching and orthogonal range searching and research into these problems has spanned decades. For many range searching problems, the best possible data structures cannot offer fast (i.e., polylogarithmic) query times if we limit ourselves to near linear storage. Even worse, it is conjectured (and proved in some cases) that only very small improvements to these might be possible. This inefficiency has encouraged many researchers to seek alternatives through approximations. In this thesis we continue this line of research and focus on relative approximation of range counting problems. One important problem where it is possible to achieve significant speedup through approximation is halfspace range counting in 3D. Here we continue the previous research done and obtain the first optimal data structure for approximate halfspace range counting in 3D. Our data structure has the slight advantage of being Las Vegas (the result is always correct) in contrast to the previous methods that were Monte Carlo (the correctness holds with high probability). Another series of problems where approximation can provide us with substantial speedup comes from robust statistics. We recognize three problems here: approximate Tukey depth, regression depth and simplicial depth queries. In 2D, we obtain an optimal data structure capable of approximating the regression depth of a query hyperplane. We also offer a linear space data structure which can answer approximate Tukey depth queries efficiently in 3D. These data structures are obtained by applying our ideas for the approximate halfspace counting problem. Approximating the simplicial depth turns out to be much more difficult, however. Computing the simplicial depth of a given point is more computationally challenging than most other definitions of data depth. In 2D we obtain the first data structure which uses near linear space and can answer approximate simplicial depth queries in polylogarithmic time. As applications of this result, we provide two non-trivial methods to approximate the simplicial depth of a given point in higher dimension. Along the way, we establish a tight combinatorial relationship between the Tukey depth of any given point and its simplicial depth. Another problem investigated in this thesis is the dominance reporting problem, an important special case of orthogonal range reporting. In three dimensions, we solve this problem in the pointer machine model and the external memory model by offering the first optimal data structures in these models of computation. Also, in the RAM model and for points from an integer grid we reduce the space complexity of the fastest known data structure to optimal. Using known techniques in the literature, we can use our results to obtain solutions for the orthogonal range searching problem as well. The query complexity offered by our orthogonal range reporting data structures match the most efficient query complexities known in the literature but our space bounds are lower than the previous methods in the external memory model and RAM model where the input is a subset of an integer grid. The results also yield improved orthogonal range searching in higher dimensions (which shows the significance of the dominance reporting problem). Intersection searching is a generalization of range searching where we deal with more complicated geometric objects instead of points. We investigate the rectilinear disjoint polygon counting problem which is a specialized intersection counting problem. We provide a linear-size data structure capable of counting the number of disjoint rectilinear polygons intersecting any rectilinear polygon of constant size. The query time (as well as some other properties of our data structure) resembles the classical simplex range searching data structures

    Algorithms for learning from spatial and mobility data

    Get PDF
    Data from the numerous mobile devices, location-based applications, and collection sensors used currently can provide important insights about human and natural processes. These insights can inform decision making in designing and optimising in frastructure such as transportation or energy. However, extracting patterns related to spatial properties is challenging due to the large quantity of the data produced and the complexity of the processes it describes. We propose scalable, multi-resolution approximation and heuristic algorithms that make use of spatial proximity properties to solve fundamental data mining and optimisation problems with a better running time and accuracy. We observe that abstracting from individual data points and working with units of neighbouring points based on various measures on similarity, improves computational efficiency and diminishes the effects of noise and overfitting. We consider applications in: mobility data compression, transit network planning, and solar power output prediction. Firstly, in order to understand transportation needs, it is essential to have efficient ways to represent large amounts of travel data. In analysing spatial trajectories (for example taxis travelling in a city), one of the main challenges is computing distances between trajectories efficiently; due to their size and complexity this task is computationally expensive. We build data structures and algorithms to sketch trajectory data that make queries such as distance computation, nearest neighbour search and clustering, which are key to finding mobility patterns, more computationally efficient. We use locality sensitive hashing, a technique that associates similar objects to the same hash. Secondly, to build efficient infrastructure it is necessary to satisfy travel demand by placing resources optimally. This is difficult due to external constraints (such as limits on budget) and the complexity of existing road networks that allow for a large number of candidate locations. For this purpose, we present heuristic algorithms for efficient transit network design with a case study on cycling lane placement. The heuristic is based on a new type of clustering by projection, that is both computationally efficient and gives good results in practice. Lastly, we devise a novel method to forecast solar power output based on numerical weather predictions, clear sky predictions and persistence data. The ensemble of a multivariate linear regression model, support vector machines model, and an artificial neural network gives more accurate predictions than any of the individual models. Analysing the performance of the models in a suite of frameworks reveals that building separate models for each self-contained area based on weather patterns gives a better accuracy than a single model that predicts the total. The ensemble can be further improved by giving performance-based weights to the individual models. This suggests that the models identify different patterns in the data, which motivated the choice of an ensemble architecture
    corecore