2,438 research outputs found
Optimal randomized incremental construction for guaranteed logarithmic planar point location
Given a planar map of segments in which we wish to efficiently locate
points, we present the first randomized incremental construction of the
well-known trapezoidal-map search-structure that only requires expected preprocessing time while deterministically guaranteeing worst-case
linear storage space and worst-case logarithmic query time. This settles a long
standing open problem; the best previously known construction time of such a
structure, which is based on a directed acyclic graph, so-called the history
DAG, and with the above worst-case space and query-time guarantees, was
expected . The result is based on a deeper understanding of the
structure of the history DAG, its depth in relation to the length of its
longest search path, as well as its correspondence to the trapezoidal search
tree. Our results immediately extend to planar maps induced by finite
collections of pairwise interior disjoint well-behaved curves.Comment: The article significantly extends the theoretical aspects of the work
presented in http://arxiv.org/abs/1205.543
Improved Implementation of Point Location in General Two-Dimensional Subdivisions
We present a major revamp of the point-location data structure for general
two-dimensional subdivisions via randomized incremental construction,
implemented in CGAL, the Computational Geometry Algorithms Library. We can now
guarantee that the constructed directed acyclic graph G is of linear size and
provides logarithmic query time. Via the construction of the Voronoi diagram
for a given point set S of size n, this also enables nearest-neighbor queries
in guaranteed O(log n) time. Another major innovation is the support of general
unbounded subdivisions as well as subdivisions of two-dimensional parametric
surfaces such as spheres, tori, cylinders. The implementation is exact,
complete, and general, i.e., it can also handle non-linear subdivisions. Like
the previous version, the data structure supports modifications of the
subdivision, such as insertions and deletions of edges, after the initial
preprocessing. A major challenge is to retain the expected O(n log n)
preprocessing time while providing the above (deterministic) space and
query-time guarantees. We describe an efficient preprocessing algorithm, which
explicitly verifies the length L of the longest query path in O(n log n) time.
However, instead of using L, our implementation is based on the depth D of G.
Although we prove that the worst case ratio of D and L is Theta(n/log n), we
conjecture, based on our experimental results, that this solution achieves
expected O(n log n) preprocessing time.Comment: 21 page
Small Extended Formulation for Knapsack Cover Inequalities from Monotone Circuits
Initially developed for the min-knapsack problem, the knapsack cover
inequalities are used in the current best relaxations for numerous
combinatorial optimization problems of covering type. In spite of their
widespread use, these inequalities yield linear programming (LP) relaxations of
exponential size, over which it is not known how to optimize exactly in
polynomial time. In this paper we address this issue and obtain LP relaxations
of quasi-polynomial size that are at least as strong as that given by the
knapsack cover inequalities.
For the min-knapsack cover problem, our main result can be stated formally as
follows: for any , there is a -size LP relaxation with an integrality gap of at most ,
where is the number of items. Prior to this work, there was no known
relaxation of subexponential size with a constant upper bound on the
integrality gap.
Our construction is inspired by a connection between extended formulations
and monotone circuit complexity via Karchmer-Wigderson games. In particular,
our LP is based on -depth monotone circuits with fan-in~ for
evaluating weighted threshold functions with inputs, as constructed by
Beimel and Weinreb. We believe that a further understanding of this connection
may lead to more positive results complementing the numerous lower bounds
recently proved for extended formulations.Comment: 21 page
Circuit Complexity Meets Ontology-Based Data Access
Ontology-based data access is an approach to organizing access to a database
augmented with a logical theory. In this approach query answering proceeds
through a reformulation of a given query into a new one which can be answered
without any use of theory. Thus the problem reduces to the standard database
setting.
However, the size of the query may increase substantially during the
reformulation. In this survey we review a recently developed framework on
proving lower and upper bounds on the size of this reformulation by employing
methods and results from Boolean circuit complexity.Comment: To appear in proceedings of CSR 2015, LNCS 9139, Springe
Tree-like Queries in OWL 2 QL: Succinctness and Complexity Results
This paper investigates the impact of query topology on the difficulty of
answering conjunctive queries in the presence of OWL 2 QL ontologies. Our first
contribution is to clarify the worst-case size of positive existential (PE),
non-recursive Datalog (NDL), and first-order (FO) rewritings for various
classes of tree-like conjunctive queries, ranging from linear queries to
bounded treewidth queries. Perhaps our most surprising result is a
superpolynomial lower bound on the size of PE-rewritings that holds already for
linear queries and ontologies of depth 2. More positively, we show that
polynomial-size NDL-rewritings always exist for tree-shaped queries with a
bounded number of leaves (and arbitrary ontologies), and for bounded treewidth
queries paired with bounded depth ontologies. For FO-rewritings, we equate the
existence of polysize rewritings with well-known problems in Boolean circuit
complexity. As our second contribution, we analyze the computational complexity
of query answering and establish tractability results (either NL- or
LOGCFL-completeness) for a range of query-ontology pairs. Combining our new
results with those from the literature yields a complete picture of the
succinctness and complexity landscapes for the considered classes of queries
and ontologies.Comment: This is an extended version of a paper accepted at LICS'15. It
contains both succinctness and complexity results and adopts FOL notation.
The appendix contains proofs that had to be omitted from the conference
version for lack of space. The previous arxiv version (a long version of our
DL'14 workshop paper) only contained the succinctness results and used
description logic notatio
Characterizing the entanglement of symmetric many-particle spin-1/2 systems
Analyzing the properties of entanglement in many-particle spin-1/2 systems is
generally difficult because the system's Hilbert space grows exponentially with
the number of constituent particles, . Fortunately, it is still possible to
investigate many-particle entanglement when the state of the system possesses
sufficient symmetry. In this paper, we present a practical method for
efficiently computing various bipartite entanglement measures for states in the
symmetric subspace and perform these calculations for . By
considering all possible bipartite splits, we construct a picture of the
multiscale entanglement in large symmetric systems. In particular, we
characterize dynamically generated spin-squeezed states by comparing them to
known reference states (e.g., GHZ and Dicke states) and new families of states
with near-maximal bipartite entropy. We quantify the trade-off between the
degree of entanglement and its robustness to particle loss, emphasizing that
substantial entanglement need not be fragile.Comment: Updated version reflects changes made in January 200
- …