620 research outputs found
On Online Labeling with Polynomially Many Labels
In the online labeling problem with parameters n and m we are presented with
a sequence of n keys from a totally ordered universe U and must assign each
arriving key a label from the label set {1,2,...,m} so that the order of labels
(strictly) respects the ordering on U. As new keys arrive it may be necessary
to change the labels of some items; such changes may be done at any time at
unit cost for each change. The goal is to minimize the total cost. An
alternative formulation of this problem is the file maintenance problem, in
which the items, instead of being labeled, are maintained in sorted order in an
array of length m, and we pay unit cost for moving an item.
For the case m=cn for constant c>1, there are known algorithms that use at
most O(n log(n)^2) relabelings in total [Itai, Konheim, Rodeh, 1981], and it
was shown recently that this is asymptotically optimal [Bul\'anek, Kouck\'y,
Saks, 2012]. For the case of m={\Theta}(n^C) for C>1, algorithms are known that
use O(n log n) relabelings. A matching lower bound was claimed in [Dietz,
Seiferas, Zhang, 2004]. That proof involved two distinct steps: a lower bound
for a problem they call prefix bucketing and a reduction from prefix bucketing
to online labeling. The reduction seems to be incorrect, leaving a (seemingly
significant) gap in the proof. In this paper we close the gap by presenting a
correct reduction to prefix bucketing. Furthermore we give a simplified and
improved analysis of the prefix bucketing lower bound. This improvement allows
us to extend the lower bounds for online labeling to the case where the number
m of labels is superpolynomial in n. In particular, for superpolynomial m we
get an asymptotically optimal lower bound {\Omega}((n log n) / (log log m - log
log n)).Comment: 15 pages, Presented at European Symposium on Algorithms 201
Cross-Document Pattern Matching
We study a new variant of the string matching problem called cross-document
string matching, which is the problem of indexing a collection of documents to
support an efficient search for a pattern in a selected document, where the
pattern itself is a substring of another document. Several variants of this
problem are considered, and efficient linear-space solutions are proposed with
query time bounds that either do not depend at all on the pattern size or
depend on it in a very limited way (doubly logarithmic). As a side result, we
propose an improved solution to the weighted level ancestor problem
Entanglement Efficiencies in PT-Symmetric Quantum Mechanics
The degree of entanglement is determined for an arbitrary state of a broad
class of PT-symmetric bipartite composite systems. Subsequently we quantify the
rate with which entangled states are generated and show that this rate can be
characterized by a small set of parameters. These relations allow one in
principle to improve the ability of these systems to entangle states. It is
also noticed that many relations resemble corresponding ones in conventional
quantum mechanics.Comment: Published version with improved figures, 5 pages, 2 figure
Weighted ancestors in suffix trees
The classical, ubiquitous, predecessor problem is to construct a data
structure for a set of integers that supports fast predecessor queries. Its
generalization to weighted trees, a.k.a. the weighted ancestor problem, has
been extensively explored and successfully reduced to the predecessor problem.
It is known that any solution for both problems with an input set from a
polynomially bounded universe that preprocesses a weighted tree in O(n
polylog(n)) space requires \Omega(loglogn) query time. Perhaps the most
important and frequent application of the weighted ancestors problem is for
suffix trees. It has been a long-standing open question whether the weighted
ancestors problem has better bounds for suffix trees. We answer this question
positively: we show that a suffix tree built for a text w[1..n] can be
preprocessed using O(n) extra space, so that queries can be answered in O(1)
time. Thus we improve the running times of several applications. Our
improvement is based on a number of data structure tools and a
periodicity-based insight into the combinatorial structure of a suffix tree.Comment: 27 pages, LNCS format. A condensed version will appear in ESA 201
PT Invariant Complex E (8) Root Spaces
We provide a construction procedure for complex root spaces invariant under antilinear transformations, which may be applied to any Coxeter group. The procedure is based on the factorisation of a chosen element of the Coxeter group into two factors. Each of the factors constitutes an involution and may therefore be deformed in an antilinear fashion. Having the importance of the E(8)-Coxeter group in mind, such as underlying a particular perturbation of the Ising model and the fact that for it no solution could be found previously, we exemplify the procedure for this particular case. As a concrete application of this construction we propose new generalisations of Calogero-Moser Sutherland models and affine Toda field theories based on the invariant complex root spaces and deformed complex simple roots, respectively
Cache-Oblivious Persistence
Partial persistence is a general transformation that takes a data structure
and allows queries to be executed on any past state of the structure. The
cache-oblivious model is the leading model of a modern multi-level memory
hierarchy.We present the first general transformation for making
cache-oblivious model data structures partially persistent
Blame Trees
We consider the problem of merging individual text documents, motivated by the single-file merge algorithms of document-based version control systems. Abstracting away the merging of conflicting edits to an external conflict resolution function (possibly implemented by a human), we consider the efficient identification of conflicting regions. We show how to implement tree-based document representation to quickly answer a data structure inspired by the “blame” query of some version control systems. A “blame” query associates every line of a document with the revision in which it was last edited. Our tree uses this idea to quickly identify conflicting edits. We show how to perform a merge operation in time proportional to the sum of the logarithms of the shared regions of the documents, plus the cost of conflict resolution. Our data structure is functional and therefore confluently persistent, allowing arbitrary version DAGs as in real version-control systems. Our results rely on concurrent traversal of two trees with short circuiting when shared subtrees are encountered.United States. Defense Advanced Research Projects Agency (Clean-Slate Design of Resilient, Adaptive, Secure Hosts (CRASH) program, BAA10-70)United States. Defense Advanced Research Projects Agency (contract #N66001-10-2-4088 (Bridging the Security Gap with Decentralized Information Flow Control))Danish National Research Foundation (Center for Massive Data Algorithmics (MADALGO)
Tree-based Coarsening and Partitioning of Complex Networks
Many applications produce massive complex networks whose analysis would
benefit from parallel processing. Parallel algorithms, in turn, often require a
suitable network partition. For solving optimization tasks such as graph
partitioning on large networks, multilevel methods are preferred in practice.
Yet, complex networks pose challenges to established multilevel algorithms, in
particular to their coarsening phase.
One way to specify a (recursive) coarsening of a graph is to rate its edges
and then contract the edges as prioritized by the rating. In this paper we (i)
define weights for the edges of a network that express the edges' importance
for connectivity, (ii) compute a minimum weight spanning tree with
respect to these weights, and (iii) rate the network edges based on the
conductance values of 's fundamental cuts. To this end, we also (iv)
develop the first optimal linear-time algorithm to compute the conductance
values of \emph{all} fundamental cuts of a given spanning tree. We integrate
the new edge rating into a leading multilevel graph partitioner and equip the
latter with a new greedy postprocessing for optimizing the maximum
communication volume (MCV). Experiments on bipartitioning frequently used
benchmark networks show that the postprocessing already reduces MCV by 11.3%.
Our new edge rating further reduces MCV by 10.3% compared to the previously
best rating with the postprocessing in place for both ratings. In total, with a
modest increase in running time, our new approach reduces the MCV of complex
network partitions by 20.4%
Analysis of a quenched lattice-QCD dressed-quark propagator
Quenched lattice-QCD data on the dressed-quark Schwinger function can be
correlated with dressed-gluon data via a rainbow gap equation so long as that
equation's kernel possesses enhancement at infrared momenta above that
exhibited by the gluon alone. The required enhancement can be ascribed to a
dressing of the quark-gluon vertex. The solutions of the rainbow gap equation
exhibit dynamical chiral symmetry breaking and are consistent with confinement.
The gap equation and related, symmetry-preserving ladder Bethe-Salpeter
equation yield estimates for chiral and physical pion observables that suggest
these quantities are materially underestimated in the quenched theory: |<bar-q
q>| by a factor of two and f_pi by 30%.Comment: 9 pages, LaTeX2e, REVTEX4, 6 figure
Scheduling Jobs in Flowshops with the Introduction of Additional Machines in the Future
This is the author's peer-reviewed final manuscript, as accepted by the publisher. The published article is copyrighted by Elsevier and can be found at: http://www.journals.elsevier.com/expert-systems-with-applications/.The problem of scheduling jobs to minimize total weighted tardiness in flowshops,\ud
with the possibility of evolving into hybrid flowshops in the future, is investigated in\ud
this paper. As this research is guided by a real problem in industry, the flowshop\ud
considered has considerable flexibility, which stimulated the development of an\ud
innovative methodology for this research. Each stage of the flowshop currently has\ud
one or several identical machines. However, the manufacturing company is planning\ud
to introduce additional machines with different capabilities in different stages in the\ud
near future. Thus, the algorithm proposed and developed for the problem is not only\ud
capable of solving the current flow line configuration but also the potential new\ud
configurations that may result in the future. A meta-heuristic search algorithm based\ud
on Tabu search is developed to solve this NP-hard, industry-guided problem. Six\ud
different initial solution finding mechanisms are proposed. A carefully planned\ud
nested split-plot design is performed to test the significance of different factors and\ud
their impact on the performance of the different algorithms. To the best of our\ud
knowledge, this research is the first of its kind that attempts to solve an industry-guided\ud
problem with the concern for future developments
- …