36,986 research outputs found
Non-Local Probes Do Not Help with Graph Problems
This work bridges the gap between distributed and centralised models of
computing in the context of sublinear-time graph algorithms. A priori, typical
centralised models of computing (e.g., parallel decision trees or centralised
local algorithms) seem to be much more powerful than distributed
message-passing algorithms: centralised algorithms can directly probe any part
of the input, while in distributed algorithms nodes can only communicate with
their immediate neighbours. We show that for a large class of graph problems,
this extra freedom does not help centralised algorithms at all: for example,
efficient stateless deterministic centralised local algorithms can be simulated
with efficient distributed message-passing algorithms. In particular, this
enables us to transfer existing lower bound results from distributed algorithms
to centralised local algorithms
Algorithmic patterns for -matrices on many-core processors
In this work, we consider the reformulation of hierarchical ()
matrix algorithms for many-core processors with a model implementation on
graphics processing units (GPUs). matrices approximate specific
dense matrices, e.g., from discretized integral equations or kernel ridge
regression, leading to log-linear time complexity in dense matrix-vector
products. The parallelization of matrix operations on many-core
processors is difficult due to the complex nature of the underlying algorithms.
While previous algorithmic advances for many-core hardware focused on
accelerating existing matrix CPU implementations by many-core
processors, we here aim at totally relying on that processor type. As main
contribution, we introduce the necessary parallel algorithmic patterns allowing
to map the full matrix construction and the fast matrix-vector
product to many-core hardware. Here, crucial ingredients are space filling
curves, parallel tree traversal and batching of linear algebra operations. The
resulting model GPU implementation hmglib is the, to the best of the authors
knowledge, first entirely GPU-based Open Source matrix library of
this kind. We conclude this work by an in-depth performance analysis and a
comparative performance study against a standard matrix library,
highlighting profound speedups of our many-core parallel approach
Data Structures in Classical and Quantum Computing
This survey summarizes several results about quantum computing related to
(mostly static) data structures. First, we describe classical data structures
for the set membership and the predecessor search problems: Perfect Hash tables
for set membership by Fredman, Koml\'{o}s and Szemer\'{e}di and a data
structure by Beame and Fich for predecessor search. We also prove results about
their space complexity (how many bits are required) and time complexity (how
many bits have to be read to answer a query). After that, we turn our attention
to classical data structures with quantum access. In the quantum access model,
data is stored in classical bits, but they can be accessed in a quantum way: We
may read several bits in superposition for unit cost. We give proofs for lower
bounds in this setting that show that the classical data structures from the
first section are, in some sense, asymptotically optimal - even in the quantum
model. In fact, these proofs are simpler and give stronger results than
previous proofs for the classical model of computation. The lower bound for set
membership was proved by Radhakrishnan, Sen and Venkatesh and the result for
the predecessor problem by Sen and Venkatesh. Finally, we examine fully quantum
data structures. Instead of encoding the data in classical bits, we now encode
it in qubits. We allow any unitary operation or measurement in order to answer
queries. We describe one data structure by de Wolf for the set membership
problem and also a general framework using fully quantum data structures in
quantum walks by Jeffery, Kothari and Magniez
- …