75 research outputs found
A lower bound for area-universal graphs
We establish a lower bound on the efficiency of area--universal circuits. The area of every graph that can host any graph of area (at most) with dilation , and congestion satisfies the tradeoff In particular, if then
Distributed Symmetry Breaking in Hypergraphs
Fundamental local symmetry breaking problems such as Maximal Independent Set
(MIS) and coloring have been recognized as important by the community, and
studied extensively in (standard) graphs. In particular, fast (i.e.,
logarithmic run time) randomized algorithms are well-established for MIS and
-coloring in both the LOCAL and CONGEST distributed computing
models. On the other hand, comparatively much less is known on the complexity
of distributed symmetry breaking in {\em hypergraphs}. In particular, a key
question is whether a fast (randomized) algorithm for MIS exists for
hypergraphs.
In this paper, we study the distributed complexity of symmetry breaking in
hypergraphs by presenting distributed randomized algorithms for a variety of
fundamental problems under a natural distributed computing model for
hypergraphs. We first show that MIS in hypergraphs (of arbitrary dimension) can
be solved in rounds ( is the number of nodes of the
hypergraph) in the LOCAL model. We then present a key result of this paper ---
an -round hypergraph MIS algorithm in
the CONGEST model where is the maximum node degree of the hypergraph
and is any arbitrarily small constant.
To demonstrate the usefulness of hypergraph MIS, we present applications of
our hypergraph algorithm to solving problems in (standard) graphs. In
particular, the hypergraph MIS yields fast distributed algorithms for the {\em
balanced minimal dominating set} problem (left open in Harris et al. [ICALP
2013]) and the {\em minimal connected dominating set problem}. We also present
distributed algorithms for coloring, maximal matching, and maximal clique in
hypergraphs.Comment: Changes from the previous version: More references adde
Worst case and probabilistic analysis of the 2-Opt algorithm for the TSP
2-Opt is probably the most basic local search heuristic for the TSP. This heuristic achieves amazingly good results on âreal worldâ Euclidean instances both with respect to running time and approximation ratio. There are numerous experimental studies on the performance of 2-Opt. However, the theoretical knowledge about this heuristic is still very limited. Not even its worst case running time on 2-dimensional Euclidean instances was known so far. We clarify this issue by presenting, for every pâN , a family of L p instances on which 2-Opt can take an exponential number of steps.
Previous probabilistic analyses were restricted to instances in which n points are placed uniformly at random in the unit square [0,1]2, where it was shown that the expected number of steps is bounded by O~(n10) for Euclidean instances. We consider a more advanced model of probabilistic instances in which the points can be placed independently according to general distributions on [0,1] d , for an arbitrary dâ„2. In particular, we allow different distributions for different points. We study the expected number of local improvements in terms of the number n of points and the maximal density Ï of the probability distributions. We show an upper bound on the expected length of any 2-Opt improvement path of O~(n4+1/3â
Ï8/3) . When starting with an initial tour computed by an insertion heuristic, the upper bound on the expected number of steps improves even to O~(n4+1/3â1/dâ
Ï8/3) . If the distances are measured according to the Manhattan metric, then the expected number of steps is bounded by O~(n4â1/dâ
Ï) . In addition, we prove an upper bound of O(Ïâd) on the expected approximation factor with respect to all L p metrics.
Let us remark that our probabilistic analysis covers as special cases the uniform input model with Ï=1 and a smoothed analysis with Gaussian perturbations of standard deviation Ï with ÏâŒ1/Ï d
Revisit Sparse Polynomial Interpolation based on Randomized Kronecker Substitution
In this paper, a new reduction based interpolation algorithm for black-box
multivariate polynomials over finite fields is given. The method is based on
two main ingredients. A new Monte Carlo method is given to reduce black-box
multivariate polynomial interpolation to black-box univariate polynomial
interpolation over any ring. The reduction algorithm leads to multivariate
interpolation algorithms with better or the same complexities most cases when
combining with various univariate interpolation algorithms. We also propose a
modified univariate Ben-or and Tiwarri algorithm over the finite field, which
has better total complexity than the Lagrange interpolation algorithm.
Combining our reduction method and the modified univariate Ben-or and Tiwarri
algorithm, we give a Monte Carlo multivariate interpolation algorithm, which
has better total complexity in most cases for sparse interpolation of black-box
polynomial over finite fields
Maximum-Reward Motion in a Stochastic Environment: The Nonequilibrium Statistical Mechanics Perspective
We consider the problem of computing the maximum-reward motion in a reward field in an online setting. We assume that the robot has a limited perception range, and it discovers the reward field on the fly. We analyze the performance of a simple, practical lattice-based algorithm with respect to the perception range. Our main result is that, with very little perception range, the robot can collect as much reward as if it could see the whole reward field, under certain assumptions. Along the way, we establish novel connections between this class of problems and certain fundamental problems of nonequilibrium statistical mechanics . We demonstrate our results in simulation examples
LEGaTO: first steps towards energy-efficient toolset for heterogeneous computing
LEGaTO is a three-year EU H2020 project which started in December 2017. The LEGaTO project will leverage task-based programming models to provide a software ecosystem for Made-in-Europe heterogeneous hardware composed of CPUs, GPUs, FPGAs and dataflow engines. The aim is to attain one order of magnitude energy savings from the edge to the converged cloud/HPC.Peer ReviewedPostprint (author's final draft
Searchable Encryption with Optimal Locality: Achieving Sublogarithmic Read Efficiency
We propose the first linear-space searchable encryption scheme with constant locality and \emph{sublogarithmic} read efficiency, strictly improving the previously best known read efficiency bound (Asharov et al., STOC 2016) from to where for any fixed . Our scheme employs four different allocation algorithms for storing the keyword lists, depending on the size of the list considered each time. For our construction we develop (i) new probability bounds for the offline two-choice allocation problem; (ii) and a new I/O-efficient oblivious RAM with bandwidth overhead and zero failure probability, both of which can be of independent interest
LEGaTO: towards energy-efficient, secure, fault-tolerant toolset for heterogeneous computing
LEGaTO is a three-year EU H2020 project which started in December 2017. The LEGaTO project will leverage task-based programming models to provide a software ecosystem for Made-in-Europe heterogeneous hardware composed of CPUs, GPUs, FPGAs and dataflow engines. The aim is to attain one order of magnitude energy savings from the edge to the converged cloud/HPC.Peer ReviewedPostprint (author's final draft
- âŠ