4,717 research outputs found
Worst case and probabilistic analysis of the 2-Opt algorithm for the TSP
2-Opt is probably the most basic local search heuristic for the TSP. This heuristic achieves amazingly good results on âreal worldâ Euclidean instances both with respect to running time and approximation ratio. There are numerous experimental studies on the performance of 2-Opt. However, the theoretical knowledge about this heuristic is still very limited. Not even its worst case running time on 2-dimensional Euclidean instances was known so far. We clarify this issue by presenting, for every pâN , a family of L p instances on which 2-Opt can take an exponential number of steps.
Previous probabilistic analyses were restricted to instances in which n points are placed uniformly at random in the unit square [0,1]2, where it was shown that the expected number of steps is bounded by O~(n10) for Euclidean instances. We consider a more advanced model of probabilistic instances in which the points can be placed independently according to general distributions on [0,1] d , for an arbitrary dâ„2. In particular, we allow different distributions for different points. We study the expected number of local improvements in terms of the number n of points and the maximal density Ï of the probability distributions. We show an upper bound on the expected length of any 2-Opt improvement path of O~(n4+1/3â
Ï8/3) . When starting with an initial tour computed by an insertion heuristic, the upper bound on the expected number of steps improves even to O~(n4+1/3â1/dâ
Ï8/3) . If the distances are measured according to the Manhattan metric, then the expected number of steps is bounded by O~(n4â1/dâ
Ï) . In addition, we prove an upper bound of O(Ïâd) on the expected approximation factor with respect to all L p metrics.
Let us remark that our probabilistic analysis covers as special cases the uniform input model with Ï=1 and a smoothed analysis with Gaussian perturbations of standard deviation Ï with ÏâŒ1/Ï d
Online Multi-Coloring with Advice
We consider the problem of online graph multi-coloring with advice.
Multi-coloring is often used to model frequency allocation in cellular
networks. We give several nearly tight upper and lower bounds for the most
standard topologies of cellular networks, paths and hexagonal graphs. For the
path, negative results trivially carry over to bipartite graphs, and our
positive results are also valid for bipartite graphs. The advice given
represents information that is likely to be available, studying for instance
the data from earlier similar periods of time.Comment: IMADA-preprint-c
Optimal Online Edge Coloring of Planar Graphs with Advice
Using the framework of advice complexity, we study the amount of knowledge
about the future that an online algorithm needs to color the edges of a graph
optimally, i.e., using as few colors as possible. For graphs of maximum degree
, it follows from Vizing's Theorem that bits of
advice suffice to achieve optimality, where is the number of edges. We show
that for graphs of bounded degeneracy (a class of graphs including e.g. trees
and planar graphs), only bits of advice are needed to compute an optimal
solution online, independently of how large is. On the other hand, we
show that bits of advice are necessary just to achieve a
competitive ratio better than that of the best deterministic online algorithm
without advice. Furthermore, we consider algorithms which use a fixed number of
advice bits per edge (our algorithm for graphs of bounded degeneracy belongs to
this class of algorithms). We show that for bipartite graphs, any such
algorithm must use at least bits of advice to achieve
optimality.Comment: CIAC 201
On smoothed analysis of quicksort and Hoare's find
We provide a smoothed analysis of Hoare's find algorithm, and we revisit the smoothed analysis of quicksort. Hoare's find algorithm - often called quickselect or one-sided quicksort - is an easy-to-implement algorithm for finding the k-th smallest element of a sequence. While the worst-case number of comparisons that Hoareâs find needs is Theta(n^2), the average-case number is Theta(n). We analyze what happens between these two extremes by providing a smoothed analysis. In the first perturbation model, an adversary specifies a sequence of n numbers of [0,1], and then, to each number of the sequence, we add a random number drawn independently from the interval [0,d]. We prove that Hoare's find needs Theta(n/(d+1) sqrt(n/d) + n) comparisons in expectation if the adversary may also specify the target element (even after seeing the perturbed sequence) and slightly fewer comparisons for finding the median. In the second perturbation model, each element is marked with a probability of p, and then a random permutation is applied to the marked elements. We prove that the expected number of comparisons to find the median is Omega((1âp)n/p log n). Finally, we provide lower bounds for the smoothed number of comparisons of quicksort and Hoareâs find for the median-of-three pivot rule, which usually yields faster algorithms than always selecting the first element: The pivot is the median of the first, middle, and last element of the sequence. We show that median-of-three does not yield a significant improvement over the classic rule
Fast Algorithm for Partial Covers in Words
A factor of a word is a cover of if every position in lies
within some occurrence of in . A word covered by thus
generalizes the idea of a repetition, that is, a word composed of exact
concatenations of . In this article we introduce a new notion of
-partial cover, which can be viewed as a relaxed variant of cover, that
is, a factor covering at least positions in . We develop a data
structure of size (where ) that can be constructed in time which we apply to compute all shortest -partial covers for a
given . We also employ it for an -time algorithm computing
a shortest -partial cover for each
Fast branching algorithm for Cluster Vertex Deletion
In the family of clustering problems, we are given a set of objects (vertices
of the graph), together with some observed pairwise similarities (edges). The
goal is to identify clusters of similar objects by slightly modifying the graph
to obtain a cluster graph (disjoint union of cliques). Hueffner et al. [Theory
Comput. Syst. 2010] initiated the parameterized study of Cluster Vertex
Deletion, where the allowed modification is vertex deletion, and presented an
elegant O(2^k * k^9 + n * m)-time fixed-parameter algorithm, parameterized by
the solution size. In our work, we pick up this line of research and present an
O(1.9102^k * (n + m))-time branching algorithm
- âŠ