315,358 research outputs found
Optimization of Partial Search
Quantum Grover search algorithm can find a target item in a database faster
than any classical algorithm. One can trade accuracy for speed and find a part
of the database (a block) containing the target item even faster, this is
partial search. A partial search algorithm was recently suggested by Grover and
Radhakrishnan. Here we optimize it. Efficiency of the search algorithm is
measured by number of queries to the oracle. The author suggests new version of
Grover-Radhakrishnan algorithm which uses minimal number of queries to the
oracle. The algorithm can run on the same hardware which is used for the usual
Grover algorithm.Comment: 5 page
An Efficient Local Search for Partial Latin Square Extension Problem
A partial Latin square (PLS) is a partial assignment of n symbols to an nxn
grid such that, in each row and in each column, each symbol appears at most
once. The partial Latin square extension problem is an NP-hard problem that
asks for a largest extension of a given PLS. In this paper we propose an
efficient local search for this problem. We focus on the local search such that
the neighborhood is defined by (p,q)-swap, i.e., removing exactly p symbols and
then assigning symbols to at most q empty cells. For p in {1,2,3}, our
neighborhood search algorithm finds an improved solution or concludes that no
such solution exists in O(n^{p+1}) time. We also propose a novel swap
operation, Trellis-swap, which is a generalization of (1,q)-swap and
(2,q)-swap. Our Trellis-neighborhood search algorithm takes O(n^{3.5}) time to
do the same thing. Using these neighborhood search algorithms, we design a
prototype iterated local search algorithm and show its effectiveness in
comparison with state-of-the-art optimization solvers such as IBM ILOG CPLEX
and LocalSolver.Comment: 17 pages, 2 figure
Interior Point Decoding for Linear Vector Channels
In this paper, a novel decoding algorithm for low-density parity-check (LDPC)
codes based on convex optimization is presented. The decoding algorithm, called
interior point decoding, is designed for linear vector channels. The linear
vector channels include many practically important channels such as inter
symbol interference channels and partial response channels. It is shown that
the maximum likelihood decoding (MLD) rule for a linear vector channel can be
relaxed to a convex optimization problem, which is called a relaxed MLD
problem. The proposed decoding algorithm is based on a numerical optimization
technique so called interior point method with barrier function. Approximate
variations of the gradient descent and the Newton methods are used to solve the
convex optimization problem. In a decoding process of the proposed algorithm, a
search point always lies in the fundamental polytope defined based on a
low-density parity-check matrix. Compared with a convectional joint message
passing decoder, the proposed decoding algorithm achieves better BER
performance with less complexity in the case of partial response channels in
many cases.Comment: 18 pages, 17 figures, The paper has been submitted to IEEE
Transaction on Information Theor
Perron vector optimization applied to search engines
In the last years, Google's PageRank optimization problems have been
extensively studied. In that case, the ranking is given by the invariant
measure of a stochastic matrix. In this paper, we consider the more general
situation in which the ranking is determined by the Perron eigenvector of a
nonnegative, but not necessarily stochastic, matrix, in order to cover
Kleinberg's HITS algorithm. We also give some results for Tomlin's HOTS
algorithm. The problem consists then in finding an optimal outlink strategy
subject to design constraints and for a given search engine.
We study the relaxed versions of these problems, which means that we should
accept weighted hyperlinks. We provide an efficient algorithm for the
computation of the matrix of partial derivatives of the criterion, that uses
the low rank property of this matrix. We give a scalable algorithm that couples
gradient and power iterations and gives a local minimum of the Perron vector
optimization problem. We prove convergence by considering it as an approximate
gradient method.
We then show that optimal linkage stategies of HITS and HOTS optimization
problems verify a threshold property. We report numerical results on fragments
of the real web graph for these search engine optimization problems.Comment: 28 pages, 5 figure
An Evolutionary Strategy based on Partial Imitation for Solving Optimization Problems
In this work we introduce an evolutionary strategy to solve combinatorial
optimization tasks, i.e. problems characterized by a discrete search space. In
particular, we focus on the Traveling Salesman Problem (TSP), i.e. a famous
problem whose search space grows exponentially, increasing the number of
cities, up to becoming NP-hard. The solutions of the TSP can be codified by
arrays of cities, and can be evaluated by fitness, computed according to a cost
function (e.g. the length of a path). Our method is based on the evolution of
an agent population by means of an imitative mechanism, we define `partial
imitation'. In particular, agents receive a random solution and then,
interacting among themselves, may imitate the solutions of agents with a higher
fitness. Since the imitation mechanism is only partial, agents copy only one
entry (randomly chosen) of another array (i.e. solution). In doing so, the
population converges towards a shared solution, behaving like a spin system
undergoing a cooling process, i.e. driven towards an ordered phase. We
highlight that the adopted `partial imitation' mechanism allows the population
to generate solutions over time, before reaching the final equilibrium. Results
of numerical simulations show that our method is able to find, in a finite
time, both optimal and suboptimal solutions, depending on the size of the
considered search space.Comment: 18 pages, 6 figure
- …