12,838 research outputs found
A Greedy Hypercube-Labeling Algorithm
Due to its attractive topological properties, the hypercube multiprocessor has emerged as one of the architectures of choice when it comes to implementing a large number of computational problems. In many such applications, Gray-code labelings of the hypercube are a crucial prerequisite for obtaining efficient algorithms. We propose a greedy algorithm that, given an n-dimensional hypercube H with N=22 nodes, returns a Gray-code labeling of H, that is, a labeling of the nodes with binary strings of length n such that two nodes are neighbors in the hypercube if, and only if, their labels differ in exactly one bit. Our algorithm is conceptually very simple and runs in O(N log N) time being, therefore, optimal. As it turns out, with a few modifications our labeling algorithm can be used to recognize hypercubes as well
Embedding cube-connected cycles graphs into faulty hypercubes
We consider the problem of embedding a cube-connected cycles graph (CCC) into a hypercube with edge faults. Our main result is an algorithm that, given a list of faulty edges, computes an embedding of the CCC that spans all of the nodes and avoids all of the faulty edges. The algorithm has optimal running time and tolerates the maximum number of faults (in a worst-case setting). Because ascend-descend algorithms can be implemented efficiently on a CCC, this embedding enables the implementation of ascend-descend algorithms, such as bitonic sort, on hypercubes with edge faults. We also present a number of related results, including an algorithm for embedding a CCC into a hypercube with edge and node faults and an algorithm for embedding a spanning torus into a hypercube with edge faults
Multiresolution vector quantization
Multiresolution source codes are data compression algorithms yielding embedded source descriptions. The decoder of a multiresolution code can build a source reproduction by decoding the embedded bit stream in part or in whole. All decoding procedures start at the beginning of the binary source description and decode some fraction of that string. Decoding a small portion of the binary string gives a low-resolution reproduction; decoding more yields a higher resolution reproduction; and so on. Multiresolution vector quantizers are block multiresolution source codes. This paper introduces algorithms for designing fixed- and variable-rate multiresolution vector quantizers. Experiments on synthetic data demonstrate performance close to the theoretical performance limit. Experiments on natural images demonstrate performance improvements of up to 8 dB over tree-structured vector quantizers. Some of the lessons learned through multiresolution vector quantizer design lend insight into the design of more sophisticated multiresolution codes
New constructions for covering designs
A {\em covering design}, or {\em covering}, is a family of
-subsets, called blocks, chosen from a -set, such that each -subset is
contained in at least one of the blocks. The number of blocks is the covering's
{\em size}, and the minimum size of such a covering is denoted by .
This paper gives three new methods for constructing good coverings: a greedy
algorithm similar to Conway and Sloane's algorithm for lexicographic
codes~\cite{lex}, and two methods that synthesize new coverings from
preexisting ones. Using these new methods, together with results in the
literature, we build tables of upper bounds on for , , and .
On connectivity-dependent resource requirements for digital quantum simulation of -level particles
A primary objective of quantum computation is to efficiently simulate quantum
physics. Scientifically and technologically important quantum Hamiltonians
include those with spin-, vibrational, photonic, and other bosonic degrees
of freedom, i.e. problems composed of or approximated by -level particles
(qudits). Recently, several methods for encoding these systems into a set of
qubits have been introduced, where each encoding's efficiency was studied in
terms of qubit and gate counts. Here, we build on previous results by including
effects of hardware connectivity. To study the number of SWAP gates required to
Trotterize commonly used quantum operators, we use both analytical arguments
and automatic tools that optimize the schedule in multiple stages. We study the
unary (or one-hot), Gray, standard binary, and block unary encodings, with
three connectivities: linear array, ladder array, and square grid. Among other
trends, we find that while the ladder array leads to substantial efficiencies
over the linear array, the advantage of the square over the ladder array is
less pronounced. These results are applicable in hardware co-design and in
choosing efficient qudit encodings for a given set of near-term quantum
hardware. Additionally, this work may be relevant to the scheduling of other
quantum algorithms for which matrix exponentiation is a subroutine.Comment: Accepted to QCE20 (IEEE Quantum Week). Corrected erroneous circuits
in Figure
- …