6,083 research outputs found
Randomized benchmarking of single and multi-qubit control in liquid-state NMR quantum information processing
Being able to quantify the level of coherent control in a proposed device
implementing a quantum information processor (QIP) is an important task for
both comparing different devices and assessing a device's prospects with
regards to achieving fault-tolerant quantum control. We implement in a
liquid-state nuclear magnetic resonance QIP the randomized benchmarking
protocol presented by Knill et al (PRA 77: 012307 (2008)). We report an error
per randomized pulse of with a
single qubit QIP and show an experimentally relevant error model where the
randomized benchmarking gives a signature fidelity decay which is not possible
to interpret as a single error per gate. We explore and experimentally
investigate multi-qubit extensions of this protocol and report an average error
rate for one and two qubit gates of for a three
qubit QIP. We estimate that these error rates are still not decoherence limited
and thus can be improved with modifications to the control hardware and
software.Comment: 10 pages, 6 figures, submitted versio
Online Permutation Routing in Partitioned Optical Passive Star Networks
This paper establishes the state of the art in both deterministic and
randomized online permutation routing in the POPS network. Indeed, we show that
any permutation can be routed online on a POPS network either with
deterministic slots, or, with high probability, with
randomized slots, where constant
. When , that we claim to be the
"interesting" case, the randomized algorithm is exponentially faster than any
other algorithm in the literature, both deterministic and randomized ones. This
is true in practice as well. Indeed, experiments show that it outperforms its
rivals even starting from as small a network as a POPS(2,2), and the gap grows
exponentially with the size of the network. We can also show that, under proper
hypothesis, no deterministic algorithm can asymptotically match its
performance
The Genomic HyperBrowser: inferential genomics at the sequence level
The immense increase in the generation of genomic scale data poses an unmet
analytical challenge, due to a lack of established methodology with the
required flexibility and power. We propose a first principled approach to
statistical analysis of sequence-level genomic information. We provide a
growing collection of generic biological investigations that query pairwise
relations between tracks, represented as mathematical objects, along the
genome. The Genomic HyperBrowser implements the approach and is available at
http://hyperbrowser.uio.no
Recommended from our members
Preparing sparse solvers for exascale computing.
Sparse solvers provide essential functionality for a wide variety of scientific applications. Highly parallel sparse solvers are essential for continuing advances in high-fidelity, multi-physics and multi-scale simulations, especially as we target exascale platforms. This paper describes the challenges, strategies and progress of the US Department of Energy Exascale Computing project towards providing sparse solvers for exascale computing platforms. We address the demands of systems with thousands of high-performance node devices where exposing concurrency, hiding latency and creating alternative algorithms become essential. The efforts described here are works in progress, highlighting current success and upcoming challenges. This article is part of a discussion meeting issue 'Numerical algorithms for high-performance computational science'
PF-OLA: A High-Performance Framework for Parallel On-Line Aggregation
Online aggregation provides estimates to the final result of a computation
during the actual processing. The user can stop the computation as soon as the
estimate is accurate enough, typically early in the execution. This allows for
the interactive data exploration of the largest datasets. In this paper we
introduce the first framework for parallel online aggregation in which the
estimation virtually does not incur any overhead on top of the actual
execution. We define a generic interface to express any estimation model that
abstracts completely the execution details. We design a novel estimator
specifically targeted at parallel online aggregation. When executed by the
framework over a massive TPC-H instance, the estimator provides
accurate confidence bounds early in the execution even when the cardinality of
the final result is seven orders of magnitude smaller than the dataset size and
without incurring overhead.Comment: 36 page
- …