9 research outputs found
DARIO: a ncRNA detection and analysis tool for next-generation sequencing experiments
Small non-coding RNAs (ncRNAs) such as microRNAs, snoRNAs and tRNAs are a diverse collection of molecules with several important biological functions. Current methods for high-throughput sequencing for the first time offer the opportunity to investigate the entire ncRNAome in an essentially unbiased way. However, there is a substantial need for methods that allow a convenient analysis of these overwhelmingly large data sets. Here, we present DARIO, a free web service that allows to study short read data from small RNA-seq experiments. It provides a wide range of analysis features, including quality control, read normalization, ncRNA quantification and prediction of putative ncRNA candidates. The DARIO web site can be accessed at http://dario.bioinf.uni-leipzig.de/
Suppressing quantum errors by scaling a surface code logical qubit
Practical quantum computing will require error rates that are well below what
is achievable with physical qubits. Quantum error correction offers a path to
algorithmically-relevant error rates by encoding logical qubits within many
physical qubits, where increasing the number of physical qubits enhances
protection against physical errors. However, introducing more qubits also
increases the number of error sources, so the density of errors must be
sufficiently low in order for logical performance to improve with increasing
code size. Here, we report the measurement of logical qubit performance scaling
across multiple code sizes, and demonstrate that our system of superconducting
qubits has sufficient performance to overcome the additional errors from
increasing qubit number. We find our distance-5 surface code logical qubit
modestly outperforms an ensemble of distance-3 logical qubits on average, both
in terms of logical error probability over 25 cycles and logical error per
cycle ( compared to ). To investigate
damaging, low-probability error sources, we run a distance-25 repetition code
and observe a logical error per round floor set by a single
high-energy event ( when excluding this event). We are able
to accurately model our experiment, and from this model we can extract error
budgets that highlight the biggest challenges for future systems. These results
mark the first experimental demonstration where quantum error correction begins
to improve performance with increasing qubit number, illuminating the path to
reaching the logical error rates required for computation.Comment: Main text: 6 pages, 4 figures. v2: Update author list, references,
Fig. S12, Table I
Measurement-induced entanglement and teleportation on a noisy quantum processor
Measurement has a special role in quantum theory: by collapsing the
wavefunction it can enable phenomena such as teleportation and thereby alter
the "arrow of time" that constrains unitary evolution. When integrated in
many-body dynamics, measurements can lead to emergent patterns of quantum
information in space-time that go beyond established paradigms for
characterizing phases, either in or out of equilibrium. On present-day NISQ
processors, the experimental realization of this physics is challenging due to
noise, hardware limitations, and the stochastic nature of quantum measurement.
Here we address each of these experimental challenges and investigate
measurement-induced quantum information phases on up to 70 superconducting
qubits. By leveraging the interchangeability of space and time, we use a
duality mapping, to avoid mid-circuit measurement and access different
manifestations of the underlying phases -- from entanglement scaling to
measurement-induced teleportation -- in a unified way. We obtain finite-size
signatures of a phase transition with a decoding protocol that correlates the
experimental measurement record with classical simulation data. The phases
display sharply different sensitivity to noise, which we exploit to turn an
inherent hardware limitation into a useful diagnostic. Our work demonstrates an
approach to realize measurement-induced physics at scales that are at the
limits of current NISQ processors
Non-Abelian braiding of graph vertices in a superconducting processor
Indistinguishability of particles is a fundamental principle of quantum
mechanics. For all elementary and quasiparticles observed to date - including
fermions, bosons, and Abelian anyons - this principle guarantees that the
braiding of identical particles leaves the system unchanged. However, in two
spatial dimensions, an intriguing possibility exists: braiding of non-Abelian
anyons causes rotations in a space of topologically degenerate wavefunctions.
Hence, it can change the observables of the system without violating the
principle of indistinguishability. Despite the well developed mathematical
description of non-Abelian anyons and numerous theoretical proposals, the
experimental observation of their exchange statistics has remained elusive for
decades. Controllable many-body quantum states generated on quantum processors
offer another path for exploring these fundamental phenomena. While efforts on
conventional solid-state platforms typically involve Hamiltonian dynamics of
quasi-particles, superconducting quantum processors allow for directly
manipulating the many-body wavefunction via unitary gates. Building on
predictions that stabilizer codes can host projective non-Abelian Ising anyons,
we implement a generalized stabilizer code and unitary protocol to create and
braid them. This allows us to experimentally verify the fusion rules of the
anyons and braid them to realize their statistics. We then study the prospect
of employing the anyons for quantum computation and utilize braiding to create
an entangled state of anyons encoding three logical qubits. Our work provides
new insights about non-Abelian braiding and - through the future inclusion of
error correction to achieve topological protection - could open a path toward
fault-tolerant quantum computing
Quantum supremacy using a programmable superconducting processor
The promise of quantum computers is that certain computational tasks might be executed exponentially faster on a quantum processor than on a classical processor1. A fundamental challenge is to build a high-fidelity processor capable of running quantum algorithms in an exponentially large computational space. Here we report the use of a processor with programmable superconducting qubits to create quantum states on 53 qubits, corresponding to a computational state-space of dimension 2^53 (about 10^16). Measurements from repeated experiments sample the resulting probability distribution, which we verify using classical simulations. Our Sycamore processor takes about 200 seconds to sample one instance of a quantum circuit a million times—our benchmarks currently indicate that the equivalent task for a state-of-the-art classical supercomputer would take approximately 10,000 years. This dramatic increase in speed compared to all known classical algorithms is an experimental realization of quantum supremacy for this specific computational task, heralding a much-anticipated computing paradigm
A map of human genome variation from population-scale sequencing
The 1000 Genomes Project aims to provide a deep characterization of human genome sequence variation as a foundation for investigating the relationship between genotype and phenotype. Here we present results of the pilot phase of the project, designed to develop and compare different strategies for genome-wide sequencing with high-throughput platforms. We undertook three projects: low-coverage whole-genome sequencing of 179 individuals from four populations; high-coverage sequencing of two mother-father-child trios; and exon-targeted sequencing of 697 individuals from seven populations. We describe the location, allele frequency and local haplotype structure of approximately 15 million single nucleotide polymorphisms, 1 million short insertions and deletions, and 20,000 structural variants, most of which were previously undescribed. We show that, because we have catalogued the vast majority of common variation, over 95% of the currently accessible variants found in any individual are present in this data set. On average, each person is found to carry approximately 250 to 300 loss-of-function variants in annotated genes and 50 to 100 variants previously implicated in inherited disorders. We demonstrate how these results can be used to inform association and functional studies. From the two trios, we directly estimate the rate of de novo germline base substitution mutations to be approximately 10−8 per base pair per generation. We explore the data with regard to signatures of natural selection, and identify a marked reduction of genetic variation in the neighbourhood of genes, due to selection at linked sites. These methods and public data will support the next phase of human genetic researc