69 research outputs found
Palaeogenomic analysis of black rat (Rattus rattus) reveals multiple European introductions associated with human economic history
The distribution of the black rat (Rattus rattus) has been heavily influenced by its association with humans. The dispersal history of this non-native commensal rodent across Europe, however, remains poorly understood, and different introductions may have occurred during the Roman and medieval periods. Here, in order to reconstruct the population history of European black rats, we generated a de novo genome assembly of the black rat, 67 ancient black rat mitogenomes and 36 ancient nuclear genomes from sites spanning the 1st-17th centuries CE in Europe and North Africa. Analyses of mitochondrial DNA confirm that black rats were introduced into the Mediterranean and Europe from Southwest Asia. Genomic analyses of the ancient rats reveal a population turnover in temperate Europe between the 6th and 10th centuries CE, coincident with an archaeologically attested decline in the black rat population. The near disappearance and re-emergence of black rats in Europe may have been the result of the breakdown of the Roman Empire, the First Plague Pandemic, and/or post-Roman climatic cooling.Competing Interest StatementThe authors have declared no competing interest.- Results and Discussion -- The demographic history of Rattus rattus and its closely related species -- A global phylogeography of the black rat based on mitochondrial DNA -- Ancient genomes reveal the relationships of European black rats over space and time - Discussion - Method
Phase transition in Random Circuit Sampling
Quantum computers hold the promise of executing tasks beyond the capability
of classical computers. Noise competes with coherent evolution and destroys
long-range correlations, making it an outstanding challenge to fully leverage
the computation power of near-term quantum processors. We report Random Circuit
Sampling (RCS) experiments where we identify distinct phases driven by the
interplay between quantum dynamics and noise. Using cross-entropy benchmarking,
we observe phase boundaries which can define the computational complexity of
noisy quantum evolution. We conclude by presenting an RCS experiment with 70
qubits at 24 cycles. We estimate the computational cost against improved
classical methods and demonstrate that our experiment is beyond the
capabilities of existing classical supercomputers
Overcoming leakage in scalable quantum error correction
Leakage of quantum information out of computational states into higher energy
states represents a major challenge in the pursuit of quantum error correction
(QEC). In a QEC circuit, leakage builds over time and spreads through
multi-qubit interactions. This leads to correlated errors that degrade the
exponential suppression of logical error with scale, challenging the
feasibility of QEC as a path towards fault-tolerant quantum computation. Here,
we demonstrate the execution of a distance-3 surface code and distance-21
bit-flip code on a Sycamore quantum processor where leakage is removed from all
qubits in each cycle. This shortens the lifetime of leakage and curtails its
ability to spread and induce correlated errors. We report a ten-fold reduction
in steady-state leakage population on the data qubits encoding the logical
state and an average leakage population of less than
throughout the entire device. The leakage removal process itself efficiently
returns leakage population back to the computational basis, and adding it to a
code circuit prevents leakage from inducing correlated error across cycles,
restoring a fundamental assumption of QEC. With this demonstration that leakage
can be contained, we resolve a key challenge for practical QEC at scale.Comment: Main text: 7 pages, 5 figure
Suppressing quantum errors by scaling a surface code logical qubit
Practical quantum computing will require error rates that are well below what
is achievable with physical qubits. Quantum error correction offers a path to
algorithmically-relevant error rates by encoding logical qubits within many
physical qubits, where increasing the number of physical qubits enhances
protection against physical errors. However, introducing more qubits also
increases the number of error sources, so the density of errors must be
sufficiently low in order for logical performance to improve with increasing
code size. Here, we report the measurement of logical qubit performance scaling
across multiple code sizes, and demonstrate that our system of superconducting
qubits has sufficient performance to overcome the additional errors from
increasing qubit number. We find our distance-5 surface code logical qubit
modestly outperforms an ensemble of distance-3 logical qubits on average, both
in terms of logical error probability over 25 cycles and logical error per
cycle ( compared to ). To investigate
damaging, low-probability error sources, we run a distance-25 repetition code
and observe a logical error per round floor set by a single
high-energy event ( when excluding this event). We are able
to accurately model our experiment, and from this model we can extract error
budgets that highlight the biggest challenges for future systems. These results
mark the first experimental demonstration where quantum error correction begins
to improve performance with increasing qubit number, illuminating the path to
reaching the logical error rates required for computation.Comment: Main text: 6 pages, 4 figures. v2: Update author list, references,
Fig. S12, Table I
Measurement-induced entanglement and teleportation on a noisy quantum processor
Measurement has a special role in quantum theory: by collapsing the
wavefunction it can enable phenomena such as teleportation and thereby alter
the "arrow of time" that constrains unitary evolution. When integrated in
many-body dynamics, measurements can lead to emergent patterns of quantum
information in space-time that go beyond established paradigms for
characterizing phases, either in or out of equilibrium. On present-day NISQ
processors, the experimental realization of this physics is challenging due to
noise, hardware limitations, and the stochastic nature of quantum measurement.
Here we address each of these experimental challenges and investigate
measurement-induced quantum information phases on up to 70 superconducting
qubits. By leveraging the interchangeability of space and time, we use a
duality mapping, to avoid mid-circuit measurement and access different
manifestations of the underlying phases -- from entanglement scaling to
measurement-induced teleportation -- in a unified way. We obtain finite-size
signatures of a phase transition with a decoding protocol that correlates the
experimental measurement record with classical simulation data. The phases
display sharply different sensitivity to noise, which we exploit to turn an
inherent hardware limitation into a useful diagnostic. Our work demonstrates an
approach to realize measurement-induced physics at scales that are at the
limits of current NISQ processors
Non-Abelian braiding of graph vertices in a superconducting processor
Indistinguishability of particles is a fundamental principle of quantum
mechanics. For all elementary and quasiparticles observed to date - including
fermions, bosons, and Abelian anyons - this principle guarantees that the
braiding of identical particles leaves the system unchanged. However, in two
spatial dimensions, an intriguing possibility exists: braiding of non-Abelian
anyons causes rotations in a space of topologically degenerate wavefunctions.
Hence, it can change the observables of the system without violating the
principle of indistinguishability. Despite the well developed mathematical
description of non-Abelian anyons and numerous theoretical proposals, the
experimental observation of their exchange statistics has remained elusive for
decades. Controllable many-body quantum states generated on quantum processors
offer another path for exploring these fundamental phenomena. While efforts on
conventional solid-state platforms typically involve Hamiltonian dynamics of
quasi-particles, superconducting quantum processors allow for directly
manipulating the many-body wavefunction via unitary gates. Building on
predictions that stabilizer codes can host projective non-Abelian Ising anyons,
we implement a generalized stabilizer code and unitary protocol to create and
braid them. This allows us to experimentally verify the fusion rules of the
anyons and braid them to realize their statistics. We then study the prospect
of employing the anyons for quantum computation and utilize braiding to create
an entangled state of anyons encoding three logical qubits. Our work provides
new insights about non-Abelian braiding and - through the future inclusion of
error correction to achieve topological protection - could open a path toward
fault-tolerant quantum computing
Realizing topologically ordered states on a quantum processor
The discovery of topological order has revolutionized the understanding of
quantum matter in modern physics and provided the theoretical foundation for
many quantum error correcting codes. Realizing topologically ordered states has
proven to be extremely challenging in both condensed matter and synthetic
quantum systems. Here, we prepare the ground state of the toric code
Hamiltonian using an efficient quantum circuit on a superconducting quantum
processor. We measure a topological entanglement entropy near the expected
value of , and simulate anyon interferometry to extract the braiding
statistics of the emergent excitations. Furthermore, we investigate key aspects
of the surface code, including logical state injection and the decay of the
non-local order parameter. Our results demonstrate the potential for quantum
processors to provide key insights into topological quantum matter and quantum
error correction.Comment: 6 pages 4 figures, plus supplementary material
- âŠ