1,384 research outputs found
Quantum read-out for cold atomic quantum simulators
Quantum simulators allow to explore static and dynamical properties of otherwise intractable quantum many-body systems. In many instances, however, the read-out limits such quantum simulations. In this work, we introduce an innovative experimental read-out exploiting coherent non-interacting dynamics. Specifically, we present a tomographic recovery method allowing to indirectly measure the second moments of the relative density fluctuations between two one-dimensional superfluids, which until now eluded direct measurements. Applying methods from signal processing, we show that we can reconstruct the relative density fluctuations from non-equilibrium data of the relative phase fluctuations. We employ the method to investigate equilibrium states, the dynamics of phonon occupation numbers and even to predict recurrences. The method opens a new window for quantum simulations with one-dimensional superfluids, enabling a deeper analysis of their equilibration and thermalization dynamics
Quantum Computing
Quantum mechanics---the theory describing the fundamental workings of
nature---is famously counterintuitive: it predicts that a particle can be in
two places at the same time, and that two remote particles can be inextricably
and instantaneously linked. These predictions have been the topic of intense
metaphysical debate ever since the theory's inception early last century.
However, supreme predictive power combined with direct experimental observation
of some of these unusual phenomena leave little doubt as to its fundamental
correctness. In fact, without quantum mechanics we could not explain the
workings of a laser, nor indeed how a fridge magnet operates. Over the last
several decades quantum information science has emerged to seek answers to the
question: can we gain some advantage by storing, transmitting and processing
information encoded in systems that exhibit these unique quantum properties?
Today it is understood that the answer is yes. Many research groups around the
world are working towards one of the most ambitious goals humankind has ever
embarked upon: a quantum computer that promises to exponentially improve
computational power for particular tasks. A number of physical systems,
spanning much of modern physics, are being developed for this task---ranging
from single particles of light to superconducting circuits---and it is not yet
clear which, if any, will ultimately prove successful. Here we describe the
latest developments for each of the leading approaches and explain what the
major challenges are for the future.Comment: 26 pages, 7 figures, 291 references. Early draft of Nature 464, 45-53
(4 March 2010). Published version is more up-to-date and has several
corrections, but is half the length with far fewer reference
Beyond the Cosmological Standard Model
After a decade and a half of research motivated by the accelerating universe,
theory and experiment have a reached a certain level of maturity. The
development of theoretical models beyond \Lambda, or smooth dark energy, often
called modified gravity, has led to broader insights into a path forward, and a
host of observational and experimental tests have been developed. In this
review we present the current state of the field and describe a framework for
anticipating developments in the next decade. We identify the guiding
principles for rigorous and consistent modifications of the standard model, and
discuss the prospects for empirical tests. We begin by reviewing attempts to
consistently modify Einstein gravity in the infrared, focusing on the notion
that additional degrees of freedom introduced by the modification must screen
themselves from local tests of gravity. We categorize screening mechanisms into
three broad classes: mechanisms which become active in regions of high
Newtonian potential, those in which first derivatives become important, and
those for which second derivatives are important. Examples of the first class,
such as f(R) gravity, employ the familiar chameleon or symmetron mechanisms,
whereas examples of the last class are galileon and massive gravity theories,
employing the Vainshtein mechanism. In each case, we describe the theories as
effective theories. We describe experimental tests, summarizing laboratory and
solar system tests and describing in some detail astrophysical and cosmological
tests. We discuss future tests which will be sensitive to different signatures
of new physics in the gravitational sector. Parts that are more relevant to
theorists vs. observers/experimentalists are clearly indicated, in the hope
that this will serve as a useful reference for both audiences, as well as
helping those interested in bridging the gap between them.Comment: 175 pages, 24 figures. v2: Minor corrections, added references.
Review article, comments welcom
Recommended from our members
Dynamic time management for improved accuracy and speed in host-compiled multi-core platform models
textWith increasing complexity and software content, modern embedded platforms employ a heterogeneous mix of multi-core processors along with hardware accelerators in order to provide high performance in limited power budgets. Due to complex interactions and highly dynamic behavior, static analysis of real-time performance and other constraints is challenging. As an alternative, full-system simulations have been widely accepted by designers. With traditional approaches being either slow or inaccurate, so-called host-compiled simulators have recently emerged as a solution for rapid evaluation of complete systems at early design stages. In such approaches, a faster simulation is achieved by natively executing application code at the source level, abstracting execution behavior of target platforms, and thus increasing simulation granularity. However, most existing host-compiled simulators often focus on application behavior only while neglecting effects of hardware/software interactions and associated speed and accuracy tradeoffs in platform modeling. In this dissertation, we focus on host-compiled operating system (OS) and processor modeling techniques, and we introduce novel dynamic timing model management approaches that efficiently improve both accuracy and speed of such models via automatically calibrating the simulation granularity. The contributions of this dissertation are twofold: We first establish an infrastructure for efficient host-compiled multi-core platform simulation by developing (a) abstract models of both real-time OSs and processors that replicate timing-accurate hardware/software interactions and enable full-system co-simulation, and (b) quantitative and analytical studies of host-compiled simulation principles to analyze error bounds and investigate possible improvements. Building on this infrastructure, we further propose specific techniques for improving accuracy and speed tradeoffs in host-compiled simulation by developing (c) an automatic timing granularity adjustment technique based on dynamically observing system state to control the simulation, (d) an out-of-order cache hierarchy modeling approach to efficiently reorder memory access behavior in the presence of temporal decoupling, and (e) a synchronized timing model to align platform threads to run efficiently in parallel simulation. Results as applied to industrial-strength platforms confirm that by providing careful abstractions and dynamic timing management, our models can achieve full-system simulations at equivalent speeds of more than a thousand MIPS with less than 3% timing error. Coupled with the capability to easily adjust simulation parameters and configurations, this demonstrates the benefits of our platform models for early application development and exploration.Electrical and Computer Engineerin
- …