390 research outputs found
Quantum circuits for strongly correlated quantum systems
In recent years, we have witnessed an explosion of experimental tools by
which quantum systems can be manipulated in a controlled and coherent way. One
of the most important goals now is to build quantum simulators, which would
open up the possibility of exciting experiments probing various theories in
regimes that are not achievable under normal lab circumstances. Here we present
a novel approach to gain detailed control on the quantum simulation of strongly
correlated quantum many-body systems by constructing the explicit quantum
circuits that diagonalize their dynamics. We show that the exact quantum
circuits underlying some of the most relevant many-body Hamiltonians only need
a finite amount of local gates. As a particularly simple instance, the full
dynamics of a one-dimensional Quantum Ising model in a transverse field with
four spins is shown to be reproduced using a quantum circuit of only six local
gates. This opens up the possibility of experimentally producing strongly
correlated states, their time evolution at zero time and even thermal
superpositions at zero temperature. Our method also allows to uncover the exact
circuits corresponding to models that exhibit topological order and to
stabilizer states
Magic-State Functional Units: Mapping and Scheduling Multi-Level Distillation Circuits for Fault-Tolerant Quantum Architectures
Quantum computers have recently made great strides and are on a long-term
path towards useful fault-tolerant computation. A dominant overhead in
fault-tolerant quantum computation is the production of high-fidelity encoded
qubits, called magic states, which enable reliable error-corrected computation.
We present the first detailed designs of hardware functional units that
implement space-time optimized magic-state factories for surface code
error-corrected machines. Interactions among distant qubits require surface
code braids (physical pathways on chip) which must be routed. Magic-state
factories are circuits comprised of a complex set of braids that is more
difficult to route than quantum circuits considered in previous work [1]. This
paper explores the impact of scheduling techniques, such as gate reordering and
qubit renaming, and we propose two novel mapping techniques: braid repulsion
and dipole moment braid rotation. We combine these techniques with graph
partitioning and community detection algorithms, and further introduce a
stitching algorithm for mapping subgraphs onto a physical machine. Our results
show a factor of 5.64 reduction in space-time volume compared to the best-known
previous designs for magic-state factories.Comment: 13 pages, 10 figure
Architecting Noisy Intermediate-Scale Trapped Ion Quantum Computers
Trapped ions (TI) are a leading candidate for building Noisy
Intermediate-Scale Quantum (NISQ) hardware. TI qubits have fundamental
advantages over other technologies such as superconducting qubits, including
high qubit quality, coherence and connectivity. However, current TI systems are
small in size, with 5-20 qubits and typically use a single trap architecture
which has fundamental scalability limitations. To progress towards the next
major milestone of 50-100 qubits, a modular architecture termed the Quantum
Charge Coupled Device (QCCD) has been proposed. In a QCCD-based TI device,
small traps are connected through ion shuttling. While the basic hardware
components for such devices have been demonstrated, building a 50-100 qubit
system is challenging because of a wide range of design possibilities for trap
sizing, communication topology and gate implementations and the need to match
diverse application resource requirements.
Towards realizing QCCD systems with 50-100 qubits, we perform an extensive
architectural study evaluating the key design choices of trap sizing,
communication topology and operation implementation methods. We built a design
toolflow which takes a QCCD architecture's parameters as input, along with a
set of applications and realistic hardware performance models. Our toolflow
maps the applications onto the target device and simulates their execution to
compute metrics such as application run time, reliability and device noise
rates. Using six applications and several hardware design points, we show that
trap sizing and communication topology choices can impact application
reliability by up to three orders of magnitude. Microarchitectural gate
implementation choices influence reliability by another order of magnitude.
From these studies, we provide concrete recommendations to tune these choices
to achieve highly reliable and performant application executions.Comment: Published in ISCA 2020 https://www.iscaconf.org/isca2020/program/
(please cite the ISCA version
Execution Integrity with In-Place Encryption
Instruction set randomization (ISR) was initially proposed with the main goal
of countering code-injection attacks. However, ISR seems to have lost its
appeal since code-injection attacks became less attractive because protection
mechanisms such as data execution prevention (DEP) as well as code-reuse
attacks became more prevalent.
In this paper, we show that ISR can be extended to also protect against
code-reuse attacks while at the same time offering security guarantees similar
to those of software diversity, control-flow integrity, and information hiding.
We present Scylla, a scheme that deploys a new technique for in-place code
encryption to hide the code layout of a randomized binary, and restricts the
control flow to a benign execution path. This allows us to i) implicitly
restrict control-flow targets to basic block entries without requiring the
extraction of a control-flow graph, ii) achieve execution integrity within
legitimate basic blocks, and iii) hide the underlying code layout under
malicious read access to the program. Our analysis demonstrates that Scylla is
capable of preventing state-of-the-art attacks such as just-in-time
return-oriented programming (JIT-ROP) and crash-resistant oriented programming
(CROP). We extensively evaluate our prototype implementation of Scylla and show
feasible performance overhead. We also provide details on how this overhead can
be significantly reduced with dedicated hardware support
Recommended from our members
Dynamic time management for improved accuracy and speed in host-compiled multi-core platform models
textWith increasing complexity and software content, modern embedded platforms employ a heterogeneous mix of multi-core processors along with hardware accelerators in order to provide high performance in limited power budgets. Due to complex interactions and highly dynamic behavior, static analysis of real-time performance and other constraints is challenging. As an alternative, full-system simulations have been widely accepted by designers. With traditional approaches being either slow or inaccurate, so-called host-compiled simulators have recently emerged as a solution for rapid evaluation of complete systems at early design stages. In such approaches, a faster simulation is achieved by natively executing application code at the source level, abstracting execution behavior of target platforms, and thus increasing simulation granularity. However, most existing host-compiled simulators often focus on application behavior only while neglecting effects of hardware/software interactions and associated speed and accuracy tradeoffs in platform modeling. In this dissertation, we focus on host-compiled operating system (OS) and processor modeling techniques, and we introduce novel dynamic timing model management approaches that efficiently improve both accuracy and speed of such models via automatically calibrating the simulation granularity. The contributions of this dissertation are twofold: We first establish an infrastructure for efficient host-compiled multi-core platform simulation by developing (a) abstract models of both real-time OSs and processors that replicate timing-accurate hardware/software interactions and enable full-system co-simulation, and (b) quantitative and analytical studies of host-compiled simulation principles to analyze error bounds and investigate possible improvements. Building on this infrastructure, we further propose specific techniques for improving accuracy and speed tradeoffs in host-compiled simulation by developing (c) an automatic timing granularity adjustment technique based on dynamically observing system state to control the simulation, (d) an out-of-order cache hierarchy modeling approach to efficiently reorder memory access behavior in the presence of temporal decoupling, and (e) a synchronized timing model to align platform threads to run efficiently in parallel simulation. Results as applied to industrial-strength platforms confirm that by providing careful abstractions and dynamic timing management, our models can achieve full-system simulations at equivalent speeds of more than a thousand MIPS with less than 3% timing error. Coupled with the capability to easily adjust simulation parameters and configurations, this demonstrates the benefits of our platform models for early application development and exploration.Electrical and Computer Engineerin
- …