28 research outputs found
Random circuit block-encoded matrix and a proposal of quantum LINPACK benchmark
The LINPACK benchmark reports the performance of a computer for solving a
system of linear equations with dense random matrices. Although this task was
not designed with a real application directly in mind, the LINPACK benchmark
has been used to define the list of TOP500 supercomputers since the debut of
the list in 1993. We propose that a similar benchmark, called the quantum
LINPACK benchmark, could be used to measure the whole machine performance of
quantum computers. The success of the quantum LINPACK benchmark should be
viewed as the minimal requirement for a quantum computer to perform a useful
task of solving linear algebra problems, such as linear systems of equations.
We propose an input model called the RAndom Circuit Block-Encoded Matrix
(RACBEM), which is a proper generalization of a dense random matrix in the
quantum setting. The RACBEM model is efficient to be implemented on a quantum
computer, and can be designed to optimally adapt to any given quantum
architecture, with relying on a black-box quantum compiler. Besides solving
linear systems, the RACBEM model can be used to perform a variety of linear
algebra tasks relevant to many physical applications, such as computing
spectral measures, time series generated by a Hamiltonian simulation, and
thermal averages of the energy. We implement these linear algebra operations on
IBM Q quantum devices as well as quantum virtual machines, and demonstrate
their performance in solving scientific computing problems.Comment: 22 pages, 18 figure
Benchmarking quantum co-processors in an application-centric, hardware-agnostic and scalable way
Existing protocols for benchmarking current quantum co-processors fail to
meet the usual standards for assessing the performance of
High-Performance-Computing platforms. After a synthetic review of these
protocols -- whether at the gate, circuit or application level -- we introduce
a new benchmark, dubbed Atos Q-score (TM), that is application-centric,
hardware-agnostic and scalable to quantum advantage processor sizes and beyond.
The Q-score measures the maximum number of qubits that can be used effectively
to solve the MaxCut combinatorial optimization problem with the Quantum
Approximate Optimization Algorithm. We give a robust definition of the notion
of effective performance by introducing an improved approximation ratio based
on the scaling of random and optimal algorithms. We illustrate the behavior of
Q-score using perfect and noisy simulations of quantum processors. Finally, we
provide an open-source implementation of Q-score that makes it easy to compute
the Q-score of any quantum hardware
HamLib: A library of Hamiltonians for benchmarking quantum algorithms and hardware
In order to characterize and benchmark computational hardware, software, and
algorithms, it is essential to have many problem instances on-hand. This is no
less true for quantum computation, where a large collection of real-world
problem instances would allow for benchmarking studies that in turn help to
improve both algorithms and hardware designs. To this end, here we present a
large dataset of qubit-based quantum Hamiltonians. The dataset, called HamLib
(for Hamiltonian Library), is freely available online and contains problem
sizes ranging from 2 to 1000 qubits. HamLib includes problem instances of the
Heisenberg model, Fermi-Hubbard model, Bose-Hubbard model, molecular electronic
structure, molecular vibrational structure, MaxCut, Max-k-SAT, Max-k-Cut,
QMaxCut, and the traveling salesperson problem. The goals of this effort are
(a) to save researchers time by eliminating the need to prepare problem
instances and map them to qubit representations, (b) to allow for more thorough
tests of new algorithms and hardware, and (c) to allow for reproducibility and
standardization across research studies
The QPACE Supercomputer : Applications of Random Matrix Theory in Two-Colour Quantum Chromodynamics
QPACE is a massively parallel and scalable supercomputer designed to meet the requirements of applications in Lattice Quantum Chromodynamics. The project was carried out by several academic institutions in collaboration with IBM Germany and other industrial partners. In November 2009 and June 2010
QPACE was the leading architecture on the Green 500 list of the most energy efficient supercomputers in the world
OpenISA, um conjunto de instruções híbrido
Orientador: Edson BorinTese (doutorado) - Universidade Estadual de Campinas, Instituto de ComputaçãoResumo: OpenISA é concebido como a interface de processadores que pretendem ser altamente flexíveis. Isto é conseguido por meio de três estratégias: em primeiro lugar, o ISA é empiricamente escolhido para ser facilmente traduzido para outros, possibilitando flexibilidade do software no caso de um processador OpenISA físico não estar disponível. Neste caso, não há nenhuma necessidade de aplicar um processador virtual OpenISA em software. O ISA está preparado para ser estaticamente traduzido para outros ISAs. Segundo, o ISA não é um ISA concreto nem um ISA virtual, mas um híbrido com a capacidade de admitir modificações nos opcodes sem afetar a compatibilidade retroativa. Este mecanismo permite que as futuras versões do ISA possam sofrer modificações em vez de extensões simples das versões anteriores, um problema comum com ISA concretos, como o x86. Em terceiro lugar, a utilização de uma licença permissiva permite o ISA ser usado livremente por qualquer parte interessada no projeto. Nesta tese de doutorado, concentramo-nos nas instruções de nível de usuário do OpenISA. A tese discute (1) alternativas para ISAs, alternativas para distribuição de programas e o impacto de cada opção, (2) características importantes de OpenISA para atingir seus objetivos e (3) fornece uma completa avaliação do ISA escolhido com respeito a emulação de desempenho em duas CPUs populares, uma projetada pela Intel e outra pela ARM. Concluímos que a versão do OpenISA apresentada aqui pode preservar desempenho próximo do nativo quando traduzida para outros hospedeiros, funcionando como um modelo promissor para ISAs flexíveis da próxima geração que podem ser facilmente estendidos preservando a compatibilidade. Ainda, também mostramos como isso pode ser usado como um formato de distribuição de programas no nível de usuárioAbstract: OpenISA is designed as the interface of processors that aim to be highly flexible. This is achieved by means of three strategies: first, the ISA is empirically chosen to be easily translated to others, providing software flexibility in case a physical OpenISA processor is not available. Second, the ISA is not a concrete ISA nor a virtual ISA, but a hybrid one with the capability of admitting modifications to opcodes without impacting backwards compatibility. This mechanism allows future versions of the ISA to have real changes instead of simple extensions of previous versions, a common problem with concrete ISAs such as the x86. Third, the use of a permissive license allows the ISA to be freely used by any party interested in the project. In this PhD. thesis, we focus on the user-level instructions of OpenISA. The thesis discusses (1) ISA alternatives, program distribution alternatives and the impact of each choice, (2) important features of OpenISA to achieve its goals and (3) provides a thorough evaluation of the chosen ISA with respect to emulation performance on two popular host CPUs, one from Intel and another from ARM. We conclude that the version of OpenISA presented here can preserve close-to-native performance when translated to other hosts, working as a promising model for next-generation, flexible ISAs that can be easily extended while preserving backwards compatibility. Furthermore, we show how this can also be a program distribution format at user-levelDoutoradoCiência da ComputaçãoDoutor em Ciência da Computação2011/09630-1FAPES
Understanding Quantum Technologies 2022
Understanding Quantum Technologies 2022 is a creative-commons ebook that
provides a unique 360 degrees overview of quantum technologies from science and
technology to geopolitical and societal issues. It covers quantum physics
history, quantum physics 101, gate-based quantum computing, quantum computing
engineering (including quantum error corrections and quantum computing
energetics), quantum computing hardware (all qubit types, including quantum
annealing and quantum simulation paradigms, history, science, research,
implementation and vendors), quantum enabling technologies (cryogenics, control
electronics, photonics, components fabs, raw materials), quantum computing
algorithms, software development tools and use cases, unconventional computing
(potential alternatives to quantum and classical computing), quantum
telecommunications and cryptography, quantum sensing, quantum technologies
around the world, quantum technologies societal impact and even quantum fake
sciences. The main audience are computer science engineers, developers and IT
specialists as well as quantum scientists and students who want to acquire a
global view of how quantum technologies work, and particularly quantum
computing. This version is an extensive update to the 2021 edition published in
October 2021.Comment: 1132 pages, 920 figures, Letter forma
Scaling and Resilience in Numerical Algorithms for Exascale Computing
The first Petascale supercomputer, the IBM Roadrunner, went online in 2008. Ten years later, the community is now looking ahead to a new generation of Exascale machines. During the decade that has passed, several hundred Petascale capable machines have been installed worldwide, yet despite the abundance of machines, applications that scale to their full size remain rare. Large clusters now routinely have 50.000+ cores, some have several million. This extreme level of parallelism, that has allowed a theoretical compute capacity in excess of a million billion operations per second, turns out to be difficult to use in many applications of practical interest. Processors often end up spending more time waiting for synchronization, communication, and other coordinating operations to complete, rather than actually computing. Component reliability is another challenge facing HPC developers. If even a single processor fail, among many thousands, the user is forced to restart traditional applications, wasting valuable compute time. These issues collectively manifest themselves as low parallel efficiency, resulting in waste of energy and computational resources. Future performance improvements are expected to continue to come in large part due to increased parallelism. One may therefore speculate that the difficulties currently faced, when scaling applications to Petascale machines, will progressively worsen, making it difficult for scientists to harness the full potential of Exascale computing.
The thesis comprises two parts. Each part consists of several chapters discussing modifications of numerical algorithms to make them better suited for future Exascale machines. In the first part, the use of Parareal for Parallel-in-Time integration techniques for scalable numerical solution of partial differential equations is considered. We propose a new adaptive scheduler that optimize the parallel efficiency by minimizing the time-subdomain length without making communication of time-subdomains too costly. In conjunction with an appropriate preconditioner, we demonstrate that it is possible to obtain time-parallel speedup on the nonlinear shallow water equation, beyond what is possible using conventional spatial domain-decomposition techniques alone. The part is concluded with the proposal of a new method for constructing Parallel-in-Time integration schemes better suited for convection dominated problems.
In the second part, new ways of mitigating the impact of hardware failures are developed and presented. The topic is introduced with the creation of a new fault-tolerant variant of Parareal. In the chapter that follows, a C++ Library for multi-level checkpointing is presented. The library uses lightweight in-memory checkpoints, protected trough the use of erasure codes, to mitigate the impact of failures by decreasing the overhead of checkpointing and minimizing the compute work lost. Erasure codes have the unfortunate property that if more data blocks are lost than parity codes created, the data is effectively considered unrecoverable. The final chapter contains a preliminary study on partial information recovery for incomplete checksums. Under the assumption that some meta knowledge exists on the structure of the data encoded, we show that the data lost may be recovered, at least partially. This result is of interest not only in HPC but also in data centers where erasure codes are widely used to protect data efficiently
Technology 2003: The Fourth National Technology Transfer Conference and Exposition, volume 2
Proceedings from symposia of the Technology 2003 Conference and Exposition, Dec. 7-9, 1993, Anaheim, CA, are presented. Volume 2 features papers on artificial intelligence, CAD&E, computer hardware, computer software, information management, photonics, robotics, test and measurement, video and imaging, and virtual reality/simulation