18,609 research outputs found
A Review on Software Architectures for Heterogeneous Platforms
The increasing demands for computing performance have been a reality
regardless of the requirements for smaller and more energy efficient devices.
Throughout the years, the strategy adopted by industry was to increase the
robustness of a single processor by increasing its clock frequency and mounting
more transistors so more calculations could be executed. However, it is known
that the physical limits of such processors are being reached, and one way to
fulfill such increasing computing demands has been to adopt a strategy based on
heterogeneous computing, i.e., using a heterogeneous platform containing more
than one type of processor. This way, different types of tasks can be executed
by processors that are specialized in them. Heterogeneous computing, however,
poses a number of challenges to software engineering, especially in the
architecture and deployment phases. In this paper, we conduct an empirical
study that aims at discovering the state-of-the-art in software architecture
for heterogeneous computing, with focus on deployment. We conduct a systematic
mapping study that retrieved 28 studies, which were critically assessed to
obtain an overview of the research field. We identified gaps and trends that
can be used by both researchers and practitioners as guides to further
investigate the topic
Physical-depth architectural requirements for generating universal photonic cluster states
Most leading proposals for linear-optical quantum computing (LOQC) use
cluster states, which act as a universal resource for measurement-based
(one-way) quantum computation (MBQC). In ballistic approaches to LOQC, cluster
states are generated passively from small entangled resource states using
so-called fusion operations. Results from percolation theory have previously
been used to argue that universal cluster states can be generated in the
ballistic approach using schemes which exceed the critical threshold for
percolation, but these results consider cluster states with unbounded size.
Here we consider how successful percolation can be maintained using a physical
architecture with fixed physical depth, assuming that the cluster state is
continuously generated and measured, and therefore that only a finite portion
of it is visible at any one point in time. We show that universal LOQC can be
implemented using a constant-size device with modest physical depth, and that
percolation can be exploited using simple pathfinding strategies without the
need for high-complexity algorithms.Comment: 18 pages, 10 figure
The future of computing beyond Moore's Law.
Moore's Law is a techno-economic model that has enabled the information technology industry to double the performance and functionality of digital electronics roughly every 2 years within a fixed cost, power and area. Advances in silicon lithography have enabled this exponential miniaturization of electronics, but, as transistors reach atomic scale and fabrication costs continue to rise, the classical technological driver that has underpinned Moore's Law for 50 years is failing and is anticipated to flatten by 2025. This article provides an updated view of what a post-exascale system will look like and the challenges ahead, based on our most recent understanding of technology roadmaps. It also discusses the tapering of historical improvements, and how it affects options available to continue scaling of successors to the first exascale machine. Lastly, this article covers the many different opportunities and strategies available to continue computing performance improvements in the absence of historical technology drivers. This article is part of a discussion meeting issue 'Numerical algorithms for high-performance computational science'
Recommended from our members
FABRIC: A National-Scale Programmable Experimental Network Infrastructure
FABRIC is a unique national research infrastructure to enable cutting-edge and exploratory research at-scale in networking, cybersecurity, distributed computing and storage systems, machine learning, and science applications. It is an everywhere-programmable nationwide instrument comprised of novel extensible network elements equipped with large amounts of compute and storage, interconnected by high speed, dedicated optical links. It will connect a number of specialized testbeds for cloud research (NSF Cloud testbeds CloudLab and Chameleon), for research beyond 5G technologies (Platforms for Advanced Wireless Research or PAWR), as well as production high-performance computing facilities and science instruments to create a rich fabric for a wide variety of experimental activities
Visualising the structure of architectural open spaces based on shape analysis
This paper proposes the application of some well known two-dimensional
geometrical shape descriptors for the visualisation of the structure of
architectural open spaces. The paper demonstrates the use of visibility
measures such as distance to obstacles and amount of visible space to calculate
shape descriptors such as convexity and skeleton of the open space. The aim of
the paper is to indicate a simple, objective and quantifiable approach to
understand the structure of open spaces otherwise impossible due to the complex
construction of built structures.Comment: 10 pages, 9 figure
Will SDN be part of 5G?
For many, this is no longer a valid question and the case is considered
settled with SDN/NFV (Software Defined Networking/Network Function
Virtualization) providing the inevitable innovation enablers solving many
outstanding management issues regarding 5G. However, given the monumental task
of softwarization of radio access network (RAN) while 5G is just around the
corner and some companies have started unveiling their 5G equipment already,
the concern is very realistic that we may only see some point solutions
involving SDN technology instead of a fully SDN-enabled RAN. This survey paper
identifies all important obstacles in the way and looks at the state of the art
of the relevant solutions. This survey is different from the previous surveys
on SDN-based RAN as it focuses on the salient problems and discusses solutions
proposed within and outside SDN literature. Our main focus is on fronthaul,
backward compatibility, supposedly disruptive nature of SDN deployment,
business cases and monetization of SDN related upgrades, latency of general
purpose processors (GPP), and additional security vulnerabilities,
softwarization brings along to the RAN. We have also provided a summary of the
architectural developments in SDN-based RAN landscape as not all work can be
covered under the focused issues. This paper provides a comprehensive survey on
the state of the art of SDN-based RAN and clearly points out the gaps in the
technology.Comment: 33 pages, 10 figure
Scalability of quantum computation with addressable optical lattices
We make a detailed analysis of error mechanisms, gate fidelity, and
scalability of proposals for quantum computation with neutral atoms in
addressable (large lattice constant) optical lattices. We have identified
possible limits to the size of quantum computations, arising in 3D optical
lattices from current limitations on the ability to perform single qubit gates
in parallel and in 2D lattices from constraints on laser power. Our results
suggest that 3D arrays as large as 100 x 100 x 100 sites (i.e.,
qubits) may be achievable, provided two-qubit gates can be performed with
sufficiently high precision and degree of parallelizability. Parallelizability
of long range interaction-based two-qubit gates is qualitatively compared to
that of collisional gates. Different methods of performing single qubit gates
are compared, and a lower bound of is determined on the
error rate for the error mechanisms affecting Cs in a blue-detuned
lattice with Raman transition-based single qubit gates, given reasonable limits
on experimental parameters.Comment: 17 pages, 5 figures. Accepted for publication in Physical Review
- …