10,597 research outputs found
A NASA family of minicomputer systems, Appendix A
This investigation was undertaken to establish sufficient specifications, or standards, for minicomputer hardware and software to provide NASA with realizable economics in quantity purchases, interchangeability of minicomputers, software, storage and peripherals, and a uniformly high quality. The standards will define minicomputer system component types, each specialized to its intended NASA application, in as many levels of capacity as required
Emulating Digital Logic using Transputer Networks (Very High Parallelism = Simplicity = Performance)
Modern VLSI technology has changed the economic rules by which the balance between processing
power, memory and communications is decided in computing systems. This will have a profound
impact on the design rules for the controlling software. In particular, the criteria for judging efficiency
of the algorithms will be somewhat different. This paper explores some of these implications through
the development of highly parallel and highly distributable algorithms based on occam and transputer
networks. The major results reported are a new simplicity for software designs, a corresponding ability
to reason (formally and informally) about their properties, the reusability of their components and some
real performance figures which demonstrate their practicality. Some guidelines to assist in these designs
are also given. As a vehicle for discussion, an interactive simulator is developed for checking the
functional and timing characteristics of digital logic circuits of arbitrary complexity
Optimized Compilation of Aggregated Instructions for Realistic Quantum Computers
Recent developments in engineering and algorithms have made real-world
applications in quantum computing possible in the near future. Existing quantum
programming languages and compilers use a quantum assembly language composed of
1- and 2-qubit (quantum bit) gates. Quantum compiler frameworks translate this
quantum assembly to electric signals (called control pulses) that implement the
specified computation on specific physical devices. However, there is a
mismatch between the operations defined by the 1- and 2-qubit logical ISA and
their underlying physical implementation, so the current practice of directly
translating logical instructions into control pulses results in inefficient,
high-latency programs. To address this inefficiency, we propose a universal
quantum compilation methodology that aggregates multiple logical operations
into larger units that manipulate up to 10 qubits at a time. Our methodology
then optimizes these aggregates by (1) finding commutative intermediate
operations that result in more efficient schedules and (2) creating custom
control pulses optimized for the aggregate (instead of individual 1- and
2-qubit operations). Compared to the standard gate-based compilation, the
proposed approach realizes a deeper vertical integration of high-level quantum
software and low-level, physical quantum hardware. We evaluate our approach on
important near-term quantum applications on simulations of superconducting
quantum architectures. Our proposed approach provides a mean speedup of
, with a maximum of . Because latency directly affects the
feasibility of quantum computation, our results not only improve performance
but also have the potential to enable quantum computation sooner than otherwise
possible.Comment: 13 pages, to apper in ASPLO
Can geocomputation save urban simulation? Throw some agents into the mixture, simmer and wait ...
There are indications that the current generation of simulation models in practical,
operational uses has reached the limits of its usefulness under existing specifications.
The relative stasis in operational urban modeling contrasts with simulation efforts in
other disciplines, where techniques, theories, and ideas drawn from computation and
complexity studies are revitalizing the ways in which we conceptualize, understand,
and model real-world phenomena. Many of these concepts and methodologies are
applicable to operational urban systems simulation. Indeed, in many cases, ideas from
computation and complexity studies—often clustered under the collective term of
geocomputation, as they apply to geography—are ideally suited to the simulation of
urban dynamics. However, there exist several obstructions to their successful use in
operational urban geographic simulation, particularly as regards the capacity of these
methodologies to handle top-down dynamics in urban systems.
This paper presents a framework for developing a hybrid model for urban geographic
simulation and discusses some of the imposing barriers against innovation in this
field. The framework infuses approaches derived from geocomputation and
complexity with standard techniques that have been tried and tested in operational
land-use and transport simulation. Macro-scale dynamics that operate from the topdown
are handled by traditional land-use and transport models, while micro-scale
dynamics that work from the bottom-up are delegated to agent-based models and
cellular automata. The two methodologies are fused in a modular fashion using a
system of feedback mechanisms. As a proof-of-concept exercise, a micro-model of
residential location has been developed with a view to hybridization. The model
mixes cellular automata and multi-agent approaches and is formulated so as to
interface with meso-models at a higher scale
Scalable Spin Amplification with a Gain over a Hundred
We propose a scalable and practical implementation of spin amplification
which does not require individual addressing nor a specially tailored spin
network. We have demonstrated a gain of 140 in a solid-state nuclear spin
system of which the spin polarization has been increased to 0.12 using dynamic
nuclear polarization with photoexcited triplet electron spins. Spin
amplification scalable to a higher gain opens the door to the single spin
measurement for a readout of quantum computers as well as practical
applications of nuclear magnetic resonance (NMR) spectroscopy to infinitesimal
samples which have been concealed by thermal noise.Comment: 6 pages, 7 figure
The Turing Test and the Zombie Argument
In this paper I shall try to put some implications concerning the Turing's test and the so-called
Zombie arguments into the context of philosophy of mind. My intention is not to compose a review
of relevant concepts, but to discuss central problems, which originate from the Turing's test - as a
paradigm of computational theory of mind - with the arguments, which refute sustainability of this
thesis.
In the first section (Section I), I expose the basic computationalist presuppositions; by
examining the premises of the Turing Test (TT) I argue that the TT, as a functionalist paradigm
concept, underlies the computational theory of mind. I treat computationalism as a thesis that
defines the human cognitive system as a physical, symbolic and semantic system, in such a
manner that the description of its physical states is isomorphic with the description of its symbolic
conditions, so that this isomorphism is semantically interpretable. In the second section (Section
II), I discuss the Zombie arguments, and the epistemological-modal problems connected with them,
which refute sustainability of computationalism. The proponents of the Zombie arguments build their
attack on the computationalism on the basis of thought experiments with creatures behaviorally,
functionally and physically indistinguishable from human beings, though these creatures do not
have phenomenal experiences. According to the consequences of these thought experiments - if
zombies are possible, then, the computationalism doesn't offer a satisfying explanation of
consciousness. I compare my thesis from Section 1, with recent versions of Zombie arguments,
which claim that computationalism fails to explain qualitative phenomenal experience. I conclude
that despite the weaknesses of computationalism, which are made obvious by zombie-arguments,
these arguments are not the last word when it comes to explanatory force of computationalism
Army/NASA small turboshaft engine digital controls research program
The emphasis of a program to conduct digital controls research for small turboshaft engines is on engine test evaluation of advanced control logic using a flexible microprocessor based digital control system designed specifically for research on advanced control logic. Control software is stored in programmable memory. New control algorithms may be stored in a floppy disk and loaded directly into memory. This feature facilitates comparative evaluation of different advanced control modes. The central processor in the digital control is an Intel 8086 16 bit microprocessor. Control software is programmed in assembly language. Software checkout is accomplished prior to engine test by connecting the digital control to a real time hybrid computer simulation of the engine. The engine currently installed in the facility has a hydromechanical control modified to allow electrohydraulic fuel metering and VG actuation by the digital control. Simulation results are presented which show that the modern control reduces the transient rotor speed droop caused by unanticipated load changes such as cyclic pitch or wind gust transients
Implementing the conjugate gradient algorithm on multi-core systems
In linear solvers, like the conjugate gradient algorithm, sparse-matrix vector multiplication is an important kernel. Due to the sparseness of the matrices, the solver runs relatively slow. For digital optical tomography (DOT), a large set of linear equations have to be solved which currently takes in the order of hours on desktop computers. Our goal was to speed up the conjugate gradient solver. In this paper we present the results of applying multiple optimization techniques and exploiting multi-core solutions offered by two recently introduced architectures: Intel’s Woodcrest\ud
general purpose processor and NVIDIA’s G80 graphical processing unit. Using these techniques for these architectures, a speedup of a factor three\ud
has been achieved
- …