400 research outputs found
Recommended from our members
Computing infrastructure issues in distributed communications systems : a survey of operating system transport system architectures
The performance of distributed applications (such as file transfer, remote login, tele-conferencing, full-motion video, and scientific visualization) is influenced by several factors that interact in complex ways. In particular, application performance is significantly affected both by communication infrastructure factors and computing infrastructure factors. Several communication infrastructure factors include channel speed, bit-error rate, and congestion at intermediate switching nodes. Computing infrastructure factors include (among other things) both protocol processing activities (such as connection management, flow control, error detection, and retransmission) and general operating system factors (such as memory latency, CPU speed, interrupt and context switching overhead, process architecture, and message buffering). Due to a several orders of magnitude increase in network channel speed and an increase in application diversity, performance bottlenecks are shifting from the network factors to the transport system factors.This paper defines an abstraction called an "Operating System Transport System Architecture" (OSTSA) that is used to classify the major components and services in the computing infrastructure. End-to-end network protocols such as TCP, TP4, VMTP, XTP, and Delta-t typically run on general-purpose computers, where they utilize various operating system resources such as processors, virtual memory, and network controllers. The OSTSA provides services that integrate these resources to support distributed applications running on local and wide area networks.A taxonomy is presented to evaluate OSTSAs in terms of their support for protocol processing activities. We use this taxonomy to compare and contrast five general-purpose commercial and experimental operating systems including System V UNIX, BSD UNIX, the x-kernel, Choices, and Xinu
A Survey of Asynchronous Programming Using Coroutines in the Internet of Things and Embedded Systems
Many Internet of Things and embedded projects are event-driven, and therefore
require asynchronous and concurrent programming. Current proposals for C++20
suggest that coroutines will have native language support. It is timely to
survey the current use of coroutines in embedded systems development. This
paper investigates existing research which uses or describes coroutines on
resource-constrained platforms. The existing research is analysed with regard
to: software platform, hardware platform and capacity; use cases and intended
benefits; and the application programming interface design used for coroutines.
A systematic mapping study was performed, to select studies published between
2007 and 2018 which contained original research into the application of
coroutines on resource-constrained platforms. An initial set of 566 candidate
papers were reduced to only 35 after filters were applied, revealing the
following taxonomy. The C & C++ programming languages were used by 22 studies
out of 35. As regards hardware, 16 studies used 8- or 16-bit processors while
13 used 32-bit processors. The four most common use cases were concurrency (17
papers), network communication (15), sensor readings (9) and data flow (7). The
leading intended benefits were code style and simplicity (12 papers),
scheduling (9) and efficiency (8). A wide variety of techniques have been used
to implement coroutines, including native macros, additional tool chain steps,
new language features and non-portable assembly language. We conclude that
there is widespread demand for coroutines on resource-constrained devices. Our
findings suggest that there is significant demand for a formalised, stable,
well-supported implementation of coroutines in C++, designed with consideration
of the special needs of resource-constrained devices, and further that such an
implementation would bring benefits specific to such devices.Comment: 22 pages, 8 figures, to be published in ACM Transactions on Embedded
Computing Systems (TECS
Mirroring Mobile Phone in the Clouds
This paper presents a framework of Mirroring Mobile Phone in the Clouds (MMPC) to speed up data/computing intensive applications on a mobile phone by taking full advantage of the super computing power of the clouds. An application on the mobile phone is dynamically partitioned in such a way that the heavy-weighted part is always running on a mirrored server in the clouds while the light-weighted part remains on the mobile phone. A performance improvement (an energy consumption reduction of 70% and a speed-up of 15x) is achieved at the cost of the communication overhead between the mobile phone and the clouds (to transfer the application codes and intermediate results) of a desired application. Our original contributions include a dynamic profiler and a dynamic partitioning algorithm compared with traditional approaches of either statically partitioning a mobile application or modifying a mobile application to support the required partitioning
A Historical Context for Data Streams
Machine learning from data streams is an active and growing research area.
Research on learning from streaming data typically makes strict assumptions
linked to computational resource constraints, including requirements for stream
mining algorithms to inspect each instance not more than once and be ready to
give a prediction at any time. Here we review the historical context of data
streams research placing the common assumptions used in machine learning over
data streams in their historical context.Comment: 9 page
Asynchronous programming in the abstract behavioural specification language
Chip manufacturers are rapidly moving towards so-called manycore chips with thousands of independent processors on the same silicon real estate. Current programming languages can only leverage the potential power by inserting code with low level concurrency constructs, sacrificing clarity. Alternatively, a programming language can integrate a thread of execution with a stable notion of identity, e.g., in active objects.Abstract Behavioural Specification (ABS) is a language for designing executable models of parallel and distributed object-oriented systems based on active objects, and is defined in terms of a formal operational semantics which enables a variety of static and dynamic analysis techniques for the ABS models.The overall goal of this thesis is to extend the asynchronous programming model and the corresponding analysis techniques in ABS.Algorithms and the Foundations of Software technolog
Concurrent Systems Need Both Sequences And Serializers
This report describes research done at the Artificial Intelligence Laboratory of the Massachusetts Institute of Technology. Support for the laboratory's artificial intelligence research is provided in part by the Office of Naval Research of the Department of Defense under contract N00014-75-C-0522.Contemporary concurrent programming languages fall roughly into two classes. Languages in the first class support the notion of a sequence of values and some kind of pipelining operation over the sequence of values. Languages in the second class support the notion of transactions and some way to serialize transactions. In terms of the actor model of computation this distinction corresponds to the difference between serialized and unserialized actors. In this paper the utility of modeling both serialized and unserialized actors in a coherent formalism is demonstrated.MIT Artificial Intelligence Laboratory
Department of Defense Office of Naval Researc
Virtual Machine Support for Many-Core Architectures: Decoupling Abstract from Concrete Concurrency Models
The upcoming many-core architectures require software developers to exploit
concurrency to utilize available computational power. Today's high-level
language virtual machines (VMs), which are a cornerstone of software
development, do not provide sufficient abstraction for concurrency concepts. We
analyze concrete and abstract concurrency models and identify the challenges
they impose for VMs. To provide sufficient concurrency support in VMs, we
propose to integrate concurrency operations into VM instruction sets.
Since there will always be VMs optimized for special purposes, our goal is to
develop a methodology to design instruction sets with concurrency support.
Therefore, we also propose a list of trade-offs that have to be investigated to
advise the design of such instruction sets.
As a first experiment, we implemented one instruction set extension for
shared memory and one for non-shared memory concurrency. From our experimental
results, we derived a list of requirements for a full-grown experimental
environment for further research
- …