12,788 research outputs found
An efficient processor allocation strategy that maintains a high degree of contiguity among processors in 2D mesh connected multicomputers
Two strategies are used for the allocation of jobs to processors connected by mesh topologies: contiguous allocation and non-contiguous allocation. In non-contiguous allocation, a job request can be split into smaller parts that are allocated to non-adjacent free sub-meshes rather than always waiting until a single sub-mesh of the requested size and shape is available. Lifting the contiguity condition is expected to reduce processor fragmentation and increase system utilization. However, the distances traversed by messages can be long, and as a result the communication overhead, especially contention, is increased. The extra communication overhead depends on how the allocation request is partitioned and assigned to free sub-meshes. This paper presents a new Non-contiguous allocation algorithm, referred to as Greedy-Available-Busy-List (GABL for short), which can decrease the communication overhead among processors allocated to a given job. The simulation results show that the new strategy can reduce the communication overhead and substantially improve performance in terms of parameters such as job turnaround time and system utilization. Moreover, the results reveal that the Shortest-Service-Demand-First (SSD) scheduling strategy is much better than the First-Come-First-Served (FCFS) scheduling strategy
A performance comparison of the contiguous allocation strategies in 3D mesh connected multicomputers
The performance of contiguous allocation strategies can be significantly affected by the distribution of job execution times. In this paper, the performance of the existing contiguous allocation strategies for 3D mesh multicomputers is re-visited in the context of heavy-tailed distributions (e.g., a Bounded Pareto distribution). The strategies are evaluated and compared using simulation experiments for both First-Come-First-Served (FCFS) and Shortest-Service-Demand (SSD) scheduling strategies under a variety of system loads and system sizes. The results show that the performance of the allocation strategies degrades considerably when job execution times follow a heavy-tailed distribution. Moreover, SSD copes much better than FCFS scheduling strategy in the presence of heavy-tailed job execution times. The results also show that the strategies that depend on a list of allocated sub-meshes for both allocation and deallocation have lower allocation overhead and deliver good system performance in terms of average turnaround time and mean system utilization
A Parallel Mesh-Adaptive Framework for Hyperbolic Conservation Laws
We report on the development of a computational framework for the parallel,
mesh-adaptive solution of systems of hyperbolic conservation laws like the
time-dependent Euler equations in compressible gas dynamics or
Magneto-Hydrodynamics (MHD) and similar models in plasma physics. Local mesh
refinement is realized by the recursive bisection of grid blocks along each
spatial dimension, implemented numerical schemes include standard
finite-differences as well as shock-capturing central schemes, both in
connection with Runge-Kutta type integrators. Parallel execution is achieved
through a configurable hybrid of POSIX-multi-threading and MPI-distribution
with dynamic load balancing. One- two- and three-dimensional test computations
for the Euler equations have been carried out and show good parallel scaling
behavior. The Racoon framework is currently used to study the formation of
singularities in plasmas and fluids.Comment: late submissio
The effect of real workloads and stochastic workloads on the performance of allocation and scheduling algorithms in 2D mesh multicomputers
The performance of the existing non-contiguous processor allocation strategies has been traditionally carried out by means of simulation based on a stochastic workload model to generate a stream of incoming jobs. To validate the performance of the existing algorithms, there has been a need to evaluate the algorithms' performance based on a real workload trace. In this paper, we evaluate the performance of several well-known processor allocation and job scheduling strategies based on a real workload trace and compare the results against those obtained from using a stochastic workload. Our results reveal that the conclusions reached on the relative performance merits of the allocation strategies when a real workload trace is used are in general compatible with those obtained when a stochastic workload is used
Cross-layer design of multi-hop wireless networks
MULTI -hop wireless networks are usually defined as a collection of nodes
equipped with radio transmitters, which not only have the capability to
communicate each other in a multi-hop fashion, but also to route each othersâ data
packets. The distributed nature of such networks makes them suitable for a variety of
applications where there are no assumed reliable central entities, or controllers, and
may significantly improve the scalability issues of conventional single-hop wireless
networks.
This Ph.D. dissertation mainly investigates two aspects of the research issues
related to the efficient multi-hop wireless networks design, namely: (a) network
protocols and (b) network management, both in cross-layer design paradigms to
ensure the notion of service quality, such as quality of service (QoS) in wireless mesh
networks (WMNs) for backhaul applications and quality of information (QoI) in
wireless sensor networks (WSNs) for sensing tasks. Throughout the presentation of
this Ph.D. dissertation, different network settings are used as illustrative examples,
however the proposed algorithms, methodologies, protocols, and models are not
restricted in the considered networks, but rather have wide applicability.
First, this dissertation proposes a cross-layer design framework integrating
a distributed proportional-fair scheduler and a QoS routing algorithm, while using
WMNs as an illustrative example. The proposed approach has significant performance
gain compared with other network protocols. Second, this dissertation proposes
a generic admission control methodology for any packet network, wired and
wireless, by modeling the network as a black box, and using a generic mathematical
0. Abstract 3
function and Taylor expansion to capture the admission impact. Third, this dissertation
further enhances the previous designs by proposing a negotiation process,
to bridge the applicationsâ service quality demands and the resource management,
while using WSNs as an illustrative example. This approach allows the negotiation
among different service classes and WSN resource allocations to reach the optimal
operational status. Finally, the guarantees of the service quality are extended to
the environment of multiple, disconnected, mobile subnetworks, where the question
of how to maintain communications using dynamically controlled, unmanned data
ferries is investigated
Design and performance evaluation of migration-based submesh allocation strategies in mesh multicomputers
Master'sMASTER OF ENGINEERIN
A parallel interaction potential approach coupled with the immersed boundary method for fully resolved simulations of deformable interfaces and membranes
In this paper we show and discuss the use of a versatile interaction
potential approach coupled with an immersed boundary method to simulate a
variety of flows involving deformable bodies. In particular, we focus on two
kinds of problems, namely (i) deformation of liquid-liquid interfaces and (ii)
flow in the left ventricle of the heart with either a mechanical or a natural
valve. Both examples have in common the two-way interaction of the flow with a
deformable interface or a membrane. The interaction potential approach (de
Tullio & Pascazio, Jou. Comp. Phys., 2016; Tanaka, Wada and Nakamura,
Computational Biomechanics, 2016) with minor modifications can be used to
capture the deformation dynamics in both classes of problems. We show that the
approach can be used to replicate the deformation dynamics of liquid-liquid
interfaces through the use of ad-hoc elastic constants. The results from our
simulations agree very well with previous studies on the deformation of drops
in standard flow configurations such as deforming drop in a shear flow or a
cross flow. We show that the same potential approach can also be used to study
the flow in the left ventricle of the heart. The flow imposed into the
ventricle interacts dynamically with the mitral valve (mechanical or natural)
and the ventricle which are simulated using the same model. Results from these
simulations are compared with ad- hoc in-house experimental measurements.
Finally, a parallelisation scheme is presented, as parallelisation is
unavoidable when studying large scale problems involving several thousands of
simultaneously deforming bodies on hundreds of distributed memory computing
processors
Redundancy management for efficient fault recovery in NASA's distributed computing system
The management of redundancy in computer systems was studied and guidelines were provided for the development of NASA's fault-tolerant distributed systems. Fault recovery and reconfiguration mechanisms were examined. A theoretical foundation was laid for redundancy management by efficient reconfiguration methods and algorithmic diversity. Algorithms were developed to optimize the resources for embedding of computational graphs of tasks in the system architecture and reconfiguration of these tasks after a failure has occurred. The computational structure represented by a path and the complete binary tree was considered and the mesh and hypercube architectures were targeted for their embeddings. The innovative concept of Hybrid Algorithm Technique was introduced. This new technique provides a mechanism for obtaining fault tolerance while exhibiting improved performance
- âŠ