47,690 research outputs found
A Lazy Bailout Approach for Dual-Criticality Systems on Uniprocessor Platforms
© 2019 by the authors. Licensee MDPI, Basel, Switzerland.A challenge in the design of cyber-physical systems is to integrate the scheduling of tasks of different criticality, while still providing service guarantees for the higher critical tasks in case of resource-shortages caused by faults. While standard real-time scheduling is agnostic to the criticality of tasks, the scheduling of tasks with different criticalities is called mixed-criticality scheduling. In this paper we present the Lazy Bailout Protocol (LBP), a mixed-criticality scheduling method where low-criticality jobs overrunning their time budget cannot threaten the timeliness of high-criticality jobs while at the same time the method tries to complete as many low-criticality jobs as possible. The key principle of LBP is instead of immediately abandoning low-criticality jobs when a high-criticality job overruns its optimistic WCET estimate, to put them in a low-priority queue for later execution. To compare mixed-criticality scheduling methods we introduce a formal quality criterion for mixed-criticality scheduling, which, above all else, compares schedulability of high-criticality jobs and only afterwards the schedulability of low-criticality jobs. Based on this criterion we prove that LBP behaves better than the original {\em Bailout Protocol} (BP). We show that LBP can be further improved by slack time exploitation and by gain time collection at runtime, resulting in LBPSG. We also show that these improvements of LBP perform better than the analogous improvements based on BP.Peer reviewedFinal Published versio
Deadline-Aware Reservation-Based Scheduling
The ever-growing need to improve return-on-investment (ROI) for cluster infrastructure that processes data which is being continuously generated at a higher rate than ever before introduces new challenges for big-data processing frameworks. Highly complex mixed workload arriving at modern clusters along with a growing number of time-sensitive critical production jobs necessitates cluster management systems to evolve. Most big-data systems are not only required to guarantee that production jobs will complete before their deadline, but also minimize the latency for best-effort jobs to increase ROI.
This research presents DARSS, a deadline-aware reservation-based scheduling system. DARSS addresses the above-stated problem by using a reservation-based approach to scheduling that supports temporal requirements of production jobs while keeping the latency for best-effort jobs low. Fined-grained resource allocation enables DARSS to schedule more tasks than a coarser-grained approach would. Furthermore, DARSS schedules production jobs as close to their deadlines as possible. This scheduling policy allows the system to maximize the number of low-priority tasks that can be scheduled opportunistically. DARSS is a scalable system that can be integrated with YARN.
DARSS is evaluated on a simulated cluster of 300 nodes against a workload derived from Google Borg's trace. DARSS is compared with Microsoft's Rayon and YARN's built-in scheduler. DARSS achieves better production job acceptance rate than both YARN and Rayon. The experiments show that all of the production jobs accepted by DARSS complete before their deadlines. Furthermore, DARSS has a higher number of best-effort jobs serviced than Rayon. And finally, DARSS has lower latency for best-effort jobs than Rayon
Framework for sustainable TVET-Teacher Education Program in Malaysia Public Universities
Studies had stated that less attention was given to the education aspect, such as
teaching and learning in planning for improving the TVET system. Due to the 21st
Century context, the current paradigm of teaching for the TVET educators also has
been reported to be fatal and need to be shifted. All these disadvantages reported
hindering the country from achieving the 5th strategy in the Strategic Plan for
Vocational Education Transformation to transform TVET system as a whole.
Therefore, this study aims to develop a framework for sustainable TVET Teacher
Education program in Malaysia. This study had adopted an Exploratory Sequential
Mix-Method design, which involves a semi-structured interview (phase one) and
survey method (phase two). Nine experts had involved in phase one chosen by using
Purposive Sampling Technique. As in phase two, 118 TVET-TE program lecturers
were selected as the survey sample chosen through random sampling method. After
data analysis in phase one (thematic analysis) and phase two (Principal Component
Analysis), eight domains and 22 elements have been identified for the framework for
sustainable TVET-TE program in Malaysia. This framework was identified to embed
the elements of 21st Century Education, thus filling the gap in this research. The
research findings also indicate that the developed framework was unidimensional and
valid for the development and research regarding TVET-TE program in Malaysia.
Lastly, it is in the hope that this research can be a guide for the nations in producing a
quality TVET teacher in the future
Mixed-Criticality Scheduling with I/O
This paper addresses the problem of scheduling tasks with different
criticality levels in the presence of I/O requests. In mixed-criticality
scheduling, higher criticality tasks are given precedence over those of lower
criticality when it is impossible to guarantee the schedulability of all tasks.
While mixed-criticality scheduling has gained attention in recent years, most
approaches typically assume a periodic task model. This assumption does not
always hold in practice, especially for real-time and embedded systems that
perform I/O. For example, many tasks block on I/O requests until devices signal
their completion via interrupts; both the arrival of interrupts and the waking
of blocked tasks can be aperiodic. In our prior work, we developed a scheduling
technique in the Quest real-time operating system, which integrates the
time-budgeted management of I/O operations with Sporadic Server scheduling of
tasks. This paper extends our previous scheduling approach with support for
mixed-criticality tasks and I/O requests on the same processing core. Results
show the effective schedulability of different task sets in the presence of I/O
requests is superior in our approach compared to traditional methods that
manage I/O using techniques such as Sporadic Servers.Comment: Second version has replaced simulation experiments with real machine
experiments, third version fixed minor error in Equation 5 (missing a plus
sign
Learning Scheduling Algorithms for Data Processing Clusters
Efficiently scheduling data processing jobs on distributed compute clusters
requires complex algorithms. Current systems, however, use simple generalized
heuristics and ignore workload characteristics, since developing and tuning a
scheduling policy for each workload is infeasible. In this paper, we show that
modern machine learning techniques can generate highly-efficient policies
automatically. Decima uses reinforcement learning (RL) and neural networks to
learn workload-specific scheduling algorithms without any human instruction
beyond a high-level objective such as minimizing average job completion time.
Off-the-shelf RL techniques, however, cannot handle the complexity and scale of
the scheduling problem. To build Decima, we had to develop new representations
for jobs' dependency graphs, design scalable RL models, and invent RL training
methods for dealing with continuous stochastic job arrivals. Our prototype
integration with Spark on a 25-node cluster shows that Decima improves the
average job completion time over hand-tuned scheduling heuristics by at least
21%, achieving up to 2x improvement during periods of high cluster load
Hybrid Meta-Heuristics for Robust Scheduling
The production and delivery of rapidly perishable goods in distributed supply networks involves a number of tightly coupled decision and optimization problems regarding the just-in-time production scheduling and the routing of the delivery vehicles in order to satisfy strict customer specified time-windows. Besides dealing with the typical combinatorial complexity related to activity assignment and synchronization, effective methods must also provide robust schedules, coping with the stochastic perturbations (typically transportation delays) affecting the distribution process. In this paper, we propose a novel metaheuristic approach for robust scheduling. Our approach integrates mathematical programming, multi-objective evolutionary computation, and problem-specific constructive heuristics. The optimization algorithm returns a set of solutions with different cost and risk tradeoffs, allowing the analyst to adapt the planning depending on the attitude to risk. The effectiveness of the approach is demonstrated by a real-world case concerning the production and distribution of ready-mixed concrete.Meta-Heuristics;Multi-Objective Genetic Optimization;Robust Scheduling;Supply Networks
- âŠ