1,560 research outputs found
Learning Scheduling Algorithms for Data Processing Clusters
Efficiently scheduling data processing jobs on distributed compute clusters
requires complex algorithms. Current systems, however, use simple generalized
heuristics and ignore workload characteristics, since developing and tuning a
scheduling policy for each workload is infeasible. In this paper, we show that
modern machine learning techniques can generate highly-efficient policies
automatically. Decima uses reinforcement learning (RL) and neural networks to
learn workload-specific scheduling algorithms without any human instruction
beyond a high-level objective such as minimizing average job completion time.
Off-the-shelf RL techniques, however, cannot handle the complexity and scale of
the scheduling problem. To build Decima, we had to develop new representations
for jobs' dependency graphs, design scalable RL models, and invent RL training
methods for dealing with continuous stochastic job arrivals. Our prototype
integration with Spark on a 25-node cluster shows that Decima improves the
average job completion time over hand-tuned scheduling heuristics by at least
21%, achieving up to 2x improvement during periods of high cluster load
Scheduling Algorithms: Challenges Towards Smart Manufacturing
Collecting, processing, analyzing, and driving knowledge from large-scale real-time data is now realized with the emergence of Artificial Intelligence (AI) and Deep Learning (DL). The breakthrough of Industry 4.0 lays a foundation for intelligent manufacturing. However, implementation challenges of scheduling algorithms in the context of smart manufacturing are not yet comprehensively studied. The purpose of this study is to show the scheduling No.s that need to be considered in the smart manufacturing paradigm. To attain this objective, the literature review is conducted in five stages using publish or perish tools from different sources such as Scopus, Pubmed, Crossref, and Google Scholar. As a result, the first contribution of this study is a critical analysis of existing production scheduling algorithms\u27 characteristics and limitations from the viewpoint of smart manufacturing. The other contribution is to suggest the best strategies for selecting scheduling algorithms in a real-world scenario
A Comparative Analysis of Deep Reinforcement Learning-based xApps in O-RAN
The highly heterogeneous ecosystem of Next Generation (NextG) wireless
communication systems calls for novel networking paradigms where
functionalities and operations can be dynamically and optimally reconfigured in
real time to adapt to changing traffic conditions and satisfy stringent and
diverse Quality of Service (QoS) demands. Open Radio Access Network (RAN)
technologies, and specifically those being standardized by the O-RAN Alliance,
make it possible to integrate network intelligence into the once monolithic RAN
via intelligent applications, namely, xApps and rApps. These applications
enable flexible control of the network resources and functionalities, network
management, and orchestration through data-driven control loops. Despite recent
work demonstrating the effectiveness of Deep Reinforcement Learning (DRL) in
controlling O-RAN systems, how to design these solutions in a way that does not
create conflicts and unfair resource allocation policies is still an open
challenge. In this paper, we perform a comparative analysis where we dissect
the impact of different DRL-based xApp designs on network performance.
Specifically, we benchmark 12 different xApps that embed DRL agents trained
using different reward functions, with different action spaces and with the
ability to hierarchically control different network parameters. We prototype
and evaluate these xApps on Colosseum, the world's largest O-RAN-compliant
wireless network emulator with hardware-in-the-loop. We share the lessons
learned and discuss our experimental results, which demonstrate how certain
design choices deliver the highest performance while others might result in a
competitive behavior between different classes of traffic with similar
objectives.Comment: 6 pages, 16 figure
A Deep Q-Learning based Smart Scheduling of EVs for Demand Response in Smart Grids
Economic and policy factors are driving the continuous increase in the
adoption and usage of electrical vehicles (EVs). However, despite being a
cleaner alternative to combustion engine vehicles, EVs have negative impacts on
the lifespan of microgrid equipment and energy balance due to increased power
demand and the timing of their usage. In our view grid management should
leverage on EVs scheduling flexibility to support local network balancing
through active participation in demand response programs. In this paper, we
propose a model-free solution, leveraging Deep Q-Learning to schedule the
charging and discharging activities of EVs within a microgrid to align with a
target energy profile provided by the distribution system operator. We adapted
the Bellman Equation to assess the value of a state based on specific rewards
for EV scheduling actions and used a neural network to estimate Q-values for
available actions and the epsilon-greedy algorithm to balance exploitation and
exploration to meet the target energy profile. The results are promising
showing that the proposed solution can effectively schedule the EVs charging
and discharging actions to align with the target profile with a Person
coefficient of 0.99, handling effective EVs scheduling situations that involve
dynamicity given by the e-mobility features, relying only on data with no
knowledge of EVs and microgrid dynamics.Comment: Submitted to journa
NICE: Robust Scheduling through Reinforcement Learning-Guided Integer Programming
Integer programs provide a powerful abstraction for representing a wide range
of real-world scheduling problems. Despite their ability to model general
scheduling problems, solving large-scale integer programs (IP) remains a
computational challenge in practice. The incorporation of more complex
objectives such as robustness to disruptions further exacerbates the
computational challenge. We present NICE (Neural network IP Coefficient
Extraction), a novel technique that combines reinforcement learning and integer
programming to tackle the problem of robust scheduling. More specifically, NICE
uses reinforcement learning to approximately represent complex objectives in an
integer programming formulation. We use NICE to determine assignments of pilots
to a flight crew schedule so as to reduce the impact of disruptions. We compare
NICE with (1) a baseline integer programming formulation that produces a
feasible crew schedule, and (2) a robust integer programming formulation that
explicitly tries to minimize the impact of disruptions. Our experiments show
that, across a variety of scenarios, NICE produces schedules resulting in 33\%
to 48\% fewer disruptions than the baseline formulation. Moreover, in more
severely constrained scheduling scenarios in which the robust integer program
fails to produce a schedule within 90 minutes, NICE is able to build robust
schedules in less than 2 seconds on average.Comment: Accepted in 36th AAAI Conference. 7 pages + 2 pages appendix, 1
figure. Code available at https://github.com/nsidn98/NIC
- …