90 research outputs found
A Review of Reinforcement Learning for Natural Language Processing, and Applications in Healthcare
Reinforcement learning (RL) has emerged as a powerful approach for tackling
complex medical decision-making problems such as treatment planning,
personalized medicine, and optimizing the scheduling of surgeries and
appointments. It has gained significant attention in the field of Natural
Language Processing (NLP) due to its ability to learn optimal strategies for
tasks such as dialogue systems, machine translation, and question-answering.
This paper presents a review of the RL techniques in NLP, highlighting key
advancements, challenges, and applications in healthcare. The review begins by
visualizing a roadmap of machine learning and its applications in healthcare.
And then it explores the integration of RL with NLP tasks. We examined dialogue
systems where RL enables the learning of conversational strategies, RL-based
machine translation models, question-answering systems, text summarization, and
information extraction. Additionally, ethical considerations and biases in
RL-NLP systems are addressed
A review of reinforcement learning based approaches for industrial demand response
Industrial demand response plays a key role in mitigating the operational challenges of smart grid brought by massive proliferation of distributed energy resources. However, industrial plants have complex and intertwined processes, which provides barriers for their participation in industrial demand response programs. This is in part due to the complexity and uncertainties of approximating systems models. More recently, reinforcement learning has emerged as a data-driven control technique for sequential decision-making under uncertainty. This emergence is strongly coupled with the abundance of data offered by advanced information technologies. The potential of applying reinforcement learning in industrial demand response is identified in this work by comparing pivotal aspects of reinforcement learning with the requirements of industrial demand response schemes
Recommended from our members
State-of-the-art on research and applications of machine learning in the building life cycle
Fueled by big data, powerful and affordable computing resources, and advanced algorithms, machine learning has been explored and applied to buildings research for the past decades and has demonstrated its potential to enhance building performance. This study systematically surveyed how machine learning has been applied at different stages of building life cycle. By conducting a literature search on the Web of Knowledge platform, we found 9579 papers in this field and selected 153 papers for an in-depth review. The number of published papers is increasing year by year, with a focus on building design, operation, and control. However, no study was found using machine learning in building commissioning. There are successful pilot studies on fault detection and diagnosis of HVAC equipment and systems, load prediction, energy baseline estimate, load shape clustering, occupancy prediction, and learning occupant behaviors and energy use patterns. None of the existing studies were adopted broadly by the building industry, due to common challenges including (1) lack of large scale labeled data to train and validate the model, (2) lack of model transferability, which limits a model trained with one data-rich building to be used in another building with limited data, (3) lack of strong justification of costs and benefits of deploying machine learning, and (4) the performance might not be reliable and robust for the stated goals, as the method might work for some buildings but could not be generalized to others. Findings from the study can inform future machine learning research to improve occupant comfort, energy efficiency, demand flexibility, and resilience of buildings, as well as to inspire young researchers in the field to explore multidisciplinary approaches that integrate building science, computing science, data science, and social science
Learning Algorithms for Minimizing Queue Length Regret
We consider a system consisting of a single transmitter/receiver pair and
channels over which they may communicate. Packets randomly arrive to the
transmitter's queue and wait to be successfully sent to the receiver. The
transmitter may attempt a frame transmission on one channel at a time, where
each frame includes a packet if one is in the queue. For each channel, an
attempted transmission is successful with an unknown probability. The
transmitter's objective is to quickly identify the best channel to minimize the
number of packets in the queue over time slots. To analyze system
performance, we introduce queue length regret, which is the expected difference
between the total queue length of a learning policy and a controller that knows
the rates, a priori. One approach to designing a transmission policy would be
to apply algorithms from the literature that solve the closely-related
stochastic multi-armed bandit problem. These policies would focus on maximizing
the number of successful frame transmissions over time. However, we show that
these methods have queue length regret. On the other hand, we
show that there exists a set of queue-length based policies that can obtain
order optimal queue length regret. We use our theoretical analysis to
devise heuristic methods that are shown to perform well in simulation.Comment: 28 Pages, 11 figure
BEAR: Physics-Principled Building Environment for Control and Reinforcement Learning
Recent advancements in reinforcement learning algorithms have opened doors
for researchers to operate and optimize building energy management systems
autonomously. However, the lack of an easily configurable building dynamical
model and energy management task simulation and evaluation platform has
arguably slowed the progress in developing advanced and dedicated reinforcement
learning (RL) and control algorithms for building operation tasks. Here we
propose "BEAR", a physics-principled Building Environment for Control And
Reinforcement Learning. The platform allows researchers to benchmark both
model-based and model-free controllers using a broad collection of standard
building models in Python without co-simulation using external building
simulators. In this paper, we discuss the design of this platform and compare
it with other existing building simulation frameworks. We demonstrate the
compatibility and performance of BEAR with different controllers, including
both model predictive control (MPC) and several state-of-the-art RL methods
with two case studies.Comment: Accepted at ACM e-Energy 2023; Code available at
https://github.com/chz056/BEA
Designing a generalised reward for Building Energy Management Reinforcement Learning agents
The reduction of the carbon footprint of buildings is a challenging task, partly due to the conflicting goals of maximising occupant comfort and minimising energy consumption. An intelligent management of Heating, Ventilation and Air Conditioning (HVAC) systems is creating a promising research line in which the creation of suitable algorithms could reduce energy consumption maintaining occupants' comfort. In this regard, Reinforcement Learning (RL) approaches are giving a good balance between data requirements and intelligent operations to control building systems. However, there is a gap concerning how to create a generalised reward signal that can train RL agents without delimiting the problem to a specific or controlled scenario. To tackle it, an analysis and discussion is presented about the necessary requirements for the creation of generalist rewards, with the objective of laying the foundations that allow the creation of generalist intelligent agents for building energy management.The work described in this paper was partially supported by the Basque Government under ELKARTEK project (LANTEGI4.0 KK-2020/00072)
Energy-Flexible Job-Shop Scheduling Using Deep Reinforcement Learning
Considering its high energy demand, the manufacturing industry has grand potential for demand response studies to increase the use of clean energy while reducing its own electricity cost. Production scheduling, driven by smart demand response services, plays a major role in adjusting the manufacturing sector to the volatile energy market. As a state-of-the-art method for scheduling problems, reinforcement learning has not yet been applied to the job-shop scheduling problem with demand response objectives. To address this gap, we conceptualize and implement deep reinforcement learning as a single-agent approach, combining energy cost and makespan minimization objectives. We consider makespan as an ancillary objective in order not to entirely abandon the timely completion of production operations while assigning different weights to both objectives and analyzing the resulting trade-offs between them. Our main contribution is the integration of the energy cost-related objective. We present two innovative reward functions, which consider the dynamic energy prices to select a job for the machine or allow the machine idle. The reinforcement learning agent finds optimal schedules determined by cumulative energy costs for benchmark scheduling cases from the literature
- …