4,726 research outputs found
Submodular Function Maximization for Group Elevator Scheduling
We propose a novel approach for group elevator scheduling by formulating it
as the maximization of submodular function under a matroid constraint. In
particular, we propose to model the total waiting time of passengers using a
quadratic Boolean function. The unary and pairwise terms in the function denote
the waiting time for single and pairwise allocation of passengers to elevators,
respectively. We show that this objective function is submodular. The matroid
constraints ensure that every passenger is allocated to exactly one elevator.
We use a greedy algorithm to maximize the submodular objective function, and
derive provable guarantees on the optimality of the solution. We tested our
algorithm using Elevate 8, a commercial-grade elevator simulator that allows
simulation with a wide range of elevator settings. We achieve significant
improvement over the existing algorithms.Comment: 10 pages; 2017 International Conference on Automated Planning and
Scheduling (ICAPS
Reinforcement Learning: A Survey
This paper surveys the field of reinforcement learning from a
computer-science perspective. It is written to be accessible to researchers
familiar with machine learning. Both the historical basis of the field and a
broad selection of current work are summarized. Reinforcement learning is the
problem faced by an agent that learns behavior through trial-and-error
interactions with a dynamic environment. The work described here has a
resemblance to work in psychology, but differs considerably in the details and
in the use of the word ``reinforcement.'' The paper discusses central issues of
reinforcement learning, including trading off exploration and exploitation,
establishing the foundations of the field via Markov decision theory, learning
from delayed reinforcement, constructing empirical models to accelerate
learning, making use of generalization and hierarchy, and coping with hidden
state. It concludes with a survey of some implemented systems and an assessment
of the practical utility of current methods for reinforcement learning.Comment: See http://www.jair.org/ for any accompanying file
Learning obstacle avoidance with an operant behavioral model
Artificial intelligence researchers have been attracted by the idea of having robots learn how to accomplish a task, rather than being told explicitly. Reinforcement learning has been proposed as an appealing framework to be used in controlling mobile agents. Robot learning research, as well as research in biological systems, face many similar problems in order to display high flexibility in performing a variety of tasks. In this work, the controlling of a vehicle in an avoidance task by a previously developed operant learning model (a form of animal learning) is studied. An environment in which a mobile robot with proximity sensors has to minimize the punishment for colliding against obstacles is simulated. The results were compared with the Q-Learning algorithm, and the proposed model had better performance. In this way a new artificial intelligence agent inspired by neurobiology, psychology, and ethology research is proposed.Fil: Gutnisky, D. A.. Universidad de Buenos Aires. Facultad de Ingeniería.Instituto de Ingeniería Biomédica; ArgentinaFil: Zanutto, Bonifacio Silvano. Consejo Nacional de Investigaciones Científicas y Técnicas. Instituto de Biología y Medicina Experimental. Fundación de Instituto de Biología y Medicina Experimental. Instituto de Biología y Medicina Experimental; Argentina. Universidad de Buenos Aires. Facultad de Ingeniería.Instituto de Ingeniería Biomédica; Argentin
- …