93,067 research outputs found
Large-Scale Solution Approaches for Healthcare and Supply Chain Scheduling
This research proposes novel solution techniques for two real world problems. We first consider a patient scheduling problem in a proton therapy facility with deterministic patient arrivals. In order to assess the impacts of several operational constraints, we propose single and multi-criteria linear programming models. In addition, we ensure that the strategic patient mix restrictions predetermined by the decision makers are also enforced within the planning horizon. We study the mathematical structures of the single criteria model with strict patient mix restrictions and derive analytical equations for the optimal solutions under several operational restrictions. These efforts lead to a set of rule of thumbs that can be utilized to assess the impacts of several input parameters and patient mix levels on the capacity utilization without solving optimization problems. The necessary and sufficient conditions to analytically generate exact efficient frontiers of the bicriteria problem without any additional side constraint are also explored. In a follow up study, we investigate the solution techniques for the same patient scheduling problem with stochastic patient arrivals. We propose two Markov Decision Process (MDP) models that are capable of tackling the stochasticity.
The second problem of interest is a variant of the parallel machine scheduling problem. We propose constraint programming (CP) and logic-based Benders decomposition algorithms in order to make the best decisions for scheduling nonidentical jobs with time windows and sequence dependent setup times on dissimilar parallel machines in a fixed planning horizon. This problem is formulated with (i) maximizing total profit and (ii) minimizing makespan objectives. We conduct several sensitivity analysis to test the quality and robustness of the solutions on a real life case study
Position-Based Multi-Agent Dynamics for Real-Time Crowd Simulation (MiG paper)
Exploiting the efficiency and stability of Position-Based Dynamics (PBD), we
introduce a novel crowd simulation method that runs at interactive rates for
hundreds of thousands of agents. Our method enables the detailed modeling of
per-agent behavior in a Lagrangian formulation. We model short-range and
long-range collision avoidance to simulate both sparse and dense crowds. On the
particles representing agents, we formulate a set of positional constraints
that can be readily integrated into a standard PBD solver. We augment the
tentative particle motions with planning velocities to determine the preferred
velocities of agents, and project the positions onto the constraint manifold to
eliminate colliding configurations. The local short-range interaction is
represented with collision and frictional contact between agents, as in the
discrete simulation of granular materials. We incorporate a cohesion model for
modeling collective behaviors and propose a new constraint for dealing with
potential future collisions. Our new method is suitable for use in interactive
games.Comment: 9 page
Recommended from our members
Understanding Model-Based Reinforcement Learning and its Application in Safe Reinforcement Learning
Model-based reinforcement learning algorithms have been shown to achieve successful results on various continuous control benchmarks, but the understanding of model-based methods is limited. We try to interpret how model-based method works through novel experiments on state-of-the-art algorithms with an emphasis on the model learning part. We evaluate the role of the model learning in policy optimization and propose methods to learn a more accurate model. With a better understanding of model-based reinforcement learning, we then apply model-based methods to solve safe reinforcement learning (RL) problems with near-zero violation of hard constraints throughout training. Drawing an analogy with how humans and animals learn to perform safe actions, we break down the safe RL problem into three stages. First, we train agents in a constraint-free environment to learn a performant policy for reaching high rewards, and simultaneously learn a model of the dynamics. Second, we use model-based methods to plan safe actions and train a safeguarding policy from these actions through imitation. Finally, we propose a factored framework to train an overall policy that mixes the performant policy and the safeguarding policy. This three-step curriculum ensures near-zero violation of safety constraints at all times. As an advantage of model-based method, the sample complexity required at the second and third steps of the process is significantly lower than model-free methods and can enable online safe learning. We demonstrate the effectiveness of our methods in various continuous control problems and analyze the advantages over state-of-the-art approaches
- …