67 research outputs found
Towards Omni-generalizable Neural Methods for Vehicle Routing Problems
Learning heuristics for vehicle routing problems (VRPs) has gained much
attention due to the less reliance on hand-crafted rules. However, existing
methods are typically trained and tested on the same task with a fixed size and
distribution (of nodes), and hence suffer from limited generalization
performance. This paper studies a challenging yet realistic setting, which
considers generalization across both size and distribution in VRPs. We propose
a generic meta-learning framework, which enables effective training of an
initialized model with the capability of fast adaptation to new tasks during
inference. We further develop a simple yet efficient approximation method to
reduce the training overhead. Extensive experiments on both synthetic and
benchmark instances of the traveling salesman problem (TSP) and capacitated
vehicle routing problem (CVRP) demonstrate the effectiveness of our method. The
code is available at: https://github.com/RoyalSkye/Omni-VRP.Comment: Accepted at ICML 202
Neural Multi-Objective Combinatorial Optimization with Diversity Enhancement
Most of existing neural methods for multi-objective combinatorial
optimization (MOCO) problems solely rely on decomposition, which often leads to
repetitive solutions for the respective subproblems, thus a limited Pareto set.
Beyond decomposition, we propose a novel neural heuristic with diversity
enhancement (NHDE) to produce more Pareto solutions from two perspectives. On
the one hand, to hinder duplicated solutions for different subproblems, we
propose an indicator-enhanced deep reinforcement learning method to guide the
model, and design a heterogeneous graph attention mechanism to capture the
relations between the instance graph and the Pareto front graph. On the other
hand, to excavate more solutions in the neighborhood of each subproblem, we
present a multiple Pareto optima strategy to sample and preserve desirable
solutions. Experimental results on classic MOCO problems show that our NHDE is
able to generate a Pareto front with higher diversity, thereby achieving
superior overall performance. Moreover, our NHDE is generic and can be applied
to different neural methods for MOCO.Comment: Accepted at NeurIPS 202
Graph Neural Networks for Job Shop Scheduling Problems:A Survey
Job shop scheduling problems (JSSPs) represent a critical and challenging class of combinatorial optimization problems. Recent years have witnessed a rapid increase in the application of graph neural networks (GNNs) to solve JSSPs, albeit lacking a systematic survey of the relevant literature. This paper aims to thoroughly review prevailing GNN methods for different types of JSSPs and the closely related flow-shop scheduling problems (FSPs), especially those leveraging deep reinforcement learning (DRL). We begin by presenting the graph representations of various JSSPs, followed by an introduction to the most commonly used GNN architectures. We then review current GNN-based methods for each problem type, highlighting key technical elements such as graph representations, GNN architectures, GNN tasks, and training algorithms. Finally, we summarize and analyze the advantages and limitations of GNNs in solving JSSPs and provide potential future research opportunities. We hope this survey can motivate and inspire innovative approaches for more powerful GNN-based approaches in tackling JSSPs and other scheduling problems
Two types of zero Hall phenomena in few-layer MnBiTe
The van der Waals antiferromagnetic topological insulator MnBiTe
represents a promising platform for exploring the layer-dependent magnetism and
topological states of matter. Despite the realization of several quantized
phenomena, such as the quantum anomalous Hall effect and the axion insulator
state, the recently observed discrepancies between magnetic and transport
properties have aroused controversies concerning the topological nature of
MnBiTe in the ground state. Here, we demonstrate the existence of two
distinct types of zero Hall phenomena in few-layer MnBiTe. In addition
to the robust zero Hall plateau associated with the axion insulator state, an
unexpected zero Hall phenomenon also occurs in some odd-number-septuple layer
devices. Importantly, a statistical survey of the optical contrast in more than
200 MnBiTe reveals that such accidental zero Hall phenomenon arises
from the reduction of effective thickness during fabrication process, a factor
that was rarely noticed in previous studies of 2D materials. Our finding not
only resolves the controversies on the relation between magnetism and anomalous
Hall effect in MnBiTe, but also highlights the critical issues
concerning the fabrication and characterization of devices based on 2D
materials.Comment: 21 pages, 4 figure
Neural combinatorial optimization: from routing to integer programming
Combinatorial optimization problems (COPs) are an important branch of mathematical optimization. It covers practical applications in many fields such as communication, transportation, manufacturing and has received massive attention in domains of computer science, operations research, economics, etc. However, most COPs are difficult to solve exactly owing to their NP-hardness. Heuristics thus are often leveraged as a popular class of optimization methods to solve COPs. However, classic heuristic methods severely depend on the design of hand-crafted rules for every specific COP, which could limit their performance in practice. Also, finding effective rules in heuristics is non-trivial and often needs massive expertise and tuning work. To overcome the above issues, the neural combinatorial optimization (NCO) domain comes into being, which refers to the research and applications of deep learning for solving COPs. Although there is already abundant literature in such domain, some challenging aspects are still not tackled well. For example, the current deep models are generally still inferior to highly-specialized solvers in some specific COPs, and meanwhile it is still difficult for NCO methods to solve general COPs, especially the ones with uncertainty and multiple objectives. In this doctoral thesis, we propose several NCO methods to automatically learn heuristics to solve 1) routing problems, which are a specific and important class of COPs, 2) integer programming, which is a general formulation to model various COPs in different domains, 3) stochastic integer programming, which can model various COPs with uncertainty, and 4) multi-objective integer programming, which is used to optimize multiple objectives for general COPs. All the proposed NCO methods obviate the need of considerable trial-and-error and domain knowledge. They also manifest themselves in the desirable generalization capacities across problem sizes or distributions.
Firstly, we propose a deep reinforcement learning (DRL) framework to learn improvement heuristics for routing problems. We describe the solving process of improvement heuristics as a Markov decision process (MDP), and then design a self-attention based neural architecture for the policy in MDP so that it is used to select the next solution at each step of the local search. We deploy the method to solve two classic routing problems, i.e., travelling salesman problem (TSP) and capacitated vehicle routing problem (CVRP). Results show that the proposed method outperforms state-of-the-art learning based methods. The learned policies for improvement heuristics are more effective than conventional hand-crafted rules and they are able to well generalize to varying problem sizes and initial solutions.
However, the above method resorts to simplified pairwise local operators, which limit the efficiency to improve the solution, and it is not applicable to general COPs. To tackle these issues, we then employ DRL to learn large neighborhood search (LNS) algorithm for solving integer programs (IPs), which can be used to model general COPs. The policy in MDP is trained as the destroy operator to pick a subset of variables at each LNS step, which is reoptimized by an IP solver as the repair operator. Since the combinatorial number of variable subsets hinders the direct application of RL algorithms, we propose action factorization to represent all variable subsets with binary decisions on each variable. The policy is parameterized via a neural network and trained by the actor-critic algorithm to select actions for each variable in parallel. Our method is evaluated on four IP benchmark problems. Results show that it significantly boosts the bounded-time performance of SCIP and Gurobi, and the learned policies generalize well to larger problems.
Despite that the learned LNS can well solve general COPs by modeling them as IPs, some vital aspects in practical COPs are still not involved, e.g., the uncertainty and multiple objectives. Therefore, we propose two NCO methods to solve stochastic and multi-objective IPs. To solve two-stage stochastic integer programs (2SIPs), we use conditional variational autoencoder (CVAE) to learn scenario representations, where a graph convolutional network (GCN) based encoder embeds each scenario with the deterministic part (i.e. context) of a 2SIP instance into the latent space, from which a decoder reconstructs the scenario from its latent representation conditioned on the context. We apply the trained encoder to scenario reduction and objective prediction, which assist in the search of scenario representatives for attaining approximate solutions to original 2SIPs. Results show that our method achieves high-quality solutions in short runtime and the trained encoder generalizes well to larger problems, more scenarios and varying distributions.
To solve multi-objective integer programs (MOIPs), we design a NCO method to refine objective-space decomposition algorithms (ODAs). Since ODAs often encounter difficulties in handling scalarized problems, which could cause infeasibility or repetitive nondominated points and thus induce redundant runtime, we present a graph neural network (GNN) based method to learn the reduction rule in the ODA. We formulate the algorithmic procedure of generic ODAs as a Markov decision process and parameterize the policy (reduction rule) with a novel two-stage GNN for state representation. We train our model with imitation learning and deploy it on a state-of-the-art ODA. Results show that our method significantly improves the solving efficiency of the ODA. The learned policy generalizes fairly well to larger problems or more objectives, and the proposed GNN outperforms existing ones for integer programming in terms of test and generalization accuracy.Doctor of Philosoph
- …