Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exact Combinatorial Optimization with Graph Convolutional Neural Networks (1906.01629v3)

Published 4 Jun 2019 in cs.LG, math.OC, and stat.ML

Abstract: Combinatorial optimization problems are typically tackled by the branch-and-bound paradigm. We propose a new graph convolutional neural network model for learning branch-and-bound variable selection policies, which leverages the natural variable-constraint bipartite graph representation of mixed-integer linear programs. We train our model via imitation learning from the strong branching expert rule, and demonstrate on a series of hard problems that our approach produces policies that improve upon state-of-the-art machine-learning methods for branching and generalize to instances significantly larger than seen during training. Moreover, we improve for the first time over expert-designed branching rules implemented in a state-of-the-art solver on large problems. Code for reproducing all the experiments can be found at https://github.com/ds4dm/learn2branch.

Citations (433)

Summary

  • The paper introduces a GCNN model that learns branch-and-bound variable selection policies for MILPs through imitation learning.
  • The approach employs a natural bipartite graph representation to reduce feature engineering and enhance solution accuracy.
  • Experimental results show robust generalization and competitive performance on NP-hard benchmarks, outperforming traditional methods.

Graph Convolutional Neural Networks for Combinatorial Optimization

The paper presents a novel approach to combinatorial optimization by leveraging Graph Convolutional Neural Networks (GCNNs) for learning branch-and-bound variable selection policies in Mixed-Integer Linear Programming (MILP). It addresses major challenges in this domain, such as efficiently encoding variable states and generalizing to larger instances, by exploiting the natural bipartite graph representation of MILPs.

Problem Context

Branch-and-bound is the prevalent method for solving combinatorial optimization problems. The challenge lies in efficiently selecting branching variables to minimize search tree size, which directly impacts solving time. Traditional methods rely on heuristic-driven approaches. The authors propose using imitation learning to train GCNN models, thereby utilizing statistical learning methods for automated tuning.

Methodology

  • State Representation: The MILP is encoded as a bipartite graph with constraint and variable nodes. This reduces manual feature engineering and captures the problem's inherent sparsity.
  • Graph Convolutional Neural Networks: The GCNN model performs graph convolutions to disseminate information across the MILP's bipartite graph. This is executed through two interleaved half-convolutions, from variables to constraints and vice versa, leveraging its permutation-invariance and sparsity-adaptiveness.
  • Imitation Learning: The GCNN is trained through a behavioral cloning approach, which learns from a strong branching expert to produce effective and computationally efficient policies.

Experimental Setup

The experiments involve four NP-hard problem benchmarks: set covering, combinatorial auctions, capacitated facility locations, and maximum independent sets. Instances are categorized by complexity (Easy, Medium, Hard) based on size. The GCNN's performance is compared against state-of-the-art branching rules and various machine learning methods like ExtraTrees and LambdaMART.

Results

  • Accuracy and Generalization: The GCNN outperforms existing machine learning approaches in terms of solution accuracy across all benchmarks. Notably, it demonstrates strong generalization capabilities to larger instances beyond the training distribution.
  • Solving Time and Node Count: The GCNN achieves competitive solving times and reduces the number of branch-and-bound nodes compared to both human-engineered and machine learning baselines. It significantly outperforms SCIP’s default branching rule on medium and hard set covering instances.

Implications and Future Work

The paper indicates that GCNNs provide a robust architectural prior for addressing the branching problem in MILPs. The success of imitation learning with graph-structured data encourages further exploration into integrating such models within existing combinatorial optimization solvers.

Future research could extend this approach to additional combinatorial domains and investigate hybrid methods that combine traditional and machine learning strategies for branching rules. Exploring reinforcement learning as a means to refine learned policies could yield further improvements.

Conclusion

The authors contribute to the field by innovatively applying GCNNs to a central issue in combinatorial optimization: branching in MILPs. By reducing feature engineering and enhancing solution efficiency, this approach marks a substantial advance in leveraging machine learning for optimization problems, providing a foundation for future developments in AI-driven solvers.