Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to branch with Tree MDPs (2205.11107v3)

Published 23 May 2022 in cs.LG and math.OC

Abstract: State-of-the-art Mixed Integer Linear Program (MILP) solvers combine systematic tree search with a plethora of hard-coded heuristics, such as the branching rule. The idea of learning branching rules from data has received increasing attention recently, and promising results have been obtained by learning fast approximations of the strong branching expert. In this work, we instead propose to learn branching rules from scratch via Reinforcement Learning (RL). We revisit the work of Etheve et al. (2020) and propose tree Markov Decision Processes, or tree MDPs, a generalization of temporal MDPs that provides a more suitable framework for learning to branch. We derive a tree policy gradient theorem, which exhibits a better credit assignment compared to its temporal counterpart. We demonstrate through computational experiments that tree MDPs improve the learning convergence, and offer a promising framework for tackling the learning-to-branch problem in MILPs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Lara Scavuzzo (6 papers)
  2. Feng Yang Chen (1 paper)
  3. Didier Chételat (32 papers)
  4. Maxime Gasse (18 papers)
  5. Andrea Lodi (85 papers)
  6. Neil Yorke-Smith (20 papers)
  7. Karen Aardal (6 papers)
Citations (40)

Summary

We haven't generated a summary for this paper yet.