Reinforcement Learning for Node Selection in Branch-and-Bound
Abstract
A big challenge in branch and bound lies in identifying the optimal node within the search tree from which to proceed. Current state-of-the-art selectors utilize either hand-crafted ensembles that automatically switch between naive sub-node selectors, or learned node selectors that rely on individual node data. In contrast to existing approaches that only consider isolated nodes, we propose a novel simulation technique that uses reinforcement learning (RL) while considering the entire tree state. To achieve this, we train a graph neural network that produces a probability distribution based on the path from the model's root to its ``to-be-selected'' leaves. Representing node-selection as a probability distribution allows us to train a decision-making policy using state-of-the-art RL techniques that capture both intrinsic node-quality and node-evaluation costs. Our method induces a high quality node selection policy on a set of varied and complex problem sets, despite only being trained on specially designed, synthetic travelling salesmen problem (TSP) instances. Using such a \emph{fixed pretrained} policy shows significant improvements on several benchmarks in optimality gap reductions and per-node efficiency under strict time constraints.
Cite
Text
Mattick and Mutschler. "Reinforcement Learning for Node Selection in Branch-and-Bound." Transactions on Machine Learning Research, 2024.Markdown
[Mattick and Mutschler. "Reinforcement Learning for Node Selection in Branch-and-Bound." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/mattick2024tmlr-reinforcement/)BibTeX
@article{mattick2024tmlr-reinforcement,
title = {{Reinforcement Learning for Node Selection in Branch-and-Bound}},
author = {Mattick, Alexander Julian and Mutschler, Christopher},
journal = {Transactions on Machine Learning Research},
year = {2024},
url = {https://mlanthology.org/tmlr/2024/mattick2024tmlr-reinforcement/}
}