Scale-Conditioned Adaptation for Large Scale Combinatorial Optimization

Abstract

Deep reinforcement learning (DRL) for combinatorial optimization has drawn attention as an alternative for human-designed solvers. However, training DRL solvers for large-scale tasks remains challenging due to combinatorial optimization problems' NP-hardness. This paper proposes a novel \textit{scale-conditioned adaptation} (SCA) scheme that improves the transferability of the pre-trained solvers on larger-scale tasks. The main idea is to design a scale-conditioned policy by plugging a simple deep neural network, denoted as \textit{scale-conditioned network} (SCN), into the existing DRL model. SCN extracts a hidden vector from a scale value, and then we add it to the representation vector of the pre-trained DRL model. The increment of the representation vector captures the context of scale information and helps the pre-trained model effectively adapt the policy to larger-scale tasks. Our method is verified to improve the zero-shot and few-shot performance of DRL-based solvers in various large-scale combinatorial optimization tasks.

Cite

Text

Kim et al. "Scale-Conditioned Adaptation for Large Scale Combinatorial Optimization." NeurIPS 2022 Workshops: DistShift, 2022.

Markdown

[Kim et al. "Scale-Conditioned Adaptation for Large Scale Combinatorial Optimization." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/kim2022neuripsw-scaleconditioned/)

BibTeX

@inproceedings{kim2022neuripsw-scaleconditioned,
  title     = {{Scale-Conditioned Adaptation for Large Scale Combinatorial Optimization}},
  author    = {Kim, Minsu and Son, Jiwoo and Kim, Hyeonah and Park, Jinkyoo},
  booktitle = {NeurIPS 2022 Workshops: DistShift},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/kim2022neuripsw-scaleconditioned/}
}