A Game-Theoretic Approach for Improving Generalization Ability of TSP Solvers

Abstract

In this paper, we introduce a two-player zero-sum framework between a trainable \emph{Solver} and a \emph{Data Generator} to improve the generalization ability of deep learning-based solvers for Traveling Salesman Problems (TSP). Grounded in \textsl{Policy Space Response Oracle} (PSRO) methods, our two-player framework outputs a population of best-responding Solvers, over which we can mix and output a combined model that achieves the least exploitability against the Generator, and thereby the most generalizable performance on different TSP tasks. We conduct experiments on a variety of TSP instances with different types and sizes. Results suggest that our Solvers achieve the state-of-the-art performance even on tasks the Solver never meets, whilst the performance of other deep learning-based Solvers drops sharply due to over-fitting. To demonstrate the principle of our framework, we study the learning outcome of the proposed two-player game and demonstrate that the exploitability of the Solver population decreases during training, and it eventually approximates the Nash equilibrium along with the Generator.

Cite

Text

Wang et al. "A Game-Theoretic Approach for Improving Generalization Ability of TSP Solvers." ICLR 2022 Workshops: GMS, 2022.

Markdown

[Wang et al. "A Game-Theoretic Approach for Improving Generalization Ability of TSP Solvers." ICLR 2022 Workshops: GMS, 2022.](https://mlanthology.org/iclrw/2022/wang2022iclrw-gametheoretic/)

BibTeX

@inproceedings{wang2022iclrw-gametheoretic,
  title     = {{A Game-Theoretic Approach for Improving Generalization Ability of TSP Solvers}},
  author    = {Wang, Chenguang and Yang, Yaodong and Slumbers, Oliver and Han, Congying and Guo, Tiande and Zhang, Haifeng and Wang, Jun},
  booktitle = {ICLR 2022 Workshops: GMS},
  year      = {2022},
  url       = {https://mlanthology.org/iclrw/2022/wang2022iclrw-gametheoretic/}
}