VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning

Abstract

The Vertex Coloring Problem (VCP) is a fundamental NP-hard problem with applications in wireless networks, compiler design, scheduling, etc. We present VColRL, a deep reinforcement learning (DRL) framework that learns to color graphs quickly by leveraging a reduction-based approach that progressively reduces the graph at each step. The core novelty of VColRL is a new Markov Decision Process (MDP) formulation tailored for VCP that assigns colors to multiple vertices at each step, incorporates a rollback mechanism to revert all conflicting vertices to the undecided state, and employs a reward function designed to minimize the highest-indexed color used. Experiments on synthetic and benchmark graphs show that VColRL improves color usage over optimization solvers and prior learning-based methods, remains competitive with search-based heuristics and metaheuristics, and achieves fast runtime, while generalizing well to diverse graph families despite being trained only on synthetic graphs from a single family.

Cite

Text

Anand et al. "VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning." Transactions on Machine Learning Research, 2025.

Markdown

[Anand et al. "VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/anand2025tmlr-vcolrl/)

BibTeX

@article{anand2025tmlr-vcolrl,
  title     = {{VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning}},
  author    = {Anand, Abhinav and Peruru, Subrahmanya Swamy and Pal, Amitangshu},
  journal   = {Transactions on Machine Learning Research},
  year      = {2025},
  url       = {https://mlanthology.org/tmlr/2025/anand2025tmlr-vcolrl/}
}