VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning
Abstract
The Vertex Coloring Problem (VCP) is a fundamental NP-hard problem with applications in wireless networks, compiler design, scheduling, etc. We present VColRL, a deep reinforcement learning (DRL) framework that learns to color graphs quickly by leveraging a reduction-based approach that progressively reduces the graph at each step. The core novelty of VColRL is a new Markov Decision Process (MDP) formulation tailored for VCP that assigns colors to multiple vertices at each step, incorporates a rollback mechanism to revert all conflicting vertices to the undecided state, and employs a reward function designed to minimize the highest-indexed color used. Experiments on synthetic and benchmark graphs show that VColRL improves color usage over optimization solvers and prior learning-based methods, remains competitive with search-based heuristics and metaheuristics, and achieves fast runtime, while generalizing well to diverse graph families despite being trained only on synthetic graphs from a single family.
Cite
Text
Anand et al. "VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning." Transactions on Machine Learning Research, 2025.Markdown
[Anand et al. "VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/anand2025tmlr-vcolrl/)BibTeX
@article{anand2025tmlr-vcolrl,
title = {{VColRL: Learn to Solve the Vertex Coloring Problem Using Reinforcement Learning}},
author = {Anand, Abhinav and Peruru, Subrahmanya Swamy and Pal, Amitangshu},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/anand2025tmlr-vcolrl/}
}