Preference-Driven Multi-Objective Combinatorial Optimization with Conditional Computation
Abstract
Recent deep reinforcement learning methods have achieved remarkable success in solving multi-objective combinatorial optimization problems (MOCOPs) by decomposing them into multiple subproblems, each associated with a specific weight vector. However, these methods typically treat all subproblems equally and solve them using a single model, hindering the effective exploration of the solution space and thus leading to suboptimal performance. To overcome the limitation, we propose POCCO, a novel plug-and-play framework that enables adaptive selection of model structures for subproblems, which are subsequently optimized based on preference signals rather than explicit reward values. Specifically, we design a conditional computation block that routes subproblems to specialized neural architectures. Moreover, we propose a preference-driven optimization algorithm that learns pairwise preferences between winning and losing solutions. We evaluate the efficacy and versatility of POCCO by applying it to two state-of-the-art neural methods for MOCOPs. Experimental results across four classic MOCOP benchmarks demonstrate its significant superiority and strong generalization.
Cite
Text
Fan et al. "Preference-Driven Multi-Objective Combinatorial Optimization with Conditional Computation." Advances in Neural Information Processing Systems, 2025.Markdown
[Fan et al. "Preference-Driven Multi-Objective Combinatorial Optimization with Conditional Computation." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/fan2025neurips-preferencedriven/)BibTeX
@inproceedings{fan2025neurips-preferencedriven,
title = {{Preference-Driven Multi-Objective Combinatorial Optimization with Conditional Computation}},
author = {Fan, Mingfeng and Zhou, Jianan and Zhang, Yifeng and Wu, Yaoxin and Chen, Jinbiao and Sartoretti, Guillaume Adrien},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/fan2025neurips-preferencedriven/}
}