Multi-Objective Online Learning

Abstract

This paper presents a systematic study of multi-objective online learning. We first formulate the framework of Multi-Objective Online Convex Optimization, which encompasses a novel multi-objective regret. This regret is built upon a sequence-wise extension of the commonly used discrepancy metric Pareto suboptimality gap in zero-order multi-objective bandits. We then derive an equivalent form of the regret, making it amenable to be optimized via first-order iterative methods. To motivate the algorithm design, we give an explicit example in which equipping OMD with the vanilla min-norm solver for gradient composition will incur a linear regret, which shows that merely regularizing the iterates, as in single-objective online learning, is not enough to guarantee sublinear regrets in the multi-objective setting. To resolve this issue, we propose a novel min-regularized-norm solver that regularizes the composite weights. Combining min-regularized-norm with OMD results in the Doubly Regularized Online Mirror Multiple Descent algorithm. We further derive the multi-objective regret bound for the proposed algorithm, which matches the optimal bound in the single-objective setting. Extensive experiments on several real-world datasets verify the effectiveness of the proposed algorithm.

Cite

Text

Jiang et al. "Multi-Objective Online Learning." International Conference on Learning Representations, 2023.

Markdown

[Jiang et al. "Multi-Objective Online Learning." International Conference on Learning Representations, 2023.](https://mlanthology.org/iclr/2023/jiang2023iclr-multiobjective/)

BibTeX

@inproceedings{jiang2023iclr-multiobjective,
  title     = {{Multi-Objective Online Learning}},
  author    = {Jiang, Jiyan and Zhang, Wenpeng and Zhou, Shiji and Gu, Lihong and Zeng, Xiaodong and Zhu, Wenwu},
  booktitle = {International Conference on Learning Representations},
  year      = {2023},
  url       = {https://mlanthology.org/iclr/2023/jiang2023iclr-multiobjective/}
}