Mitigating Gradient Bias in Multi-Objective Learning: A Provably Convergent Approach

Abstract

Many machine learning problems today have multiple objective functions. They appear either in learning with multiple criteria where learning has to make a trade-off between multiple performance metrics such as fairness, safety and accuracy; or, in multi-task learning where multiple tasks are optimized jointly, sharing inductive bias between them. This problems are often tackled by the multi-objective optimization framework. However, existing stochastic multi-objective gradient methods and its variants (e.g., MGDA, PCGrad, CAGrad, etc.) all adopt a biased noisy gradient direction, which leads to degraded empirical performance. To this end, we develop a stochastic multi-objective gradient correction (MoCo) method for multi-objective optimization. The unique feature of our method is that it can guarantee convergence without increasing the batch size even in the nonconvex setting. Simulations on multi-task supervised and reinforcement learning demonstrate the effectiveness of our method relative to the state-of-the-art methods.

Cite

Text

Fernando et al. "Mitigating Gradient Bias in Multi-Objective Learning: A Provably Convergent Approach." International Conference on Learning Representations, 2023.

Markdown

[Fernando et al. "Mitigating Gradient Bias in Multi-Objective Learning: A Provably Convergent Approach." International Conference on Learning Representations, 2023.](https://mlanthology.org/iclr/2023/fernando2023iclr-mitigating/)

BibTeX

@inproceedings{fernando2023iclr-mitigating,
  title     = {{Mitigating Gradient Bias in Multi-Objective Learning: A Provably Convergent Approach}},
  author    = {Fernando, Heshan Devaka and Shen, Han and Liu, Miao and Chaudhury, Subhajit and Murugesan, Keerthiram and Chen, Tianyi},
  booktitle = {International Conference on Learning Representations},
  year      = {2023},
  url       = {https://mlanthology.org/iclr/2023/fernando2023iclr-mitigating/}
}