Dueling Convex Optimization with General Preferences

Abstract

We address the problem of convex optimization with dueling feedback, where the goal is to minimize a convex function given a weaker form of dueling feedback. Each query consists of two points and the dueling feedback returns a (noisy) single-bit binary comparison of the function values of the two queried points. The translation of the function values to the single comparison bit is through a transfer function. This problem has been addressed previously for some restricted classes of transfer functions, but here we consider a very general transfer function class which includes all functions that admit a series expansion about the origin. Our main contribution is an efficient algorithm with convergence rate of $O(\epsilon^{-4p})$ for smooth convex functions, and an optimal rate of $\widetilde O(\epsilon^{-2p})$ when the objective is both smooth and strongly convex, where $p$ is the minimal degree (with a non-zero coefficient) in the transfer’s series expansion about the origin.

Cite

Text

Saha et al. "Dueling Convex Optimization with General Preferences." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Saha et al. "Dueling Convex Optimization with General Preferences." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/saha2025icml-dueling/)

BibTeX

@inproceedings{saha2025icml-dueling,
  title     = {{Dueling Convex Optimization with General Preferences}},
  author    = {Saha, Aadirupa and Koren, Tomer and Mansour, Yishay},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {52552-52564},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/saha2025icml-dueling/}
}