Descent with Misaligned Gradients and Applications to Hidden Convexity

Abstract

We consider the problem of minimizing a convex objective given access to an oracle that outputs "misaligned" stochastic gradients, where the expected value of the output is guaranteed to be correlated with, but not necessarily equal to the true gradient of the objective. In the case where the misalignment (or bias) of the oracle changes slowly, we obtain an optimization algorithm that achieves the optimum iteration complexity of $\tilde O(\epsilon^{-2})$; for the more general case where the changes need not be slow, we obtain an algorithm with $\tilde O(\epsilon^{-3})$ iteration complexity. As an application of our framework, we consider optimization problems with a "hidden convexity" property, and obtain an algorithm with $O(\epsilon^{-3})$ iteration complexity.

Cite

Text

Bhaskara et al. "Descent with Misaligned Gradients and Applications to Hidden Convexity." International Conference on Learning Representations, 2025.

Markdown

[Bhaskara et al. "Descent with Misaligned Gradients and Applications to Hidden Convexity." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/bhaskara2025iclr-descent/)

BibTeX

@inproceedings{bhaskara2025iclr-descent,
  title     = {{Descent with Misaligned Gradients and Applications to Hidden Convexity}},
  author    = {Bhaskara, Aditya and Cutkosky, Ashok and Kumar, Ravi and Purohit, Manish},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/bhaskara2025iclr-descent/}
}