Efficient Convex Optimization Requires Superlinear Memory (Extended Abstract)

Abstract

Minimizing a convex function with access to a first order oracle---that returns the function evaluation and (sub)gradient at a query point---is a canonical optimization problem and a fundamental primitive in machine learning. Gradient-based methods are the most popular approaches used for solving the problem, owing to their simplicity and computational efficiency. These methods, however, do not achieve the information-theoretically optimal query complexity for minimizing the underlying function to small error, which are achieved by more expensive techniques based on cutting-plane methods. Is it possible to achieve the information-theoretically query complexity without using these more complex and computationally expensive methods? In this work, we use memory as a lens to understand this, and show that is is not possible to achieve optimal query complexity without using significantly more memory than that used by gradient descent.

Cite

Text

Marsden et al. "Efficient Convex Optimization Requires Superlinear Memory (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/722

Markdown

[Marsden et al. "Efficient Convex Optimization Requires Superlinear Memory (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/marsden2023ijcai-efficient/) doi:10.24963/IJCAI.2023/722

BibTeX

@inproceedings{marsden2023ijcai-efficient,
  title     = {{Efficient Convex Optimization Requires Superlinear Memory (Extended Abstract)}},
  author    = {Marsden, Annie and Sharan, Vatsal and Sidford, Aaron and Valiant, Gregory},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {6468-6473},
  doi       = {10.24963/IJCAI.2023/722},
  url       = {https://mlanthology.org/ijcai/2023/marsden2023ijcai-efficient/}
}