The Space Complexity of Approximating Logistic Loss

Abstract

We provide space complexity lower bounds for data structures that approximate logistic loss up to $\epsilon$-relative error on a logistic regression problem with data $\mathbf{X} \in \mathbb{R}^{n \times d}$ and labels $\mathbf{y} \in \\{-1,1\\}^d$. The space complexity of existing coreset constructions depend on a natural complexity measure $\mu_\mathbf{y}(\mathbf{X})$. We give an $\tilde{\Omega}(\frac{d}{\epsilon^2})$ space complexity lower bound in the regime $\mu_\mathbf{y}(\mathbf{X}) = \mathcal{O}(1)$ that shows existing coresets are optimal in this regime up to lower order factors. We also prove a general $\tilde{\Omega}(d\cdot \mu_\mathbf{y}(\mathbf{X}))$ space lower bound when $\epsilon$ is constant, showing that the dependency on $\mu_\mathbf{y}(\mathbf{X})$ is not an artifact of mergeable coresets. Finally, we refute a prior conjecture that $\mu_\mathbf{y}(\mathbf{X})$ is hard to compute by providing an efficient linear programming formulation, and we empirically compare our algorithm to prior approximate methods.

Cite

Text

Dexter et al. "The Space Complexity of Approximating Logistic Loss." Neural Information Processing Systems, 2024. doi:10.52202/079017-2888

Markdown

[Dexter et al. "The Space Complexity of Approximating Logistic Loss." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/dexter2024neurips-space/) doi:10.52202/079017-2888

BibTeX

@inproceedings{dexter2024neurips-space,
  title     = {{The Space Complexity of Approximating Logistic Loss}},
  author    = {Dexter, Gregory and Drineas, Petros and Khanna, Rajiv},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-2888},
  url       = {https://mlanthology.org/neurips/2024/dexter2024neurips-space/}
}