On-Line Variance Minimization in O(n2) per Trial?

Abstract

Consider the following canonical online learning problem with matrices [WK06]: In each trial t the algorithm chooses a density matrixWt ∈ Rn×n (i.e., a positive semi-definite matrix with trace one). Then nature chooses a symmetric loss matrix Lt ∈ Rn×n whose eigenvalues lie in the interval [0, 1] and the algorithms incurs loss tr(WtLt). The goal is to find algorithms that for any sequence of trials have small regret against the best dyad chosen in hindsight. Here a dyad is an outer product uu # of a unit vector u in Rn. More precisely the regret after T trials is defined as follows: T∑ t=1 tr(WtLt) − L∗, where L ∗ = inf u:‖u‖=1 tr uu#L≤T with L≤T =

Cite

Text

Hazan et al. "On-Line Variance Minimization in O(n2) per Trial?." Annual Conference on Computational Learning Theory, 2010.

Markdown

[Hazan et al. "On-Line Variance Minimization in O(n2) per Trial?." Annual Conference on Computational Learning Theory, 2010.](https://mlanthology.org/colt/2010/hazan2010colt-line/)

BibTeX

@inproceedings{hazan2010colt-line,
  title     = {{On-Line Variance Minimization in O(n2) per Trial?}},
  author    = {Hazan, Elad and Kale, Satyen and Warmuth, Manfred K.},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2010},
  pages     = {314-315},
  url       = {https://mlanthology.org/colt/2010/hazan2010colt-line/}
}