On-Line Variance Minimization in O(n2) per Trial?
Abstract
Consider the following canonical online learning problem with matrices [WK06]: In each trial t the algorithm chooses a density matrixWt ∈ Rn×n (i.e., a positive semi-definite matrix with trace one). Then nature chooses a symmetric loss matrix Lt ∈ Rn×n whose eigenvalues lie in the interval [0, 1] and the algorithms incurs loss tr(WtLt). The goal is to find algorithms that for any sequence of trials have small regret against the best dyad chosen in hindsight. Here a dyad is an outer product uu # of a unit vector u in Rn. More precisely the regret after T trials is defined as follows: T∑ t=1 tr(WtLt) − L∗, where L ∗ = inf u:‖u‖=1 tr uu#L≤T with L≤T =
Cite
Text
Hazan et al. "On-Line Variance Minimization in O(n2) per Trial?." Annual Conference on Computational Learning Theory, 2010.Markdown
[Hazan et al. "On-Line Variance Minimization in O(n2) per Trial?." Annual Conference on Computational Learning Theory, 2010.](https://mlanthology.org/colt/2010/hazan2010colt-line/)BibTeX
@inproceedings{hazan2010colt-line,
title = {{On-Line Variance Minimization in O(n2) per Trial?}},
author = {Hazan, Elad and Kale, Satyen and Warmuth, Manfred K.},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2010},
pages = {314-315},
url = {https://mlanthology.org/colt/2010/hazan2010colt-line/}
}