General Loss Bounds for Universal Sequence Prediction

Abstract

The Bayesian framework is ideally suited for induction problems. The probability of observing $x_t$ at time $t$, given past observations $x_1...x_{t-1}$ can be computed with Bayes'' rule if the true distribution $\mu$ of the sequences $x_1x_2x_3...$ is known. The problem, however, is that in many cases one does not even have a reasonable estimate of the true distribution. In order to overcome this problem a universal distribution $\xi$ is defined as a weighted sum of distributions $\mu_i\!\in\!M$, where $M$ is any countable set of distributions including $\mu$. This is a generalization of Solomonoff induction, in which $M$ is the set of all enumerable semi-measures. Systems which predict $y_t$, given $x_1...x_{t-1}$ and which receive loss $l_{x_t y_t}$ if $x_t$ is the true next symbol of the sequence are considered. It is proven that using the universal $\xi$ as a prior is nearly as good as using the unknown true distribution $\mu$. Furthermore, games of chance, defined as a sequence of bets, observations, and rewards are studied. The time needed to reach the winning zone is bounded in terms of the relative entropy of $\mu$ and $\xi$. Extensions to arbitrary alphabets, partial and delayed prediction, and more active systems are discussed.

Cite

Text

Hutter. "General Loss Bounds for Universal Sequence Prediction." International Conference on Machine Learning, 2001.

Markdown

[Hutter. "General Loss Bounds for Universal Sequence Prediction." International Conference on Machine Learning, 2001.](https://mlanthology.org/icml/2001/hutter2001icml-general/)

BibTeX

@inproceedings{hutter2001icml-general,
  title     = {{General Loss Bounds for Universal Sequence Prediction}},
  author    = {Hutter, Marcus},
  booktitle = {International Conference on Machine Learning},
  year      = {2001},
  pages     = {210-217},
  url       = {https://mlanthology.org/icml/2001/hutter2001icml-general/}
}