Extension of the PAC Framework to Finite and Countable Markov Chains
Abstract
We consider a model of learning in which the successive observations follow a certain Markov chain. The observations are labeled according to a membership to some unknown target set. For a Markov chain with finitely many states we show that, if the target set belongs to a family of sets with a finite Vapnik-Chervonenkis (1995) dimension, then probably approximately correct (PAC) learning of this set is possible with polynomially large samples. Specifically for observations following a random walk with a state space /spl Xscr/ and uniform stationary distribution, the sample size required is no more than /spl Omega/(t/sub 0//1-/spl lambda//sub 2/log(t/sub 0/|/spl chi/|1//spl delta/)), where /spl delta/ is the confidence level, /spl lambda//sub 2/ is the second largest eigenvalue of the transition matrix, and t/sub 0/ is the sample size sufficient for learning from independent and identically distributed (i.i.d.) observations. We then obtain similar results for Markov chains with countably many states using Lyapunov function technique and results on mixing properties of infinite state Markov chains.
Cite
Text
Gamarnik. "Extension of the PAC Framework to Finite and Countable Markov Chains." Annual Conference on Computational Learning Theory, 1999. doi:10.1145/307400.307478Markdown
[Gamarnik. "Extension of the PAC Framework to Finite and Countable Markov Chains." Annual Conference on Computational Learning Theory, 1999.](https://mlanthology.org/colt/1999/gamarnik1999colt-extension/) doi:10.1145/307400.307478BibTeX
@inproceedings{gamarnik1999colt-extension,
title = {{Extension of the PAC Framework to Finite and Countable Markov Chains}},
author = {Gamarnik, David},
booktitle = {Annual Conference on Computational Learning Theory},
year = {1999},
pages = {308-317},
doi = {10.1145/307400.307478},
url = {https://mlanthology.org/colt/1999/gamarnik1999colt-extension/}
}