Learning Linearly Separable Languages

Abstract

This paper presents a novel paradigm for learning languages that consists of mapping strings to an appropriate high-dimensional feature space and learning a separating hyperplane in that space. It initiates the study of the linear separability of automata and languages by examining the rich class of piecewise-testable languages. It introduces a high-dimensional feature map and proves piecewise-testable languages to be linearly separable in that space. The proof makes use of word combinatorial results relating to subsequences. It also shows that the positive definite kernel associated to this embedding can be computed in quadratic time. It examines the use of support vector machines in combination with this kernel to determine a separating hyperplane and the corresponding learning guarantees. It also proves that all languages linearly separable under a regular finite cover embedding, a generalization of the embedding we used, are regular.

Cite

Text

Kontorovich et al. "Learning Linearly Separable Languages." International Conference on Algorithmic Learning Theory, 2006. doi:10.1007/11894841_24

Markdown

[Kontorovich et al. "Learning Linearly Separable Languages." International Conference on Algorithmic Learning Theory, 2006.](https://mlanthology.org/alt/2006/kontorovich2006alt-learning/) doi:10.1007/11894841_24

BibTeX

@inproceedings{kontorovich2006alt-learning,
  title     = {{Learning Linearly Separable Languages}},
  author    = {Kontorovich, Leonid and Cortes, Corinna and Mohri, Mehryar},
  booktitle = {International Conference on Algorithmic Learning Theory},
  year      = {2006},
  pages     = {288-303},
  doi       = {10.1007/11894841_24},
  url       = {https://mlanthology.org/alt/2006/kontorovich2006alt-learning/}
}