Tensor Networks for Probabilistic Sequence Modeling

Abstract

Tensor networks are a powerful modeling framework developed for computational many-body physics, which have only recently been applied within machine learning. In this work we utilize a uniform matrix product state (u-MPS) model for probabilistic modeling of sequence data. We first show that u-MPS enable sequence-level parallelism, with length-n sequences able to be evaluated in depth O(log n). We then introduce a novel generative algorithm giving trained u-MPS the ability to efficiently sample from a wide variety of conditional distributions, each one defined by a regular expression. Special cases of this algorithm correspond to autoregressive and fill-in-the-blank sampling, but more complex regular expressions permit the generation of richly structured data in a manner that has no direct analogue in neural generative models. Experiments on sequence modeling with synthetic and real text data show u-MPS outperforming a variety of baselines and effectively generalizing their predictions in the presence of limited data.

Cite

Text

Miller et al. "Tensor Networks for Probabilistic Sequence Modeling." Artificial Intelligence and Statistics, 2021.

Markdown

[Miller et al. "Tensor Networks for Probabilistic Sequence Modeling." Artificial Intelligence and Statistics, 2021.](https://mlanthology.org/aistats/2021/miller2021aistats-tensor/)

BibTeX

@inproceedings{miller2021aistats-tensor,
  title     = {{Tensor Networks for Probabilistic Sequence Modeling}},
  author    = {Miller, Jacob and Rabusseau, Guillaume and Terilla, John},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2021},
  pages     = {3079-3087},
  volume    = {130},
  url       = {https://mlanthology.org/aistats/2021/miller2021aistats-tensor/}
}