Tandem Inference: An Out-of-Core Streaming Algorithm for Very Large-Scale Relational Inference

Abstract

Statistical relational learning (SRL) frameworks allow users to create large, complex graphical models using a compact, rule-based representation. However, these models can quickly become prohibitively large and not fit into machine memory. In this work we address this issue by introducing a novel technique called tandem inference (ti). The primary idea of ti is to combine grounding and inference such that both processes happen in tandem. ti uses an out-of-core streaming approach to overcome memory limitations. Even when memory is not an issue, we show that our proposed approach is able to do inference faster while using less memory than existing approaches. To show the effectiveness of ti, we use a popular SRL framework called Probabilistic Soft Logic (PSL). We implement ti for PSL by proposing a gradient-based inference engine and a streaming approach to grounding. We show that we are able to run an SRL model with over 1B cliques in under nine hours and using only 10 GB of RAM; previous approaches required more than 800 GB for this model and are infeasible on common hardware. To the best of our knowledge, this is the largest SRL model ever run.

Cite

Text

Srinivasan et al. "Tandem Inference: An Out-of-Core Streaming Algorithm for Very Large-Scale Relational Inference." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I06.6588

Markdown

[Srinivasan et al. "Tandem Inference: An Out-of-Core Streaming Algorithm for Very Large-Scale Relational Inference." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/srinivasan2020aaai-tandem/) doi:10.1609/AAAI.V34I06.6588

BibTeX

@inproceedings{srinivasan2020aaai-tandem,
  title     = {{Tandem Inference: An Out-of-Core Streaming Algorithm for Very Large-Scale Relational Inference}},
  author    = {Srinivasan, Sriram and Augustine, Eriq and Getoor, Lise},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {10259-10266},
  doi       = {10.1609/AAAI.V34I06.6588},
  url       = {https://mlanthology.org/aaai/2020/srinivasan2020aaai-tandem/}
}