Learning Graphical State Transitions

Abstract

Graph-structured data is important in modeling relationships between multiple entities, and can be used to represent states of the world as well as many data structures. Li et al. (2016) describe a model known as a Gated Graph Sequence Neural Network (GGS-NN) that produces sequences from graph-structured input. In this work I introduce the Gated Graph Transformer Neural Network (GGT-NN), an extension of GGS-NNs that uses graph-structured data as an intermediate representation. The model can learn to construct and modify graphs in sophisticated ways based on textual input, and also to use the graphs to produce a variety of outputs. For example, the model successfully learns to solve almost all of the bAbI tasks (Weston et al., 2016), and also discovers the rules governing graphical formulations of a simple cellular automaton and a family of Turing machines.

Cite

Text

Johnson. "Learning Graphical State Transitions." International Conference on Learning Representations, 2017.

Markdown

[Johnson. "Learning Graphical State Transitions." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/johnson2017iclr-learning/)

BibTeX

@inproceedings{johnson2017iclr-learning,
  title     = {{Learning Graphical State Transitions}},
  author    = {Johnson, Daniel D.},
  booktitle = {International Conference on Learning Representations},
  year      = {2017},
  url       = {https://mlanthology.org/iclr/2017/johnson2017iclr-learning/}
}