Interaction Asymmetry: A General Principle for Learning Composable Abstractions

Abstract

Learning disentangled representations of concepts and re-composing them in unseen ways is crucial for generalizing to out-of-domain situations. However, the underlying properties of concepts that enable such disentanglement and compositional generalization remain poorly understood. In this work, we propose the principle of interaction asymmetry which states: "Parts of the same concept have more complex interactions than parts of different concepts". We formalize this via block diagonality conditions on the $(n+1)$th order derivatives of the generator mapping concepts to observed data, where different orders of "complexity" correspond to different $n$. Using this formalism, we prove that interaction asymmetry enables both disentanglement and compositional generalization. Our results unify recent theoretical results for learning concepts of objects, which we show are recovered as special cases with $n=0$ or $1$. We provide results for up to $n=2$, thus extending these prior works to more flexible generator functions, and conjecture that the same proof strategies generalize to larger $n$.

Cite

Text

Brady et al. "Interaction Asymmetry: A General Principle for Learning Composable Abstractions." NeurIPS 2024 Workshops: CRL, 2024.

Markdown

[Brady et al. "Interaction Asymmetry: A General Principle for Learning Composable Abstractions." NeurIPS 2024 Workshops: CRL, 2024.](https://mlanthology.org/neuripsw/2024/brady2024neuripsw-interaction/)

BibTeX

@inproceedings{brady2024neuripsw-interaction,
  title     = {{Interaction Asymmetry: A General Principle for Learning Composable Abstractions}},
  author    = {Brady, Jack and von Kügelgen, Julius and Lachapelle, Sebastien and Buchholz, Simon and Brendel, Wieland},
  booktitle = {NeurIPS 2024 Workshops: CRL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/brady2024neuripsw-interaction/}
}