A General Framework for Equivariant Neural Networks on Reductive Lie Groups

Abstract

Reductive Lie Groups, such as the orthogonal groups, the Lorentz group, or the unitary groups, play essential roles across scientific fields as diverse as high energy physics, quantum mechanics, quantum chromodynamics, molecular dynamics, computer vision, and imaging. In this paper, we present a general Equivariant Neural Network architecture capable of respecting the symmetries of the finite-dimensional representations of any reductive Lie Group. Our approach generalizes the successful ACE and MACE architectures for atomistic point clouds to any data equivariant to a reductive Lie group action. We also introduce the lie-nn software library, which provides all the necessary tools to develop and implement such general G-equivariant neural networks. It implements routines for the reduction of generic tensor products of representations into irreducible representations, making it easy to apply our architecture to a wide range of problems and groups. The generality and performance of our approach are demonstrated by applying it to the tasks of top quark decay tagging (Lorentz group) and shape recognition (orthogonal group).

Cite

Text

Batatia et al. "A General Framework for Equivariant Neural Networks on Reductive Lie Groups." Neural Information Processing Systems, 2023.

Markdown

[Batatia et al. "A General Framework for Equivariant Neural Networks on Reductive Lie Groups." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/batatia2023neurips-general/)

BibTeX

@inproceedings{batatia2023neurips-general,
  title     = {{A General Framework for Equivariant Neural Networks on Reductive Lie Groups}},
  author    = {Batatia, Ilyes and Geiger, Mario and Munoz, Jose and Smidt, Tess and Silberman, Lior and Ortner, Christoph},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/batatia2023neurips-general/}
}