Positional Encodings as Group Representations: A Unified Framework
Abstract
Positional encodings are ubiquitous as an input featurization tool in language modeling, computer vision, and graph representation learning, enabling neural networks to capture important geometric structure of the input. Traditionally, positional encodings have been defined anew for each data domain. In this work, we reinterpret positional encodings for disparate data types --- including sequences, grids, graphs, and manifolds --- in the unifying framework of group representations. We show how to express existing positional encodings as group representations, and conversely, propose new positional encodings by choosing suitable groups and representations. We validate our framework with experiments on implicit neural representations of images and vector fields, highlighting the practical utility of such positional encodings for encouraging approximate equivariance and capturing geometric structure.
Cite
Text
Lim et al. "Positional Encodings as Group Representations: A Unified Framework." ICML 2023 Workshops: TAGML, 2023.Markdown
[Lim et al. "Positional Encodings as Group Representations: A Unified Framework." ICML 2023 Workshops: TAGML, 2023.](https://mlanthology.org/icmlw/2023/lim2023icmlw-positional/)BibTeX
@inproceedings{lim2023icmlw-positional,
title = {{Positional Encodings as Group Representations: A Unified Framework}},
author = {Lim, Derek and Lawrence, Hannah and Huang, Ningyuan Teresa and Thiede, Erik Henning},
booktitle = {ICML 2023 Workshops: TAGML},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/lim2023icmlw-positional/}
}