A Coding-Theoretic Analysis of Hyperspherical Prototypical Learning Geometry

Abstract

Hyperspherical Prototypical Learning (HPL) is a supervised approach to representation learning that designs class prototypes on the unit hypersphere. The prototypes bias the representations to class separation in a scale invariant and known geometry. Previous approaches to HPL have either of the following shortcomings: (i) they follow an unprincipled optimisation procedure; or (ii) they are theoretically sound, but are constrained to only one possible latent dimension. In this paper, we address both shortcomings. To address (i), we present a principled optimisation procedure whose solution we show is optimal. To address (ii), we construct well-separated prototypes in a wide range of dimensions using linear block codes. Additionally, we give a full characterisation of the optimal prototype placement in terms of achievable and converse bounds, showing that our proposed methods are near-optimal.

Cite

Text

Lindström et al. "A Coding-Theoretic Analysis of Hyperspherical Prototypical Learning Geometry." ICML 2024 Workshops: GRaM, 2024.

Markdown

[Lindström et al. "A Coding-Theoretic Analysis of Hyperspherical Prototypical Learning Geometry." ICML 2024 Workshops: GRaM, 2024.](https://mlanthology.org/icmlw/2024/lindstrom2024icmlw-codingtheoretic/)

BibTeX

@inproceedings{lindstrom2024icmlw-codingtheoretic,
  title     = {{A Coding-Theoretic Analysis of Hyperspherical Prototypical Learning Geometry}},
  author    = {Lindström, Martin and Gálvez, Borja Rodríguez and Thobaben, Ragnar and Skoglund, Mikael},
  booktitle = {ICML 2024 Workshops: GRaM},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/lindstrom2024icmlw-codingtheoretic/}
}