Sparse Network Initialization Using Deterministic Ramanujan Graphs

Abstract

We introduce a sparsely connected neural network architecture inspired by Ramanujan graphs, which achieves performance comparable to dense networks. They are constructed from Cayley graphs of specific algebraic groups or as Ramanujan $r$-coverings of the full $(k,l)$ bi-regular bipartite graph with $k + l$ vertices. This novel method employs zero-shot, data-independent, deterministic pruning at initialization, facilitating early identification of winning lottery tickets. Unlike traditional methods that rely on iterative processes to find these tickets, our technique identifies them at the outset. Our ultimate goal is to construct sparse, scalable Foundation Models. Experimental results demonstrate that our proposed architecture achieves competitive accuracy and sparsity ratios comparable to those obtained by previous pre-training pruning algorithms.

Cite

Text

Biswas et al. "Sparse Network Initialization Using Deterministic Ramanujan Graphs." ICML 2024 Workshops: TF2M, 2024.

Markdown

[Biswas et al. "Sparse Network Initialization Using Deterministic Ramanujan Graphs." ICML 2024 Workshops: TF2M, 2024.](https://mlanthology.org/icmlw/2024/biswas2024icmlw-sparse/)

BibTeX

@inproceedings{biswas2024icmlw-sparse,
  title     = {{Sparse Network Initialization Using Deterministic Ramanujan Graphs}},
  author    = {Biswas, Arindam and Kalra, Suryam Arnav and Mitra, Pabitra and Basu, Biswajit},
  booktitle = {ICML 2024 Workshops: TF2M},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/biswas2024icmlw-sparse/}
}