Set-Based Neural Network Encoding Without Weight Tying

Abstract

We propose a neural network weight encoding method for network property prediction that utilizes set-to-set and set-to-vector functionsto efficiently encode neural network parameters. Our approach is capable of encoding neural networks in a model zoo of mixed architecture and different parameter sizes as opposed to previous approaches that require custom encoding models for different architectures. Furthermore, our \textbf{S}et-based \textbf{N}eural network \textbf{E}ncoder (SNE) takes into consideration the hierarchical computational structure of neural networks. To respect symmetries inherent in network weight space, we utilize Logit Invariance to learn the required minimal invariance properties. Additionally, we introduce a \textit{pad-chunk-encode} pipeline to efficiently encode neural network layers that is adjustable to computational and memory constraints. We also introduce two new tasks for neural network property prediction: cross-dataset and cross-architecture. In cross-dataset property prediction, we evaluate how well property predictors generalize across model zoos trained on different datasets but of the same architecture. In cross-architecture property prediction, we evaluate how well property predictors transfer to model zoos of different architecture not seen during training. We show that SNE outperforms the relevant baselines on standard benchmarks.

Cite

Text

Andreis et al. "Set-Based Neural Network Encoding Without Weight Tying." Neural Information Processing Systems, 2024. doi:10.52202/079017-2436

Markdown

[Andreis et al. "Set-Based Neural Network Encoding Without Weight Tying." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/andreis2024neurips-setbased/) doi:10.52202/079017-2436

BibTeX

@inproceedings{andreis2024neurips-setbased,
  title     = {{Set-Based Neural Network Encoding Without Weight Tying}},
  author    = {Andreis, Bruno and Bedionita, Soro and Torr, Philip H.S. and Hwang, Sung Ju},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-2436},
  url       = {https://mlanthology.org/neurips/2024/andreis2024neurips-setbased/}
}