The Impact of Model Zoo Size and Composition on Weight Space Learning

Abstract

Re-using trained neural network models is a common strategy to reduce training cost and transfer knowledge. Weight space learning - using the weights of trained models as data modality - is a promising new field to re-use populations of pre-trained models for future tasks. Approaches in this field have demonstrated high performance both on model analysis and weight generation tasks. However, until now their learning setup requires homogeneous model zoos where all models share the same exact architecture, limiting their capability to generalize beyond the population of models they saw during training. In this work, we remove this constraint and propose a modification to a common weight space learning method to accommodate training on heterogeneous populations of models. We further investigate the resulting impact of model diversity on generating unseen neural network model weights for zero-shot knowledge transfer. Our extensive experimental evaluation shows that including models with varying underlying image datasets has a high impact on performance and generalization, for both in- and out-of-distribution settings. Code is available on github.com/HSG-AIML/MultiZoo-SANE.

Cite

Text

Falk et al. "The Impact of Model Zoo Size and Composition on Weight Space Learning." ICLR 2025 Workshops: WSL, 2025. doi:10.48550/arxiv.2504.10141

Markdown

[Falk et al. "The Impact of Model Zoo Size and Composition on Weight Space Learning." ICLR 2025 Workshops: WSL, 2025.](https://mlanthology.org/iclrw/2025/falk2025iclrw-impact/) doi:10.48550/arxiv.2504.10141

BibTeX

@inproceedings{falk2025iclrw-impact,
  title     = {{The Impact of Model Zoo Size and Composition on Weight Space Learning}},
  author    = {Falk, Damian and Schürholt, Konstantin and Borth, Damian},
  booktitle = {ICLR 2025 Workshops: WSL},
  year      = {2025},
  doi       = {10.48550/arxiv.2504.10141},
  url       = {https://mlanthology.org/iclrw/2025/falk2025iclrw-impact/}
}