Interpolating Compressed Parameter Subspaces

Abstract

Though distribution shifts have caused growing concern for machine learning scalability, solutions tend to specialize towards a specific type of distribution shift. We learn that constructing a Compressed Parameter Subspaces (CPS), a geometric structure representing distance-regularized parameters mapped to a set of train-time distributions, can maximize average accuracy over a broad range of distribution shifts concurrently. We show sampling parameters within a CPS can mitigate backdoor, adversarial, permutation, stylization and rotation perturbations. Regularizing a hypernetwork with CPS can also reduce task forgetting.

Cite

Text

Datta and Shadbolt. "Interpolating Compressed Parameter Subspaces." NeurIPS 2022 Workshops: INTERPOLATE, 2022.

Markdown

[Datta and Shadbolt. "Interpolating Compressed Parameter Subspaces." NeurIPS 2022 Workshops: INTERPOLATE, 2022.](https://mlanthology.org/neuripsw/2022/datta2022neuripsw-interpolating/)

BibTeX

@inproceedings{datta2022neuripsw-interpolating,
  title     = {{Interpolating Compressed Parameter Subspaces}},
  author    = {Datta, Siddhartha and Shadbolt, Nigel},
  booktitle = {NeurIPS 2022 Workshops: INTERPOLATE},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/datta2022neuripsw-interpolating/}
}