On Scale-Invariant Sharpness Measures

Abstract

Recently, there has been a substantial surge of interest in the development of optimization algorithms tailored for overparameterized models. This interest centers around the objective of minimizing a concept of sharpness in conjunction with the original loss function, e.g., the Sharpness-Aware Minimization (SAM) algorithm shown effective in practice. Nevertheless, the majority of sharpness measures exhibit sensitivity to parameter scaling in neural networks, and they may even experience significant magnification when subjected to rescaling operations. Motivated by this issue, in this paper, we introduce a new class of scale-invariant sharpness measures, that gives rise to a new class of scale-invariant sharpness-aware objective functions. Furthermore, we prove that the newly introduced objective functions are explicitly biased towards the minimization of our scale-invariant sharpness measures.

Cite

Text

Tahmasebi et al. "On Scale-Invariant Sharpness Measures." NeurIPS 2023 Workshops: M3L, 2023.

Markdown

[Tahmasebi et al. "On Scale-Invariant Sharpness Measures." NeurIPS 2023 Workshops: M3L, 2023.](https://mlanthology.org/neuripsw/2023/tahmasebi2023neuripsw-scaleinvariant/)

BibTeX

@inproceedings{tahmasebi2023neuripsw-scaleinvariant,
  title     = {{On Scale-Invariant Sharpness Measures}},
  author    = {Tahmasebi, Behrooz and Soleymani, Ashkan and Jegelka, Stefanie and Jaillet, Patrick},
  booktitle = {NeurIPS 2023 Workshops: M3L},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/tahmasebi2023neuripsw-scaleinvariant/}
}