Controlling the Inductive Bias of Wide Neural Networks by Modifying the Kernel’s Spectrum

Abstract

Wide neural networks are biased towards learning certain functions, influencing both the rate of convergence of gradient descent (GD) and the functions that are reachable with GD in finite training time. As such, there is a great need for methods that can modify this bias according to the task at hand. To that end, we introduce Modified Spectrum Kernels (MSKs), a novel family of constructed kernels that can be used to approximate kernels with desired eigenvalues for which no closed form is known. We leverage the duality between wide neural networks and Neural Tangent Kernels and propose a preconditioned gradient descent method, which alters the trajectory of GD. As a result, this allows for a polynomial and, in some cases, exponential training speedup without changing the final solution. Our method is both computationally efficient and simple to implement.

Cite

Text

Geifman et al. "Controlling the Inductive Bias of Wide Neural Networks by Modifying the Kernel’s Spectrum." Transactions on Machine Learning Research, 2024.

Markdown

[Geifman et al. "Controlling the Inductive Bias of Wide Neural Networks by Modifying the Kernel’s Spectrum." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/geifman2024tmlr-controlling/)

BibTeX

@article{geifman2024tmlr-controlling,
  title     = {{Controlling the Inductive Bias of Wide Neural Networks by Modifying the Kernel’s Spectrum}},
  author    = {Geifman, Amnon and Barzilai, Daniel and Basri, Ronen and Galun, Meirav},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/geifman2024tmlr-controlling/}
}