A New Neural Kernel Regime: The Inductive Bias of Multi-Task Learning
Abstract
This paper studies the properties of solutions to multi-task shallow ReLU neural network learning problems, wherein the network is trained to fit a dataset with minimal sum of squared weights. Remarkably, the solutions learned for each individual task resemble those obtained by solving a kernel regression problem, revealing a novel connection between neural networks and kernel methods. It is known that single-task neural network learning problems are equivalent to a minimum norm interpolation problem in a non-Hilbertian Banach space, and that the solutions of such problems are generally non-unique. In contrast, we prove that the solutions to univariate-input, multi-task neural network interpolation problems are almost always unique, and coincide with the solution to a minimum-norm interpolation problem in a Sobolev (Reproducing Kernel) Hilbert Space. We also demonstrate a similar phenomenon in the multivariate-input case; specifically, we show that neural network learning problems with large numbers of tasks are approximately equivalent to an $\ell^2$ (Hilbert space) minimization problem over a fixed kernel determined by the optimal neurons.
Cite
Text
Nakhleh et al. "A New Neural Kernel Regime: The Inductive Bias of Multi-Task Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-4458Markdown
[Nakhleh et al. "A New Neural Kernel Regime: The Inductive Bias of Multi-Task Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/nakhleh2024neurips-new/) doi:10.52202/079017-4458BibTeX
@inproceedings{nakhleh2024neurips-new,
title = {{A New Neural Kernel Regime: The Inductive Bias of Multi-Task Learning}},
author = {Nakhleh, Julia and Shenouda, Joseph and Nowak, Robert D.},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-4458},
url = {https://mlanthology.org/neurips/2024/nakhleh2024neurips-new/}
}