An Adaptive Tangent Feature Perspective of Neural Networks
Abstract
In order to better understand feature learning in neural networks, we propose and study linear models in tangent feature space where the features are allowed to be transformed during training. We consider linear feature transformations, resulting in a joint optimization over parameters and transformations with a bilinear interpolation constraint. We show that this relaxed optimization problem has an equivalent linearly constrained optimization with structured regularization that encourages approximately low rank solutions. Specializing to structures arising in neural networks, we gain insights into how the features and thus the kernel function change, providing additional nuance to the phenomenon of kernel alignment when the target function is poorly represented by tangent features. We verify our theoretical observations in the kernel alignment of real neural networks.
Cite
Text
LeJeune and Alemohammad. "An Adaptive Tangent Feature Perspective of Neural Networks." Conference on Parsimony and Learning, 2024.Markdown
[LeJeune and Alemohammad. "An Adaptive Tangent Feature Perspective of Neural Networks." Conference on Parsimony and Learning, 2024.](https://mlanthology.org/cpal/2024/lejeune2024cpal-adaptive/)BibTeX
@inproceedings{lejeune2024cpal-adaptive,
title = {{An Adaptive Tangent Feature Perspective of Neural Networks}},
author = {LeJeune, Daniel and Alemohammad, Sina},
booktitle = {Conference on Parsimony and Learning},
year = {2024},
pages = {379-394},
volume = {234},
url = {https://mlanthology.org/cpal/2024/lejeune2024cpal-adaptive/}
}