Rethinking Fine-Tuning Through Geometric Perspective
Abstract
Fine-tuning pre-trained neural networks has become a cornerstone of transfer learning. However, the practical success of existing methods like low-rank adaptation (LoRA) lacks theoretical explanation. We introduce geometry-guided fine-tuning, a novel paradigm that models the fine-tuning process as the subtle movement of pre-trained weights on a low-dimensional manifold. Our approach formalizes this process through a learnable ordinary differential equation (ODE) - based framework that controls the search space of the weights, bridging existing methods with geometric principles. We empirically evaluate our method in the context of multi-task learning (MTL) fine-tuning of hierarchical vision transformers in computer vision. We propose a parameter-efficient ODE and evaluate it on the PASCAL-Context MTL benchmark. Our approach, dubbed DeLoRAoffers competitive performance across multiple dense prediction tasks, reducing trainable parameters by up to 4$\times$ compared to the best-performing baseline. This work advances both the theoretical understanding and practical application of fine-tuning, promoting efficient learning in resource-constrained environments.
Cite
Text
Mantri et al. "Rethinking Fine-Tuning Through Geometric Perspective." NeurIPS 2024 Workshops: UniReps, 2024.Markdown
[Mantri et al. "Rethinking Fine-Tuning Through Geometric Perspective." NeurIPS 2024 Workshops: UniReps, 2024.](https://mlanthology.org/neuripsw/2024/mantri2024neuripsw-rethinking/)BibTeX
@inproceedings{mantri2024neuripsw-rethinking,
title = {{Rethinking Fine-Tuning Through Geometric Perspective}},
author = {Mantri, Krishna Sri Ipsit and Eliasof, Moshe and Schönlieb, Carola-Bibiane and Ribeiro, Bruno},
booktitle = {NeurIPS 2024 Workshops: UniReps},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/mantri2024neuripsw-rethinking/}
}