XoRA: Expander Adapted LoRA Finetuning
Abstract
Parameter-efficient fine-tuning aims to reduce the computational cost of adapting foundational models to downstream tasks. Low-rank matrix based adaptation (LoRA) techniques are popular for this purpose. We propose XoRA, an efficient fine-tuning scheme, which sparsifies the low-rank matrices even further using expander masks. The mask is generated using extremal expander graphs (Ramanujan graphs) to maintain high edge connectivity even at a very high sparsity. Experimental results demonstrate that this method has comparable performance with the LoRA fine-tuning method while retaining much fewer number of parameters.
Cite
Text
Ev et al. "XoRA: Expander Adapted LoRA Finetuning." NeurIPS 2024 Workshops: FITML, 2024.Markdown
[Ev et al. "XoRA: Expander Adapted LoRA Finetuning." NeurIPS 2024 Workshops: FITML, 2024.](https://mlanthology.org/neuripsw/2024/ev2024neuripsw-xora/)BibTeX
@inproceedings{ev2024neuripsw-xora,
title = {{XoRA: Expander Adapted LoRA Finetuning}},
author = {Ev, Amaljith and Biswas, Arindam and Kalra, Suryam Arnav and Mitra, Pabitra and Basu, Biswajit},
booktitle = {NeurIPS 2024 Workshops: FITML},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/ev2024neuripsw-xora/}
}