UnoLoRA: Single Low-Rank Adaptation for Efficient Multitask Fine-Tuning

Abstract

Recent advances in Parameter-Efficient Fine-Tuning (PEFT) have shown Low- Rank Adaptation (LoRA) to be an effective implicit regularizer for large language models. Building on these findings, we propose UnoLoRA, a novel approach that leverages a single shared LoRA module for efficient multi-task learning. While ex- isting methods typically use separate LoRA adaptations for each task, our approach demonstrates that a single shared adapter can effectively capture both task-specific and task-agnostic knowledge. We further introduce UnoLoRA*, an enhanced variant that employs a shared hypernetwork to generate task-specific embeddings, improving convergence and task adaptation. Our method significantly reduces train- able parameters to just 0.05% per task while maintaining competitive performance on the GLUE benchmark. Our analysis reveals that the A and B matrices in our shared LoRA adapter naturally develop complementary roles: A matrices capture generalizable features across tasks, while B matrices specialize in task-specific representations. Our results show that sharing a single LoRA adapter can achieve efficient multi-task learning while significantly reducing memory requirements, making it particularly valuable for resource-constrained applications.

Cite

Text

Kamalesh et al. "UnoLoRA: Single Low-Rank Adaptation for Efficient Multitask Fine-Tuning." NeurIPS 2024 Workshops: FITML, 2024.

Markdown

[Kamalesh et al. "UnoLoRA: Single Low-Rank Adaptation for Efficient Multitask Fine-Tuning." NeurIPS 2024 Workshops: FITML, 2024.](https://mlanthology.org/neuripsw/2024/kamalesh2024neuripsw-unolora/)

BibTeX

@inproceedings{kamalesh2024neuripsw-unolora,
  title     = {{UnoLoRA: Single Low-Rank Adaptation for Efficient Multitask Fine-Tuning}},
  author    = {Kamalesh, Akash and Lakhotia, Anirudh and Nischal, H S and Kulkarni, Prerana Sanjay and Srinivasa, Gowri},
  booktitle = {NeurIPS 2024 Workshops: FITML},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/kamalesh2024neuripsw-unolora/}
}