Unlocking the Global Synergies in Low-Rank Adapters

Abstract

Low-rank Adaption (LoRA) has been the de-facto parameter-efficient fine-tuning technique for large language models. We present HeteroLoRA, a lightweight search algorithm that leverages zero-cost proxies to allocate the limited LoRA trainable parameters across the model for better fine-tuned performance. In addition to the allocation for the standard LoRA-adapted models, we also demonstrate the efficacy of HeteroLoRA by performing the allocation in a more challenging search space that includes LoRA modules and LoRA-adapted shortcut connections. Experiments show that HeteroLoRA enables improvements in model performance given the same parameter budget. For example, on MRPC, we see an improvement of 1.6% in accuracy with similar training parameter budget. We have open-sourced our algorithm.

Cite

Text

Zhang et al. "Unlocking the Global Synergies in Low-Rank Adapters." ICML 2024 Workshops: ES-FoMo-II, 2024.

Markdown

[Zhang et al. "Unlocking the Global Synergies in Low-Rank Adapters." ICML 2024 Workshops: ES-FoMo-II, 2024.](https://mlanthology.org/icmlw/2024/zhang2024icmlw-unlocking/)

BibTeX

@inproceedings{zhang2024icmlw-unlocking,
  title     = {{Unlocking the Global Synergies in Low-Rank Adapters}},
  author    = {Zhang, Zixi and Zhang, Cheng and Gao, Xitong and Mullins, Robert D. and Constantinides, George Anthony and Zhao, Yiren},
  booktitle = {ICML 2024 Workshops: ES-FoMo-II},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/zhang2024icmlw-unlocking/}
}