Flexora: Flexible Low-Rank Adaptation for Large Language Models

Abstract

Large language models (LLMs) have revolutionized artificial intelligence, but their performance on specific tasks is often limited by knowledge boundaries. While fine-tuning techniques like low-rank adaptation (LoRA) aim to address this, they can suffer from overfitting. We propose flexible low-rank adaptation (Flexora), a novel method that automatically selects the most critical layers for fine-tuning to optimize performance across diverse downstream tasks. Flexora formulates layer selection as a hyperparameter optimization problem, employs unrolled differentiation for efficient solving, and identifies the most impactful layers based on optimized hyperparameters. Extensive experiments across various pre-trained models and natural language tasks demonstrate that Flexora consistently outperforms existing baselines. We provide theoretical insights and comprehensive ablation studies to elucidate the effectiveness of Flexora. Therefore, Flexora offers a robust solution to enhance LoRA fine-tuning for LLMs, potentially advancing the field of adaptive language model optimization.

Cite

Text

Wei et al. "Flexora: Flexible Low-Rank Adaptation for Large Language Models." NeurIPS 2024 Workshops: FITML, 2024.

Markdown

[Wei et al. "Flexora: Flexible Low-Rank Adaptation for Large Language Models." NeurIPS 2024 Workshops: FITML, 2024.](https://mlanthology.org/neuripsw/2024/wei2024neuripsw-flexora/)

BibTeX

@inproceedings{wei2024neuripsw-flexora,
  title     = {{Flexora: Flexible Low-Rank Adaptation for Large Language Models}},
  author    = {Wei, Chenxing and Shu, Yao and He, Ying Tiffany and Yu, Fei},
  booktitle = {NeurIPS 2024 Workshops: FITML},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/wei2024neuripsw-flexora/}
}