A Comprehensive Analysis of Adapter Efficiency

Abstract

Adapters have been positioned as a parameter-efficient fine-tuning (PEFT) approach. However, adapters have not been sufficiently analyzed to understand if PEFT translates to benefits in training/deployment efficiency and maintainability/extensibility. Through extensive experiments on many adapters, tasks, and languages in supervised and cross-lingual zero-shot settings, we clearly show that for Natural Language Understanding (NLU) tasks, the parameter efficiency in adapters does not translate to efficiency gains compared to full fine-tuning of models. More precisely, adapters are relatively expensive to train and have slightly higher deployment latency. Furthermore, the maintainability/extensibility benefits of adapters can be achieved with simpler approaches like multi-task training via full fine-tuning, which also provide relatively faster training times. We, therefore, recommend that for moderately sized models for NLU tasks, practitioners should rely on full fine-tuning or multi-task training rather than using adapters. Our code is available at https://github.com/AI4Bharat/adapter-efficiency.

Cite

Text

Mundra et al. "A Comprehensive Analysis of Adapter Efficiency." ICML 2023 Workshops: ES-FoMO, 2023.

Markdown

[Mundra et al. "A Comprehensive Analysis of Adapter Efficiency." ICML 2023 Workshops: ES-FoMO, 2023.](https://mlanthology.org/icmlw/2023/mundra2023icmlw-comprehensive/)

BibTeX

@inproceedings{mundra2023icmlw-comprehensive,
  title     = {{A Comprehensive Analysis of Adapter Efficiency}},
  author    = {Mundra, Nandini and Doddapaneni, Sumanth and Dabre, Raj and Kunchukuttan, Anoop and Puduppully, Ratish and Khapra, Mitesh M},
  booktitle = {ICML 2023 Workshops: ES-FoMO},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/mundra2023icmlw-comprehensive/}
}