MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model

Abstract

The emergence of foundation models, such as the Segment Anything Model (SAM), has sparked interest in Parameter-Efficient Fine-Tuning (PEFT) methods that tailor these large models to specific application domains. Various PEFT techniques modify the representation of a model differently, making it a non-trivial task to select the most appropriate method for the domain of interest. We propose a new framework, Mixture-of-PEFT methods (MoPEFT), that is inspired by traditional Mixture-of-Experts (MoE) methodologies and is utilized for fine-tuning SAM. Our MoPEFT framework incorporates three different PEFT techniques as submodules and dynamically learns to activate the ones that are best suited for a given data-task setup. We use MoPEFT to fine-tune the Segment Anything Model and show that MoPEFT consistently outperforms other fine-tuning methods on the MESS benchmark.

Cite

Text

Sahay and Savakis. "MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.

Markdown

[Sahay and Savakis. "MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.](https://mlanthology.org/cvprw/2025/sahay2025cvprw-mopeft/)

BibTeX

@inproceedings{sahay2025cvprw-mopeft,
  title     = {{MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model}},
  author    = {Sahay, Rajat and Savakis, Andreas E.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2025},
  pages     = {6500-6510},
  url       = {https://mlanthology.org/cvprw/2025/sahay2025cvprw-mopeft/}
}