MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model
Abstract
The emergence of foundation models, such as the Segment Anything Model (SAM), has sparked interest in Parameter-Efficient Fine-Tuning (PEFT) methods that tailor these large models to specific application domains. Various PEFT techniques modify the representation of a model differently, making it a non-trivial task to select the most appropriate method for the domain of interest. We propose a new framework, Mixture-of-PEFT methods (MoPEFT), that is inspired by traditional Mixture-of-Experts (MoE) methodologies and is utilized for fine-tuning SAM. Our MoPEFT framework incorporates three different PEFT techniques as submodules and dynamically learns to activate the ones that are best suited for a given data-task setup. We use MoPEFT to fine-tune the Segment Anything Model and show that MoPEFT consistently outperforms other fine-tuning methods on the MESS benchmark.
Cite
Text
Sahay and Savakis. "MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.Markdown
[Sahay and Savakis. "MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.](https://mlanthology.org/cvprw/2025/sahay2025cvprw-mopeft/)BibTeX
@inproceedings{sahay2025cvprw-mopeft,
title = {{MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model}},
author = {Sahay, Rajat and Savakis, Andreas E.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2025},
pages = {6500-6510},
url = {https://mlanthology.org/cvprw/2025/sahay2025cvprw-mopeft/}
}