Compositional Subspace Representation Fine-Tuning for Adaptive Large Language Models
Abstract
Adapting large language models to multiple tasks can cause cross-skill interference, where improvements for one skill degrade another. While methods such as LoRA impose orthogonality constraints at the weight level, they do not fully address interference in hidden-state representations. We propose Compositional Subspace Representation Fine-tuning (CS-ReFT), a novel representation-based approach that learns multiple orthonormal subspace transformations, each specializing in a distinct skill, and composes them via a lightweight router. By isolating these subspace edits in the hidden state, rather than weight matrices, CS-ReFT prevents cross-task conflicts more effectively. On the AlpacaEval benchmark, applying CS-ReFT to Llama-2-7B achieves a 93.94% win rate, surpassing GPT-3.5 Turbo (86.30%) while requiring only 0.0098% of model parameters. These findings show that specialized representation edits, composed via a simple router, significantly enhance multi-task instruction following with minimal overhead.
Cite
Text
Zhou and Arel. "Compositional Subspace Representation Fine-Tuning for Adaptive Large Language Models." ICLR 2025 Workshops: SCOPE, 2025.Markdown
[Zhou and Arel. "Compositional Subspace Representation Fine-Tuning for Adaptive Large Language Models." ICLR 2025 Workshops: SCOPE, 2025.](https://mlanthology.org/iclrw/2025/zhou2025iclrw-compositional/)BibTeX
@inproceedings{zhou2025iclrw-compositional,
title = {{Compositional Subspace Representation Fine-Tuning for Adaptive Large Language Models}},
author = {Zhou, Andy and Arel, Ron},
booktitle = {ICLR 2025 Workshops: SCOPE},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/zhou2025iclrw-compositional/}
}