LaX: Boosting Low-Rank Training of Foundation Models via Latent Crossing
Abstract
Training foundation models such as ViTs and LLMs requires tremendous computing cost. Low-rank matrix or tensor factorization offers a parameter-efficient alternative, but often downgrades performance due to the restricted parameter space. In this work, we introduce ${\textbf{Latent Crossing (LaX)}}$ -- a simple yet effective plug-and-play module that enhances the capacity of low-rank models by enabling information flow across low-rank subspaces. We extensively validate the benefits of LaX on pre-training tasks with ViT-Base/Large and LLaMA-like models ranging from 60M to 1B parameters. LaX boosts low-rank model performance to match or exceed the full-rank baselines while using 2-3$\times$ fewer parameters. When equipped with low-rank adapters (i.e., LoRA) for fine-tuning LLaMA-7/13B, LaX consistently improves performance on arithmetic and common sense reasoning tasks with negligible cost.
Cite
Text
Zhang et al. "LaX: Boosting Low-Rank Training of Foundation Models via Latent Crossing." Advances in Neural Information Processing Systems, 2025.Markdown
[Zhang et al. "LaX: Boosting Low-Rank Training of Foundation Models via Latent Crossing." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhang2025neurips-lax/)BibTeX
@inproceedings{zhang2025neurips-lax,
title = {{LaX: Boosting Low-Rank Training of Foundation Models via Latent Crossing}},
author = {Zhang, Ruijie and Liu, Ziyue and Wang, Zhengyang and Zhang, Zheng},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/zhang2025neurips-lax/}
}