Disentangling Latent Shifts of In-Context Learning with Weak Supervision
Abstract
In-context learning (ICL) enables large language models to perform few-shot learning by conditioning on labeled examples in the prompt. Despite its flexibility, ICL suffers from instability -- especially as prompt length increases with more demonstrations. To address this, we treat ICL as a source of weak supervision and propose a parameter-efficient method that disentangles demonstration-induced latent shifts from those of the query. An ICL-based teacher generates pseudo-labels on unlabeled queries, while a student predicts them using only the query input, updating a lightweight adapter. This captures demonstration effects in a compact, reusable form, enabling efficient inference while remaining composable with new demonstrations. Although trained on noisy teacher outputs, the student often outperforms its teacher through pseudo-label correction and coverage expansion, consistent with the weak-to-strong generalization effect. Empirically, our method improves generalization, stability, and efficiency across both in-domain and out-of-domain tasks, surpassing standard ICL and prior disentanglement methods.
Cite
Text
Jukić and Šnajder. "Disentangling Latent Shifts of In-Context Learning with Weak Supervision." Advances in Neural Information Processing Systems, 2025.Markdown
[Jukić and Šnajder. "Disentangling Latent Shifts of In-Context Learning with Weak Supervision." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/jukic2025neurips-disentangling/)BibTeX
@inproceedings{jukic2025neurips-disentangling,
title = {{Disentangling Latent Shifts of In-Context Learning with Weak Supervision}},
author = {Jukić, Josip and Šnajder, Jan},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/jukic2025neurips-disentangling/}
}