On the Power of Context-Enhanced Learning in LLMs
Abstract
We formalize a new concept for LLMs, **context-enhanced learning**. It involves standard gradient-based learning on text except that the context is enhanced with additional data on which no auto-regressive gradients are computed. This setting is a gradient-based analog of usual in-context learning (ICL) and appears in some recent works. Using a multi-step reasoning task, we prove in a simplified setting that context-enhanced learning can be **exponentially more sample-efficient** than standard learning when the model is capable of ICL. At a mechanistic level, we find that the benefit of context-enhancement arises from a more accurate gradient learning signal. We also experimentally demonstrate that **it appears hard to detect or recover learning materials that were used in the context during training**. This may have implications for data security as well as copyright.
Cite
Text
Zhu et al. "On the Power of Context-Enhanced Learning in LLMs." ICLR 2025 Workshops: Data_Problems, 2025.Markdown
[Zhu et al. "On the Power of Context-Enhanced Learning in LLMs." ICLR 2025 Workshops: Data_Problems, 2025.](https://mlanthology.org/iclrw/2025/zhu2025iclrw-power/)BibTeX
@inproceedings{zhu2025iclrw-power,
title = {{On the Power of Context-Enhanced Learning in LLMs}},
author = {Zhu, Xingyu and Panigrahi, Abhishek and Arora, Sanjeev},
booktitle = {ICLR 2025 Workshops: Data_Problems},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/zhu2025iclrw-power/}
}