On the Power of Context-Enhanced Learning in LLMs

Abstract

We formalize a new concept for LLMs, **context-enhanced learning**. It involves standard gradient-based learning on text except that the context is enhanced with additional data on which no auto-regressive gradients are computed. This setting is a gradient-based analog of usual in-context learning (ICL) and appears in some recent works. Using a multi-step reasoning task, we prove in a simplified setting that context-enhanced learning can be **exponentially more sample-efficient** than standard learning when the model is capable of ICL. At a mechanistic level, we find that the benefit of context-enhancement arises from a more accurate gradient learning signal. We also experimentally demonstrate that **it appears hard to detect or recover learning materials that were used in the context during training**. This may have implications for data security as well as copyright.

Cite

Text

Zhu et al. "On the Power of Context-Enhanced Learning in LLMs." ICLR 2025 Workshops: Data_Problems, 2025.

Markdown

[Zhu et al. "On the Power of Context-Enhanced Learning in LLMs." ICLR 2025 Workshops: Data_Problems, 2025.](https://mlanthology.org/iclrw/2025/zhu2025iclrw-power/)

BibTeX

@inproceedings{zhu2025iclrw-power,
  title     = {{On the Power of Context-Enhanced Learning in LLMs}},
  author    = {Zhu, Xingyu and Panigrahi, Abhishek and Arora, Sanjeev},
  booktitle = {ICLR 2025 Workshops: Data_Problems},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/zhu2025iclrw-power/}
}