Explicit Knowledge Factorization Meets In-Context Learning: What Do We Gain?
Abstract
Transformer models have shown considerable success in modeling predictive problems in diverse domains. It has been shown that they can efficiently learn in-context, i.e., solve new tasks without any further training when provided some examples as context (ICL). While first observed in language, consequent studies show that ICL also generalizes to a variety of algorithmic tasks. Recent research highlights that transformers might be implicitly modeling the posterior predictive distribution over latent variables which are important to solve different tasks. However, ICL diverges from standard Bayesian methods as it foregoes defining explicit latent variable model and consequently doing inference on it, in favor of an implicit mechanism. This begs a natural question: is there any benefit in explicitly factorizing knowledge or are we better off letting the model implicitly decide an appropriate solution space. We conduct a thorough analysis to uncover both the advantages and limitations of the current ICL setting and show that models which explicitly factorize knowledge can be more readily augmented with inductive biases which significantly boosts performance when domain knowledge is present.
Cite
Text
Mittal et al. "Explicit Knowledge Factorization Meets In-Context Learning: What Do We Gain?." ICLR 2024 Workshops: R2-FM, 2024.Markdown
[Mittal et al. "Explicit Knowledge Factorization Meets In-Context Learning: What Do We Gain?." ICLR 2024 Workshops: R2-FM, 2024.](https://mlanthology.org/iclrw/2024/mittal2024iclrw-explicit/)BibTeX
@inproceedings{mittal2024iclrw-explicit,
title = {{Explicit Knowledge Factorization Meets In-Context Learning: What Do We Gain?}},
author = {Mittal, Sarthak and Elmoznino, Eric and Gagnon, Leo and Bhardwaj, Sangnie and Sridhar, Dhanya and Lajoie, Guillaume},
booktitle = {ICLR 2024 Workshops: R2-FM},
year = {2024},
url = {https://mlanthology.org/iclrw/2024/mittal2024iclrw-explicit/}
}