Can Models Learn Skill Composition from Examples?
Abstract
As large language models (LLMs) become increasingly capable, their ability to exhibit *compositional generalization* of skills has garnered significant attention. Yu et al. 2023 recently introduced Skill-Mix evaluation, where models are tasked with composing a short paragraph demonstrating the use of a specified $k$-tuple of language skills. While small models struggled with even $k=3$, larger models like GPT-4 showed reasonable performance with $k=5$ and $6$. In this paper, we employ a setup akin to Skill-Mix to evaluate the capacity of smaller models to learn compositional generalization from examples. Utilizing a diverse set of language skills—including rhetorical, literary, reasoning, and theory of mind—GPT-4 was used to generate text samples that exhibit random subsets of $k$ skills. Subsequent fine-tuning of 7B and 13B parameter models on these combined skill texts, for increasing values of $k$, revealed the following findings: 1) Training on combinations of $k=2$ and $3$ skills results in noticeable improvements in the ability to compose texts with $k=4$ and $5$ skills, despite models never having seen such examples during training. 2) When skill categories are split into training and held-out groups, models significantly improve at composing texts with held-out skills despite having only seen training skills during fine-tuning, illustrating the efficacy of the training approach even with previously unseen skills.
Cite
Text
Zhao et al. "Can Models Learn Skill Composition from Examples?." ICML 2024 Workshops: LLMs_and_Cognition, 2024.Markdown
[Zhao et al. "Can Models Learn Skill Composition from Examples?." ICML 2024 Workshops: LLMs_and_Cognition, 2024.](https://mlanthology.org/icmlw/2024/zhao2024icmlw-models/)BibTeX
@inproceedings{zhao2024icmlw-models,
title = {{Can Models Learn Skill Composition from Examples?}},
author = {Zhao, Haoyu and Kaur, Simran and Yu, Dingli and Goyal, Anirudh and Arora, Sanjeev},
booktitle = {ICML 2024 Workshops: LLMs_and_Cognition},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/zhao2024icmlw-models/}
}