Generalization and Translatability in Emergent Communication via Informational Constraints

Abstract

Traditional emergent communication (EC) methods often fail to generalize to novel settings or align with representations of natural language. Here, we show how controlling the Information Bottleneck (IB) tradeoff between complexity and informativeness (a principle thought to guide human languages) helps to address both of these problems in EC. Using VQ-VIB, a recent method for training EC agents while controlling the IB tradeoff, we find that: (1) increasing pressure for informativeness, which encourages agents to develop a shared understanding beyond task-specific needs, leads to better generalization to more challenging tasks and novel inputs; (2) VQ-VIB agents develop an EC space that encodes some semantic similarities and facilitates open-domain communication, similar to word embeddings in natural language; and (3) when translating between English and EC, greater complexity leads to improved performance of teams of simulated English speakers and trained VQ-VIB listeners, but only up to a threshold corresponding to the English complexity. These results indicate the importance of informational constraints for improving self-play performance and human-agent interaction.

Cite

Text

Tucker et al. "Generalization and Translatability in Emergent Communication via Informational Constraints." NeurIPS 2022 Workshops: InfoCog, 2022.

Markdown

[Tucker et al. "Generalization and Translatability in Emergent Communication via Informational Constraints." NeurIPS 2022 Workshops: InfoCog, 2022.](https://mlanthology.org/neuripsw/2022/tucker2022neuripsw-generalization/)

BibTeX

@inproceedings{tucker2022neuripsw-generalization,
  title     = {{Generalization and Translatability in Emergent Communication via Informational Constraints}},
  author    = {Tucker, Mycal and Levy, Roger P. and Shah, Julie and Zaslavsky, Noga},
  booktitle = {NeurIPS 2022 Workshops: InfoCog},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/tucker2022neuripsw-generalization/}
}