In a Nutshell, the Human Asked for This: Latent Goals for Following Temporal Specifications

Abstract

We address the problem of building agents whose goal is to learn to execute out-of distribution (OOD) multi-task instructions expressed in temporal logic (TL) by using deep reinforcement learning (DRL). Recent works provided evidence that the agent's neural architecture is a key feature when DRL agents are learning to solve OOD tasks in TL. Yet, the studies on this topic are still in their infancy. In this work, we propose a new deep learning configuration with inductive biases that lead agents to generate latent representations of their current goal, yielding a stronger generalization performance. We use these latent-goal networks within a neuro-symbolic framework that executes multi-task formally-defined instructions and contrast the performance of the proposed neural networks against employing different state-of-the-art (SOTA) architectures when generalizing to unseen instructions in OOD environments.

Cite

Text

León et al. "In a Nutshell, the Human Asked for This: Latent Goals for Following Temporal Specifications." International Conference on Learning Representations, 2022.

Markdown

[León et al. "In a Nutshell, the Human Asked for This: Latent Goals for Following Temporal Specifications." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/leon2022iclr-nutshell/)

BibTeX

@inproceedings{leon2022iclr-nutshell,
  title     = {{In a Nutshell, the Human Asked for This: Latent Goals for Following Temporal Specifications}},
  author    = {León, Borja G. and Shanahan, Murray and Belardinelli, Francesco},
  booktitle = {International Conference on Learning Representations},
  year      = {2022},
  url       = {https://mlanthology.org/iclr/2022/leon2022iclr-nutshell/}
}