Goal-Conditioned Generators of Deep Policies
Abstract
Goal-conditioned Reinforcement Learning (RL) aims at learning optimal policies, given goals encoded in special command inputs. Here we study goal-conditioned neural nets (NNs) that learn to generate deep NN policies in form of context-specific weight matrices, similar to Fast Weight Programmers and other methods from the 1990s. Using context commands of the form ``generate a policy that achieves a desired expected return,'' our NN generators combine powerful exploration of parameter space with generalization across commands to iteratively find better and better policies. A form of weight-sharing HyperNetworks and policy embeddings scales our method to generate deep NNs. Experiments show how a single learned policy generator can produce policies that achieve any return seen during training. Finally, we evaluate our algorithm on a set of continuous control tasks where it exhibits competitive performance. Our code is public.
Cite
Text
Faccio et al. "Goal-Conditioned Generators of Deep Policies." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I6.25912Markdown
[Faccio et al. "Goal-Conditioned Generators of Deep Policies." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/faccio2023aaai-goal/) doi:10.1609/AAAI.V37I6.25912BibTeX
@inproceedings{faccio2023aaai-goal,
title = {{Goal-Conditioned Generators of Deep Policies}},
author = {Faccio, Francesco and Herrmann, Vincent and Ramesh, Aditya A. and Kirsch, Louis and Schmidhuber, Jürgen},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {7503-7511},
doi = {10.1609/AAAI.V37I6.25912},
url = {https://mlanthology.org/aaai/2023/faccio2023aaai-goal/}
}