Imitation Learning with Demonstrations and Shaping Rewards
Abstract
Imitation Learning (IL) is a popular approach for teaching behavior policies to agents by demonstrating the desired target policy. While the approach has lead to many successes, IL often requires a large set of demonstrations to achieve robust learning, which can be expensive for the teacher. In this paper, we consider a novel approach to improve the learning efficiency of IL by providing a shaping reward function in addition to the usual demonstrations. Shaping rewards are numeric functions of states (and possibly actions) that are generally easily specified, and capture general principles of desired behavior, without necessarily completely specifying the behavior. Shaping rewards have been used extensively in reinforcement learning, but have been seldom considered for IL, though they are often easy to specify. Our main contribution is to propose an IL approach that learns from both shaping rewards and demonstrations. We demonstrate the effectiveness of the approach across several IL problems, even when the shaping reward is not fully consistent with the demonstrations.
Cite
Text
Judah et al. "Imitation Learning with Demonstrations and Shaping Rewards." AAAI Conference on Artificial Intelligence, 2014. doi:10.1609/AAAI.V28I1.9024Markdown
[Judah et al. "Imitation Learning with Demonstrations and Shaping Rewards." AAAI Conference on Artificial Intelligence, 2014.](https://mlanthology.org/aaai/2014/judah2014aaai-imitation/) doi:10.1609/AAAI.V28I1.9024BibTeX
@inproceedings{judah2014aaai-imitation,
title = {{Imitation Learning with Demonstrations and Shaping Rewards}},
author = {Judah, Kshitij and Fern, Alan and Tadepalli, Prasad and Goetschalckx, Robby},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2014},
pages = {1890-1896},
doi = {10.1609/AAAI.V28I1.9024},
url = {https://mlanthology.org/aaai/2014/judah2014aaai-imitation/}
}