Uncertainty-Aware Action Repeating Options

Abstract

In reinforcement learning, employing temporal abstraction within the action space is a prevalent strategy for simplifying policy learning through temporally-extended actions. Recently, algorithms that repeat a primitive action for a certain number of steps, a simple method to implement temporal abstraction in practice, have demonstrated better performance than traditional algorithms. However, a significant drawback of earlier studies on action repetition is the potential for repeated sub-optimal actions to considerably degrade performance. To tackle this problem, we introduce a new algorithm that employs ensemble methods to estimate uncertainty when extending an action. Our framework offers flexibility, allowing policies to either prioritize exploration or adopt an uncertainty-averse stance based on their specific needs. We provide empirical results on various environments, highlighting the superior performance of our proposed method compared to other action-repeating algorithms. These results indicate that our uncertainty-aware strategy effectively counters the downsides of action repetition, enhancing policy learning efficiency.

Cite

Text

Lee et al. "Uncertainty-Aware Action Repeating Options." NeurIPS 2023 Workshops: GenPlan, 2023.

Markdown

[Lee et al. "Uncertainty-Aware Action Repeating Options." NeurIPS 2023 Workshops: GenPlan, 2023.](https://mlanthology.org/neuripsw/2023/lee2023neuripsw-uncertaintyaware/)

BibTeX

@inproceedings{lee2023neuripsw-uncertaintyaware,
  title     = {{Uncertainty-Aware Action Repeating Options}},
  author    = {Lee, Joongkyu and Park, Seung Joon and Tang, Yunhao and Oh, Min-hwan},
  booktitle = {NeurIPS 2023 Workshops: GenPlan},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/lee2023neuripsw-uncertaintyaware/}
}