Value Refinement Network (VRN)

Abstract

In robotic tasks, we encounter the unique strengths of (1) reinforcement learning (RL) that can handle high-dimensional observations as well as unknown, complex dynamics and (2) planning that can handle sparse and delayed rewards given a dynamics model. Combining these strengths of RL and planning, we propose the Value Refinement Network (VRN), in this work. Our VRN is an RL-trained neural network architecture that learns to locally refine an initial (value-based) plan in a simplified (2D) problem abstraction based on detailed local sensory observations. We evaluate the VRN on simulated robotic (navigation) tasks and demonstrate that it can successfully refine sub-optimal plans to match the performance of more costly planning in the non-simplified problem. Furthermore, in a dynamic environment, the VRN still enables high task completion without global re-planning.

Cite

Text

Wöhlke et al. "Value Refinement Network (VRN)." International Joint Conference on Artificial Intelligence, 2022. doi:10.24963/IJCAI.2022/494

Markdown

[Wöhlke et al. "Value Refinement Network (VRN)." International Joint Conference on Artificial Intelligence, 2022.](https://mlanthology.org/ijcai/2022/wohlke2022ijcai-value/) doi:10.24963/IJCAI.2022/494

BibTeX

@inproceedings{wohlke2022ijcai-value,
  title     = {{Value Refinement Network (VRN)}},
  author    = {Wöhlke, Jan and Schmitt, Felix and van Hoof, Herke},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2022},
  pages     = {3558-3565},
  doi       = {10.24963/IJCAI.2022/494},
  url       = {https://mlanthology.org/ijcai/2022/wohlke2022ijcai-value/}
}