Improving Generalization of Differentiable Simulator Policies with Sharpness-Aware Optimization
Abstract
This work contributes to the ongoing discussion on the trade-off between performance and generalization in reinforcement learning, particularly in the context of sim-to-real transfer in robotics. We investigate the generalization capabilities of policies learned using differentiable simulators in contact-rich robotic scenarios. While first-order optimization achieves a higher sample efficiency, it has been empirically shown to be unstable in loco-manipulation problems. We demonstrate that, while first-order methods achieve superior performance and sample efficiency in training, they exhibit less robustness to environmental variations. To address this limitation, we propose augmenting them with sharpness-aware optimization. Our simulation results show that this approach improves the generalization of learned policies over a larger magnitude of perturbation noise.
Cite
Text
Bochem et al. "Improving Generalization of Differentiable Simulator Policies with Sharpness-Aware Optimization." NeurIPS 2024 Workshops: D3S3, 2024.Markdown
[Bochem et al. "Improving Generalization of Differentiable Simulator Policies with Sharpness-Aware Optimization." NeurIPS 2024 Workshops: D3S3, 2024.](https://mlanthology.org/neuripsw/2024/bochem2024neuripsw-improving/)BibTeX
@inproceedings{bochem2024neuripsw-improving,
title = {{Improving Generalization of Differentiable Simulator Policies with Sharpness-Aware Optimization}},
author = {Bochem, Severin and Sanchez, Eduardo Gonzalez and Bicker, Yves and Fadini, Gabriele},
booktitle = {NeurIPS 2024 Workshops: D3S3},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/bochem2024neuripsw-improving/}
}