Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning

Abstract

Recently, various pre-training methods have been introduced in vision-based Reinforcement Learning (RL). However, their generalization ability remains unclear due to evaluations being limited to in-distribution environments and non-unified experimental setups. To address this, we introduce the Atari Pre-training Benchmark (Atari-PB), which pre-trains a ResNet-50 model on 10 million transitions from 50 Atari games and evaluates it across diverse environment distributions. Our experiments show that pre-training objectives focused on learning task-agnostic features (e.g., identifying objects and understanding temporal dynamics) enhance generalization across different environments. In contrast, objectives focused on learning task-specific knowledge (e.g., identifying agents and fitting reward functions) improve performance in environments similar to the pre-training dataset but not in varied ones. We publicize our codes, datasets, and model checkpoints at https://github.com/dojeon-ai/Atari-PB.

Cite

Text

Kim et al. "Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning." International Conference on Machine Learning, 2024.

Markdown

[Kim et al. "Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/kim2024icml-investigating/)

BibTeX

@inproceedings{kim2024icml-investigating,
  title     = {{Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning}},
  author    = {Kim, Donghu and Lee, Hojoon and Lee, Kyungmin and Hwang, Dongyoon and Choo, Jaegul},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {24294-24326},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/kim2024icml-investigating/}
}