MEDIRL: Predicting the Visual Attention of Drivers via Maximum Entropy Deep Inverse Reinforcement Learning

Abstract

Inspired by human visual attention, we propose a novel inverse reinforcement learning formulation using Maximum Entropy Deep Inverse Reinforcement Learning (MEDIRL) for predicting the visual attention of drivers in accident-prone situations. MEDIRL predicts fixation locations that lead to maximal rewards by learning a task-sensitive reward function from eye fixation patterns recorded from attentive drivers. Additionally, we introduce EyeCar, a new driver attention dataset in accident-prone situations. We conduct comprehensive experiments to evaluate our proposed model on three common benchmarks: (DR(eye)VE, BDD-A, DADA-2000), and our EyeCar dataset. Results indicate that MEDIRL outperforms existing models for predicting attention and achieves state-of-the-art performance. We present extensive ablation studies to provide more insights into different features of our proposed model.

Cite

Text

Baee et al. "MEDIRL: Predicting the Visual Attention of Drivers via Maximum Entropy Deep Inverse Reinforcement Learning." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.01293

Markdown

[Baee et al. "MEDIRL: Predicting the Visual Attention of Drivers via Maximum Entropy Deep Inverse Reinforcement Learning." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/baee2021iccv-medirl/) doi:10.1109/ICCV48922.2021.01293

BibTeX

@inproceedings{baee2021iccv-medirl,
  title     = {{MEDIRL: Predicting the Visual Attention of Drivers via Maximum Entropy Deep Inverse Reinforcement Learning}},
  author    = {Baee, Sonia and Pakdamanian, Erfan and Kim, Inki and Feng, Lu and Ordonez, Vicente and Barnes, Laura},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {13178-13188},
  doi       = {10.1109/ICCV48922.2021.01293},
  url       = {https://mlanthology.org/iccv/2021/baee2021iccv-medirl/}
}