Active Vision with Predictive Coding and Uncertainty Minimization

Abstract

We present an end-to-end procedure for embodied visual exploration based on two biologically inspired computations: predictive coding and uncertainty minimization. The procedure can be applied in a task-independent and intrinsically driven manner. We evaluate our approach on an active vision task, where an agent actively samples its visual environment to gather information. We show that our model builds unsupervised representations through exploration that allow it to efficiently categorize visual scenes. We further show that using these representations for downstream classification leads to superior data efficiency and learning speed compared to other baselines while maintaining lower parameter complexity. Finally, the modularity of our model allows us to probe its internal mechanisms and analyze the interaction between perception and action during exploratory behavior.

Cite

Text

Sharafeldin et al. "Active Vision with Predictive Coding and Uncertainty Minimization." NeurIPS 2023 Workshops: InfoCog, 2023.

Markdown

[Sharafeldin et al. "Active Vision with Predictive Coding and Uncertainty Minimization." NeurIPS 2023 Workshops: InfoCog, 2023.](https://mlanthology.org/neuripsw/2023/sharafeldin2023neuripsw-active/)

BibTeX

@inproceedings{sharafeldin2023neuripsw-active,
  title     = {{Active Vision with Predictive Coding and Uncertainty Minimization}},
  author    = {Sharafeldin, Abdelrahman and Imam, Nabil and Choi, Hannah},
  booktitle = {NeurIPS 2023 Workshops: InfoCog},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/sharafeldin2023neuripsw-active/}
}