Info3D: Representation Learning on 3D Objects Using Mutual Information Maximization and Contrastive Learning

Abstract

A major endeavor of computer vision is to represent, understand and extract structure from 3D data. Towards this goal, unsupervised learning is a powerful and necessary tool. Most current unsupervised methods for 3D shape analysis use datasets that are aligned, require objects to be reconstructed and suffer from deteriorated performance on downstream tasks. To solve these issues we propose to extend the InfoMax and contrastive learning principles on 3D shapes. We show that we can maximize the mutual information between 3D objects and their ""chunks"" to improve the representations in aligned datasets. Furthermore, we can achieve rotation invariance in SO(3) group by maximizing the mutual information between the 3D objects and their geometric transformed versions. Finally, we conduct several experiments such as clustering, transfer learning, shape retrieval, and achieve state of art results.

Cite

Text

Sanghi. "Info3D: Representation Learning on 3D Objects Using Mutual Information Maximization and Contrastive Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58526-6_37

Markdown

[Sanghi. "Info3D: Representation Learning on 3D Objects Using Mutual Information Maximization and Contrastive Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/sanghi2020eccv-info3d/) doi:10.1007/978-3-030-58526-6_37

BibTeX

@inproceedings{sanghi2020eccv-info3d,
  title     = {{Info3D: Representation Learning on 3D Objects Using Mutual Information Maximization and Contrastive Learning}},
  author    = {Sanghi, Aditya},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2020},
  doi       = {10.1007/978-3-030-58526-6_37},
  url       = {https://mlanthology.org/eccv/2020/sanghi2020eccv-info3d/}
}