Generalizable Implicit Neural Representations via Instance Pattern Composers
Abstract
Despite recent advances in implicit neural representations (INRs), it remains challenging for a coordinate-based multi-layer perceptron (MLP) of INRs to learn a common representation across data instances and generalize it for unseen instances. In this work, we introduce a simple yet effective framework for generalizable INRs that enables a coordinate-based MLP to represent complex data instances by modulating only a small set of weights in an early MLP layer as an instance pattern composer; the remaining MLP weights learn pattern composition rules to learn common representations across instances. Our generalizable INR framework is fully compatible with existing meta-learning and hypernetworks in learning to predict the modulated weight for unseen instances. Extensive experiments demonstrate that our method achieves high performance on a wide range of domains such as an audio, image, and 3D object, while the ablation study validates our weight modulation.
Cite
Text
Kim et al. "Generalizable Implicit Neural Representations via Instance Pattern Composers." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.01136Markdown
[Kim et al. "Generalizable Implicit Neural Representations via Instance Pattern Composers." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/kim2023cvpr-generalizable/) doi:10.1109/CVPR52729.2023.01136BibTeX
@inproceedings{kim2023cvpr-generalizable,
title = {{Generalizable Implicit Neural Representations via Instance Pattern Composers}},
author = {Kim, Chiheon and Lee, Doyup and Kim, Saehoon and Cho, Minsu and Han, Wook-Shin},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2023},
pages = {11808-11817},
doi = {10.1109/CVPR52729.2023.01136},
url = {https://mlanthology.org/cvpr/2023/kim2023cvpr-generalizable/}
}