COIN++: Neural Compression Across Modalities
Abstract
Neural compression algorithms are typically based on autoencoders that require specialized encoder and decoder architectures for different data modalities. In this paper, we propose COIN++, a neural compression framework that seamlessly handles a wide range of data modalities. Our approach is based on converting data to implicit neural representations, i.e. neural functions that map coordinates (such as pixel locations) to features (such as RGB values). Then, instead of storing the weights of the implicit neural representation directly, we store modulations applied to a meta-learned base network as a compressed code for the data. We further quantize and entropy code these modulations, leading to large compression gains while reducing encoding time by two orders of magnitude compared to baselines. We empirically demonstrate the feasibility of our method by compressing various data modalities, from images and audio to medical and climate data.
Cite
Text
Dupont et al. "COIN++: Neural Compression Across Modalities." Transactions on Machine Learning Research, 2022.Markdown
[Dupont et al. "COIN++: Neural Compression Across Modalities." Transactions on Machine Learning Research, 2022.](https://mlanthology.org/tmlr/2022/dupont2022tmlr-coin/)BibTeX
@article{dupont2022tmlr-coin,
title = {{COIN++: Neural Compression Across Modalities}},
author = {Dupont, Emilien and Loya, Hrushikesh and Alizadeh, Milad and Golinski, Adam and Teh, Yee Whye and Doucet, Arnaud},
journal = {Transactions on Machine Learning Research},
year = {2022},
url = {https://mlanthology.org/tmlr/2022/dupont2022tmlr-coin/}
}