Instance-Adaptive Video Compression: Improving Neural Codecs by Training on the Test Set
Abstract
We introduce a video compression algorithm based on instance-adaptive learning. On each video sequence to be transmitted, we finetune a pretrained compression model. The optimal parameters are transmitted to the receiver along with the latent code. By entropy-coding the parameter updates under a suitable mixture model prior, we ensure that the network parameters can be encoded efficiently. This instance-adaptive compression algorithm is agnostic about the choice of base model and has the potential to improve any neural video codec. On UVG, HEVC, and Xiph datasets, our codec improves the performance of a scale-space flow model by between 21% and 27% BD-rate savings, and that of a state-of-the-art B-frame model by 17 to 20% BD-rate savings. We also demonstrate that instance-adaptive finetuning improves the robustness to domain shift. Finally, our approach reduces the capacity requirements of compression models. We show that it enables a competitive performance even after reducing the network size by 70%.
Cite
Text
van Rozendaal et al. "Instance-Adaptive Video Compression: Improving Neural Codecs by Training on the Test Set." Transactions on Machine Learning Research, 2023.Markdown
[van Rozendaal et al. "Instance-Adaptive Video Compression: Improving Neural Codecs by Training on the Test Set." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/vanrozendaal2023tmlr-instanceadaptive/)BibTeX
@article{vanrozendaal2023tmlr-instanceadaptive,
title = {{Instance-Adaptive Video Compression: Improving Neural Codecs by Training on the Test Set}},
author = {van Rozendaal, Ties and Brehmer, Johann and Zhang, Yunfan and Pourreza, Reza and Wiggers, Auke J. and Cohen, Taco},
journal = {Transactions on Machine Learning Research},
year = {2023},
url = {https://mlanthology.org/tmlr/2023/vanrozendaal2023tmlr-instanceadaptive/}
}