Automated Multi-Stage Compression of Neural Networks
Abstract
Low-rank tensor approximations are very promising for compression of deep neural networks. We propose a new simple and efficient iterative approach, which alternates low-rank factorization with smart rank selection and fine-tuning. We demonstrate the efficiency of our method comparing to non-iterative ones. Our approach improves the compression rate while maintaining the accuracy for a variety of tasks.
Cite
Text
Gusak et al. "Automated Multi-Stage Compression of Neural Networks." IEEE/CVF International Conference on Computer Vision Workshops, 2019. doi:10.1109/ICCVW.2019.00306Markdown
[Gusak et al. "Automated Multi-Stage Compression of Neural Networks." IEEE/CVF International Conference on Computer Vision Workshops, 2019.](https://mlanthology.org/iccvw/2019/gusak2019iccvw-automated/) doi:10.1109/ICCVW.2019.00306BibTeX
@inproceedings{gusak2019iccvw-automated,
title = {{Automated Multi-Stage Compression of Neural Networks}},
author = {Gusak, Julia and Kholyavchenko, Maksym and Ponomarev, Evgeny and Markeeva, Larisa and Blagoveschensky, Philip and Cichocki, Andrzej and Oseledets, Ivan V.},
booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
year = {2019},
pages = {2501-2508},
doi = {10.1109/ICCVW.2019.00306},
url = {https://mlanthology.org/iccvw/2019/gusak2019iccvw-automated/}
}