Distilling Vision-Language Models on Millions of Videos
Abstract
The recent advance in vision-language models is largely attributed to the abundance of image-text data. We aim to replicate this success for video-language models but there simply is not enough human-curated video-text data available. We thus resort to fine-tuning a video-language model from a strong image-language baseline with synthesized instructional data. The resulting video model by video-instruction-tuning (VIIT) is then used to auto-label millions of videos to generate high-quality captions. We show the adapted video-language model performs well on a wide range of video-language benchmarks. For instance it surpasses the best prior result on open-ended NExT-QA by2.8%. Besides our model generates detailed descriptions for previously unseen videos which provide better textual supervision than existing methods. Experiments show that a video-language dual-encoder model contrastively trained on these auto-generated captions is 3.8% better than the strongest baseline that also leverages vision-language models. Our best model outperforms state-of-the-art methods on MSR-VTT zero-shot text-to-video retrieval by 6%. As a side product we generate the largest video caption dataset to date.
Cite
Text
Zhao et al. "Distilling Vision-Language Models on Millions of Videos." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.01245Markdown
[Zhao et al. "Distilling Vision-Language Models on Millions of Videos." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/zhao2024cvpr-distilling/) doi:10.1109/CVPR52733.2024.01245BibTeX
@inproceedings{zhao2024cvpr-distilling,
title = {{Distilling Vision-Language Models on Millions of Videos}},
author = {Zhao, Yue and Zhao, Long and Zhou, Xingyi and Wu, Jialin and Chu, Chun-Te and Miao, Hui and Schroff, Florian and Adam, Hartwig and Liu, Ting and Gong, Boqing and Krahenbuhl, Philipp and Yuan, Liangzhe},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2024},
pages = {13106-13116},
doi = {10.1109/CVPR52733.2024.01245},
url = {https://mlanthology.org/cvpr/2024/zhao2024cvpr-distilling/}
}