Spatio-Temporal Crop Aggregation for Video Representation Learning
Abstract
We propose Spatio-temporal Crop Aggregation for video representation LEarning (SCALE), a novel method that enjoys high scalability at both training and inference time. Our model builds long-range video features by learning from sets of video clip-level features extracted with a pre-trained backbone. To train the model, we propose a self-supervised objective consisting of masked clip feature predictions. We apply sparsity to both the input, by extracting a random set of video clips, and to the loss function, by only reconstructing the sparse inputs. Moreover, we use dimensionality reduction by working in the latent space of a pre-trained backbone applied to single video clips. These techniques make our method not only extremely efficient to train but also highly effective in transfer learning. We demonstrate that our video representation yields state-of-the-art performance with linear, nonlinear, and k-NN probing on common action classification and video understanding datasets.
Cite
Text
Sameni et al. "Spatio-Temporal Crop Aggregation for Video Representation Learning." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.00521Markdown
[Sameni et al. "Spatio-Temporal Crop Aggregation for Video Representation Learning." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/sameni2023iccv-spatiotemporal/) doi:10.1109/ICCV51070.2023.00521BibTeX
@inproceedings{sameni2023iccv-spatiotemporal,
title = {{Spatio-Temporal Crop Aggregation for Video Representation Learning}},
author = {Sameni, Sepehr and Jenni, Simon and Favaro, Paolo},
booktitle = {International Conference on Computer Vision},
year = {2023},
pages = {5664-5674},
doi = {10.1109/ICCV51070.2023.00521},
url = {https://mlanthology.org/iccv/2023/sameni2023iccv-spatiotemporal/}
}