Imbalanced Continual Learning with Partitioning Reservoir Sampling
Abstract
Continual learning from a sequential stream of data is a crucial challenge for machine learning research. Most studies have been conducted on this topic under the single-label classification setting along with an assumption of balanced label distribution. This work expands this research horizon towards multi-label classification. In doing so, we identify unanticipated adversity innately existent in many multilabel datasets, the long-tailed distribution. We jointly address the two independently solved problems, Catastropic Forgetting and the long-tailed label distribution by first empirically showing a new challenge of destructive forgetting of the minority concepts on the tail. Then, we curate two benchmark datasets, COCOseq and NUS-WIDEseq, that allow the study of both intra- and inter-task imbalances. Lastly, we propose a new sampling strategy for replay-based approach named Partitioning Reservoir Sampling (PRS), which allows the model to maintain a balanced knowledge of both head and tail classes. We publicly release the dataset and the code in our project page.
Cite
Text
Kim et al. "Imbalanced Continual Learning with Partitioning Reservoir Sampling." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58601-0_25Markdown
[Kim et al. "Imbalanced Continual Learning with Partitioning Reservoir Sampling." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/kim2020eccv-imbalanced/) doi:10.1007/978-3-030-58601-0_25BibTeX
@inproceedings{kim2020eccv-imbalanced,
title = {{Imbalanced Continual Learning with Partitioning Reservoir Sampling}},
author = {Kim, Chris Dongjoo and Jeong, Jinseo and Kim, Gunhee},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58601-0_25},
url = {https://mlanthology.org/eccv/2020/kim2020eccv-imbalanced/}
}