Deep Companion Learning: Enhancing Generalization Through Historical Consistency
Abstract
We propose Deep Companion Learning (DCL), a novel training method for Deep Neural Networks (DNNs) that enhances generalization by penalizing inconsistent model predictions compared to its historical performance. To achieve this, we train a deep-companion model (DCM), by using previous versions of the model to provide forecasts on new inputs. This companion model deciphers a meaningful latent semantic structure within the data, thereby providing targeted supervision that encourages the primary model to address the scenarios it finds most challenging. We validate our approach through both theoretical analysis and extensive experimentation, including ablation studies, on a variety of benchmark datasets (CIFAR-100, Tiny-ImageNet, ImageNet-1K) using diverse architectural models (ShuffleNetV2, ResNet, Vision Transformer, etc.), demonstrating state-of-the-art performance.
Cite
Text
Zhu and Saligrama. "Deep Companion Learning: Enhancing Generalization Through Historical Consistency." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-72913-3_22Markdown
[Zhu and Saligrama. "Deep Companion Learning: Enhancing Generalization Through Historical Consistency." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/zhu2024eccv-deep/) doi:10.1007/978-3-031-72913-3_22BibTeX
@inproceedings{zhu2024eccv-deep,
title = {{Deep Companion Learning: Enhancing Generalization Through Historical Consistency}},
author = {Zhu, Ruizhao and Saligrama, Venkatesh},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2024},
doi = {10.1007/978-3-031-72913-3_22},
url = {https://mlanthology.org/eccv/2024/zhu2024eccv-deep/}
}