Positive-Congruent Training: Towards Regression-Free Model Updates

Abstract

Reducing inconsistencies in the behavior of different versions of an AI system can be as important in practice as reducing its overall error. In image classification, sample-wise inconsistencies appear as "negative flips": A new model incorrectly predicts the output for a test sample that was correctly classified by the old (reference) model. Positive-congruent (PC) training aims at reducing error rate while at the same reducing negative flips, thus maximizing congruency with the reference model only on positive predictions, unlike model distillation. We propose a simple approach for PC training, Focal Distillation, which enforces congruence with the reference model by giving more weights to samples that were correctly classified. We also found that, if the reference model itself can be chosen as an ensemble of multiple deep neural networks, negative flips can be further reduced without affecting the new model's accuracy.

Cite

Text

Yan et al. "Positive-Congruent Training: Towards Regression-Free Model Updates." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.01407

Markdown

[Yan et al. "Positive-Congruent Training: Towards Regression-Free Model Updates." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/yan2021cvpr-positivecongruent/) doi:10.1109/CVPR46437.2021.01407

BibTeX

@inproceedings{yan2021cvpr-positivecongruent,
  title     = {{Positive-Congruent Training: Towards Regression-Free Model Updates}},
  author    = {Yan, Sijie and Xiong, Yuanjun and Kundu, Kaustav and Yang, Shuo and Deng, Siqi and Wang, Meng and Xia, Wei and Soatto, Stefano},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {14299-14308},
  doi       = {10.1109/CVPR46437.2021.01407},
  url       = {https://mlanthology.org/cvpr/2021/yan2021cvpr-positivecongruent/}
}