Robust Distillation for Worst-Class Performance: On the Interplay Between Teacher and Student Objectives
Abstract
Knowledge distillation is a popular technique that has been shown to produce remarkable gains in average accuracy. However, recent work has shown that these gains are not uniform across subgroups in the data, and can often come at the cost of accuracy on rare subgroups and classes. Robust optimization is a common remedy to improve worst-class accuracy in standard learning settings, but in distillation it is unknown whether it is best to apply robust objectives when training the teacher, the student, or both. This work studies the interplay between robust objectives for the teacher and student. Empirically, we show that that jointly modifying the teacher and student objectives can lead to better worst-class student performance and even Pareto improvement in the trade-off between worst-class and overall performance. Theoretically, we show that the per-class calibration of teacher scores is key when training a robust student. Both the theory and experiments support the surprising finding that applying a robust teacher training objective does not always yield a more robust student.
Cite
Text
Wang et al. "Robust Distillation for Worst-Class Performance: On the Interplay Between Teacher and Student Objectives." Uncertainty in Artificial Intelligence, 2023.Markdown
[Wang et al. "Robust Distillation for Worst-Class Performance: On the Interplay Between Teacher and Student Objectives." Uncertainty in Artificial Intelligence, 2023.](https://mlanthology.org/uai/2023/wang2023uai-robust/)BibTeX
@inproceedings{wang2023uai-robust,
title = {{Robust Distillation for Worst-Class Performance: On the Interplay Between Teacher and Student Objectives}},
author = {Wang, Serena and Narasimhan, Harikrishna and Zhou, Yichen and Hooker, Sara and Lukasik, Michal and Menon, Aditya Krishna},
booktitle = {Uncertainty in Artificial Intelligence},
year = {2023},
pages = {2237-2247},
volume = {216},
url = {https://mlanthology.org/uai/2023/wang2023uai-robust/}
}