Trust, but Verify: Using Self-Supervised Probing to Improve Trustworthiness

Abstract

Trustworthy machine learning is of primary importance to the practical deployment of deep learning models. While state-of-the-art models achieve astonishingly good performance in terms of accuracy, recent literature reveals that their predictive confidence scores unfortunately cannot be trusted: e.g., they are often overconfident when wrong predictions are made, or so even for obvious outliers. In this paper, we introduce a new approach of \emph{self-supervised probing}, which enables us to check and mitigate the overconfidence issue for a trained model, thereby improving its trustworthiness. We provide a simple yet effective framework, which can be flexibly applied to existing trustworthiness-related methods in a plug-and-play manner. Extensive experiments on three trustworthiness-related tasks (misclassification detection, calibration and out-of-distribution detection) across various benchmarks verify the effectiveness of our proposed probing framework.

Cite

Text

Deng et al. "Trust, but Verify: Using Self-Supervised Probing to Improve Trustworthiness." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-19778-9_21

Markdown

[Deng et al. "Trust, but Verify: Using Self-Supervised Probing to Improve Trustworthiness." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/deng2022eccv-trust/) doi:10.1007/978-3-031-19778-9_21

BibTeX

@inproceedings{deng2022eccv-trust,
  title     = {{Trust, but Verify: Using Self-Supervised Probing to Improve Trustworthiness}},
  author    = {Deng, Ailin and Li, Shen and Xiong, Miao and Chen, Zhirui and Hooi, Bryan},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2022},
  doi       = {10.1007/978-3-031-19778-9_21},
  url       = {https://mlanthology.org/eccv/2022/deng2022eccv-trust/}
}