DOS: Diverse Outlier Sampling for Out-of-Distribution Detection

Abstract

Modern neural networks are known to give overconfident predictions for out-of-distribution inputs when deployed in the open world. It is common practice to leverage a surrogate outlier dataset to regularize the model during training, and recent studies emphasize the role of uncertainty in designing the sampling strategy for outlier datasets. However, the OOD samples selected solely based on predictive uncertainty can be biased towards certain types, which may fail to capture the full outlier distribution. In this work, we empirically show that diversity is critical in sampling outliers for OOD detection performance. Motivated by the observation, we propose a straightforward and novel sampling strategy named DOS (Diverse Outlier Sampling) to select diverse and informative outliers. Specifically, we cluster the normalized features at each iteration, and the most informative outlier from each cluster is selected for model training with absent category loss. With DOS, the sampled outliers efficiently shape a globally compact decision boundary between ID and OOD data. Extensive experiments demonstrate the superiority of DOS, reducing the average FPR95 by up to 25.79% on CIFAR-100 with TI-300K.

Cite

Text

Jiang et al. "DOS: Diverse Outlier Sampling for Out-of-Distribution Detection." International Conference on Learning Representations, 2024.

Markdown

[Jiang et al. "DOS: Diverse Outlier Sampling for Out-of-Distribution Detection." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/jiang2024iclr-dos/)

BibTeX

@inproceedings{jiang2024iclr-dos,
  title     = {{DOS: Diverse Outlier Sampling for Out-of-Distribution Detection}},
  author    = {Jiang, Wenyu and Cheng, Hao and Chen, MingCai and Wang, Chongjun and Wei, Hongxin},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/jiang2024iclr-dos/}
}