Density-SoftMax: Efficient Test-Time Model for Uncertainty Estimation and Robustness Under Distribution Shifts
Abstract
Sampling-based methods, e.g., Deep Ensembles and Bayesian Neural Nets have become promising approaches to improve the quality of uncertainty estimation and robust generalization. However, they suffer from a large model size and high latency at test time, which limits the scalability needed for low-resource devices and real-time applications. To resolve these computational issues, we propose Density-Softmax, a sampling-free deterministic framework via combining a density function built on a Lipschitz-constrained feature extractor with the softmax layer. Theoretically, we show that our model is the solution of minimax uncertainty risk and is distance-aware on feature space, thus reducing the over-confidence of the standard softmax under distribution shifts. Empirically, our method enjoys competitive results with state-of-the-art techniques in terms of uncertainty and robustness, while having a lower number of model parameters and a lower latency at test time.
Cite
Text
Bui and Liu. "Density-SoftMax: Efficient Test-Time Model for Uncertainty Estimation and Robustness Under Distribution Shifts." International Conference on Machine Learning, 2024.Markdown
[Bui and Liu. "Density-SoftMax: Efficient Test-Time Model for Uncertainty Estimation and Robustness Under Distribution Shifts." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/bui2024icml-densitysoftmax/)BibTeX
@inproceedings{bui2024icml-densitysoftmax,
title = {{Density-SoftMax: Efficient Test-Time Model for Uncertainty Estimation and Robustness Under Distribution Shifts}},
author = {Bui, Ha Manh and Liu, Anqi},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {4822-4853},
volume = {235},
url = {https://mlanthology.org/icml/2024/bui2024icml-densitysoftmax/}
}