Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution

Abstract

Many tasks in explainable machine learning, such as data valuation and feature attribution, perform expensive computation for each data point and are intractable for large datasets. These methods require efficient approximations, and although amortizing the process by learning a network to directly predict the desired output is a promising solution, training such models with exact labels is often infeasible. We therefore explore training amortized models with noisy labels, and we find that this is inexpensive and surprisingly effective. Through theoretical analysis of the label noise and experiments with various models and datasets, we show that this approach tolerates high noise levels and significantly accelerates several feature attribution and data valuation methods, often yielding an order of magnitude speedup over existing approaches.

Cite

Text

Covert et al. "Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution." Neural Information Processing Systems, 2024. doi:10.52202/079017-0143

Markdown

[Covert et al. "Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/covert2024neurips-stochastic/) doi:10.52202/079017-0143

BibTeX

@inproceedings{covert2024neurips-stochastic,
  title     = {{Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution}},
  author    = {Covert, Ian and Kim, Chanwoo and Lee, Su-In and Zou, James and Hashimoto, Tatsunori},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-0143},
  url       = {https://mlanthology.org/neurips/2024/covert2024neurips-stochastic/}
}