On Scalable Testing of Samplers
Abstract
In this paper we study the problem of testing of constrained samplers over high-dimensional distributions with $(\varepsilon,\eta,\delta)$ guarantees. Samplers are increasingly used in a wide range of safety-critical ML applications, and hence the testing problem has gained importance. For $n$-dimensional distributions, the existing state-of-the-art algorithm, $\mathsf{Barbarik2}$, has a worst case query complexity of exponential in $n$ and hence is not ideal for use in practice. Our primary contribution is an exponentially faster algorithm, $\mathsf{Barbarik3}$, that has a query complexity linear in $n$ and hence can easily scale to larger instances. We demonstrate our claim by implementing our algorithm and then comparing it against $\mathsf{Barbarik2}$. Our experiments on the samplers $\mathsf{wUnigen3}$ and $\mathsf{wSTS}$, find that $\mathsf{Barbarik3}$ requires $10\times$ fewer samples for $\mathsf{wUnigen3}$ and $450\times$ fewer samples for $\mathsf{wSTS}$ as compared to $\mathsf{Barbarik2}$.
Cite
Text
Pote and Meel. "On Scalable Testing of Samplers." Neural Information Processing Systems, 2022.Markdown
[Pote and Meel. "On Scalable Testing of Samplers." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/pote2022neurips-scalable/)BibTeX
@inproceedings{pote2022neurips-scalable,
title = {{On Scalable Testing of Samplers}},
author = {Pote, Yash and Meel, Kuldeep S},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/pote2022neurips-scalable/}
}