Binary Classification Under Local Label Differential Privacy Using Randomized Response Mechanisms
Abstract
Label differential privacy is a popular branch of $\epsilon$-differential privacy for protecting labels in training datasets with non-private features. In this paper, we study the generalization performance of a binary classifier trained on a dataset privatized under the label differential privacy achieved by the randomized response mechanism. Particularly, we establish minimax lower bounds for the excess risks of the deep neural network plug-in classifier, theoretically quantifying how privacy guarantee $\epsilon$ affects its generalization performance. Our theoretical result shows: (1) the randomized response mechanism slows down the convergence of excess risk by lessening the multiplicative constant term compared with the non-private case $(\epsilon=\infty)$; (2) as $\epsilon$ decreases, the optimal structure of the neural network should be smaller for better generalization performance; (3) the convergence of its excess risk is guaranteed even if $\epsilon$ is adaptive to the size of training sample $n$ at a rate slower than $O(n^{-1/2})$. Our theoretical results are validated by extensive simulated examples and two real applications.
Cite
Text
Xu et al. "Binary Classification Under Local Label Differential Privacy Using Randomized Response Mechanisms." Transactions on Machine Learning Research, 2023.Markdown
[Xu et al. "Binary Classification Under Local Label Differential Privacy Using Randomized Response Mechanisms." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/xu2023tmlr-binary/)BibTeX
@article{xu2023tmlr-binary,
title = {{Binary Classification Under Local Label Differential Privacy Using Randomized Response Mechanisms}},
author = {Xu, Shirong and Wang, Chendi and Sun, Will Wei and Cheng, Guang},
journal = {Transactions on Machine Learning Research},
year = {2023},
url = {https://mlanthology.org/tmlr/2023/xu2023tmlr-binary/}
}