Optimal Learning from Verified Training Data
Abstract
Standard machine learning algorithms typically assume that data is sampled independently from the distribution of interest. In attempts to relax this assumption, fields such as adversarial learning typically assume that data is provided by an adversary, whose sole objective is to fool a learning algorithm. However, in reality, it is often the case that data comes from self-interested agents, with less malicious goals and intentions which lie somewhere between the two settings described above. To tackle this problem, we present a Stackelberg competition model for least squares regression, in which data is provided by agents who wish to achieve specific predictions for their data. Although the resulting optimisation problem is nonconvex, we derive an algorithm which converges globally, outperforming current approaches which only guarantee convergence to local optima. We also provide empirical results on two real-world datasets, the medical personal costs dataset and the red wine dataset, showcasing the performance of our algorithm relative to algorithms which are optimal under adversarial assumptions, outperforming the state of the art.
Cite
Text
Bishop et al. "Optimal Learning from Verified Training Data." Neural Information Processing Systems, 2020.Markdown
[Bishop et al. "Optimal Learning from Verified Training Data." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/bishop2020neurips-optimal/)BibTeX
@inproceedings{bishop2020neurips-optimal,
title = {{Optimal Learning from Verified Training Data}},
author = {Bishop, Nicholas and Tran-Thanh, Long and Gerding, Enrico},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/bishop2020neurips-optimal/}
}