Generalised Lipschitz Regularisation Equals Distributional Robustness
Abstract
The problem of adversarial examples has highlighted the need for a theory of regularisation that is general enough to apply to exotic function classes, such as universal approximators. In response, we have been able to significantly sharpen existing results regarding the relationship between distributional robustness and regularisation, when defined with a transportation cost uncertainty set. The theory allows us to characterise the conditions under which the distributional robustness equals a Lipschitz-regularised model, and to tightly quantify, for the first time, the slackness under very mild assumptions. As a theoretical application we show a new result explicating the connection between adversarial learning and distributional robustness. We then give new results for how to achieve Lipschitz regularisation of kernel classifiers, which are demonstrated experimentally.
Cite
Text
Cranko et al. "Generalised Lipschitz Regularisation Equals Distributional Robustness." International Conference on Machine Learning, 2021.Markdown
[Cranko et al. "Generalised Lipschitz Regularisation Equals Distributional Robustness." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/cranko2021icml-generalised/)BibTeX
@inproceedings{cranko2021icml-generalised,
title = {{Generalised Lipschitz Regularisation Equals Distributional Robustness}},
author = {Cranko, Zac and Shi, Zhan and Zhang, Xinhua and Nock, Richard and Kornblith, Simon},
booktitle = {International Conference on Machine Learning},
year = {2021},
pages = {2178-2188},
volume = {139},
url = {https://mlanthology.org/icml/2021/cranko2021icml-generalised/}
}