Evasion and Hardening of Tree Ensemble Classifiers

Abstract

Classifier evasion consists in finding for a given instance x the “nearest” instance x’ such that the classifier predictions of x and x’ are different. We present two novel algorithms for systematically computing evasions for tree ensembles such as boosted trees and random forests. Our first algorithm uses a Mixed Integer Linear Program solver and finds the optimal evading instance under an expressive set of constraints. Our second algorithm trades off optimality for speed by using symbolic prediction, a novel algorithm for fast finite differences on tree ensembles. On a digit recognition task, we demonstrate that both gradient boosted trees and random forests are extremely susceptible to evasions. Finally, we harden a boosted tree model without loss of predictive accuracy by augmenting the training set of each boosting round with evading instances, a technique we call adversarial boosting.

Cite

Text

Kantchelian et al. "Evasion and Hardening of Tree Ensemble Classifiers." International Conference on Machine Learning, 2016.

Markdown

[Kantchelian et al. "Evasion and Hardening of Tree Ensemble Classifiers." International Conference on Machine Learning, 2016.](https://mlanthology.org/icml/2016/kantchelian2016icml-evasion/)

BibTeX

@inproceedings{kantchelian2016icml-evasion,
  title     = {{Evasion and Hardening of Tree Ensemble Classifiers}},
  author    = {Kantchelian, Alex and Tygar, J. D. and Joseph, Anthony},
  booktitle = {International Conference on Machine Learning},
  year      = {2016},
  pages     = {2387-2396},
  volume    = {48},
  url       = {https://mlanthology.org/icml/2016/kantchelian2016icml-evasion/}
}