Feature Selection Filters Based on the Permutation Test

Abstract

We investigate the problem of supervised feature selection within the filtering framework. In our approach, applicable to the two-class problems, the feature strength is inversely proportional to the p -value of the null hypothesis that its class-conditional densities, p ( X | Y = 0) and p ( X | Y = 1), are identical. To estimate the p-values, we use Fisher’s permutation test combined with the four simple filtering criteria in the roles of test statistics: sample mean difference, symmetric Kullback-Leibler distance, information gain, and chi-square statistic. The experimental results of our study, performed using naive Bayes classifier and support vector machines, strongly indicate that the permutation test improves the above-mentioned filters and can be used effectively when sample size is relatively small and number of features relatively large.

Cite

Text

Radivojac et al. "Feature Selection Filters Based on the Permutation Test." European Conference on Machine Learning, 2004. doi:10.1007/978-3-540-30115-8_32

Markdown

[Radivojac et al. "Feature Selection Filters Based on the Permutation Test." European Conference on Machine Learning, 2004.](https://mlanthology.org/ecmlpkdd/2004/radivojac2004ecml-feature/) doi:10.1007/978-3-540-30115-8_32

BibTeX

@inproceedings{radivojac2004ecml-feature,
  title     = {{Feature Selection Filters Based on the Permutation Test}},
  author    = {Radivojac, Predrag and Obradovic, Zoran and Dunker, A. Keith and Vucetic, Slobodan},
  booktitle = {European Conference on Machine Learning},
  year      = {2004},
  pages     = {334-346},
  doi       = {10.1007/978-3-540-30115-8_32},
  url       = {https://mlanthology.org/ecmlpkdd/2004/radivojac2004ecml-feature/}
}