Feature Selection at the Discrete Limit

Abstract

Feature selection plays an important role in many machine learning and data mining applications. In this paper, we propose to use L2,p norm for feature selection with emphasis on small p. As p approaches 0, feature selection becomes discrete feature selection problem. We provide two algorithms, proximal gradient algorithm and rank one update algorithm, which is more efficient at large regularization. We provide closed form solutions of the proximal operator at p = 0, 1/2. Experiments onreal life datasets show that features selected at small p consistently outperform features selected at p = 1, the standard L2,1 approach and other popular feature selection methods.

Cite

Text

Zhang et al. "Feature Selection at the Discrete Limit." AAAI Conference on Artificial Intelligence, 2014. doi:10.1609/AAAI.V28I1.8919

Markdown

[Zhang et al. "Feature Selection at the Discrete Limit." AAAI Conference on Artificial Intelligence, 2014.](https://mlanthology.org/aaai/2014/zhang2014aaai-feature/) doi:10.1609/AAAI.V28I1.8919

BibTeX

@inproceedings{zhang2014aaai-feature,
  title     = {{Feature Selection at the Discrete Limit}},
  author    = {Zhang, Miao and Ding, Chris H. Q. and Zhang, Ya and Nie, Feiping},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2014},
  pages     = {1355-1361},
  doi       = {10.1609/AAAI.V28I1.8919},
  url       = {https://mlanthology.org/aaai/2014/zhang2014aaai-feature/}
}