Feature Selection in a Kernel Space

Abstract

We address the problem of feature selection in a kernel space to select the most discriminative and informative features for classification and data analysis. This is a difficult problem because the dimension of a kernel space may be infinite. In the past, little work has been done on feature selection in a kernel space. To solve this problem, we derive a basis set in the kernel space as a first step for feature selection. Using the basis set, we then extend the margin-based feature selection algorithms that are proven effective even when many features are dependent. The selected features form a subspace of the kernel space, in which different state-of-the-art classification algorithms can be applied for classification. We conduct extensive experiments over real and simulated data to compare our proposed method with four baseline algorithms. Both theoretical analysis and experimental results validate the effectiveness of our proposed method.

Cite

Text

Cao et al. "Feature Selection in a Kernel Space." International Conference on Machine Learning, 2007. doi:10.1145/1273496.1273512

Markdown

[Cao et al. "Feature Selection in a Kernel Space." International Conference on Machine Learning, 2007.](https://mlanthology.org/icml/2007/cao2007icml-feature/) doi:10.1145/1273496.1273512

BibTeX

@inproceedings{cao2007icml-feature,
  title     = {{Feature Selection in a Kernel Space}},
  author    = {Cao, Bin and Shen, Dou and Sun, Jian-Tao and Yang, Qiang and Chen, Zheng},
  booktitle = {International Conference on Machine Learning},
  year      = {2007},
  pages     = {121-128},
  doi       = {10.1145/1273496.1273512},
  url       = {https://mlanthology.org/icml/2007/cao2007icml-feature/}
}