A Probabilistic Model for Dirty Multi-Task Feature Selection

Abstract

Multi-task feature selection methods often make the hypothesis that learning tasks share relevant and irrelevant features. However, this hypothesis may be too restrictive in practice. For example, there may be a few tasks with specific relevant and irrelevant features (outlier tasks). Similarly, a few of the features may be relevant for only some of the tasks (outlier features). To account for this, we propose a model for multi-task feature selection based on a robust prior distribution that introduces a set of binary latent variables to identify outlier tasks and outlier features. Expectation propagation can be used for efficient approximate inference under the proposed prior. Several experiments show that a model based on the new robust prior provides better predictive performance than other benchmark methods.

Cite

Text

Hernandez-Lobato et al. "A Probabilistic Model for Dirty Multi-Task Feature Selection." International Conference on Machine Learning, 2015.

Markdown

[Hernandez-Lobato et al. "A Probabilistic Model for Dirty Multi-Task Feature Selection." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/hernandezlobato2015icml-probabilistic-a/)

BibTeX

@inproceedings{hernandezlobato2015icml-probabilistic-a,
  title     = {{A Probabilistic Model for Dirty Multi-Task Feature Selection}},
  author    = {Hernandez-Lobato, Daniel and Hernandez-Lobato, Jose Miguel and Ghahramani, Zoubin},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {1073-1082},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/hernandezlobato2015icml-probabilistic-a/}
}