Learning with Marginalized Corrupted Features and Labels Together

Abstract

Tagging has become increasingly important in many real-world applications noticeably including web applications, such as web blogs and resource sharing systems. Despite this importance, tagging methods often face difficult challenges such as limited training samples and incomplete labels, which usually lead to degenerated performance on tag prediction. To improve the generalization performance, in this paper, we propose Regularized Marginalized Cross-View learning (RMCV) by jointly modeling on attribute noise and label noise. In more details, the proposed model constructs infinite training examples with attribute noises from known exponential-family distributions and exploits label noise via marginalized denoising autoencoder. Therefore, the model benefits from its robustness and alleviates the problem of tag sparsity. While RMCV is a general method for learning tagging, in the evaluations we focus on the specific application of multi-label text tagging. Extensive evaluations on three benchmark data sets demonstrate that RMCV outstands with a superior performance in comparison with state-of-the-art methods.

Cite

Text

Li et al. "Learning with Marginalized Corrupted Features and Labels Together." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10152

Markdown

[Li et al. "Learning with Marginalized Corrupted Features and Labels Together." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/li2016aaai-learning/) doi:10.1609/AAAI.V30I1.10152

BibTeX

@inproceedings{li2016aaai-learning,
  title     = {{Learning with Marginalized Corrupted Features and Labels Together}},
  author    = {Li, Yingming and Yang, Ming and Xu, Zenglin and Zhang, Zhongfei (Mark)},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {1251-1257},
  doi       = {10.1609/AAAI.V30I1.10152},
  url       = {https://mlanthology.org/aaai/2016/li2016aaai-learning/}
}