Distribution-Aware Online Classifiers

Abstract

We propose a family of Passive-Aggressive Mahalanobis (PAM) algorithms, which are incremental (online) binary classifiers that consider the distribution of data. PAM is in fact a generalization of the Passive-Aggressive (PA) algorithms to handle data distributions that can be represented by a covariance matrix. The update equations for PAM are derived and theoretical error loss bounds computed. We benchmarked PAM against the original PA-I, PA-II, and Confidence Weighted (CW) learning. Although PAM somewhat resembles CW in its update equations, PA minimizes differences in the weights while CW minimizes differences in weight distributions. Results on 8 classification datasets, which include a real-life micro-blog sentiment classification task, show that PAM consistently outperformed its competitors, most notably CW. This shows that a simple approach like PAM is more practical in real-life classification tasks, compared to more elegant and sophisticated approaches like CW.

Cite

Text

Nguyen et al. "Distribution-Aware Online Classifiers." International Joint Conference on Artificial Intelligence, 2011. doi:10.5591/978-1-57735-516-8/IJCAI11-241

Markdown

[Nguyen et al. "Distribution-Aware Online Classifiers." International Joint Conference on Artificial Intelligence, 2011.](https://mlanthology.org/ijcai/2011/nguyen2011ijcai-distribution/) doi:10.5591/978-1-57735-516-8/IJCAI11-241

BibTeX

@inproceedings{nguyen2011ijcai-distribution,
  title     = {{Distribution-Aware Online Classifiers}},
  author    = {Nguyen, Tam T. and Chang, Kuiyu and Hui, Siu Cheung},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2011},
  pages     = {1427-1432},
  doi       = {10.5591/978-1-57735-516-8/IJCAI11-241},
  url       = {https://mlanthology.org/ijcai/2011/nguyen2011ijcai-distribution/}
}