Forward and Backward Knowledge Transfer for Sentiment Classification
Abstract
This paper studies the problem of learning a sequence of sentiment classification tasks. The learned knowledge from each task is retained and later used to help future or subsequent task learning. This learning paradigm is called \textit{lifelong learning}. However, existing lifelong learning methods either only transfer knowledge forward to help future learning and do not go back to improve the model of a previous task or require the training data of the previous task to retrain its model to exploit backward/reverse knowledge transfer. This paper studies reverse knowledge transfer of lifelong learning. It aims to improve the model of a previous task by leveraging future knowledge without retraining using its training data, which is challenging now. In this work, this is done by exploiting a key characteristic of the generative model of naïve Bayes. That is, it is possible to improve the naïve Bayesian classifier for a task by improving its model parameters directly using the retained knowledge from other tasks. Experimental results show that the proposed method markedly outperforms existing lifelong learning baselines.
Cite
Text
Wang et al. "Forward and Backward Knowledge Transfer for Sentiment Classification." Proceedings of The Eleventh Asian Conference on Machine Learning, 2019.Markdown
[Wang et al. "Forward and Backward Knowledge Transfer for Sentiment Classification." Proceedings of The Eleventh Asian Conference on Machine Learning, 2019.](https://mlanthology.org/acml/2019/wang2019acml-forward/)BibTeX
@inproceedings{wang2019acml-forward,
title = {{Forward and Backward Knowledge Transfer for Sentiment Classification}},
author = {Wang, Hao and Liu, Bing and Wang, Shuai and Ma, Nianzu and Yang, Yan},
booktitle = {Proceedings of The Eleventh Asian Conference on Machine Learning},
year = {2019},
pages = {457-472},
volume = {101},
url = {https://mlanthology.org/acml/2019/wang2019acml-forward/}
}