Learning by Interpreting

Abstract

This paper introduces a novel way of enhancing NLP prediction accuracy by incorporating model interpretation insights. Conventional efforts often focus on balancing the trade-offs between accuracy and interpretability, for instance, sacrificing model performance to increase the explainability. Here, we take a unique approach and show that model interpretation can ultimately help improve NLP quality. Specifically, we employ our learned interpretability results using attention mechanisms, LIME, and SHAP to train our model. We demonstrate a significant increase in accuracy of up to +3.4 BLEU points on NMT and up to +4.8 points on GLUE tasks, verifying our hypothesis that it is possible to achieve better model learning by incorporating model interpretation knowledge.

Cite

Text

Tang et al. "Learning by Interpreting." International Joint Conference on Artificial Intelligence, 2022. doi:10.24963/IJCAI.2022/609

Markdown

[Tang et al. "Learning by Interpreting." International Joint Conference on Artificial Intelligence, 2022.](https://mlanthology.org/ijcai/2022/tang2022ijcai-learning/) doi:10.24963/IJCAI.2022/609

BibTeX

@inproceedings{tang2022ijcai-learning,
  title     = {{Learning by Interpreting}},
  author    = {Tang, Xuting and Khan, Abdul Rafae and Wang, Shusen and Xu, Jia},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2022},
  pages     = {4390-4396},
  doi       = {10.24963/IJCAI.2022/609},
  url       = {https://mlanthology.org/ijcai/2022/tang2022ijcai-learning/}
}