Adaptive Pre-Training of Language Models for Better Logical Reasoning

Abstract

Logical reasoning of text is an important ability that requires understanding the logical information present in the text and reasoning through them to infer new conclusions. Prior works on improving the logical reasoning ability of language models require complex processing of training data (e.g., aligning symbolic knowledge to text), yielding task-specific data augmentation solutions that restrict the learning of general logical reasoning skills. In this work, we propose AERIE, an adaptively pre-trained language model that has improved logical reasoning abilities. We select a subset of Wikipedia, based on a set of logical inference keywords, for continued pretraining of a language model. We use two self-supervised loss functions: a modified masked language modeling loss where only specific parts-of-speech words, that would likely require more reasoning than basic language understanding, are masked, and a sentence classification loss that teaches the model to distinguish between entailment and contradiction types of sentences. The proposed training paradigm is both simple and generalizable across tasks. We demonstrate the effectiveness of AERIE by comparing it with prior baselines on two logical reasoning datasets. AERIE performs comparably on ReClor and outperforms baselines on LogiQA.

Cite

Text

Sanyal et al. "Adaptive Pre-Training of Language Models for Better Logical Reasoning." NeurIPS 2022 Workshops: DistShift, 2022.

Markdown

[Sanyal et al. "Adaptive Pre-Training of Language Models for Better Logical Reasoning." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/sanyal2022neuripsw-adaptive/)

BibTeX

@inproceedings{sanyal2022neuripsw-adaptive,
  title     = {{Adaptive Pre-Training of Language Models for Better Logical Reasoning}},
  author    = {Sanyal, Soumya and Xu, Yichong and Wang, Shuohang and Yang, Ziyi and Pryzant, Reid and Yu, Wenhao and Zhu, Chenguang and Ren, Xiang},
  booktitle = {NeurIPS 2022 Workshops: DistShift},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/sanyal2022neuripsw-adaptive/}
}