Passage Ranking with Weak Supervision

Abstract

In this paper, we propose a \textit{weak supervision} framework for neural ranking tasks based on the data programming paradigm \citep{Ratner2016}, which enables us to leverage multiple weak supervision signals from different sources. Empirically, we consider two sources of weak supervision signals, unsupervised ranking functions and semantic feature similarities. We train a BERT-based passage-ranking model (which achieves new state-of-the-art performances on two benchmark datasets with full supervision) in our weak supervision framework. Without using ground-truth training labels, BERT-PR models outperform BM25 baseline by a large margin on all three datasets and even beat the previous state-of-the-art results with full supervision on two of datasets.

Cite

Text

Xu et al. "Passage Ranking with Weak Supervision." ICLR 2019 Workshops: LLD, 2019.

Markdown

[Xu et al. "Passage Ranking with Weak Supervision." ICLR 2019 Workshops: LLD, 2019.](https://mlanthology.org/iclrw/2019/xu2019iclrw-passage/)

BibTeX

@inproceedings{xu2019iclrw-passage,
  title     = {{Passage Ranking with Weak Supervision}},
  author    = {Xu, Peng and Ma, Xiaofei and Nallapati, Ramesh and Xiang, Bing},
  booktitle = {ICLR 2019 Workshops: LLD},
  year      = {2019},
  url       = {https://mlanthology.org/iclrw/2019/xu2019iclrw-passage/}
}