Passage Ranking with Weak Supervision
Abstract
In this paper, we propose a \textit{weak supervision} framework for neural ranking tasks based on the data programming paradigm \citep{Ratner2016}, which enables us to leverage multiple weak supervision signals from different sources. Empirically, we consider two sources of weak supervision signals, unsupervised ranking functions and semantic feature similarities. We train a BERT-based passage-ranking model (which achieves new state-of-the-art performances on two benchmark datasets with full supervision) in our weak supervision framework. Without using ground-truth training labels, BERT-PR models outperform BM25 baseline by a large margin on all three datasets and even beat the previous state-of-the-art results with full supervision on two of datasets.
Cite
Text
Xu et al. "Passage Ranking with Weak Supervision." ICLR 2019 Workshops: LLD, 2019.Markdown
[Xu et al. "Passage Ranking with Weak Supervision." ICLR 2019 Workshops: LLD, 2019.](https://mlanthology.org/iclrw/2019/xu2019iclrw-passage/)BibTeX
@inproceedings{xu2019iclrw-passage,
title = {{Passage Ranking with Weak Supervision}},
author = {Xu, Peng and Ma, Xiaofei and Nallapati, Ramesh and Xiang, Bing},
booktitle = {ICLR 2019 Workshops: LLD},
year = {2019},
url = {https://mlanthology.org/iclrw/2019/xu2019iclrw-passage/}
}