Fine-Grained Distillation for Long Document Retrieval

Abstract

Long document retrieval aims to fetch query-relevant documents from a large-scale collection, where knowledge distillation has become de facto to improve a retriever by mimicking a heterogeneous yet powerful cross-encoder. However, in contrast to passages or sentences, retrieval on long documents suffers from the \textit{scope hypothesis} that a long document may cover multiple topics. This maximizes their structure heterogeneity and poses a granular-mismatch issue, leading to an inferior distillation efficacy. In this work, we propose a new learning framework, fine-grained distillation (FGD), for long-document retrievers. While preserving the conventional dense retrieval paradigm, it first produces global-consistent representations crossing different fine granularity and then applies multi-granular aligned distillation merely during training. In experiments, we evaluate our framework on two long-document retrieval benchmarks, which show state-of-the-art performance.

Cite

Text

Zhou et al. "Fine-Grained Distillation for Long Document Retrieval." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I17.29947

Markdown

[Zhou et al. "Fine-Grained Distillation for Long Document Retrieval." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/zhou2024aaai-fine/) doi:10.1609/AAAI.V38I17.29947

BibTeX

@inproceedings{zhou2024aaai-fine,
  title     = {{Fine-Grained Distillation for Long Document Retrieval}},
  author    = {Zhou, Yucheng and Shen, Tao and Geng, Xiubo and Tao, Chongyang and Shen, Jianbing and Long, Guodong and Xu, Can and Jiang, Daxin},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {19732-19740},
  doi       = {10.1609/AAAI.V38I17.29947},
  url       = {https://mlanthology.org/aaai/2024/zhou2024aaai-fine/}
}