Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models
Abstract
Generation of plausible but incorrect factual information, often termed hallucination, has attracted significant research interest. Retrieval-augmented language model (RALM)—which enhances models with up-to-date knowledge—emerges as a promising method to reduce hallucination. However, existing RALMs may instead exacerbate hallucination when retrieving lengthy contexts. To address this challenge, we propose COFT, a novel COarse-to-Fine highlighTing method to focus on different granularity-level key texts, thereby avoiding getting lost in lengthy contexts. Specifically, COFT consists of three components: recaller, scorer, and selector. First, recaller applies a knowledge graph to extract potential key entities in a given context. Second, scorer measures the importance of each entity by calculating its contextual weight. Finally, selector selects high contextual weight entities with a dynamic threshold algorithm and highlights the corresponding paragraphs, sentences, or words in a coarse-to-fine manner. Extensive experiments on knowledge hallucination benchmark demonstrate the effectiveness of COFT, leading to a superior performance over 30% in F1 score metric. Moreover, COFT also exhibits remarkable versatility across various long-form tasks, such as reading comprehension and question answering.
Cite
Text
Lv et al. "Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models." International Conference on Machine Learning, 2024.Markdown
[Lv et al. "Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/lv2024icml-coarsetofine/)BibTeX
@inproceedings{lv2024icml-coarsetofine,
title = {{Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models}},
author = {Lv, Qitan and Wang, Jie and Chen, Hanzhu and Li, Bin and Zhang, Yongdong and Wu, Feng},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {33594-33623},
volume = {235},
url = {https://mlanthology.org/icml/2024/lv2024icml-coarsetofine/}
}