On Incorporating Prior Knowledge Extracted from Pre-Trained Language Models into Causal Discovery
Abstract
Pre-trained Language Models (PLMs) can reason about causality by leveraging vast pre-trained knowledge and text descriptions of datasets, proving their effectiveness even when data is scarce. However, there are crucial limitations in current PLM-based causal reasoning methods: i) PLM cannot utilize large datasets in prompt due to the limits of context length, and ii) the methods are not adept at comprehending the whole interconnected causal structures. On the other hand, data-driven causal discovery can discover the causal structure as a whole, although it works well only when the number of data observations is sufficiently large enough. To overcome each other approaches' limitations, we propose a new framework that integrates PLMs-based causal reasoning into data-driven causal discovery, resulting in improved and robust performance. Furthermore, our framework extends to the time-series data and exhibits superior performance.
Cite
Text
Lee et al. "On Incorporating Prior Knowledge Extracted from Pre-Trained Language Models into Causal Discovery." NeurIPS 2024 Workshops: CALM, 2024.Markdown
[Lee et al. "On Incorporating Prior Knowledge Extracted from Pre-Trained Language Models into Causal Discovery." NeurIPS 2024 Workshops: CALM, 2024.](https://mlanthology.org/neuripsw/2024/lee2024neuripsw-incorporating/)BibTeX
@inproceedings{lee2024neuripsw-incorporating,
title = {{On Incorporating Prior Knowledge Extracted from Pre-Trained Language Models into Causal Discovery}},
author = {Lee, Chanhui and Kim, Juhyeon and Jeong, YongJun and Yeom, Yoonseok and Lyu, Juhyun and Kim, Jung-Hee and Lee, Sangmin and Han, Sangjun and Choe, Hyeokjun and Park, Soyeon and Lim, Woohyung and Bae, Kyunghoon and Lim, Sungbin and Lee, Sanghack},
booktitle = {NeurIPS 2024 Workshops: CALM},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/lee2024neuripsw-incorporating/}
}