Logically Consistent Language Models via Neuro-Symbolic Integration
Abstract
Large language models (LLMs) are a promising venue for natural language understanding and generation. However, current LLMs are far from reliable: they are prone to generating non-factual information and, more crucially, to contradicting themselves when prompted to reason about relations between entities of the world. These problems are currently addressed with large scale fine-tuning or by delegating reasoning to external tools. In this work, we strive for a middle ground and introduce a loss based on neuro-symbolic reasoning that teaches an LLM to be logically consistent with an external set of facts and rules and improves self-consistency even when the LLM is fine-tuned on a limited set of facts. Our approach also allows to easily combine multiple logical constraints at once in a principled way, delivering LLMs that are more consistent w.r.t. all constraints and improve over several baselines w.r.t. a given constraint. Moreover, our method allows LLMs to extrapolate to unseen but semantically similar factual knowledge, represented in unseen datasets, more systematically.
Cite
Text
Calanzone et al. "Logically Consistent Language Models via Neuro-Symbolic Integration." NeurIPS 2024 Workshops: Sys2-Reasoning, 2024.Markdown
[Calanzone et al. "Logically Consistent Language Models via Neuro-Symbolic Integration." NeurIPS 2024 Workshops: Sys2-Reasoning, 2024.](https://mlanthology.org/neuripsw/2024/calanzone2024neuripsw-logically/)BibTeX
@inproceedings{calanzone2024neuripsw-logically,
title = {{Logically Consistent Language Models via Neuro-Symbolic Integration}},
author = {Calanzone, Diego and Teso, Stefano and Vergari, Antonio},
booktitle = {NeurIPS 2024 Workshops: Sys2-Reasoning},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/calanzone2024neuripsw-logically/}
}