ThinkEval: Practical Evaluation of Knowledge Leakage in LLM Editing Using Thought-Based Knowledge Graphs
Abstract
Robust model-editing techniques are essential for deploying large language models (LLMs) in practical applications, as they enable cost-effective ways to deal with challenges such as privacy breaches, bias mitigation and misinformation spread. For example, an LLM-based healthcare assistance may need to update out-dated or incorrect knowledge to prevent harmful recommendations. However, many editing techniques focus on isolated facts, which critically fail to prevent indirect knowledge leakage---the unintended reconstruction of edited-out information through persistent causal links and contextual relationships. To assist users in selecting the right editing technique, we develop and present ThinkEval, a framework to systematically quantify indirect knowledge leakage and ripple effects in model-editing. ThinkEval builds and employs specialized knowledge graphs to analyze the causal structure of facts before and after editing. To support this approach, we present KnowGIC, a benchmark dataset comprising multi-step reasoning paths that precisely measure these complex knowledge transformation effects. We evaluate five editing techniques: AlphaEdit, RECT, ROME, MEMIT, and PRUNE across multiple LLMs. Our results show that these techniques struggle to balance indirect fact suppression with the preservation of related knowledge, compromising the contextual integrity of a model's knowledge. Our dataset is available at: https://github.com/manitbaser/KnowGIC.
Cite
Text
Baser et al. "ThinkEval: Practical Evaluation of Knowledge Leakage in LLM Editing Using Thought-Based Knowledge Graphs." Transactions on Machine Learning Research, 2026.Markdown
[Baser et al. "ThinkEval: Practical Evaluation of Knowledge Leakage in LLM Editing Using Thought-Based Knowledge Graphs." Transactions on Machine Learning Research, 2026.](https://mlanthology.org/tmlr/2026/baser2026tmlr-thinkeval/)BibTeX
@article{baser2026tmlr-thinkeval,
title = {{ThinkEval: Practical Evaluation of Knowledge Leakage in LLM Editing Using Thought-Based Knowledge Graphs}},
author = {Baser, Manit and Divakaran, Dinil Mon and Gurusamy, Mohan},
journal = {Transactions on Machine Learning Research},
year = {2026},
url = {https://mlanthology.org/tmlr/2026/baser2026tmlr-thinkeval/}
}