Towards Graph Foundation Models: Training on Knowledge Graphs Enables Transferability to General Graphs
Abstract
Inspired by the success of large language models, there is a trend toward developing graph foundation models to conduct diverse downstream tasks in various domains. However, current models often require extra fine-tuning to apply their learned structural and semantic representations to new graphs, which limits their versatility. Recent breakthroughs in zero-shot inductive reasoning on knowledge graphs (KGs), offer us a new perspective on extending KG reasoning to general graph applications. In this paper, we introduce SCR, a unified graph reasoning framework designed to train on knowledge graphs and effectively generalize across a wide range of graph tasks and domains. We begin by designing the task-specific KG structures to establish a unified topology for different task formats. Then we propose semantic-conditioned message passing, a novel mechanism addressing the inherent semantic isolation in traditional KG reasoning, by jointly modeling structural and semantic invariance patterns in graph representations. Evaluated on 38 diverse datasets spanning node-, link-, and graph-level tasks, SCR achieves substantial performance gains over existing foundation models and supervised baselines, demonstrating its remarkable efficacy and adaptability.
Cite
Text
Wang et al. "Towards Graph Foundation Models: Training on Knowledge Graphs Enables Transferability to General Graphs." Advances in Neural Information Processing Systems, 2025.Markdown
[Wang et al. "Towards Graph Foundation Models: Training on Knowledge Graphs Enables Transferability to General Graphs." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/wang2025neurips-graph/)BibTeX
@inproceedings{wang2025neurips-graph,
title = {{Towards Graph Foundation Models: Training on Knowledge Graphs Enables Transferability to General Graphs}},
author = {Wang, Kai and Luo, Siqiang and Shan, Caihua and Shen, Yifei},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/wang2025neurips-graph/}
}