Bi-Level Contrastive Learning for Knowledge-Enhanced Molecule Representations

Abstract

Molecular representation learning is vital for various downstream applications, including the analysis and prediction of molecular properties and side effects. While Graph Neural Networks (GNNs) have been a popular framework for modeling molecular data, they often struggle to capture the full complexity of molecular representations. In this paper, we introduce a novel method called Gode, which accounts for the dual-level structure inherent in molecules. Molecules possess an intrinsic graph structure and simultaneously function as nodes within a broader molecular knowledge graph. Gode integrates individual molecular graph representations with multi-domain biochemical data from knowledge graphs. By pre-training two GNNs on different graph structures and employing contrastive learning, Gode effectively fuses molecular structures with their corresponding knowledge graph substructures. This fusion yields a more robust and informative representation, enhancing molecular property predictions by leveraging both chemical and biological information. When fine-tuned across 11 chemical property tasks, our model significantly outperforms existing benchmarks, achieving an average ROC-AUC improvement of 12.7% for classification tasks and an average RMSE/MAE improvement of 34.4% for regression tasks. Notably, Gode surpasses the current leading model in property prediction, with advancements of 2.2% in classification and 7.2% in regression tasks.

Cite

Text

Jiang et al. "Bi-Level Contrastive Learning for Knowledge-Enhanced Molecule Representations." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I1.32013

Markdown

[Jiang et al. "Bi-Level Contrastive Learning for Knowledge-Enhanced Molecule Representations." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/jiang2025aaai-bi/) doi:10.1609/AAAI.V39I1.32013

BibTeX

@inproceedings{jiang2025aaai-bi,
  title     = {{Bi-Level Contrastive Learning for Knowledge-Enhanced Molecule Representations}},
  author    = {Jiang, Pengcheng and Xiao, Cao and Fu, Tianfan and Bhatia, Parminder and Kass-Hout, Taha A. and Sun, Jimeng and Han, Jiawei},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {352-360},
  doi       = {10.1609/AAAI.V39I1.32013},
  url       = {https://mlanthology.org/aaai/2025/jiang2025aaai-bi/}
}