Co-Supervised Pre-Training of Pocket and Ligand

Abstract

Can we inject the pocket-ligand complementarity knowledge into the pre-trained model and jointly learn their chemical space? Pre-training molecules and proteins have attracted considerable attention in recent years, while most of these approaches focus on learning one of the chemical spaces and lack the consideration of their complementarity. We propose a co-supervised pre-training (CoSP) framework to learn 3D pocket and ligand representations simultaneously. We use a gated geometric message passing layer to model 3D pockets and ligands, where each node’s chemical features, geometric position, and direction are considered. To learn meaningful biological embeddings, we inject the pocket-ligand complementarity into the pre-training model via ChemInfoNCE loss, cooperating with a chemical similarity-enhanced negative sampling strategy to improve the representation learning. Through extensive experiments, we conclude that CoSP can achieve competitive results in pocket matching, molecule property prediction, and virtual screening.

Cite

Text

Gao et al. "Co-Supervised Pre-Training of Pocket and Ligand." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2023. doi:10.1007/978-3-031-43412-9_24

Markdown

[Gao et al. "Co-Supervised Pre-Training of Pocket and Ligand." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2023.](https://mlanthology.org/ecmlpkdd/2023/gao2023ecmlpkdd-cosupervised/) doi:10.1007/978-3-031-43412-9_24

BibTeX

@inproceedings{gao2023ecmlpkdd-cosupervised,
  title     = {{Co-Supervised Pre-Training of Pocket and Ligand}},
  author    = {Gao, Zhangyang and Tan, Cheng and Xia, Jun and Li, Stan Z.},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2023},
  pages     = {405-421},
  doi       = {10.1007/978-3-031-43412-9_24},
  url       = {https://mlanthology.org/ecmlpkdd/2023/gao2023ecmlpkdd-cosupervised/}
}