Substructure-Atom Cross Attention for Molecular Representation Learning

Abstract

Designing a neural network architecture for molecular representation is crucial for AI-driven drug discovery and molecule design. In this work, we propose a new framework for molecular representation learning. Our contribution is threefold: (a) demonstrating the usefulness of incorporating substructures to node-wise features from molecules, (b) designing two branch networks consisting of a transformer and a graph neural network so that the networks fused with asymmetric attention, and (c) not requiring heuristic features and computationally-expensive information from molecules. Using 1.8 million molecules collected from ChEMBL and PubChem database, we pretrain our network to learn a general representation of molecules with minimal supervision. The experimental results show that our pretrained network achieves competitive performance on 11 downstream tasks for molecular property prediction.

Cite

Text

Kim et al. "Substructure-Atom Cross Attention for Molecular Representation Learning." NeurIPS 2022 Workshops: AI4Science, 2022.

Markdown

[Kim et al. "Substructure-Atom Cross Attention for Molecular Representation Learning." NeurIPS 2022 Workshops: AI4Science, 2022.](https://mlanthology.org/neuripsw/2022/kim2022neuripsw-substructureatom/)

BibTeX

@inproceedings{kim2022neuripsw-substructureatom,
  title     = {{Substructure-Atom Cross Attention for Molecular Representation Learning}},
  author    = {Kim, Jiye and Lee, Seungbeom and Kim, Dongwoo and Ahn, Sungsoo and Park, Jaesik},
  booktitle = {NeurIPS 2022 Workshops: AI4Science},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/kim2022neuripsw-substructureatom/}
}