MORE: Molecule Pretraining with Multi-Level Pretext Task
Abstract
Foundation models, serving as pretrained fundamental bases for a variety of downstream tasks, try to learn versatile, rich, and generalizable representations that can be quickly adopted through fine-tuning or even in a zero-shot manner for specific applications. Foundation models for molecular representation are no exception. Various pretext tasks have been proposed for pretraining molecular representations, but these approaches have focused on only single or partial properties. Molecules are complicated and require different perspectives depending on purposes: insights from local- or global-level, 2D-topology or 3D-spatial arrangement, and low- or high-level semantics. We propose Multi-level mOlecule gRaph prE-train (MORE) to consider these multiple aspects of molecules simultaneously. Experimental results demonstrate that our proposed method effectively learns comprehensive representations by showing outstanding performance in both linear probing and full fine-tuning. Notably, in quantification experiments of forgetting the pretrained models, MORE consistently exhibits minimal and stable parameter changes with the smallest performance gap, whereas other methods show substantial and inconsistent fluctuations with larger gaps. The effectiveness of individual pretext tasks varies depending on the problems being solved, which again highlights the need for a multi-level perspective. Scalability experiments reveal steady improvements of MORE as the dataset size increases, suggesting potential gains with larger datasets as well.
Cite
Text
Son et al. "MORE: Molecule Pretraining with Multi-Level Pretext Task." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I19.34262Markdown
[Son et al. "MORE: Molecule Pretraining with Multi-Level Pretext Task." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/son2025aaai-more/) doi:10.1609/AAAI.V39I19.34262BibTeX
@inproceedings{son2025aaai-more,
title = {{MORE: Molecule Pretraining with Multi-Level Pretext Task}},
author = {Son, Yeongyeong and Noh, Dasom and Heo, Gyoungyoung and Park, Gyoung Jin and Kwon, Sunyoung},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {20531-20539},
doi = {10.1609/AAAI.V39I19.34262},
url = {https://mlanthology.org/aaai/2025/son2025aaai-more/}
}