Flaky Performances When Pretraining on Relational Databases (Student Abstract)
Abstract
We explore the downstream task performances for graph neural network (GNN) self-supervised learning (SSL) methods trained on subgraphs extracted from relational databases (RDBs). Intuitively, this joint use of SSL and GNNs should allow to leverage more of the available data, which could translate to better results. However, we found that naively porting contrastive SSL techniques can cause ``negative transfer'': linear evaluation on fixed representation from a pretrained model performs worse than on representations from the randomly-initialized model. Based on the conjecture that contrastive SSL conflicts with the message passing layers of the GNN, we propose InfoNode: a contrastive loss aiming to maximize the mutual information between a node's initial- and final-layer representation. The primary empirical results support our conjecture and the effectiveness of InfoNode.
Cite
Text
Liu et al. "Flaky Performances When Pretraining on Relational Databases (Student Abstract)." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I13.26993Markdown
[Liu et al. "Flaky Performances When Pretraining on Relational Databases (Student Abstract)." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/liu2023aaai-flaky/) doi:10.1609/AAAI.V37I13.26993BibTeX
@inproceedings{liu2023aaai-flaky,
title = {{Flaky Performances When Pretraining on Relational Databases (Student Abstract)}},
author = {Liu, Shengchao and Vázquez, David and Tang, Jian and Noël, Pierre-André},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {16266-16267},
doi = {10.1609/AAAI.V37I13.26993},
url = {https://mlanthology.org/aaai/2023/liu2023aaai-flaky/}
}