Calibrate and Debias Layer-Wise Sampling for Graph Convolutional Networks
Abstract
Multiple sampling-based methods have been developed for approximating and accelerating node embedding aggregation in graph convolutional networks (GCNs) training. Among them, a layer-wise approach recursively performs importance sampling to select neighbors jointly for existing nodes in each layer. This paper revisits the approach from a matrix approximation perspective, and identifies two issues in the existing layer-wise sampling methods: suboptimal sampling probabilities and estimation biases induced by sampling without replacement. To address these issues, we accordingly propose two remedies: a new principle for constructing sampling probabilities and an efficient debiasing algorithm. The improvements are demonstrated by extensive analyses of estimation variance and experiments on common benchmarks. Code and algorithm implementations are publicly available at \url{https://github.com/ychen-stat-ml/GCN-layer-wise-sampling}.
Cite
Text
Chen et al. "Calibrate and Debias Layer-Wise Sampling for Graph Convolutional Networks." Transactions on Machine Learning Research, 2023.Markdown
[Chen et al. "Calibrate and Debias Layer-Wise Sampling for Graph Convolutional Networks." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/chen2023tmlr-calibrate/)BibTeX
@article{chen2023tmlr-calibrate,
title = {{Calibrate and Debias Layer-Wise Sampling for Graph Convolutional Networks}},
author = {Chen, Yifan and Xu, Tianning and Hakkani-Tur, Dilek and Jin, Di and Yang, Yun and Zhu, Ruoqing},
journal = {Transactions on Machine Learning Research},
year = {2023},
url = {https://mlanthology.org/tmlr/2023/chen2023tmlr-calibrate/}
}