Towards Deeper GCNs: Alleviating Over-Smoothing via Iterative Training and Fine-Tuning
Abstract
Graph Convolutional Networks (GCNs) suffer from severe performance degradation in deep architectures due to over-smoothing. While existing studies primarily attribute the over-smoothing to repeated applications of graph Laplacian operators, our empirical analysis reveals a critical yet overlooked factor: trainable linear transformations in GCNs significantly exacerbate feature collapse, even at moderate depths (e.g., 8 layers). In contrast, Simplified Graph Convolution (SGC), which removes these transformations, maintains stable feature diversity up to 32 layers, highlighting linear transformations’ dual role in facilitating expressive power and inducing over-smoothing. However, completely removing linear transformations weakens the model’s expressive capacity. To address this trade-off, we propose Layer-wise Gradual Training (LGT) , a novel training strategy that progressively builds deep GCNs while preserving their expressiveness. LGT integrates three complementary components: (1) layer-wise training to stabilize optimization from shallow to deep layers, (2) low-rank adaptation to fine-tune shallow layers and accelerate training, and (3) identity initialization to ensure smooth integration of new layers and accelerate convergence. Extensive experiments on benchmark datasets demonstrate that LGT achieves state-of-the-art performance on vanilla GCN, significantly improving accuracy even in 32-layer settings. Moreover, as a training method, LGT can be seamlessly combined with existing methods such as PairNorm and ContraNorm, further enhancing their performance in deeper networks. LGT offers a general, architecture-agnostic training framework for scalable deep GCNs. The code is available at https://github.com/jfklasdfj/LGT_GCN .
Cite
Text
Peng et al. "Towards Deeper GCNs: Alleviating Over-Smoothing via Iterative Training and Fine-Tuning." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2025. doi:10.1007/978-3-032-06066-2_13Markdown
[Peng et al. "Towards Deeper GCNs: Alleviating Over-Smoothing via Iterative Training and Fine-Tuning." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2025.](https://mlanthology.org/ecmlpkdd/2025/peng2025ecmlpkdd-deeper/) doi:10.1007/978-3-032-06066-2_13BibTeX
@inproceedings{peng2025ecmlpkdd-deeper,
title = {{Towards Deeper GCNs: Alleviating Over-Smoothing via Iterative Training and Fine-Tuning}},
author = {Peng, Furong and Gao, Jinzhen and Lu, Xuan and Liu, Kang and Huo, Yifan and Wang, Sheng},
booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
year = {2025},
pages = {211-227},
doi = {10.1007/978-3-032-06066-2_13},
url = {https://mlanthology.org/ecmlpkdd/2025/peng2025ecmlpkdd-deeper/}
}