DropEdge: Towards Deep Graph Convolutional Networks on Node Classification

ICLR 2020  ·  Yu Rong, Wenbing Huang, Tingyang Xu, Junzhou Huang ·

\emph{Over-fitting} and \emph{over-smoothing} are two main obstacles of developing deep Graph Convolutional Networks (GCNs) for node classification. In particular, over-fitting weakens the generalization ability on small dataset, while over-smoothing impedes model training by isolating output representations from the input features with the increase in network depth. This paper proposes DropEdge, a novel and flexible technique to alleviate both issues. At its core, DropEdge randomly removes a certain number of edges from the input graph at each training epoch, acting like a data augmenter and also a message passing reducer. Furthermore, we theoretically demonstrate that DropEdge either reduces the convergence speed of over-smoothing or relieves the information loss caused by it. More importantly, our DropEdge is a general skill that can be equipped with many other backbone models (e.g. GCN, ResGCN, GraphSAGE, and JKNet) for enhanced performance. Extensive experiments on several benchmarks verify that DropEdge consistently improves the performance on a variety of both shallow and deep GCNs. The effect of DropEdge on preventing over-smoothing is empirically visualized and validated as well. Codes are released on~\url{https://github.com/DropEdge/DropEdge}.

PDF Abstract ICLR 2020 PDF ICLR 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Node Classification Citeseer Full-supervised IncepGCN+DropEdge Accuracy 80.50% # 1
Node Classification Cora Full-supervised IncepGCN+DropEdge Accuracy 88.2% # 2
Node Classification Pubmed Full-supervised GraphSAGE+DropEdge Accuracy 91.70% # 1
Node Classification Reddit JKNet+DropEdge Accuracy 97.02% # 5

Methods