Graph Random Neural Networks for
Semi-Supervised Learning on Graphs
Presented by:
Gauransh Sawhney 2018A3PS0325P
Utkarsh Kumar Singh 2018A3PS0368P
In partial fulfillment of the requirements of the course:
BITS F464 Machine Learning
Submitted to: Dr. Kamlesh Tiwari
Semi-Supervised Learning
Semi-Supervised Learning on Graphs
Graph Neural Networks
Graph Neural Network
Existing Issues
GRAND: Graph Random Neural Network
GRAND
GRAND
Algorithm
Loss Functions
Results
Some of the results presented in the paper:
Dataset Description
Dataset | Nodes | Edges | Train/Valid/Test Nodes | Classes | Features | Default Label Rate |
Cora | 2708 | 5429 | 140/500/1000 | 7 | 1433 | 0.052 |
Citeseer | 3327 | 4732 | 120/500/1000 | 6 | 3703 | 0.036 |
Pubmed | 19717 | 44338 | 60/500/1000 | 3 | 500 | 0.003 |
3 datasets were used to benchmark results
Comparison with existing architectures
Generalization Analysis
(b) Without CR
(c) GRAND(with RP and CR)
Robustness Analysis
Over-smoothing analysis
Other results presented in the paper
Over-smoothness of GRAND and its variants(on Cora)
Other results presented in the paper
Classification Accuracy of GRAND on large datasets
Experiments
The following experiments were conducted:
Effect of using an MLP vs GCN as the classification network
GRAND very clearly outperforms
GRAND_GCN in terms of classification accuracy
2. Classification Accuracy v/s {K, S}
(i) Effect of K(propagation order) and S(number of data augmentations) on
Classification Accuracy on GRAND(DropNode data augmentation)
2. Classification Accuracy v/s {K, S}
2. Classification Accuracy v/s {K, S}
(ii) Effect of K(propagation order) and S(number of data augmentations) on
Classification Accuracy on GRAND_dropout and GRAND_dropedge(alternative data augmentation techniques)
3. Sensitivity wrt ƛ
Classification Accuracy v/s ƛ
MLP classification network
GCN classification network
Both using DropNode for data augmentation
Consistency Regularization Loss Coefficient
3. Sensitivity wrt ƛ
Classification Accuracy v/s ƛ
Both using MLPs as classification networks
DropEdge data augmentation
Dropout data augmentation
3. Sensitivity wrt ƛ
Ablation Study
The effect of the absence of the following parameters was studied:
Ablation Study
Method | Cora | Pubmed | Citeseer |
w/o CR (λ=0) | 0.841 | 0.811 | 0.728 |
w/o mDN (S=1) | 0.85 | 0.80 | 0.744 |
w/o sharpening (T=1) | 0.844 | 0.816 | 0.578 |
w/o CR & DN (λ=0, δ=0) | 0.835 | 0.787 | 0.597 |
these are classification accuracies
END
Thank You!