Understanding GNN Computational Graph: A Coordinated Computation, IO, and Memory Perspective
Speedup
DOI:
10.48550/arxiv.2110.09524
Publication Date:
2021-01-01
AUTHORS (7)
ABSTRACT
Graph Neural Networks (GNNs) have been widely used in various domains, and GNNs with sophisticated computational graph lead to higher latency larger memory consumption. Optimizing the GNN suffers from: (1) Redundant neural operator computation. The same data are propagated through structure perform operation multiple times GNNs, leading redundant computation which accounts for 92.4% of total operators. (2) Inconsistent thread mapping. Efficient mapping schemes vertex-centric edge-centric operators different. This inconsistency prohibits fusion reduce IO. (3) Excessive intermediate data. For training is usually performed concurrently inference, must be stored backward pass, consuming 91.9% requirement. To tackle these challenges, we propose following designs optimize from a novel coordinated computation, IO, perspective: Propagation-postponed reorganization. We reorganize operations before propagation, thus eliminated. Unified fusion. unified scheme both vertex- enable Intermediate recomputation. recomputed during pass Extensive experimental results on three typical models show that, achieve up 2.75x end-to-end speedup, 6.89x less 7.73x consumption over state-of-the-art frameworks.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....