Distributed Training of Large Graph Neural Networks with Variable Communication Rates
Juan Cervino, Md Asadullah Turja, Hesham Mostafa, Nageen Himayat, Alejandro Ribeiro
TL;DR
The paper tackles the bottleneck of inter-machine communication in distributed GNN training on large graphs. It introduces VARCO, a variable compression scheme that progressively reduces communication by compressing boundary activations, with a convergence analysis showing it can reach the full-communication solution as compression errors decay. Empirically, VARCO matches or surpasses full communication in accuracy while using significantly fewer communicated bytes across random and METIS graph partitions on large datasets, and it outperforms fixed compression strategies in efficiency. This work provides a practical, partition-agnostic approach to scalable graph learning with strong theoretical and empirical guarantees.
Abstract
Training Graph Neural Networks (GNNs) on large graphs presents unique challenges due to the large memory and computing requirements. Distributed GNN training, where the graph is partitioned across multiple machines, is a common approach to training GNNs on large graphs. However, as the graph cannot generally be decomposed into small non-interacting components, data communication between the training machines quickly limits training speeds. Compressing the communicated node activations by a fixed amount improves the training speeds, but lowers the accuracy of the trained GNN. In this paper, we introduce a variable compression scheme for reducing the communication volume in distributed GNN training without compromising the accuracy of the learned model. Based on our theoretical analysis, we derive a variable compression method that converges to a solution equivalent to the full communication case, for all graph partitioning schemes. Our empirical results show that our method attains a comparable performance to the one obtained with full communication. We outperform full communication at any fixed compression ratio for any communication budget.
