GSplit: Scaling Graph Neural Network Training on Large Graphs via Split-Parallelism

Sandeep Polisetty, Juelin Liu, Yi Fung, Hui Guan, Marco Serafini

Conference on Machine Learning and Systems 2025 · Day 2 · Session 2: Parallel and Distributed Systems

Graph Neural Networks (GNNs) have emerged as a pivotal technology for extracting insights from graph-structured data across diverse fields, from social network analysis and personalized recommendations to molecular biology and materials science. By learning powerful representations for nodes within a graph, GNNs enable critical downstream tasks such as node classification, link prediction, and graph classification. However, the burgeoning scale of real-world graphs presents significant challenges to GNN training. As graphs grow in size and complexity, the computational and memory demands of GNNs often exceed the capabilities of single-GPU systems, necessitating distributed training approaches.

AI review

GSplit presents a legitimate systems contribution to distributed GNN training — the load-aware partitioning idea is well-motivated and the speedup numbers are real. But this article reads like a cleaned-up abstract, not an engineering talk summary. The 'how to actually build it' layer is mostly missing, and the experimental section is thin enough that I can't evaluate whether the results generalize.