TY - GEN
T1 - FGI
T2 - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
AU - Ji, Binglin
AU - Zhao, Chenfeng
AU - Chamberlain, Roger D.
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Graph Neural Networks (GNNs) are becoming increasingly popular, with their applications expanding across diverse domains. As the scale of graph data continues to grow, including larger numbers of nodes, edges, and higher embedding dimensions, standardized libraries such as DGL and PyG have been developed to facilitate GNN computation. However, with the rapid increase in the number of processor cores and the evolution of multi-core architectures, these libraries often show poor scalability and fail to execute GNN inference efficiently on the latest multi-core systems, particularly those with upwards of a hundred cores. To address this limitation, we present FGI, a fast GNN inference system for large-scale graph data (over 100,000 vertices). FGI employs different parallelization strategies optimized for diverse graph structures, maximizing the utilization of multi-level cache hierarchies in multi-core systems. We evaluate the Graph Convolutional Network (GCN) model with FGI on a 128-core AMD EPYC system. FGI achieves up to 2.64× inference speedup compared to state-of-the-art libraries across a range of large-scale, high-dimensional graph datasets with different properties.
AB - Graph Neural Networks (GNNs) are becoming increasingly popular, with their applications expanding across diverse domains. As the scale of graph data continues to grow, including larger numbers of nodes, edges, and higher embedding dimensions, standardized libraries such as DGL and PyG have been developed to facilitate GNN computation. However, with the rapid increase in the number of processor cores and the evolution of multi-core architectures, these libraries often show poor scalability and fail to execute GNN inference efficiently on the latest multi-core systems, particularly those with upwards of a hundred cores. To address this limitation, we present FGI, a fast GNN inference system for large-scale graph data (over 100,000 vertices). FGI employs different parallelization strategies optimized for diverse graph structures, maximizing the utilization of multi-level cache hierarchies in multi-core systems. We evaluate the Graph Convolutional Network (GCN) model with FGI on a 128-core AMD EPYC system. FGI achieves up to 2.64× inference speedup compared to state-of-the-art libraries across a range of large-scale, high-dimensional graph datasets with different properties.
KW - graph convolutional network
KW - graph neural network
KW - multi-core
UR - https://www.scopus.com/pages/publications/105015359742
U2 - 10.1109/IPDPSW66978.2025.00119
DO - 10.1109/IPDPSW66978.2025.00119
M3 - Conference contribution
AN - SCOPUS:105015359742
T3 - Proceedings - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
SP - 748
EP - 757
BT - Proceedings - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 3 June 2025 through 7 June 2025
ER -