Degree-Based Scheduling and Memory Management for Large-Scale Exact Online GNN Inference. Namazi, A., Shen, H., Sen, T., & Zhang, M. In 2025 IEEE International Conference on Big Data (BigData), pages 1–10, Macau, China, December, 2025. IEEE.
Degree-Based Scheduling and Memory Management for Large-Scale Exact Online GNN Inference [link]Paper  doi  abstract   bibtex   
Graph Neural Networks (GNNs) have demonstrated remarkable performance on tasks operating on graph-structured data. Some applications require exact online inference, which processes all neighbors rather than sampling a random subset to avoid stochasticity and accuracy degradation, while still meeting tight latency constraints. In this paper, we analyze exact online inference and identify two key challenges: headof-line blocking caused by heterogeneous request processing times, and highly variable memory usage that complicates cache size optimization for balancing computation and CPU–GPU communication. To reduce the average request latency, we propose GSM, an exact online GNN inference serving system with efficient scheduling and memory management. GSM integrates three main components. The parameter tuner optimizes the tradeoff between computation and CPU–GPU communication overhead by finding near-optimal parameters, such as batch in-degree sum and batch subgraph size, for the target workload. The degree-based scheduler groups requests with similar in-degrees into batches, ensuring each batch’s in-degree sum approaches the optimal value and prioritizing batches with shorter processing times. The batch divider partitions a batch subgraph to achieve the optimal subgraph size and improve processing efficiency. Largescale experiments on graphs with billions of edges (Papers100M and Friendster) demonstrate that GSM achieves up to 89% lower latency and 71% higher throughput compared with state-of-theart systems. The implementation of GSM is publicly available.
@inproceedings{namazi_degree-based_2025,
	address = {Macau, China},
	title = {Degree-{Based} {Scheduling} and {Memory} {Management} for {Large}-{Scale} {Exact} {Online} {GNN} {Inference}},
	copyright = {https://doi.org/10.15223/policy-029},
	isbn = {979-8-3315-9447-3},
	url = {https://ieeexplore.ieee.org/document/11402115/},
	doi = {10.1109/BigData66926.2025.11402115},
	abstract = {Graph Neural Networks (GNNs) have demonstrated remarkable performance on tasks operating on graph-structured data. Some applications require exact online inference, which processes all neighbors rather than sampling a random subset to avoid stochasticity and accuracy degradation, while still meeting tight latency constraints. In this paper, we analyze exact online inference and identify two key challenges: headof-line blocking caused by heterogeneous request processing times, and highly variable memory usage that complicates cache size optimization for balancing computation and CPU–GPU communication. To reduce the average request latency, we propose GSM, an exact online GNN inference serving system with efficient scheduling and memory management. GSM integrates three main components. The parameter tuner optimizes the tradeoff between computation and CPU–GPU communication overhead by finding near-optimal parameters, such as batch in-degree sum and batch subgraph size, for the target workload. The degree-based scheduler groups requests with similar in-degrees into batches, ensuring each batch’s in-degree sum approaches the optimal value and prioritizing batches with shorter processing times. The batch divider partitions a batch subgraph to achieve the optimal subgraph size and improve processing efficiency. Largescale experiments on graphs with billions of edges (Papers100M and Friendster) demonstrate that GSM achieves up to 89\% lower latency and 71\% higher throughput compared with state-of-theart systems. The implementation of GSM is publicly available.},
	language = {en},
	urldate = {2026-03-19},
	booktitle = {2025 {IEEE} {International} {Conference} on {Big} {Data} ({BigData})},
	publisher = {IEEE},
	author = {Namazi, Alireza and Shen, Haiying and Sen, Tanmoy and Zhang, Minjia},
	month = dec,
	year = {2025},
	pages = {1--10},
}

Downloads: 0