Survey on Characterizing and Understanding GNNs from a Computer Architecture Perspective

📅 2024-08-04
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
Graph Neural Networks (GNNs) suffer from low execution efficiency and poor training stability on parallel and distributed systems due to unoptimized computational, memory access, and communication patterns. Method: This paper adopts a computer architecture–centric perspective to systematically analyze these patterns and identify critical performance bottlenecks. It proposes the first hardware-software co-design–oriented, three-level GNN workload taxonomy—unifying architectural characterization across diverse GNN models. Leveraging architecture-aware analysis, fine-grained performance modeling, and cross-platform empirical evaluation, it quantifies behavioral disparities of GNN workloads across CPUs, GPUs, and distributed clusters. Contribution/Results: The work delivers a reusable modeling framework and actionable optimization guidelines for efficient, architecture-agnostic GNN implementation. It significantly improves hardware resource utilization and training stability while enabling systematic, principled design of high-performance GNN systems.

Technology Category

Application Category

📝 Abstract
Characterizing and understanding graph neural networks (GNNs) is essential for identifying performance bottlenecks and facilitating their deployment in parallel and distributed systems. Despite substantial work in this area, a comprehensive survey on characterizing and understanding GNNs from a computer architecture perspective is lacking. This work presents a comprehensive survey, proposing a triple-level classification method to categorize, summarize, and compare existing efforts, particularly focusing on their implications for parallel architectures and distributed systems. We identify promising future directions for GNN characterization that align with the challenges of optimizing hardware and software in parallel and distributed systems. Our survey aims to help scholars systematically understand GNN performance bottlenecks and execution patterns from a computer architecture perspective, thereby contributing to the development of more efficient GNN implementations across diverse parallel architectures and distributed systems.
Problem

Research questions and friction points this paper is trying to address.

Graph Neural Networks
Performance Optimization
Parallel Processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Graph Neural Networks
Hardware Design
Parallel and Distributed Systems
🔎 Similar Papers
No similar papers found.
Meng Wu
Meng Wu
Department of Electrical Engineering, Stanford University
Medical ImagingMachine LearningComputer Vision
M
Mingyu Yan
Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China, and University of Chinese Academy of Sciences, Beijing, China
W
Wenming Li
Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China, and University of Chinese Academy of Sciences, Beijing, China
X
Xiaochun Ye
Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China, and University of Chinese Academy of Sciences, Beijing, China
Dongrui Fan
Dongrui Fan
Institute of Computing Technology, Chinese Academy of Sciences
Computer ArchitectureProcessor DesignMany-core Design
Y
Yuan Xie
Department of Electronic and Computer Engineering, Hong Kong University of Science and Technology, Hong Kong, SAR, China