Core Concepts
GNNBENCH is a standardized benchmarking platform that enables fair and productive evaluation of single-GPU GNN systems by providing stable system APIs, minimizing framework overhead, and automatically identifying and correcting accuracy issues in integrated systems.
Abstract
The paper proposes GNNBENCH, a standardized benchmarking platform for evaluating single-GPU Graph Neural Network (GNN) systems. GNNBENCH addresses several challenges in GNN system design and evaluation that the community has overlooked:
Stable System APIs: GNNBENCH introduces a producer-only DLPack protocol to enable stable system APIs that are independent of the underlying deep learning framework. This allows GNNBENCH-System to accept custom data structures like graphs, unlike the limitations of framework-specific plugin environments.
Productivity and Fairness: GNNBENCH provides a common GNN model front-end and automatically generates integration code, enabling researchers to quickly prototype and evaluate their system innovations. It also ensures fair comparisons by minimizing framework overhead.
Accuracy Issue Identification: GNNBENCH's well-tested workflow and auxiliary flags help identify and correct accuracy issues in integrated GNN systems, which often suffer from various pitfalls.
New Artifacts: GNNBENCH is used to integrate several original GNN systems that had integration issues, unknown memory corruption, or missing backward computation. The resulting artifacts serve as useful baselines for future research.
Insights from Evaluation: The evaluation on GNNBENCH provides several new insights, such as the reliance on smaller datasets being a poor practice, the significant framework-memory overhead in popular baselines like DGL, and the need to revisit the advantages of kernel fusion techniques.
Stats
The paper does not contain any key metrics or important figures to support the author's key logics.
Quotes
The paper does not contain any striking quotes supporting the author's key logics.