GNNs & Hardware
Graph Neural Networks (GNNs) are booming across domains, and efficient execution is
becoming as important as the algorithms themselves. Below is a categorized list of
notable publications — surveys, FPGA frameworks, quantization methods, and core GNN models —
each linked to its official source.
Surveys & Reviews
Focused on FPGA-targeted GNN accelerators; provides taxonomy and representative architectures.
Comprehensive survey covering algorithmic optimizations, system-level strategies, and custom hardware for GNN acceleration.
Focuses on hardware-level techniques and trade-offs when accelerating GNNs.
Explores algorithmic strategies that make GNNs efficient on reconfigurable platforms such as FPGAs.
FPGA Frameworks & Accelerators
A modular, high-level synthesis (HLS) framework for accelerating multiple GNN models on FPGAs.
Targets dynamic graphs with changing topology; optimizes inference for time-varying data structures.
Explores various hardware optimization strategies to improve the performance of GNN computation.
Quantization & Hardware-Aware Training
Introduces training-time approximations to maintain performance when deploying low-precision GNNs.
Explores deep quantization and training methods that adapt to runtime hardware constraints.
GNN Architectures & Theory
Introduced attention mechanisms to GNNs, allowing learnable weighting of neighboring nodes.
Extends GAT to heterogeneous graphs with multiple node and edge types.
Incorporates edge features directly into the attention mechanism for richer representations.
Highlights the structural equivalence between Transformers and GNNs, unifying both under a message-passing view.
Applications & Learning Paradigms
Demonstrates how GNNs can perform few-shot learning by representing examples as a graph and propagating relational information.
Quick Takeaways
- Frameworks: GenGNN and DGNN-Booster provide flexible FPGA infrastructures for static and dynamic graphs.
- Attention bottlenecks: GAT and its variants remain core compute hot spots for accelerators.
- Quantization: Hardware-aware training is key for deploying efficient, accurate models.
- Transformers & GNNs: Increasing convergence means hardware insights can transfer across domains.