Core Concepts
Efficient mini-batch training with personalized PageRank tokenization in graph transformers.
Abstract
The paper introduces VCR-Graphormer, a novel approach for mini-batch training in graph transformers using personalized PageRank tokenization. It addresses the computational complexity issue of dense attention mechanisms by leveraging virtual connections to encode local and global contexts efficiently. The proposed method achieves competitive performance on both small-scale and large-scale datasets, demonstrating its effectiveness and scalability.
Abstract:
- Graph transformer efficiency through mini-batch training.
- Personalized PageRank tokenization for node representation.
- Virtual connections for encoding local and global contexts.
- Competitive performance on small-scale and large-scale datasets.
Introduction:
- Transformer architectures' success in various tasks.
- Importance of effective graph transformers for non-grid data.
- Previous works like GT, Gophormer, Coarformer, etc., using dense attention.
Data Extraction:
- "VCR-Graphormer needs O(m+klogk) complexity for graph tokenization as compared to O(n3) of previous works."
Stats
"VCR-Graphormer needs O(m+klogk) complexity for graph tokenization as compared to O(n3) of previous works."
Quotes
"The logic of this paradigm is easy to follow, which tokenizes the input graph by assigning each target node a token list Tu."
"We further prove this PPR tokenization is viable as a graph convolution network with a fixed polynomial filter and jumping knowledge."