TY - GEN
T1 - DyFormer
T2 - 2023 SIAM International Conference on Data Mining, SDM 2023
AU - Cong, Weilin
AU - Wu, Yanhong
AU - Tian, Yuandong
AU - Gu, Mengting
AU - Xia, Yinglong
AU - Jason Chen, Chun Cheng
AU - Mahdavi, Mehrdad
N1 - Publisher Copyright:
Copyright © 2023 by SIAM.
PY - 2023
Y1 - 2023
N2 - Transformers have achieved great success in several domains, including Natural Language Processing and Computer Vision. However, their application to real-world graphs is less explored, mainly due to its high computation cost and its poor generalizability caused by the lack of enough training data in the graph domain. To fill in this gap, we propose a scalable Transformer-like dynamic graph learning method named Dynamic Graph Transformer (DyFormer) with spatial-temporal encoding to effectively learn graph topology and capture implicit links. To achieve efficient and scalable training, we propose temporal-union graph structure and its associated subgraph-based node sampling strategy. To improve the generalization ability, we introduce two complementary self-supervised pre-training tasks and show that jointly optimizing the two pre-training tasks results in a smaller Bayesian error rate via an information-theoretic analysis. Extensive experiments on the real-world datasets illustrate that DyFormer achieves a consistent 1% ~ 3% AUC gain (averaged over all time steps) compared with baselines on all benchmarks.
AB - Transformers have achieved great success in several domains, including Natural Language Processing and Computer Vision. However, their application to real-world graphs is less explored, mainly due to its high computation cost and its poor generalizability caused by the lack of enough training data in the graph domain. To fill in this gap, we propose a scalable Transformer-like dynamic graph learning method named Dynamic Graph Transformer (DyFormer) with spatial-temporal encoding to effectively learn graph topology and capture implicit links. To achieve efficient and scalable training, we propose temporal-union graph structure and its associated subgraph-based node sampling strategy. To improve the generalization ability, we introduce two complementary self-supervised pre-training tasks and show that jointly optimizing the two pre-training tasks results in a smaller Bayesian error rate via an information-theoretic analysis. Extensive experiments on the real-world datasets illustrate that DyFormer achieves a consistent 1% ~ 3% AUC gain (averaged over all time steps) compared with baselines on all benchmarks.
UR - http://www.scopus.com/inward/record.url?scp=85180629063&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85180629063&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85180629063
T3 - 2023 SIAM International Conference on Data Mining, SDM 2023
SP - 442
EP - 450
BT - 2023 SIAM International Conference on Data Mining, SDM 2023
PB - Society for Industrial and Applied Mathematics Publications
Y2 - 27 April 2023 through 29 April 2023
ER -