Qifang Zhao, Weidong Ren, Tianyu Li, Xiaoxiao Xu, Hong Liu
We introduce \textit{GraphGPT}, a novel model for Graph learning by self-supervised Generative Pre-training Transformers. Our model transforms each graph or sampled subgraph into a sequence of tokens representing the node, edge and attributes reversibly using the Eulerian path first. Then we feed the tokens into a standard transformer decoder and pre-train it with the next-token-prediction (NTP) task. Lastly, we fine-tune the GraphGPT model with the supervised tasks. This intuitive, yet effective model achieves superior or close results to the state-of-the-art methods for the graph-, edge- and node-level tasks on the large scale molecular dataset PCQM4Mv2, the protein-protein association dataset ogbl-ppa and the ogbn-proteins dataset from the Open Graph Benchmark (OGB). Furthermore, the generative pre-training enables us to train GraphGPT up to 400M+ parameters with consistently increasing performance, which is beyond the capability of GNNs and previous graph transformers. The source code and pre-trained checkpoints will be released soon\footnote{\url{https://github.com/alibaba/graph-gpt}} to pave the way for the graph foundation model research, and also to assist the scientific discovery in pharmaceutical, chemistry, material and bio-informatics domains, etc.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Link Property Prediction | ogbl-citation2 | Number of params | 133096832 | GraphGPT(d1n30) |
| Link Property Prediction | ogbl-citation2 | Number of params | 46784128 | GraphGPT(SMTP) |
| Link Property Prediction | ogbl-ppa | Number of params | 145263360 | GraphGPT(SMTP) |