Graph Neural Networks (GNNs) have achieved significant success in recent years in fields such as social network analysis, recommendation systems, knowledge graphs, and drug discovery. However, unlike in the fields of computer vision and natural language processing, research on pre-training for graph neural networks is relatively lagging. Traditional GNN training usually uses random initialization followed by supervised learning on specific downstream tasks, which faces two main challenges:
- Scarcity of labeled data: It is difficult to obtain large amounts of high-quality labeled data in many graph application domains
- Insufficient generalization ability: Models trained on specific tasks are hard to transfer to other related tasks
Inspired by the successful experience of the GPT (Generative Pre-Training) series models in the natural language processing field, researchers have begun to explore pre-training methods for graph neural networks. GPT-GNN was born in this context, proposing a new generative pre-training paradigm that enables the model to acquire general graph representation capabilities by learning to reconstruct the attributes and structure of the graph.