Graph Prompting for Graph Learning Models: Recent Advances and Future Directions
- AI4CEVLM
Graph learning models have demonstrated great prowess in learning expressive representations from large-scale graph data in a wide variety of real-world scenarios. As a prevalent strategy for training powerful graph learning models, the "pre-training, adaptation" scheme first pre-trains graph learning models on unlabeled graph data in a self-supervised manner and then adapts them to specific downstream tasks. During the adaptation phase, graph prompting emerges as a promising approach that learns trainable prompts while keeping the pre-trained graph learning models unchanged. In this paper, we present a systematic review of recent advancements in graph prompting. First, we introduce representative graph pre-training methods that serve as the foundation step of graph prompting. Next, we review mainstream techniques in graph prompting and elaborate on how they design learnable prompts for graph prompting. Furthermore, we summarize the real-world applications of graph prompting from different domains. Finally, we discuss several open challenges in existing studies with promising future directions in this field.
View on arXiv@article{fu2025_2506.08326, title={ Graph Prompting for Graph Learning Models: Recent Advances and Future Directions }, author={ Xingbo Fu and Zehong Wang and Zihan Chen and Jiazheng Li and Yaochen Zhu and Zhenyu Lei and Cong Shen and Yanfang Ye and Chuxu Zhang and Jundong Li }, journal={arXiv preprint arXiv:2506.08326}, year={ 2025 } }