19
0

Toward a Graph Foundation Model: Pre-Training Transformers With Random Walks

Main:9 Pages
4 Figures
Bibliography:5 Pages
11 Tables
Appendix:11 Pages
Abstract

A foundation model like GPT elicits many emergent abilities, owing to the pre-training with broad inclusion of data and the use of the powerful Transformer architecture. While foundation models in natural languages are prevalent, can we build similar models for graphs? This paper describes an approach toward a graph foundation model that is pre-trained with diverse graph datasets by adapting the Transformer backbone. A central challenge toward this end is how a sequence model encodes graphs of varying sizes and from different domains. We propose representing a node as multiple random walks, such that the Transformer can extract node representations from sequences, which in turn form edge and graph representations. We develop a novel context prediction loss for these random walks and theoretically analyze their expressive power in distinguishing neighborhoods and graphs. We also demonstrate the pre-training of our model and its adaptation to downstream tasks, showcasing its potential as a foundation for processing and reasoning with graph-structured data.

View on arXiv
@article{tang2025_2506.14098,
  title={ Toward a Graph Foundation Model: Pre-Training Transformers With Random Walks },
  author={ Ziyuan Tang and Jie Chen },
  journal={arXiv preprint arXiv:2506.14098},
  year={ 2025 }
}
Comments on this paper