66
0

Adaptive Tokenization: On the Hop-Overpriority Problem in Tokenized Graph Learning Models

Main:9 Pages
6 Figures
Bibliography:3 Pages
8 Tables
Appendix:19 Pages
Abstract

Graph Transformers, leveraging the global attention to capture long-range dependencies in graph structures, have significantly advanced graph machine learning, but face prohibitive computational complexity. Tokenized Graph Learning Models (TGLMs) address this issue by converting graphs into ordered token lists for scalable processing. Besides, TGLMs also empower Large Language Models (LLMs) to handle text-attributed graphs more effectively and thus are also employed in Graph LLMs. However, existing TGLMs rely on hand-designed token lists and their adaptability to diverse graph learning scenarios remains unexplored. In this paper, we first conduct extensive empirical and theoretical preliminary studies for hand-designed token lists. Surprisingly, we identify an unexplored hop-overpriority problem: the common pre-defined token lists overemphasize nearby nodes and overwhelm the ability of TGLMs to balance local and global signals. This phenomenon is especially harmful for heterophilic graphs. To address this problem, we propose the Learnable Graph Token List (LGTL), a plug-and-play module to replace hand-designed token lists in TGLMs. Specifically, LGTL adaptively adjusts the weights across hops and prioritizes informative nodes within hops through a graph attention gate module and a selection module, respectively. In this way, contextually informative nodes can be adaptively emphasized for both homophilic and heterophilic graphs. Besides, we theoretically show that LGTL can address the hop-overpriority problem. Extensive experiments on benchmarks validate the efficacy of LGTL across both Graph Transformers and Graph LLM backbones.

View on arXiv
@article{wang2025_2505.15845,
  title={ Adaptive Tokenization: On the Hop-Overpriority Problem in Tokenized Graph Learning Models },
  author={ Zhibiao Wang and Yunlong Zhou and Ziwei Zhang and Mengmei Zhang and Shirui Pan and Chunming Hu and Xiao Wang },
  journal={arXiv preprint arXiv:2505.15845},
  year={ 2025 }
}
Comments on this paper