20
4

Fast Online Node Labeling for Very Large Graphs

Abstract

This paper studies the online node classification problem under a transductive learning setting. Current methods either invert a graph kernel matrix with O(n3)\mathcal{O}(n^3) runtime and O(n2)\mathcal{O}(n^2) space complexity or sample a large volume of random spanning trees, thus are difficult to scale to large graphs. In this work, we propose an improvement based on the \textit{online relaxation} technique introduced by a series of works (Rakhlin et al.,2012; Rakhlin and Sridharan, 2015; 2017). We first prove an effective regret O(n1+γ)\mathcal{O}(\sqrt{n^{1+\gamma}}) when suitable parameterized graph kernels are chosen, then propose an approximate algorithm FastONL enjoying O(kn1+γ)\mathcal{O}(k\sqrt{n^{1+\gamma}}) regret based on this relaxation. The key of FastONL is a \textit{generalized local push} method that effectively approximates inverse matrix columns and applies to a series of popular kernels. Furthermore, the per-prediction cost is O(vol(S)log1/ϵ)\mathcal{O}(\text{vol}({\mathcal{S}})\log 1/\epsilon) locally dependent on the graph with linear memory cost. Experiments show that our scalable method enjoys a better tradeoff between local and global consistency.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.