: Learning Efficient Visual Representations via Flexible Grid-Graphs
- SSLGNN

We propose , a novel approach for image representation learning based on Graph Convolutional Network (GCN). Existing visual representation methods suffer from several issues, such as requiring high-computation, losing in-depth structures, and being restricted to specific objects. converts an image to a low-dimensional feature vector. A key component of is Flexible Grid-Graphs, a spatially-adaptive method based on the image key-points, as a flexible grid, to generate the graph representation. It represents each image with a graph of unique node locations and edge distances. Nodes, in Flexible Grid-Graphs, describe the most representative patches in the image. We develop a multi-channel Convolutional Neural Network architecture to learn local features of each patch. We implement a hybrid node-embedding method, i.e., having spectral and non-spectral components. It aggregates the products of neighbours' features and node's eigenvector centrality score. We compare the performance of with a set of state-of-the-art representation learning and visual recognition models. has only features in comparison to a range from VGG16 with to NASNet with . We show the models' superior accuracy in both binary and multi-class image classification. Although we utilise imbalanced, low-size dataset, shows stable and superior results against the well-known base classifiers.
View on arXiv