Skip to yearly menu bar Skip to main content


Spotlight Poster

Cluster-wise Graph Transformer with Dual-granularity Kernelized Attention

Siyuan Huang · Yunchong Song · Jiayue Zhou · Zhouhan Lin

[ ]
Thu 12 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

In the realm of graph learning, there is a category of methods that conceptualize graphs as hierarchical structures, utilizing node clustering to capture broader structural information. While generally effective, these methods often rely on a fixed graph coarsening routine, leading to overly homogeneous cluster representations and loss of node-level information. In this paper, we envision the graph as a network of interconnected node sets without resorting to the graph coarsening pipeline. To enable effective information transfer among these node sets, we propose the Node-to-Cluster Attention (N2C-Attn) mechanism. N2C-Attn incorporates techniques from Multiple Kernel Learning into the kernelized attention framework, effectively capturing information at both node and cluster levels. We then devise an efficient form for N2C-Attn using the cluster-wise message-passing framework, achieving linear time complexity. We further analyze how N2C-Attn combines bi-level feature maps of queries and keys, demonstrating its capability to merge dual-granularity information. The resulting architecture, Cluster-wise Graph Transformer (Cluster-GT), which uses node clusters as tokens and employs our proposed N2C-Attn module, shows superior performance on various graph-level tasks.

Live content is unavailable. Log in and register to view live content