Skip to yearly menu bar Skip to main content


Poster

What Matters in Graph Class Incremental Learning? An Information Preservation Perspective

Jialu Li · Yu Wang · Pengfei Zhu · Wanyu Lin · Qinghua Hu

East Exhibit Hall A-C #2903
[ ] [ Project Page ]
Fri 13 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract:

Graph class incremental learning (GCIL) requires the model to classify emerging nodes of new classes while remembering old classes. Existing methods are designed to preserve effective information of old models or graph data to alleviate forgetting, but there is no clear theoretical understanding of what matters in information preservation. In this paper, we consider that present practice suffers from high semantic and structural shifts assessed by two devised shift metrics. We provide insights into information preservation in GCIL and find that maintaining graph information can preserve information of old models in theory to calibrate node semantic and graph structure shifts. We correspond graph information into low-frequency local-global parts and high-frequency parts in spatial domain. Based on the analysis, we propose a universal framework, Graph Spatial Information Preservation (GSIP). Specifically, for low-frequency information preservation, the old node representations obtained by inputting replayed nodes into the old model are aligned with the outputs of the node and its neighbors in the new model, then old and new outputs are globally matched after pooling. For high-frequency information preservation, the new node representations are encouraged to imitate the near neighbor pair similarity of old node representations. GSIP achieves a 10% increase in terms of the forgetting metric compared to prior methods on large-scale datasets. Our framework can also seamlessly integrate existing replay designs.

Live content is unavailable. Log in and register to view live content