TY - GEN
T1 - Graph few-shot class-incremental learning
AU - Tan, Zhen
AU - Ding, Kaize
AU - Guo, Ruocheng
AU - Liu, Huan
N1 - Publisher Copyright:
© 2022 ACM.
PY - 2022/2/11
Y1 - 2022/2/11
N2 - The ability to incrementally learn new classes is vital to all real-world artificial intelligence systems. A large portion of high-impact applications like social media, recommendation systems, E-commerce platforms, etc. can be represented by graph models. In this paper, we investigate the challenging yet practical problem,Graph Few-shot Class-incremental (Graph FCL) problem, where the graph model is tasked to classify both newly encountered classes and previously learned classes. Towards that purpose, we put forward a Graph Pseudo Incremental Learning paradigm by sampling tasks recurrently from the base classes, so as to produce an arbitrary number of training episodes for our model to practice the incremental learning skill. Furthermore, we design a Hierarchical-Attention-based Graph Meta-learning framework, HAG-Meta from an optimization perspective. We present a task-sensitive regularizer calculated from task-level attention and node class prototypes to mitigate overfitting onto either novel or base classes. To employ the topological knowledge, we add a node-level attention module to adjust the prototype representation. Our model not only achieves greater stability of old knowledge consolidation, but also acquires advantageous adaptability to new knowledge with very limited data samples. Extensive experiments on three real-world datasets, including Amazon-clothing, Reddit, and DBLP, show that our framework demonstrates remarkable advantages in comparison with the baseline and other related state-of-the-art methods.
AB - The ability to incrementally learn new classes is vital to all real-world artificial intelligence systems. A large portion of high-impact applications like social media, recommendation systems, E-commerce platforms, etc. can be represented by graph models. In this paper, we investigate the challenging yet practical problem,Graph Few-shot Class-incremental (Graph FCL) problem, where the graph model is tasked to classify both newly encountered classes and previously learned classes. Towards that purpose, we put forward a Graph Pseudo Incremental Learning paradigm by sampling tasks recurrently from the base classes, so as to produce an arbitrary number of training episodes for our model to practice the incremental learning skill. Furthermore, we design a Hierarchical-Attention-based Graph Meta-learning framework, HAG-Meta from an optimization perspective. We present a task-sensitive regularizer calculated from task-level attention and node class prototypes to mitigate overfitting onto either novel or base classes. To employ the topological knowledge, we add a node-level attention module to adjust the prototype representation. Our model not only achieves greater stability of old knowledge consolidation, but also acquires advantageous adaptability to new knowledge with very limited data samples. Extensive experiments on three real-world datasets, including Amazon-clothing, Reddit, and DBLP, show that our framework demonstrates remarkable advantages in comparison with the baseline and other related state-of-the-art methods.
KW - Few-shot learning
KW - Graph neural networks
KW - Incremental learning
UR - http://www.scopus.com/inward/record.url?scp=85125786371&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85125786371&partnerID=8YFLogxK
U2 - 10.1145/3488560.3498455
DO - 10.1145/3488560.3498455
M3 - Conference contribution
AN - SCOPUS:85125786371
T3 - WSDM 2022 - Proceedings of the 15th ACM International Conference on Web Search and Data Mining
SP - 987
EP - 996
BT - WSDM 2022 - Proceedings of the 15th ACM International Conference on Web Search and Data Mining
PB - Association for Computing Machinery, Inc
T2 - 15th ACM International Conference on Web Search and Data Mining, WSDM 2022
Y2 - 21 February 2022 through 25 February 2022
ER -