期刊文献+

Graph distillation with network symmetry

原文传递
导出
摘要 Graph neural networks(GNNs)have demonstrated excellent performance in graph representation learning.However,as the volume of graph data grows,issues related to cost and efficiency become increasingly prominent.Graph distillation methods address this challenge by extracting a smaller,reduced graph,ensuring that GNNs trained on both the original and reduced graphs show similar performance.Existing methods,however,primarily optimize the feature matrix of the reduced graph and rely on correlation information from GNNs,while neglecting the original graph’s structure and redundant nodes.This often results in a loss of critical information within the reduced graph.To overcome this limitation,we propose a graph distillation method guided by network symmetry.Specifically,we identify symmetric nodes with equivalent neighborhood structures and merge them into“super nodes”,thereby simplifying the network structure,reducing redundant parameter optimization and enhancing training efficiency.At the same time,instead of relying on the original node features,we employ gradient descent to match optimal features that align with the original features,thus improving downstream task performance.Theoretically,our method guarantees that the reduced graph retains the key information present in the original graph.Extensive experiments demonstrate that our approach achieves significant improvements in graph distillation,exhibiting strong generalization capability and outperforming existing graph reduction methods.
作者 Feng Lin Jia-Lin He 林峰;何嘉林
出处 《Chinese Physics B》 2025年第4期262-271,共10页 中国物理B(英文版)
基金 Project supported by the National Natural Science Foundation of China(Grant No.62176217) the Program from the Sichuan Provincial Science and Technology,China(Grant No.2018RZ0081) the Fundamental Research Funds of China West Normal University(Grant No.17E063).
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部