Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

ProtoMGAE: Prototype-Aware Masked Graph Auto-Encoder for Graph Representation Learning

Published: 12 April 2024 Publication History

Abstract

Graph self-supervised representation learning has gained considerable attention and demonstrated remarkable efficacy in extracting meaningful representations from graphs, particularly in the absence of labeled data. Two representative methods in this domain are graph auto-encoding and graph contrastive learning. However, the former methods primarily focus on global structures, potentially overlooking some fine-grained information during reconstruction. The latter methods emphasize node similarity across correlated views in the embedding space, potentially neglecting the inherent global graph information in the original input space. Moreover, handling incomplete graphs in real-world scenarios, where original features are unavailable for certain nodes, poses challenges for both types of methods. To alleviate these limitations, we integrate masked graph auto-encoding and prototype-aware graph contrastive learning into a unified model to learn node representations in graphs. In our method, we begin by masking a portion of node features and utilize a specific decoding strategy to reconstruct the masked information. This process facilitates the recovery of graphs from a global or macro level and enables handling incomplete graphs easily. Moreover, we treat the masked graph and the original one as a pair of contrasting views, enforcing the alignment and uniformity between their corresponding node representations at a local or micro level. Last, to capture cluster structures from a meso level and learn more discriminative representations, we introduce a prototype-aware clustering consistency loss that is jointly optimized with the preceding two complementary objectives. Extensive experiments conducted on several datasets demonstrate that the proposed method achieves significantly better or competitive performance on downstream tasks, especially for graph clustering, compared with the state-of-the-art methods, showcasing its superiority in enhancing graph representation learning.

References

[1]
Hangbo Bao, Li Dong, Songhao Piao, and Furu Wei. 2022. BEiT: BERT pre-training of image transformers. In Proceedings of the 10th International Conference on Learning Representations.
[2]
Piotr Bielak, Tomasz Kajdanowicz, and Nitesh V. Chawla. 2022. Graph Barlow Twins: A self-supervised representation learning framework for graphs. Knowledge-Based Systems 256 (2022), 109631. DOI:
[3]
Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. 2020. Unsupervised learning of visual features by contrasting cluster assignments. In Advances in Neural Information Processing Systems. 9912–9924.
[4]
Feihu Che, Guohua Yang, Dawei Zhang, Jianhua Tao, and Tong Liu. 2021. Self-supervised graph representation learning via bootstrapping. Neurocomputing 456 (2021), 88–96. DOI:
[5]
Jialu Chen and Gang Kou. 2023. Attribute and structure preserving graph contrastive learning. In Proceedings of the 37th AAAI Conference on Artificial Intelligence. 7024–7032. DOI:
[6]
Marco Cuturi. 2013. Sinkhorn distances: Lightspeed computation of optimal transport. In Advances in Neural Information Processing Systems. 2292–2300.
[7]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 4171–4186. DOI:
[8]
Xiaoyi Dong, Jianmin Bao, Ting Zhang, Dongdong Chen, Weiming Zhang, Lu Yuan, Dong Chen, Fang Wen, and Nenghai Yu. 2022. Bootstrapped masked autoencoders for vision BERT pretraining. In Proceedings of the European Conference on Computer Vision, Vol. 13690. 247–264. DOI:
[9]
Michelle Girvan and Mark E. J. Newman. 2002. Community structure in social and biological networks. Proceedings of the National Academy of Sciences 99, 12 (2002), 7821–7826. DOI:
[10]
Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H. Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Ávila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, Rémi Munos, and Michal Valko. 2020. Bootstrap your own latent—A new approach to self-supervised learning. In Advances in Neural Information Processing Systems. 21271–21284.
[11]
Hakim Hafidi, Mounir Ghogho, Philippe Ciblat, and Ananthram Swami. 2022. Negative sampling strategies for contrastive self-supervised learning of graph representations. Signal Processing 190 (2022), 108310. DOI:
[12]
William L. Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems. 1024–1034.
[13]
Kaveh Hassani and Amir Hosein Khas Ahmadi. 2020. Contrastive multi-view representation learning on graphs. In Proceedings of the 37th International Conference on Machine Learning, Vol. 119. 4116–4126.
[14]
Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross B. Girshick. 2022. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 15979–15988. DOI:
[15]
R. Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Philip Bachman, Adam Trischler, and Yoshua Bengio. 2019. Learning deep representations by mutual information estimation and maximization. In Proceedings of the 7th International Conference on Learning Representations.
[16]
Harold Hotelling. 1992. Relations between two sets of variates. In Breakthroughs in Statistics: Methodology and Distribution. Springer Series in Statistics. Springer, 162–190. DOI:
[17]
Zhenyu Hou, Yufei He, Yukuo Cen, Xiao Liu, Yuxiao Dong, Evgeny Kharlamov, and Jie Tang. 2023. GraphMAE2: A decoding-enhanced masked self-supervised graph learner. In Proceedings of the Web Conference. 737–746. DOI:
[18]
Zhenyu Hou, Xiao Liu, Yukuo Cen, Yuxiao Dong, Hongxia Yang, Chunjie Wang, and Jie Tang. 2022. GraphMAE: Self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 594–604. DOI:
[19]
Zhicheng Huang, Xiaojie Jin, Chengze Lu, Qibin Hou, Ming-Ming Cheng, Dongmei Fu, Xiaohui Shen, and Jiashi Feng. 2022. Contrastive masked autoencoders are stronger vision learners. arXiv preprint arXiv:2207.13532 (2022).
[20]
Jincen Jiang, Xuequan Lu, Lizhi Zhao, Richard Dazaley, and Meili Wang. 2023. Masked autoencoders in 3D point cloud representation learning. IEEE Transactions on Multimedia. Early Access, September 13, 2023. DOI:
[21]
Di Jin, Rui Wang, Tao Wang, Dongxiao He, Weiping Ding, Yuxiao Huang, Longbiao Wang, and Witold Pedrycz. 2023. Amer: A new attribute-missing network embedding approach. IEEE Transactions on Cybernetics 53, 7 (2023), 4306–4319. DOI:
[22]
Ming Jin, Yizhen Zheng, Yuan-Fang Li, Chen Gong, Chuan Zhou, and Shirui Pan. 2021. Multi-scale contrastive Siamese networks for self-supervised graph representation learning. In Proceedings of the 30th International Joint Conference on Artificial Intelligence. 1477–1483. DOI:
[23]
Baoyu Jing, Chanyoung Park, and Hanghang Tong. 2021. HDMI: High-order deep multiplex infomax. In Proceedings of the Web Conference. 2414–2424. DOI:
[24]
Zekarias Tilahun Kefato and Sarunas Girdzijauskas. 2021. Self-supervised graph neural networks without explicit negative sampling. In Proceedings of the International Workshop on Self-Supervised Learning for the Web (SSL ’21) at the 2021 World Wide Web Conference.
[25]
Thomas N. Kipf and Max Welling. 2016. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016).
[26]
Thomas N. Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In Proceedings of the 5th International Conference on Learning Representations.
[27]
Hanna Krasnova, Sarah Spiekermann, Ksenia Koroleva, and Thomas Hildebrand. 2010. Online social networks: Why we disclose. Journal of Information Technology 25, 2 (2010), 109–125. DOI:
[28]
Youngwan Lee, Jeffrey Ryan Willette, Jonghee Kim, Juho Lee, and Sung Ju Hwang. 2023. Exploring the role of mean teachers in self-supervised masked auto-encoders. In Proceedings of the 11th International Conference on Learning Representations.
[29]
Dongjie Li, Dong Li, and Guang Lian. 2023. Variational graph autoencoder with adversarial mutual information learning for network representation learning. ACM Transactions on Knowledge Discovery from Data 17, 3 (2023), Article 45, 18 pages. DOI:
[30]
Jintang Li, Ruofan Wu, Wangbin Sun, Liang Chen, Sheng Tian, Liang Zhu, Changhua Meng, Zibin Zheng, and Weiqiang Wang. 2023. What’s behind the mask: Understanding masked graph modeling for graph autoencoders. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 1268–1279. DOI:
[31]
Shuai Lin, Chen Liu, Pan Zhou, Zi-Yuan Hu, Shuojia Wang, Ruihui Zhao, Yefeng Zheng, Liang Lin, Eric Xing, and Xiaodan Liang. 2022. Prototypical graph contrastive learning. IEEE Transactions on Neural Networks and Learning Systems 35, 2 (2022), 2747–2758. DOI:
[32]
Xiao Liu, Fanjin Zhang, Zhenyu Hou, Li Mian, Zhaoyu Wang, Jing Zhang, and Jie Tang. 2023. Self-supervised learning: Generative or contrastive. IEEE Transactions on Knowledge and Data Engineering 35, 1 (2023), 857–876. DOI:
[33]
James B. MacQueen. 1967. Some methods for classification and analysis of multivariate observations. In Proceedings of the 5th Berkeley Symposium on Mathematical Statistics and Probability, Vol. 1. 281–297.
[34]
Péter Mernyei and Cătălina Cangea. 2020. Wiki-CS: A Wikipedia-based benchmark for graph neural networks. arXiv preprint arXiv:2007.02901 (2020).
[35]
Shirui Pan, Ruiqi Hu, Sai-Fu Fung, Guodong Long, Jing Jiang, and Chengqi Zhang. 2020. Learning graph embedding with adversarial training methods. IEEE Transactions on Cybernetics 50, 6 (2020), 2475–2487. DOI:
[36]
Shirui Pan, Ruiqi Hu, Guodong Long, Jing Jiang, Lina Yao, and Chengqi Zhang. 2018. Adversarially regularized graph autoencoder for graph embedding. In Proceedings of the 27th International Joint Conference on Artificial Intelligence. 2609–2615. DOI:
[37]
Yatian Pang, Wenxiao Wang, Francis E. H. Tay, Wei Liu, Yonghong Tian, and Li Yuan. 2022. Masked autoencoders for point cloud self-supervised learning. In Proceedings of the European Conference on Computer Vision, Vol. 13662. 604–621. DOI:
[38]
Jiwoong Park, Minsik Lee, Hyung Jin Chang, Kyuewang Lee, and Jin Young Choi. 2019. Symmetric graph convolutional autoencoder for unsupervised graph representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 6518–6527. DOI:
[39]
Hao Peng, Jianxin Li, Yu He, Yaopeng Liu, Mengjiao Bao, Lihong Wang, Yangqiu Song, and Qiang Yang. 2018. Large-scale hierarchical text classification with recursively regularized deep graph-CNN. In Proceedings of the 2018 World Wide Web Conference. 1063–1072. DOI:
[40]
Zhen Peng, Wenbing Huang, Minnan Luo, Qinghua Zheng, Yu Rong, Tingyang Xu, and Junzhou Huang. 2020. Graph representation learning via graphical mutual information maximization. In Proceedings of the Web Conference. 259–270. DOI:
[41]
Robi Polikar. 2012. Ensemble learning. In Ensemble Machine Learning: Methods and Applications. Springer, 1–34.
[42]
Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan Günnemann. 2018. Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868 (2018).
[43]
Weijing Shi and Raj Rajkumar. 2020. Point-GNN: Graph neural network for 3D object detection in a point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1708–1716. DOI:
[44]
Fan-Yun Sun, Jordan Hoffmann, Vikas Verma, and Jian Tang. 2020. InfoGraph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization. In Proceedings of the 8th International Conference on Learning Representations.
[45]
Heli Sun, Fang He, Jianbin Huang, Yizhou Sun, Yang Li, Chenyu Wang, Liang He, Zhongbin Sun, and Xiaolin Jia. 2020. Network embedding for community detection in attributed networks. ACM Transactions on Knowledge Discovery from Data 14, 3 (2020), Article 36, 25 pages. DOI:
[46]
Heli Sun, Yang Li, Bing Lv, Wujie Yan, Liang He, Shaojie Qiao, and Jianbin Huang. 2022. Graph community infomax. ACM Transactions on Knowledge Discovery from Data 16, 3 (2022), Article 53, 21 pages. DOI:
[47]
Qiaoyu Tan, Ninghao Liu, Xiao Huang, Soo-Hyun Choi, Li Li, Rui Chen, and Xia Hu. 2023. S2GAE: Self-supervised graph autoencoders are generalizable learners with graph masking. In Proceedings of the 16th ACM International Conference on Web Search and Data Mining. 787–795. DOI:
[48]
Mingyue Tang, Pan Li, and Carl Yang. 2022. Graph auto-encoder via neighborhood Wasserstein reconstruction. In Proceedings of the 10th International Conference on Learning Representations.
[49]
Shantanu Thakoor, Corentin Tallec, Mohammad Gheshlaghi Azar, Rémi Munos, Petar Veličković, and Michal Valko. 2021. Bootstrapped representation learning on graphs. In Proceedings of the 9th International Conference on Learning Representations Workshop on Geometrical and Topological Representation Learning.
[50]
Yijun Tian, Kaiwen Dong, Chunhui Zhang, Chuxu Zhang, and Nitesh V. Chawla. 2023. Heterogeneous graph masked autoencoders. In Proceedings of the 37th AAAI Conference on Artificial Intelligence. 9997–10005. DOI:
[51]
Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2018. Graph attention networks. In Proceedings of the 6th International Conference on Learning Representations.
[52]
Petar Velickovic, William Fedus, William L. Hamilton, Pietro Liò, Yoshua Bengio, and R. Devon Hjelm. 2019. Deep graph infomax. In Proceedings of the 7th International Conference on Learning Representations.
[53]
Chun Wang, Shirui Pan, Guodong Long, Xingquan Zhu, and Jing Jiang. 2017. MGAE: Marginalized graph autoencoder for graph clustering. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 889–898. DOI:
[54]
Luya Wang, Feng Liang, Yangguang Li, Honggang Zhang, Wanli Ouyang, and Jing Shao. 2022. RePre: Improving self-supervised vision transformer with reconstructive pre-training. In Proceedings of the 31st International Joint Conference on Artificial Intelligence. 1437–1443. DOI:
[55]
Tongzhou Wang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In Proceedings of the 37th International Conference on Machine Learning, Vol. 119. 9929–9939.
[56]
Yuting Wang, Jinpeng Wang, Bin Chen, Ziyun Zeng, and Shu-Tao Xia. 2023. Contrastive masked autoencoders for self-supervised video hashing. In Proceedings of the 37th AAAI Conference on Artificial Intelligence. 2733–2741. DOI:
[57]
Stanley Wasserman and Katherine Faust. 1994. Social Network Analysis: Methods and Applications. Cambridge University Press. DOI:
[58]
Lirong Wu, Haitao Lin, Cheng Tan, Zhangyang Gao, and Stan Z. Li. 2023. Self-supervised learning on graphs: Contrastive, generative, or predictive. IEEE Transactions on Knowledge and Data Engineering 35, 4 (2023), 4216–4235. DOI:
[59]
Yu Xie, Yanfeng Liang, Maoguo Gong, A. K. Qin, Yew-Soon Ong, and Tiantian He. 2022. Semisupervised graph neural networks for graph classification. IEEE Transactions on Cybernetics 53, 10 (2022), 6222–6235. DOI:
[60]
Yaochen Xie, Zhao Xu, Jingtun Zhang, Zhengyang Wang, and Shuiwang Ji. 2023. Self-supervised learning of graph neural networks: A unified review. IEEE Transactions on Pattern Analysis and Machine Intelligence 45, 2 (2023), 2412–2429. DOI:
[61]
Zhenda Xie, Zheng Zhang, Yue Cao, Yutong Lin, Jianmin Bao, Zhuliang Yao, Qi Dai, and Han Hu. 2022. SimMIM: A simple framework for masked image modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9643–9653. DOI:
[62]
Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2019. How powerful are graph neural networks? In Proceedings of the 7th International Conference on Learning Representations.
[63]
Xihong Yang, Yue Liu, Sihang Zhou, Siwei Wang, Wenxuan Tu, Qun Zheng, Xinwang Liu, Liming Fang, and En Zhu. 2023. Cluster-guided contrastive graph clustering network. In Proceedings of the 37th AAAI Conference on Artificial Intelligence. 10834–10842. DOI:
[64]
Zhilin Yang, William W. Cohen, and Ruslan Salakhutdinov. 2016. Revisiting semi-supervised learning with graph embeddings. In Proceedings of the 33nd International Conference on Machine Learning, Vol. 48. 40–48.
[65]
Xumin Yu, Lulu Tang, Yongming Rao, Tiejun Huang, Jie Zhou, and Jiwen Lu. 2022. Point-BERT: Pre-training 3D point cloud transformers with masked point modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 19291–19300. DOI:
[66]
Hengrui Zhang, Qitian Wu, Junchi Yan, David Wipf, and Philip S. Yu. 2021. From canonical correlation analysis to self-supervised graph neural networks. In Advances in Neural Information Processing Systems. 76–89.
[67]
Qi Zhang, Yifei Wang, and Yisen Wang. 2022. How mask matters: Towards theoretical understandings of masked autoencoders. In Advances in Neural Information Processing Systems. 27127–27139.
[68]
Yi Zhu, Lei Li, and Xindong Wu. 2021. Stacked convolutional sparse auto-encoders for representation learning. ACM Transactions on Knowledge Discovery from Data 15, 2 (2021), Article 31, 21 pages. DOI:
[69]
Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, and Liang Wang. 2020. Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131 (2020).
[70]
Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, and Liang Wang. 2021. Graph contrastive learning with adaptive augmentation. In Proceedings of the Web Conference. 2069–2080. DOI:
[71]
Jiabo Zhuang, Shunmei Meng, Jing Zhang, and Victor S. Sheng. 2023. Contrastive learning based graph convolution network for social recommendation. ACM Transactions on Knowledge Discovery from Data 17, 8 (2023), Article 120, 21 pages. DOI:

Cited By

View all
  • (2024)A review of graph neural networks and pretrained language models for knowledge graph reasoningNeurocomputing10.1016/j.neucom.2024.128490609(128490)Online publication date: Dec-2024
  • (2024)A faster deep graph clustering network based on dynamic graph weight update mechanismCluster Computing10.1007/s10586-024-04549-627:9(12123-12140)Online publication date: 1-Dec-2024

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Knowledge Discovery from Data
ACM Transactions on Knowledge Discovery from Data  Volume 18, Issue 6
July 2024
760 pages
EISSN:1556-472X
DOI:10.1145/3613684
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 April 2024
Online AM: 20 February 2024
Accepted: 14 February 2024
Revised: 13 February 2024
Received: 22 August 2023
Published in TKDD Volume 18, Issue 6

Check for updates

Author Tags

  1. Self-supervised representation learning
  2. graph auto-encoder
  3. graph contrastive learning
  4. masking

Qualifiers

  • Research-article

Funding Sources

  • National Key Research and Development Program of China

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)586
  • Downloads (Last 6 weeks)54
Reflects downloads up to 19 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)A review of graph neural networks and pretrained language models for knowledge graph reasoningNeurocomputing10.1016/j.neucom.2024.128490609(128490)Online publication date: Dec-2024
  • (2024)A faster deep graph clustering network based on dynamic graph weight update mechanismCluster Computing10.1007/s10586-024-04549-627:9(12123-12140)Online publication date: 1-Dec-2024

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media