Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/3015812.3015982guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Deep neural networks for learning graph representations

Published: 12 February 2016 Publication History

Abstract

In this paper, we propose a novel model for learning graph representations, which generates a low-dimensional vector representation for each vertex by capturing the graph structural information. Different from other previous research efforts, we adopt a random surfing model to capture graph structural information directly, instead of using the sampling-based method for generating linear sequences proposed by Perozzi et al. (2014). The advantages of our approach will be illustrated from both theorical and empirical perspectives. We also give a new perspective for the matrix factorization method proposed by Levy and Goldberg (2014), in which the pointwise mutual information (PMI) matrix is considered as an analytical solution to the objective function of the skip-gram model with negative sampling proposed by Mikolov et al. (2013). Unlike their approach which involves the use of the SVD for finding the low-dimensitonal projections from the PMI matrix, however, the stacked denoising autoencoder is introduced in our model to extract complex features and model non-linearities. To demonstrate the effectiveness of our model, we conduct experiments on clustering and visualization tasks, employing the learned vertex representations as features. Empirical results on datasets of varying sizes show that our model outperforms other stat-of-the-art models in such tasks.

References

[1]
Agirre, E.; Alfonseca, E.; Hall, K.; Kravalova, J.; Paşca, M.; and Soroa, A. 2009. A study on similarity and relatedness using distributional and wordnet-based approaches. In NAACL, 19-27.
[2]
Arthur, D., and Vassilvitskii, S. 2007. k-means++: The advantages of careful seeding. In SODA, 1027-1035.
[3]
Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H.; et al. 2007. Greedy layer-wise training of deep networks. In NIPS, 153-160.
[4]
Bourlard, H., and Kamp, Y. 1988. Auto-association by multilayer perceptrons and singular value decomposition. Biological cybernetics 59(4-5):291-294.
[5]
Bullinaria, J. A., and Levy, J. P. 2007. Extracting semantic representations from word co-occurrence statistics: A computational study. Behavior research methods 39(3):510-526.
[6]
Bullinaria, J. A., and Levy, J. P. 2012. Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and svd. Behavior research methods 44(3):890-907.
[7]
Cao, S.; Lu, W.; and Xu, Q. 2015. Grarep: Learning graph representations with global structural information. In CIKM, 891-900.
[8]
Church, K. W., and Hanks, P. 1990. Word association norms, mutual information, and lexicography. Computational linguistics 16(1):22-29.
[9]
Dahl, G. E.; Yu, D.; Deng, L.; and Acero, A. 2012. Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition. Audio, Speech, and Language Processing, IEEE Transactions on 20(1):30-42.
[10]
Eckart, C., and Young, G. 1936. The approximation of one matrix by another of lower rank. Psychometrika 1(3):211-218.
[11]
Finkelstein, L.; Gabrilovich, E.; Matias, Y.; Rivlin, E.; Solan, Z.; Wolfman, G.; and Ruppin, E. 2001. Placing search in context: The concept revisited. In WWW, 406-414.
[12]
Gutmann, M. U., and Hyvärinen, A. 2012. Noise-contrastive estimation of unnormalized statistical models, with applications to natural image statistics. JMLR 13(1):307-361.
[13]
Hinton, G. E., and Salakhutdinov, R. R. 2006. Reducing the dimensionality of data with neural networks. Science 313(5786):504-507.
[14]
Hinton, G. E., and Zemel, R. S. 1994. Autoencoders, minimum description length, and helmholtz free energy. In NIPS, 3-10.
[15]
Huang, E. H.; Socher, R.; Manning, C. D.; and Ng, A. Y. 2012. Improving word representations via global context and multiple word prototypes. In ACL, 873-882.
[16]
Krizhevsky, A.; Sutskever, I.; and Hinton, G. E. 2012. Imagenet classification with deep convolutional neural networks. In NIPS, 1097-1105.
[17]
Levy, O., and Goldberg, Y. 2014. Neural word embedding as implicit matrix factorization. In NIPS, 2177-2185.
[18]
Levy, O.; Goldberg, Y.; and Dagan, I. 2015. Improving distributional similarity with lessons learned from word embeddings. TACL 3:211-225.
[19]
Lichman, M. 2013. UCI machine learning repository.
[20]
Lund, K., and Burgess, C. 1996. Producing high-dimensional semantic spaces from lexical co-occurrence. Behavior Research Methods, Instruments, and Computers 28(2):203-208.
[21]
Macskassy, S. A., and Provost, F. 2003. A simple relational classifier. Technical report, DTIC Document.
[22]
Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G. S.; and Dean, J. 2013. Distributed representations of words and phrases and their compositionality. In NIPS, 3111-3119.
[23]
Miller, G. A., and Charles, W. G. 1991. Contextual correlates of semantic similarity. Language and cognitive processes 6(1):1-28.
[24]
Mnih, A., and Teh, Y. W. 2012. A fast and simple algorithm for training neural probabilistic language models. arXiv preprint arXiv:1206.6426.
[25]
Pennington, J.; Socher, R.; and Manning, C. D. 2014. Glove: Global vectors for word representation. In EMNLP, 1532-1543.
[26]
Perozzi, B.; Al-Rfou, R.; and Skiena, S. 2014. Deepwalk: Online learning of social representations. In SIGKDD, 701-710.
[27]
Shaoul, C. 2010. The westbury lab wikipedia corpus. Edmonton, AB: University of Alberta.
[28]
Strehl, A.; Ghosh, J.; and Mooney, R. 2000. Impact of similarity measures on web-page clustering. In AAAI, 58-64.
[29]
Tang, L., and Liu, H. 2009a. Relational learning via latent social dimensions. In SIGKDD, 817-826.
[30]
Tang, L., and Liu, H. 2009b. Scalable learning of collective behavior based on sparse social dimensions. In CIKM, 1107-1116.
[31]
Tang, L., and Liu, H. 2011. Leveraging social media networks for classification. Data Mining and Knowledge Discovery 23(3):447-478.
[32]
Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; and Mei, Q. 2015. Line: Large-scale information network embedding. In WWW, 1067-1077.
[33]
Tian, F.; Gao, B.; Cui, Q.; Chen, E.; and Liu, T.-Y. 2014. Learning deep representations for graph clustering. In AAAI.
[34]
Van der Maaten, L., and Hinton, G. 2008. Visualizing data using t-sne. JMLR 9(2579-2605):85.
[35]
Vincent, P.; Larochelle, H.; Bengio, Y.; and Manzagol, P.-A. 2008. Extracting and composing robust features with denoising autoencoders. In ICML, 1096-1103.
[36]
Zar, J. H. 1972. Significance testing of the spearman rank correlation coefficient. Journal of the American Statistical Association 67(339):578-580.
[37]
Zesch, T.; Müller, C.; and Gurevych, I. 2008. Using wiktionary for computing semantic relatedness. In AAAI, volume 8, 861-866.

Cited By

View all
  • (2022)Graph Neural Networks: Taxonomy, Advances, and TrendsACM Transactions on Intelligent Systems and Technology10.1145/349516113:1(1-54)Online publication date: 10-Jan-2022
  • (2022)Network Representation Learning: From Preprocessing, Feature Extraction to Node EmbeddingACM Computing Surveys10.1145/349120655:2(1-35)Online publication date: 18-Jan-2022
  • (2021)RWNEJournal of Artificial Intelligence Research10.1613/jair.1.1256771(237-263)Online publication date: 10-Sep-2021
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
AAAI'16: Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence
February 2016
4406 pages

Sponsors

  • Association for the Advancement of Artificial Intelligence

Publisher

AAAI Press

Publication History

Published: 12 February 2016

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2022)Graph Neural Networks: Taxonomy, Advances, and TrendsACM Transactions on Intelligent Systems and Technology10.1145/349516113:1(1-54)Online publication date: 10-Jan-2022
  • (2022)Network Representation Learning: From Preprocessing, Feature Extraction to Node EmbeddingACM Computing Surveys10.1145/349120655:2(1-35)Online publication date: 18-Jan-2022
  • (2021)RWNEJournal of Artificial Intelligence Research10.1613/jair.1.1256771(237-263)Online publication date: 10-Sep-2021
  • (2021)I Know What You Need: Investigating Document Retrieval Effectiveness with Partial Session ContextsACM Transactions on Information Systems10.1145/348866740:3(1-30)Online publication date: 17-Nov-2021
  • (2021)Graph Embedding Based on Euclidean Distance Matrix and its ApplicationsProceedings of the 30th ACM International Conference on Information & Knowledge Management10.1145/3459637.3482261(1140-1149)Online publication date: 26-Oct-2021
  • (2021)SMADProceedings of the 30th ACM International Conference on Information & Knowledge Management10.1145/3459637.3482183(3543-3547)Online publication date: 26-Oct-2021
  • (2021)Health Status Prediction with Local-Global Heterogeneous Behavior GraphACM Transactions on Multimedia Computing, Communications, and Applications10.1145/345789317:4(1-21)Online publication date: 12-Nov-2021
  • (2021)Learning Graph Neural Networks with Positive and Unlabeled NodesACM Transactions on Knowledge Discovery from Data10.1145/345031615:6(1-25)Online publication date: 28-Jun-2021
  • (2021)Inductive Entity Representations from Text via Link PredictionProceedings of the Web Conference 202110.1145/3442381.3450141(798-808)Online publication date: 19-Apr-2021
  • (2021)Learning and Updating Node Embedding on Dynamic Heterogeneous Information NetworkProceedings of the 14th ACM International Conference on Web Search and Data Mining10.1145/3437963.3441745(184-192)Online publication date: 8-Mar-2021
  • Show More Cited By

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media