Towards Robust Representations of Spatial Networks Using Graph Neural Networks
<p>Figure showing a heterogeneous graph representation of spatial entities from a map. To the left is a map snippet with streets and buildings highlighted. To the right is the graph representation with the triangles representing buildings on the maps, circles representing the streets. The edges between the shapes denote some defined relation such as distance or spatial intersection. An abstract depiction of heterogeneous graph representations is shown in <a href="#applsci-11-06918-f002" class="html-fig">Figure 2</a>.</p> "> Figure 2
<p>An abstract depiction of a heterogeneous graph representation of spatial entities. The triangles represents buildings, circles represents streets, parallelograms represents natural bodies. The relations between the entities is defined using relations such as the spatial <span class="html-italic">within</span>, spatial <span class="html-italic">intersects</span>, spatial <span class="html-italic">adjacent</span>.</p> "> Figure 3
<p>Figure showing the network schema for the homogeneous and heterogeneous graph representations used in our experiments. To the left, is the network schema of the homogeneous representations for streets and buildings. To the right, is the network schema of the heterogeneous graph representations.</p> ">
Abstract
:1. Introduction
- 1.
- Invariance test—the model results should vary across space.
- 2.
- Representation test—the model should contain spatial representations.
- 3.
- Formulation test—the model formulation should use spatial concepts.
- 4.
- Outcome test—the model inputs and outputs should differ.
2. Background
3. Preamble
3.1. Graph Neural Networks
Problem Definition
4. Methods
4.1. Constructing Heterogeneous Graph Representations
Algorithm 1: Derive Heterogeneous Graph Representation |
|
Computational Complexity
4.2. Learning on Heterogeneous Graph Representations of Spatial Networks
4.2.1. Graph Structure
4.2.2. Modelling Semantics
4.2.3. Neural Architecture
5. Experiments
5.1. Data
5.2. Model Development
Baselines
5.3. Results
- In Table 4, which refers to the macro F1-score for building type inference, the results are not significantly impacted by the training size. Particularly for the GraphSAGE and the heterogeneous models. This provides evidence that model performance is not significantly impacted by the size of the training data for building type inference. Further, the homogeneous models seem to perform best for the building type inference task, with the heterogeneous models following closely behind.
- Table 5 refers to the macro F1-score for street type inference. The heterogeneous models perform best across all training sizes for all datasets. The homogeneous models show similar performance for all training sizes. This suggests that the homogeneous models may not benefit from an increase in training data for the task of street type inference.
- We add Table 6 and Table 7 for completeness. We observe that the models perform sufficiently well and stable in most cases. However, there is no clear benefit to the impact of train size on model performance and the homogeneous and heterogeneous models perform relatively similar with no clear winner.
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Conflicts of Interest
References
- Liu, Z.; Lin, Y.; Sun, M. Representation Learning for Natural Language Processing; Springer Nature: London, UK, 2020. [Google Scholar]
- Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
- Barthélemy, M. Spatial networks. Phys. Rep. 2011, 499, 1–101. [Google Scholar] [CrossRef] [Green Version]
- Shi, C.; Li, Y.; Zhang, J.; Sun, Y.; Philip, S.Y. A survey of heterogeneous information network analysis. IEEE Trans. Knowl. Data Eng. 2016, 29, 17–37. [Google Scholar] [CrossRef]
- Iddianozie, C.; Bertolotto, M.; Mcardle, G. Exploring Budgeted Learning for Data-Driven Semantic Inference via Urban Functions. IEEE Access 2020, 8, 32258–32269. [Google Scholar] [CrossRef]
- Corcoran, P.; Jilani, M.; Mooney, P.; Bertolotto, M. Inferring semantics from geometry: The case of street networks. In Proceedings of the 23rd SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, DC, USA, 3–6 November 2015; ACM: New York, NY, USA, 2015; p. 42. [Google Scholar]
- Iddianozie, C.; McArdle, G. A transfer learning paradigm for spatial networks. In Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing, Limassol, Cyprus, 8–12 April 2019; pp. 659–666. [Google Scholar]
- Iddianozie, C.; McArdle, G. Improved Graph Neural Networks for Spatial Networks Using Structure-Aware Sampling. ISPRS Int. J. Geo-Inf. 2020, 9, 674. [Google Scholar] [CrossRef]
- Janowicz, K.; Gao, S.; McKenzie, G.; Hu, Y.; Bhaduri, B. GeoAI: Spatially Explicit Artificial Intelligence Techniques for Geographic Knowledge Discovery and Beyond; Taylor & Francis: Oxfordshire, UK, 2020. [Google Scholar]
- Mac Aodha, O.; Cole, E.; Perona, P. Presence-only geographical priors for fine-grained image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 9596–9606. [Google Scholar]
- Chu, G.; Potetz, B.; Wang, W.; Howard, A.; Song, Y.; Brucher, F.; Leung, T.; Adam, H. Geo-aware networks for fine-grained recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019. [Google Scholar]
- Yan, B.; Janowicz, K.; Mai, G.; Gao, S. From itdl to place2vec: Reasoning about place type similarity and relatedness by learning embeddings from augmented spatial contexts. In Proceedings of the 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Redondo Beach, CA, USA, 7–10 November 2017; pp. 1–10. [Google Scholar]
- Goodchild, M. Issues in spatially explicit modeling. In Agent-Based Models of Land-Use and Land-Cover Change, Report and Review of an International Workshop, 4–7 October 2001; LUCC International Project Office: Louvain-la-Neuve, Belgium, 2001; pp. 13–17. [Google Scholar]
- Hu, Y.; Gao, S.; Lunga, D.; Li, W.; Newsam, S.; Bhaduri, B. GeoAI at ACM SIGSPATIAL: Progress, challenges, and future directions. Sigspat. Spec. 2019, 11, 5–15. [Google Scholar] [CrossRef]
- Shelton, T.; Poorthuis, A.; Zook, M. Social media and the city: Rethinking urban socio-spatial inequality using user-generated geographic information. Landsc. Urban Plan. 2015, 142, 198–211. [Google Scholar] [CrossRef]
- Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Bonafilia, D.; Gill, J.; Basu, S.; Yang, D. Building High Resolution Maps for Humanitarian Aid and Development with Weakly-and Semi-Supervised Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 1–9. [Google Scholar]
- Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral networks and locally connected networks on graphs. arXiv 2013, arXiv:1312.6203. [Google Scholar]
- Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
- Duvenaud, D.K.; Maclaurin, D.; Iparraguirre, J.; Bombarell, R.; Hirzel, T.; Aspuru-Guzik, A.; Adams, R.P. Convolutional networks on graphs for learning molecular fingerprints. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 2224–2232. [Google Scholar]
- Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 3844–3852. [Google Scholar]
- Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1024–1034. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
- Battaglia, P.W.; Hamrick, J.B.; Bapst, V.; Sanchez-Gonzalez, A.; Zambaldi, V.; Malinowski, M.; Tacchetti, A.; Raposo, D.; Santoro, A.; Faulkner, R.; et al. Relational inductive biases, deep learning, and graph networks. arXiv 2018, arXiv:1806.01261. [Google Scholar]
- Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- He, S.; Bastani, F.; Jagwani, S.; Park, E.; Abbar, S.; Alizadeh, M.; Balakrishnan, H.; Chawla, S.; Madden, S.; Sadeghi, M.A. RoadTagger: Robust Road Attribute Inference with Graph Neural Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 10965–10972. [Google Scholar]
- Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the European Semantic Web Conference, Crete, Greece, 7–13 June 2018; pp. 593–607. [Google Scholar]
- Wang, X.; Ji, H.; Shi, C.; Wang, B.; Ye, Y.; Cui, P.; Yu, P.S. Heterogeneous graph attention network. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 7–13 May 2019; pp. 2022–2032. [Google Scholar]
- Haklay, M. How good is volunteered geographical information? A comparative study of OpenStreetMap and Ordnance Survey datasets. Environ. Plan. B Plan. Des. 2010, 37, 682–703. [Google Scholar] [CrossRef] [Green Version]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
- Wang, M.; Yu, L.; Zheng, D.; Gan, Q.; Gai, Y.; Ye, Z.; Li, M.; Zhou, J.; Huang, Q.; Ma, C.; et al. Deep Graph Library: Towards Efficient and Scalable Deep Learning on Graphs. arXiv 2019, arXiv:1909.01315. [Google Scholar]
Notation | Definition |
---|---|
A graph representation | |
set of nodes in | |
set of edges in | |
semantic path | |
the network schema of | |
number of nodes in | |
number of node types in | |
number of relation types in | |
node representation | |
set of class labels for nodes in | |
Neighbourhood of a node v |
Concept | Definition |
---|---|
Adjacent nodes | The nodes of a graph representation are adjacent if an edge connects them. |
Incidence | A node and an edge of a graph are incident if the edge connects the node to another node. |
Node degree | The number of edges incident to the node. |
Directed graph | A graph representation is said to be directed if the edges have a direction. |
Connected graph | A graph representation is said to be connected if all of its nodes can be accessed from any other node. |
Train | Validation | Test | ||||
---|---|---|---|---|---|---|
Dataset | Streets | Buildings | Streets | Buildings | Streets | Buildings |
Dublin | 29,253 | 3847 | 8358 | 1099 | 4179 | 549 |
Frankfurt | 6794 | 1862 | 1941 | 532 | 970 | 266 |
Toronto | 26,749 | 3691 | 7642 | 1054 | 3821 | 527 |
Manchester | 4688 | 121 | 1339 | 34 | 669 | 17 |
(a) Homogeneous Models | ||||||
Dublin | 17,479 | 12,780 | 4994 | 3651 | 2497 | 1825 |
Frankfurt | 10,874 | 2669 | 3107 | 762 | 1553 | 381 |
Toronto | 48,659 | 8911 | 13,902 | 2546 | 6951 | 1273 |
Manchester | 20,373 | 331 | 5821 | 94 | 2910 | 47 |
(b) Heterogeneous Models |
Dataset | Split | GCN | GAT | GraphSAGE | HetSpatial |
---|---|---|---|---|---|
Dublin | 20% | 0.33 ± 0.012 | 0.36 ± 0.029 | 0.32 ± 0.002 | 0.33 ± 0.010 |
40% | 0.33 ± 0.014 | 0.40 ± 0.031 | 0.34 ± 0.050 | 0.33 ± 0.004 | |
50% | 0.57 ± 0.045 | 0.57 ± 0.037 | 0.32 ± 0.002 | 0.33 ± 0.007 | |
70% | 0.33 ± 0.011 | 0.54 ± 0.047 | 0.39 ± 0.063 | 0.33 ± 0.007 | |
Frankfurt | 20% | 0.50 ± 0.018 | 0.51 ± 0.019 | 0.47 ± 0.014 | 0.41 ± 0.010 |
40% | 0.49 ± 0.025 | 0.50 ± 0.030 | 0.45 ± 0.006 | 0.41 ± 0.008 | |
50% | 0.49 ± 0.029 | 0.52 ± 0.030 | 0.45 ± 0.005 | 0.41 ± 0.012 | |
70% | 0.50 ± 0.026 | 0.48 ± 0.024 | 0.47 ± 0.022 | 0.41 ± 0.007 | |
Toronto | 20% | 0.65 ± 0.035 | 0.49 ± 0.014 | 0.48 ± 0.002 | 0.49 ± 0.001 |
40% | 0.62 ± 0.027 | 0.51 ± 0.020 | 0.48 ± 0.003 | 0.49 ± 0.002 | |
50% | 0.66 ± 0.020 | 0.47 ± 0.013 | 0.48 ± 0.003 | 0.49 ± 0.001 | |
70% | 0.63 ± 0.028 | 0.48 ± 0.002 | 0.48 ± 0.002 | 0.49 ± 0.002 | |
Manchester | 20% | 0.50 ± 0.225 | 0.36 ± 0.080 | 0.40 ± 0.163 | 0.42 ± 0.028 |
40% | 0.36 ± 0.080 | 0.58 ± 0.234 | 0.42 ± 0.162 | 0.42 ± 0.022 | |
50% | 0.74 ± 0.227 | 0.52 ± 0.215 | 0.73 ± 0.250 | 0.41 ± 0.019 | |
70% | 0.58 ± 0.252 | 0.74 ± 0.254 | 0.49 ± 0.199 | 0.41 ± 0.027 |
Dataset | Split | GCN | GAT | GraphSAGE | HetSpatial |
---|---|---|---|---|---|
Dublin | 20% | 0.29 ± 0.001 | 0.32 ± 0.006 | 0.29 ± 0.001 | 0.56 ± 0.348 |
40% | 0.29 ± 0.001 | 0.31 ± 0.005 | 0.29 ± 0.002 | 0.64 ± 0.345 | |
50% | 0.29 ± 0.001 | 0.32 ± 0.007 | 0.29 ± 0.002 | 0.52 ± 0.333 | |
70% | 0.29 ± 0.003 | 0.31 ± 0.004 | 0.29 ± 0.001 | 0.58 ± 0.290 | |
Frankfurt | 20% | 0.29 ± 0.008 | 0.33 ± 0.012 | 0.28 ± 0.004 | 0.52 ± 0.300 |
40% | 0.28 ± 0.003 | 0.33 ± 0.018 | 0.28 ± 0.003 | 0.55 ± 0.307 | |
50% | 0.28 ± 0.003 | 0.33 ± 0.015 | 0.28 ± 0.003 | 0.60 ± 0.310 | |
70% | 0.28 ± 0.003 | 0.34 ± 0.013 | 0.28 ± 0.003 | 0.50 ± 0.281 | |
Toronto | 20% | 0.30 ± 0.005 | 0.32 ± 0.005 | 0.27 ± 0.002 | 0.32 ± 0.087 |
40% | 0.29 ± 0.004 | 0.29 ± 0.003 | 0.27 ± 0.003 | 0.36 ± 0.081 | |
50% | 0.28 ± 0.004 | 0.31 ± 0.007 | 0.27 ± 0.002 | 0.33 ± 0.101 | |
70% | 0.27 ± 0.002 | 0.30 ± 0.007 | 0.27 ± 0.002 | 0.33 ± 0.179 | |
Manchester | 20% | 0.32 ± 0.013 | 0.34 ± 0.021 | 0.31 ± 0.007 | 0.75 ± 0.313 |
40% | 0.31 ± 0.002 | 0.31 ± 0.012 | 0.31 ± 0.002 | 0.65 ± 0.329 | |
50% | 0.31 ± 0.007 | 0.32 ± 0.013 | 0.30 ± 0.003 | 0.78 ± 0.317 | |
70% | 0.31 ± 0.006 | 0.35 ± 0.020 | 0.30 ± 0.003 | 0.62 ± 0.373 |
Dataset | Split | GCN | GAT | GraphSAGE | HetSpatial |
---|---|---|---|---|---|
Dublin | 20% | 0.92 ± 0.013 | 0.94 ± 0.013 | 0.93 ± 0.011 | 0.93 ± 0.007 |
40% | 0.93 ± 0.010 | 0.94 ± 0.008 | 0.93 ± 0.009 | 0.93 ± 0.006 | |
50% | 0.95 ± 0.008 | 0.96 ± 0.008 | 0.93 ± 0.009 | 0.93 ± 0.005 | |
70% | 0.92 ± 0.012 | 0.95 ± 0.008 | 0.93 ± 0.009 | 0.93 ± 0.004 | |
Frankfurt | 20% | 0.73 ± 0.023 | 0.80 ± 0.013 | 0.81 ± 0.023 | 0.69 ± 0.02 |
40% | 0.77 ± 0.022 | 0.80 ± 0.02 | 0.82 ± 0.021 | 0.69 ± 0.024 | |
50% | 0.74 ± 0.031 | 0.81 ± 0.028 | 0.82 ± 0.017 | 0.70 ± 0.026 | |
70% | 0.75 ± 0.020 | 0.83 ± 0.025 | 0.82 ± 0.017 | 0.70 ± 0.018 | |
Toronto | 20% | 0.90 ± 0.013 | 0.91 ± 0.011 | 0.92 ± 0.009 | 0.96 ± 0.006 |
40% | 0.87 ± 0.013 | 0.86 ± 0.016 | 0.93 ± 0.012 | 0.96 ± 0.006 | |
50% | 0.89 ± 0.014 | 0.84 ± 0.014 | 0.92 ± 0.012 | 0.96 ± 0.005 | |
70% | 0.86 ± 0.016 | 0.92 ± 0.009 | 0.93 ± 0.009 | 0.96 ± 0.007 | |
Manchester | 20% | 0.86 ± 0.078 | 0.84 ± 0.064 | 0.87 ± 0.077 | 0.73 ± 0.085 |
40% | 0.85 ± 0.076 | 0.90 ± 0.072 | 0.85 ± 0.077 | 0.71 ± 0.066 | |
50% | 0.95 ± 0.040 | 0.90 ± 0.070 | 0.94 ± 0.058 | 0.71 ± 0.057 | |
70% | 0.90 ± 0.088 | 0.94 ± 0.064 | 0.86 ± 0.075 | 0.70 ± 0.074 |
Dataset | Split | GCN | GAT | GraphSAGE | HetSpatial |
---|---|---|---|---|---|
Dublin | 20% | 0.75 ± 0.005 | 0.62 ± 0.008 | 0.75 ± 0.005 | 0.72 ± 0.180 |
40% | 0.75 ± 0.006 | 0.59 ± 0.008 | 0.75 ± 0.009 | 0.79 ± 0.247 | |
50% | 0.75 ± 0.006 | 0.58 ± 0.008 | 0.75 ± 0.007 | 0.70 ± 0.251 | |
70% | 0.74 ± 0.006 | 0.62 ± 0.006 | 0.75 ± 0.006 | 0.78 ± 0.180 | |
Frankfurt | 20% | 0.71 ± 0.013 | 0.57 ± 0.013 | 0.73 ± 0.017 | 0.72 ± 0.180 |
40% | 0.73 ± 0.010 | 0.59 ± 0.021 | 0.73 ± 0.013 | 0.75 ± 0.199 | |
50% | 0.72 ± 0.012 | 0.62 ± 0.013 | 0.73 ± 0.012 | 0.78 ± 0.213 | |
70% | 0.73 ± 0.016 | 0.61 ± 0.011 | 0.73 ± 0.015 | 0.71 ± 0.247 | |
Toronto | 20% | 0.62 ± 0.006 | 0.52 ± 0.008 | 0.67 ± 0.009 | 0.70 ± 0.121 |
40% | 0.64 ± 0.006 | 0.64 ± 0.007 | 0.67 ± 0.008 | 0.73 ± 0.113 | |
50% | 0.65 ± 0.006 | 0.55 ± 0.009 | 0.67 ± 0.008 | 0.68 ± 0.151 | |
70% | 0.67 ± 0.007 | 0.57 ± 0.009 | 0.67 ± 0.006 | 0.66 ± 0.153 | |
Manchester | 20% | 0.81 ± 0.017 | 0.74 ± 0.02 | 0.84 ± 0.016 | 0.85 ± 0.202 |
40% | 0.85 ± 0.011 | 0.84 ± 0.013 | 0.84 ± 0.011 | 0.78 ± 0.224 | |
50% | 0.83 ± 0.013 | 0.67 ± 0.016 | 0.84 ± 0.013 | 0.85 ± 0.229 | |
70% | 0.84 ± 0.014 | 0.67 ± 0.014 | 0.84 ± 0.013 | 0.73 ± 0.298 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Iddianozie, C.; McArdle, G. Towards Robust Representations of Spatial Networks Using Graph Neural Networks. Appl. Sci. 2021, 11, 6918. https://doi.org/10.3390/app11156918
Iddianozie C, McArdle G. Towards Robust Representations of Spatial Networks Using Graph Neural Networks. Applied Sciences. 2021; 11(15):6918. https://doi.org/10.3390/app11156918
Chicago/Turabian StyleIddianozie, Chidubem, and Gavin McArdle. 2021. "Towards Robust Representations of Spatial Networks Using Graph Neural Networks" Applied Sciences 11, no. 15: 6918. https://doi.org/10.3390/app11156918
APA StyleIddianozie, C., & McArdle, G. (2021). Towards Robust Representations of Spatial Networks Using Graph Neural Networks. Applied Sciences, 11(15), 6918. https://doi.org/10.3390/app11156918