Spatial Attention-Based Kernel Point Convolution Network for Semantic Segmentation of Transmission Corridor Scenarios in Airborne Laser Scanning Point Clouds
<p>Overall processing workflow of transmission corridor scene semantic segmentation, including two stages: data preprocessing and semantic segmentation.</p> "> Figure 2
<p>Schematic diagram of grid sampling (taking towers as an example). The left side of the diagram represents the original point cloud, while the right side illustrates the grid sampling method and its results.</p> "> Figure 3
<p>Proposed SA-KPConv network architecture. Light green rectangles represent kernel point convolution blocks, dark green rectangles indicate geometric neighborhood features, dark blue rectangles are unary blocks, and light blue rectangles show merging operations. Orange rectangles denote the spatial attention module, while purple and orange arrows indicate upsampling and downsampling processes, respectively.</p> "> Figure 4
<p>Structure of the kernel point convolution layer. This layer updates the weight of each point based on the kernel point function, facilitating the extraction of local geometric features.</p> "> Figure 5
<p>Spatial attention module. The features obtained after kernel point convolution undergo global attention updates through this module, enhancing the model’s prediction accuracy.</p> "> Figure 6
<p>Sample proportions of each category in the datasets. Blue represents important facilities within the transmission corridor, while gray indicates other categories.</p> "> Figure 7
<p>Data and annotations used in experiments, including flatland, buildings, and mountainous areas.</p> "> Figure 8
<p>Semantic segmentation results of our methods in flat and built-up areas, where (<b>a</b>) represents built-up areas and (<b>b</b>) represents flat areas. The red circles mark the parts that were incorrectly predicted.</p> "> Figure 9
<p>Semantic segmentation results in mountainous areas. The black box highlights the details of the power tower section.</p> "> Figure 10
<p>Qualitative results of semantic segmentation compared to different methods. The red circles mark the parts that were incorrectly predicted.</p> "> Figure 11
<p>Qualitative results of semantic segmentation compared to different methods. The black circles mark the parts that were incorrectly predicted.</p> ">
Abstract
:1. Introduction
- To address the issues of sparsity in airborne transmission corridor point clouds and the presence of numerous small-category targets, we employ deformable kernel point convolution (KPConv) to learn the spatial geometric features of the power scene point clouds. By updating point weights through kernel functions, we enhance the learning of point cloud convolution, thereby improving the feature extraction capability for various land cover targets.
- To achieve precise semantic segmentation of transmission corridor scenes at a large scale, we introduce a spatial attention module. This module models the complex interactions between points to enhance the model’s perception of contextual information, thereby improving point cloud segmentation accuracy from a global perspective.
- Experiments were conducted on field-collected transmission scene data, achieving an average intersection over union (IoU) of 89.62%, demonstrating superior performance compared to other methods of the same type. Additionally, experiments conducted across multiple terrains yielded mean intersection over union (mIoU) values exceeding 87%, confirming the robustness of our approach.
2. Relate Work
2.1. Hand-Craft Based Methods
2.2. Data-Driven-Based Methods
3. Methodology
3.1. Preprocess
3.1.1. Grid Sampling
3.1.2. Data Augmentation Strategy
3.2. SA-KPConv
3.2.1. Overall Network Architecture
3.2.2. Kernel Point Convolution Layer
3.2.3. Spatial Attention Module
4. Experiments and Analysis
4.1. Datasets and Metrics
4.1.1. Datasets
4.1.2. Metrics
4.2. Implementation Details
4.3. Semantic Segmentation Results and Analysis
4.4. Comparison to State-of-the-Art Methods
4.5. Effectiveness of Each Proposed Module
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Nguyen, V.N.; Jenssen, R.; Roverso, D. Automatic Autonomous Vision-Based Power Line Inspection: A Review of Current Status and the Potential Role of Deep Learning. Int. J. Electr. Power Energy Syst. 2018, 99, 107–120. [Google Scholar] [CrossRef]
- Yang, L.; Fan, J.; Liu, Y.; Li, E.; Peng, J.; Liang, Z. A Review on State-of-the-Art Power Line Inspection Techniques. IEEE Trans. Instrum. Meas. 2020, 69, 9350–9365. [Google Scholar] [CrossRef]
- Vo, A.V.; Truong-Hong, L.; Laefer, D.F.; Bertolotto, M. Octree-Based Region Growing for Point Cloud Segmentation. ISPRS J. Photogramm. Remote Sens. 2015, 104, 88–100. [Google Scholar] [CrossRef]
- Lin, C.H.; Chen, J.Y.; Su, P.L.; Chen, C.H. Eigen-Feature Analysis of Weighted Covariance Matrices for LiDAR Point Cloud Classification. ISPRS J. Photogramm. Remote Sens. 2014, 94, 70–79. [Google Scholar] [CrossRef]
- Huang, Y.; Du, Y.; Shi, W. Fast and Accurate Power Line Corridor Survey Using Spatial Line Clustering of Point Cloud. Remote Sens. 2021, 13, 1571. [Google Scholar] [CrossRef]
- Liao, L.; Tang, S.; Liao, J.; Li, X.; Wang, W.; Li, Y.; Guo, R. A Supervoxel-Based Random Forest Method for Robust and Effective Airborne LiDAR Point Cloud Classification. Remote Sens. 2022, 14, 1516. [Google Scholar] [CrossRef]
- Tang, Q.; Zhang, L.; Lan, G.; Shi, X.; Duanmu, X.; Chen, K. A Classification Method of Point Clouds of Transmission Line Corridor Based on Improved Random Forest and Multi-Scale Features. Sensors 2023, 23, 1320. [Google Scholar] [CrossRef]
- Wang, Y.; Chen, Q.; Liu, L.; Li, X.; Sangaiah, A.K.; Li, K. Systematic Comparison of Power Line Classification Methods from ALS and MLS Point Cloud Data. Remote Sens. 2018, 10, 1222. [Google Scholar] [CrossRef]
- Jiang, S.; Guo, W.; Fan, Y.; Fu, H. Fast Semantic Segmentation of 3D Lidar Point Cloud Based on Random Forest Method. In Proceedings of the China Satellite Navigation Conference (CSNC 2022), Beijing, China, 25–27 May 2022; Yang, C., Xie, J., Eds.; Springer Nature: Berlin/Heidelberg, Germany, 2022; pp. 415–424. [Google Scholar] [CrossRef]
- Shokri, D.; Rastiveis, H.; Sheikholeslami, S.M.; Shahhoseini, R.; Li, J. Fast Extraction of Power Lines from Mobile LiDAR Point Clouds Based on SVM Classification in Non-Urban Area. Earth Obs. Geomat. Eng. 2021, 5, 63–73. [Google Scholar] [CrossRef]
- Otter, D.W.; Medina, J.R.; Kalita, J.K. A Survey of the Usages of Deep Learning for Natural Language Processing. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 604–624. [Google Scholar] [CrossRef]
- Xu, C.; Leng, B.; Chen, B.; Zhang, C.; Zhou, X. Learning Discriminative and Generative Shape Embeddings for Three-Dimensional Shape Retrieval. IEEE Trans. Multimed. 2019, 22, 2234–2245. [Google Scholar] [CrossRef]
- Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
- Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar] [CrossRef]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Long Beach, CA, USA, 4–9 December 2017; pp. 5105–5114. [Google Scholar]
- Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. Randla-net: Efficient semantic segmentation of large-scale point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11108–11117. [Google Scholar]
- Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE, Seoul, Republic of Korea, 27–28 October 2019; pp. 6410–6419. [Google Scholar] [CrossRef]
- Hui, C.; Tingting, W.; Zuoxiao, D.; Weibin, L.; Menhas, M.I. Power Equipment Segmentation of 3D Point Clouds Based on Geodesic Distance with K-means Clustering. In Proceedings of the 2021 6th International Conference on Power and Renewable Energy (ICPRE), Shanghai, China, 17–20 September 2021; pp. 317–321. [Google Scholar] [CrossRef]
- Ying, S.; Xu, G.; Li, C.; Mao, Z. Point Cluster Analysis Using a 3D Voronoi Diagram with Applications in Point Cloud Segmentation. ISPRS Int. J. Geo-Inf. 2015, 4, 1480–1499. [Google Scholar] [CrossRef]
- Cao, Y.; Wang, Y.; Xue, Y.; Zhang, H.; Lao, Y. FEC: Fast Euclidean Clustering for Point Cloud Segmentation. Drones 2022, 6, 325. [Google Scholar] [CrossRef]
- Réjichi, S.; Chaabane, F. Feature Extraction Using PCA for VHR Satellite Image Time Series Spatio-Temporal Classification. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Athens, Greece, 7–12 July 2024; pp. 485–488. [Google Scholar] [CrossRef]
- Nurunnabi, A.; Belton, D.; West, G. Robust Segmentation in Laser Scanning 3D Point Cloud Data. In Proceedings of the 2012 International Conference on Digital Image Computing Techniques and Applications (DICTA), Fremantle, Australia, 3–5 December 2012; pp. 1–8. [Google Scholar] [CrossRef]
- Duan, Y.; Yang, C.; Chen, H.; Yan, W.; Li, H. Low-Complexity Point Cloud Denoising for LiDAR by PCA-based Dimension Reduction. Opt. Commun. 2021, 482, 126567. [Google Scholar] [CrossRef]
- Zafar, B.; Ashraf, R.; Ali, N.; Ahmed, M.; Jabbar, S.; Naseer, K.; Ahmad, A.; Jeon, G. Intelligent Image Classification-Based on Spatial Weighted Histograms of Concentric Circles. Comput. Sci. Inf. Syst. 2018, 15, 615–633. [Google Scholar] [CrossRef]
- Lawin, F.J.; Danelljan, M.; Tosteberg, P.; Bhat, G.; Khan, F.S.; Felsberg, M. Deep Projective 3D Semantic Segmentation. In Proceedings of the Computer Analysis of Images and Patterns, Ystad, Sweden, 22–24 August 2017; Felsberg, M., Heyden, A., Krüger, N., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 95–107. [Google Scholar] [CrossRef]
- Huang, J.; Yan, W.; Li, T.; Liu, S.; Li, G. Learning the Global Descriptor for 3-D Object Recognition Based on Multiple Views Decomposition. IEEE Trans. Multimed. 2020, 24, 188–201. [Google Scholar] [CrossRef]
- Tchapmi, L.; Choy, C.; Armeni, I.; Gwak, J.; Savarese, S. SEGCloud: Semantic Segmentation of 3D Point Clouds. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 537–547. [Google Scholar] [CrossRef]
- Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1912–1920. [Google Scholar] [CrossRef]
- Meng, H.Y.; Gao, L.; Lai, Y.K.; Manocha, D. VV-net: Voxel VAE Net with Group Convolutions for Point Cloud Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 8499–8507. [Google Scholar] [CrossRef]
- Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution on X-transformed Points. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Curran Associates, Inc.: New York, NY, USA, 2018; Volume 31. [Google Scholar]
- Ye, X.; Li, J.; Huang, H.; Du, L.; Zhang, X. 3D Recurrent Neural Networks with Context Fusion for Point Cloud Semantic Segmentation. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; Volume 11211, pp. 415–430. [Google Scholar] [CrossRef]
- Zhang, L.; Wang, J.; Shen, Y.; Liang, J.; Chen, Y.; Chen, L.; Zhou, M. A Deep Learning Based Method for Railway Overhead Wire Reconstruction from Airborne LiDAR Data. Remote Sens. 2022, 14, 5272. [Google Scholar] [CrossRef]
- Ni, P.; Li, X.; Xu, W.; Zhou, X.; Jiang, T.; Hu, W. Robust 3D Semantic Segmentation Method Based on Multi-Modal Collaborative Learning. Remote Sens. 2024, 16, 453. [Google Scholar] [CrossRef]
- Sun, T.; Zhang, Z.; Tan, X.; Qu, Y.; Xie, Y. Image Understands Point Cloud: Weakly Supervised 3D Semantic Segmentation via Association Learning. IEEE Trans. Image Process. 2024, 33, 1838–1852. [Google Scholar] [CrossRef]
- Kang, X.; Chu, L.; Li, J.; Chen, X.; Lu, Y. Hierarchical Intra-modal Correlation Learning for Label-free 3D Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 28244–28253. [Google Scholar]
- Xu, J.; Yang, W.; Kong, L.; Liu, Y.; Zhang, R.; Zhou, Q.; Fei, B. Visual Foundation Models Boost Cross-Modal Unsupervised Domain Adaptation for 3D Semantic Segmentation. arXiv 2024, arXiv:2403.10001. [Google Scholar]
- Xu, R.; Wang, C.; Zhang, D.; Zhang, M.; Xu, S.; Meng, W.; Zhang, X. DefFusion: Deformable Multimodal Representation Fusion for 3D Semantic Segmentation. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 7732–7739. [Google Scholar] [CrossRef]
- Park, J.; Lee, S.; Kim, S.; Xiong, Y.; Kim, H.J. Self-Positioning Point-Based Transformer for Point Cloud Understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 21814–21823. [Google Scholar]
- Wu, X.; Jiang, L.; Wang, P.S.; Liu, Z.; Liu, X.; Qiao, Y.; Ouyang, W.; He, T.; Zhao, H. Point Transformer V3: Simpler Faster Stronger. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 4840–4851. [Google Scholar]
- Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
- Robert, D.; Raguet, H.; Landrieu, L. Scalable 3D Panoptic Segmentation As Superpoint Graph Clustering. In Proceedings of the 2024 International Conference on 3D Vision (3DV), Davos, Switzerland, 18–21 May 2024; pp. 179–189. [Google Scholar] [CrossRef]
- Robert, D.; Raguet, H.; Landrieu, L. Efficient 3D Semantic Segmentation with Superpoint Transformer. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 17149–17158. [Google Scholar] [CrossRef]
- Landrieu, L.; Simonovsky, M. Large-Scale Point Cloud Semantic Segmentation With Superpoint Graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Su, C.; Wu, X.; Guo, Y.; Lai, C.S.; Xu, L.; Zhao, X. Automatic Multi-Source Data Fusion Technique of Powerline Corridor Using UAV Lidar. In Proceedings of the 2022 IEEE International Smart Cities Conference (ISC2), Paphos, Cyprus, 26–29 September 2022; pp. 1–5. [Google Scholar] [CrossRef]
- Wang, G.; Wang, L.; Wu, S.; Zu, S.; Song, B. Semantic Segmentation of Transmission Corridor 3D Point Clouds Based on CA-PointNet++. Electronics 2023, 12, 2829. [Google Scholar] [CrossRef]
- Zhao, W.; Dong, Q.; Zuo, Z. A Point Cloud Segmentation Method for Power Lines and Towers Based on a Combination of Multiscale Density Features and Point-Based Deep Learning. Int. J. Digit. Earth 2023, 16, 620–644. [Google Scholar] [CrossRef]
- Yu, H.; Wang, Z.; Zhou, Q.; Ma, Y.; Wang, Z.; Liu, H.; Ran, C.; Wang, S.; Zhou, X.; Zhang, X. Deep-Learning-Based Semantic Segmentation Approach for Point Clouds of Extra-High-Voltage Transmission Lines. Remote Sens. 2023, 15, 2371. [Google Scholar] [CrossRef]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
- Ni, H.; Lin, X.; Zhang, J. Classification of ALS Point Cloud with Improved Point Cloud Segmentation and Random Forests. Remote Sens. 2017, 9, 288. [Google Scholar] [CrossRef]
Sub-Datasets | IoU | mIoU | OA | |||||
---|---|---|---|---|---|---|---|---|
Ground | Building | Low Vegetation | High Vegetation | Conductor | Structure | |||
Building Area | 90.17% | 80.22% | 88.24% | 93.66% | 87.21% | 84.48% | 87.33% | 97.34% |
Mountain Area | 91.40% | 85.16% | 85.65% | 95.03% | 90.95% | 84.14% | 88.72% | 95.74% |
Flatland Area | 92.43% | 72.97% | 98.59% | 95.59% | 98.70% | 96.48% | 92.46% | 99.06% |
Methods | Ground | Building | Low Vegetation | High Vegetation | Conductor | Structure | mIoU | OA |
---|---|---|---|---|---|---|---|---|
RF | 43.94% | 1.67% | 35.00% | 84.11% | 68.92% | 67.18% | 50.14% | 72.75% |
PointNet++ | 65.32% | 31.63% | 55.76% | 88.10% | 77.63% | 56.74% | 62.53% | 78.12% |
RandLA-Net | 84.87% | 62.14% | 78.89% | 92.89% | 76.32% | 62.88% | 76.33% | 87.62% |
KPFCNN | 80.14% | 75.71% | 59.12% | 94.15% | 93.99% | 85.27% | 81.40% | 90.78% |
SPT | 85.56% | 84.05% | 79.32% | 86.50% | 95.27% | 88.56% | 86.54% | 93.83% |
Ours | 86.70% | 83.01% | 89.74% | 93.83% | 94.75% | 89.67% | 89.62% | 95.74% |
mIoU (%) | OA (%) | |
---|---|---|
WO SA | 83.78 | 93.78 |
WO FL | 87.31 | 94.53 |
WO KP | 76.05 | 82.72 |
SA-KPConv (Ours) | 88.95 | 95.74 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhou, F.; Wen, G.; Ma, Y.; Pan, H.; Wang, G.; Wang, Y. Spatial Attention-Based Kernel Point Convolution Network for Semantic Segmentation of Transmission Corridor Scenarios in Airborne Laser Scanning Point Clouds. Electronics 2024, 13, 4501. https://doi.org/10.3390/electronics13224501
Zhou F, Wen G, Ma Y, Pan H, Wang G, Wang Y. Spatial Attention-Based Kernel Point Convolution Network for Semantic Segmentation of Transmission Corridor Scenarios in Airborne Laser Scanning Point Clouds. Electronics. 2024; 13(22):4501. https://doi.org/10.3390/electronics13224501
Chicago/Turabian StyleZhou, Fangrong, Gang Wen, Yi Ma, Hao Pan, Guofang Wang, and Yifan Wang. 2024. "Spatial Attention-Based Kernel Point Convolution Network for Semantic Segmentation of Transmission Corridor Scenarios in Airborne Laser Scanning Point Clouds" Electronics 13, no. 22: 4501. https://doi.org/10.3390/electronics13224501