Automatic Extraction of Damaged Houses by Earthquake Based on Improved YOLOv5: A Case Study in Yangbi
"> Figure 1
<p>The flow for the automatic extraction of damaged houses based on YOLOv5s-ViT-BiFPN.</p> "> Figure 2
<p>The study area.</p> "> Figure 3
<p>The samples of types of damaged houses. The green dots are the locations of field investigation.</p> "> Figure 4
<p>The UAV orthophotos acquired after the Yangbi Earthquake in Cangshanxi Town, Yangbi County, Yunnan Province: (<b>a</b>) Beiyinpo; (<b>b</b>) Jinniu; (<b>c</b>) Xiajie; (<b>d</b>) Baiyang (<b>e</b>) Cunwei; (<b>f</b>) Baimu; (<b>g</b>) Hetaoyuan, and (<b>h</b>) Longjing.</p> "> Figure 5
<p>The samples of damaged houses by the Yangbi Earthquake. The red boxes are the bounding boxes for damaged houses.</p> "> Figure 6
<p>The network architecture of YOLOv5.</p> "> Figure 7
<p>Examples of Mosaic data enhancement. Mosaic scales the four different images and arranges them to fit in the desired output size. The red boxes are the bounding boxes for damaged houses.</p> "> Figure 8
<p>The Focus structure of YOLOv5.</p> "> Figure 9
<p>The structure of the Vision Transformer.</p> "> Figure 10
<p>Replacing PANet with BiFPN to improve Feature Fusion Network. For PANet, a top-down and bottom-up pathway were adopted to fuse multi-scale features; for BiFPN, the strategy for top-down and bottom-up bidirectional feature fusion was used and then repeated, applying the same block.</p> "> Figure 11
<p>The improved network architecture for YOLOv5s-ViT-BiFPN. The Vision Transformer is inserted behind the Backbone. The PANet is replaced by BiFPN to fuse the multi-scale features and, in this study, only repeated once for efficiency. The blue box is the outputs for different scales.</p> "> Figure 12
<p>Comparison of change curves of the loss function and AP for 3 models: (<b>a</b>) Change curve of loss function and (<b>b</b>) change curve of AP.</p> "> Figure 13
<p>Accuracy comparison for the performances of the four models based on the metrics Precision (%), Recall (%), F1 (%), AP (%).</p> "> Figure 14
<p>The test results of YOLOv5s-ViT-BiFPN for the 5 test area: (<b>d</b>) Baiyang (<b>e</b>) Cunwei; (<b>f</b>) Baimu; (<b>g</b>) Hetaoyuan, and (<b>h</b>) Longjing. The red blocks are the damaged houses, and the green blocks are the missing targets.</p> "> Figure 14 Cont.
<p>The test results of YOLOv5s-ViT-BiFPN for the 5 test area: (<b>d</b>) Baiyang (<b>e</b>) Cunwei; (<b>f</b>) Baimu; (<b>g</b>) Hetaoyuan, and (<b>h</b>) Longjing. The red blocks are the damaged houses, and the green blocks are the missing targets.</p> "> Figure 15
<p>The examples of detection results by YOLOv5s-ViT-BiFPN.</p> "> Figure 15 Cont.
<p>The examples of detection results by YOLOv5s-ViT-BiFPN.</p> "> Figure 16
<p>The samples for UAV images of different types of houses damaged by the Ya’an Ms7.0 earthquake on 20 April. The red vertical bounding boxes are the results of our method. The red irregular polygons are the annotations from references [<a href="#B8-remotesensing-14-00382" class="html-bibr">8</a>,<a href="#B9-remotesensing-14-00382" class="html-bibr">9</a>].</p> "> Figure 17
<p>Visualization of the feature maps. (<b>a</b>) The feature maps of the first layer; (<b>b</b>) The feature maps of the Vision Transformer layer; (<b>c</b>) The feature maps of the BiFPN structure; (<b>d</b>)The feature maps of the output layer.</p> "> Figure 17 Cont.
<p>Visualization of the feature maps. (<b>a</b>) The feature maps of the first layer; (<b>b</b>) The feature maps of the Vision Transformer layer; (<b>c</b>) The feature maps of the BiFPN structure; (<b>d</b>)The feature maps of the output layer.</p> "> Figure 18
<p>Visualization of the heatmaps. (<b>a</b>) The original images; (<b>b</b>) The heatmaps of the first layer; (<b>c</b>) The heatmaps of the Vision Transformer layer; (<b>d</b>) The heatmaps of the BiFPN structure; (<b>e</b>) The heatmaps of the output layer; (<b>f</b>) The final results.</p> ">
Abstract
:1. Introduction
2. Materials and Methods
2.1. Study Area and Data
2.2. Improvement of YOLOv5s
2.2.1. Improvements for Backbone
2.2.2. Improvements for Neck
2.3. Experimental Configuration
2.4. Evaluation Metrics
3. Results
3.1. Performance of the Model
3.2. Applicability of the Model on Test Images
3.3. Transferability of the Model in Ya’an Earthquake
4. Discussion
4.1. Visualization of the Feature Maps
4.2. Further Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Nie, G.; Gao, J.; Ma, Z.; Gao, Q.; Su, G. On the Risk of Earthquake Disaster in China in the Coming 10~15 Years. J. Nat. Disasters 2002, 1, 68–73. (In Chinese) [Google Scholar]
- Wang, Z. A preliminary report on the Great Wenchuan Earthquake. Earthq. Eng. Eng. Vib. 2008, 7, 225–234. [Google Scholar] [CrossRef]
- Chen, L.; Wang, H.; Ran, Y.; Sun, X.; Su, G.; Wang, J.; Tan, X.; Li, Z.; Zhang, X. The MS7.1 Yushu earthquake surface rupture and large historical earthquakes on the Garzê-Yushu Fault. Chin. Sci. Bull. 2010, 55, 3504–3509. [Google Scholar] [CrossRef]
- Zhou, S.; Chen, G.; Fang, L. Distribution Pattern of Landslides Triggered by the 2014 Ludian Earthquake of China: Implications for Regional Threshold Topography and the Seismogenic Fault Identification. ISPRS Int. J. Geo. Inf. 2016, 5, 46. [Google Scholar] [CrossRef] [Green Version]
- Topics on Earthquake Relief and Disaster Relief from the “5.21” Earthquake in Yangbi, Dali Prefecture. Available online: http://www.dali.gov.cn/dlrmzf/c105806/202105/413a1a71166a4209bb7e2a2b94a3e23e.shtml (accessed on 1 September 2021). (In Chinese)
- Wang, X.; Huang, S.; Ding, X.; Cui, L.; Dou, A.; Li, Y. Extraction and Analysis of Building Damage Caused by Nepal Ms8.1 Earthquake from Remote Sensing Images. Technol. Earthq. Disaster Prev. 2015, 10, 481–490. (In Chinese) [Google Scholar]
- Tinka, V.; Jacopo, M.; van den Homberg, M.; Jorma, L. Multi-Hazard and Spatial Transferability of a CNN for Automated Building Damage Assessment. Remote Sens. 2020, 12, 2839. [Google Scholar] [CrossRef]
- Li, S.; Tang, H.; He, S.; Shu, Y.; Mao, T.; Li, J.; Xu, Z. Unsupervised Detection of Earthquake-Triggered Roof-Holes From UAV Images Using Joint Color and Shape Features. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1823–1827. [Google Scholar] [CrossRef]
- Li, S.; Tang, H. Building Damage Extraction Triggered by Earthquake Using the UAV Imagery. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Beijing, China, 7–10 May 2018; Volume XLII-3, pp. 929–936. [Google Scholar] [CrossRef] [Green Version]
- Çömert, R.; Matci, D.K.; Avdan, U. Detection of Collapsed Building from Unmanned Aerial Vehicle Data with Object Based Image Classification. Eskişehir Tech. Univ. J. Sci. Technol. B—Theor. Sci. 2018, 6, 109–116. [Google Scholar] [CrossRef] [Green Version]
- So, S.; Badloe, T.; Noh, J.; Bravo-Abad, J.; Rho, J. Deep learning enabled inverse design in nanophotonics. Nanophotonics 2020, 9, 1041–1057. [Google Scholar] [CrossRef] [Green Version]
- Vetrivel, A.; Gerke, M.; Kerle, N.; Nex, F.; Vosselman, G. Disaster Damage Detection through Dynergistic Use of Deep Learning and 3D Point Cloud Features Derived from Very High Resolution Oblique Aerial Images, and multiple-kernel-learning. ISPRS J. Photogramm. Remote Sens. 2018, 140, 45–59. [Google Scholar] [CrossRef]
- Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. Available online: https://arxiv.org/pdf/1804.02767.pdf (accessed on 3 April 2020).
- Redmon, J. Darknet: Open source neural networks in c. Available online: https://pjreddie.com/darknet/ (accessed on 11 January 2021).
- Pi, Y.; Nath, N.D.; Behzadan, A.H. Convolutional neural networks for object detection in aerial imagery for disaster response and recovery. Adv. Eng. Inform. 2020, 43, 101009. [Google Scholar] [CrossRef]
- Ma, H.; Liu, Y.; Ren, Y.; Yu, J. Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3. Remote Sens. 2020, 12, 44. [Google Scholar] [CrossRef] [Green Version]
- Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet v2: Practical guidelines for efficient CNN architecture design. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 122–138. [Google Scholar] [CrossRef] [Green Version]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. Available online: https://arxiv.org/abs/2004.10934 (accessed on 28 June 2020).
- Jocher, G.; Stoken, A.; Borovec, J. Ultralytic/Yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 25 June 2021).
- Lema, D.G.; Pedrayes, O.D.; Usamentiaga, R.; García, D.F.; Alonso, Á. Cost-Performance Evaluation of a Recognition Service of Livestock Activity Using Aerial Images. Remote Sens. 2021, 13, 2318. [Google Scholar] [CrossRef]
- Zhang, H.; Ban, Y.; Guo, L.; Jin, Y.; Chen, L. Detection Method of Remote Sensing Image Ship Based on YOLOv5. Electron. Meas. Technol. 2021, 44, 87–92. (In Chinese) [Google Scholar]
- Zhao, J.; Zhang, X.; Yan, J.; Qiu, X.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W. A Wheat Spike Detection Method in UAV Images Based on Improved YOLOv5. Remote Sens. 2021, 13, 3095. [Google Scholar] [CrossRef]
- Zhang, C.M. Seismic risk-coping behavior in rural ethnic minority communities in Dali, China. Nat. Hazards 2020, 103, 3499–3522. [Google Scholar] [CrossRef]
- Wang, Y.; Shi, P.; Wang, J. The housing loss assessment of rural villages caused by earthquake disaster in Yunnan Province. Acta Seimol. Sin. 2005, 18, 590–601. [Google Scholar] [CrossRef] [Green Version]
- Gao, X.; Ji, J. Analysis of the seismic vulnerability and the structural characteristics of houses in Chinese rural areas. Nat. Hazards 2014, 70, 1099–1114. [Google Scholar] [CrossRef]
- Yu, Y.; Zhao, J.; Gong, Q.; Huang, C.; Zheng, G.; Ma, J. Real-Time Underwater Maritime Object Detection in Side-Scan Sonar Images Based on Transformer-YOLOv5. Remote Sens. 2021, 13, 3555. [Google Scholar] [CrossRef]
- Wang, C.-Y.; Mark Liao, H.-Y.; Wu, Y.-H.; Chen, P.-Y.; Hsieh, J.-W.; Yeh, I.-H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the 2020 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef] [Green Version]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS), Long Beach, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. Available online: https://arxiv.org/abs/2010.11929 (accessed on 21 August 2020).
- Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
- Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar] [CrossRef] [Green Version]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
- Nex, F.; Duarte, D.; Steenbeek, A.; Kerle, N. Towards real-time building damage mapping with low-cost UAV solutions. Remote Sens. 2019, 11, 287. [Google Scholar] [CrossRef] [Green Version]
- Tilon, S.; Nex, F.; Kerle, N.; Vosselman, G. Post-Disaster Building Damage Detection from Earth Observation Imagery Using Unsupervised and Transferable Anomaly Detecting Generative Adversarial Networks. Remote Sens. 2020, 12, 4193. [Google Scholar] [CrossRef]
- Li, Y.; Xu, W.; Chen, H.; Jiang, J.; Li, X. A Novel Framework Based on Mask R-CNN and Histogram Thresholding for Scalable Segmentation of New and Rural Buildings. Remote Sens. 2021, 13, 1070. [Google Scholar] [CrossRef]
- Bhuiyan, M.A.E.; Witharana, C.; Liljedahl, A.K. Use of Very High Spatial Resolution Commercial Satellite Imagery and Deep Learning to Automatically Map Ice-Wedge Polygons across Tundra Vegetation Types. J. Imaging 2020, 6, 137. [Google Scholar] [CrossRef]
- Mahmoud, A.S.; Mohamed, S.A.; EI-Khoribi, R.A.; AbdelSalam, H.M. Object Detection Using Adaptive Mask RCNN in Optical Remote Sensing Images. Int. J. Intell. Eng. Syst. 2020, 13, 65–76. [Google Scholar] [CrossRef]
- Zhao, K.; Kang, J.; Jung, J.; Sohn, G. Building Extraction from Satellite Images Using Mask R-CNN with Building Boundary Regularization. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 242–244. [Google Scholar]
Model | Precision (%) | Recall (%) | F1 (%) | AP (%) | FPS (f/s) | Training Time (h) | Parameter Size (MB) |
---|---|---|---|---|---|---|---|
YOLOv3 | - | - | - | 81.63 | 27 | 5.6 | 235 |
YOLOv5s | 84.97 | 84.56 | 84.76 | 89.17 | 100 | 1.417 | 14.4 |
YOLOv5s-ViT | 88.46 | 87.34 | 87.90 | 90.40 | 100 | 1.467 | 14.4 |
YOLOv5s-ViT-BiFPN | 89.01 | 89.37 | 89.19 | 90.94 | 80 | 1.203 | 16.5 |
Test Areas | The Number for Detection | The Number for Visual Interpretation | Accuracy (%) | Time (s) |
---|---|---|---|---|
(d) Baiyang | 30 | 33 | 90.91 | 36.50 |
(e) Cunwei | 28 | 31 | 90.32 | 26.46 |
(f) Baimu | 66 | 73 | 90.41 | 38.02 |
(g) Hetaoyuan | 71 | 77 | 92.20 | 54.08 |
(h) Longjing | 120 | 130 | 92.31 | 94.45 |
Average Accuracy | - | - | 91.23 | - |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jing, Y.; Ren, Y.; Liu, Y.; Wang, D.; Yu, L. Automatic Extraction of Damaged Houses by Earthquake Based on Improved YOLOv5: A Case Study in Yangbi. Remote Sens. 2022, 14, 382. https://doi.org/10.3390/rs14020382
Jing Y, Ren Y, Liu Y, Wang D, Yu L. Automatic Extraction of Damaged Houses by Earthquake Based on Improved YOLOv5: A Case Study in Yangbi. Remote Sensing. 2022; 14(2):382. https://doi.org/10.3390/rs14020382
Chicago/Turabian StyleJing, Yafei, Yuhuan Ren, Yalan Liu, Dacheng Wang, and Linjun Yu. 2022. "Automatic Extraction of Damaged Houses by Earthquake Based on Improved YOLOv5: A Case Study in Yangbi" Remote Sensing 14, no. 2: 382. https://doi.org/10.3390/rs14020382
APA StyleJing, Y., Ren, Y., Liu, Y., Wang, D., & Yu, L. (2022). Automatic Extraction of Damaged Houses by Earthquake Based on Improved YOLOv5: A Case Study in Yangbi. Remote Sensing, 14(2), 382. https://doi.org/10.3390/rs14020382