Research Article
BibTex RIS Cite

Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images

Year 2021, , 358 - 368, 25.09.2021
https://doi.org/10.28979/jarnas.911130

Abstract

Boundary extraction in remote sensing has an important task in studies such as environmental observa-tion, risk management and monitoring urban growth. Although significant progress has been made in the different calculation methods proposed, there are issues that need improvement, especially in terms of accuracy, efficiency and speed. In this study, dual stream network architecture of three different models that can obtain boundary extraction by using normalized Digital Surface Model (nDSM), Normalized Difference Vegetation Index (NDVI) and Near-Infrared (IR) band as the second stream, was explained. Model I is designed as the original HED, whereas the second stream of Model II, III, and IV use nDSM, nDSM + NDVI and nDSM + NDVI + IR, respectively. Thus, by comparing the models trained based on different data combinations, the contribution of different input data to the success of boundary extraction was revealed. For the training of the models, boundary maps produced from The International Society for Photogrammetry and Remote Sensing (ISPRS) Potsdam data set and input datasets augmented by rotation, mirroring and rotation were used. When the test results obtained from two-stream and multi-data-based models are evaluated, 11% better accuracy has achieved with Model IV compared to the original HED. The outcomes clearly revealed the importance of using multispectral band, height data and vegetation information as input data in boundary extraction beside commonly used RGB images.

Supporting Institution

TÜBİTAK

Project Number

119Y363

Thanks

Acknowledgement This research was supported by The Scientific and Technological Research Council of Turkey (TÜBİTAK), Project No: 119Y363.

References

  • Altınoluk E., Akçay Ö., Kınacı A. C., Avşar Ö., Polat A. B. & Aydar U. (2020). Effects of Orthophoto Band Combinations on Semantic Segmentation. Intercontinental Geoinformation Days (IGD), (pp. 9-12).
  • Chen, L. C., Papandreou, G., Kokkinos, I., Murphy, K., & Yuille, A. L. (2014). Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv preprint arXiv:1412.7062.
  • Chen, L. C., Barron, J. T., Papandreou, G., Murphy, K., & Yuille, A. L. (2016). Semantic image segmentation with task-specific edge detection using cnns and a discriminatively trained domain transform. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4545-4554).
  • Cheng, G., & Han, J. (2016). A survey on object detection in optical remote sensing images. ISPRS Journal of Photogrammetry and Remote Sensing, 117, 11-28.
  • Dai, J., He, K., & Sun, J. (2016). Instance-aware semantic segmentation via multi-task network cascades. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3150-3158).
  • Guo, Y., Liu, Y., Georgiou, T., & Lew, M. S. (2018). A review of semantic segmentation using deep neural networks. International journal of multimedia information retrieval, 7(2), 87-93.
  • Han, X., Zhong, Y., & Zhang, L. (2017). An efficient and robust integrated geospatial object detection framework for high spatial resolution remote sensing imagery. Remote Sensing, 9(7), 666.
  • Hariharan, B., Arbelaez, P., Bourdev, L., Maji, S., & Malik, J. (2011). Semantic Contours from Inverse Detectors, International conference on computer vision (pp. 991-998). IEEE.
  • Kinzie, J., & Kuh, G. D. (2004). Going DEEP: Learning from campuses that share responsibility for student success. About Campus, 9(5), 2-8.
  • Kokkinos, I. (2015). Pushing the boundaries of boundary detection using deep learning. arXiv preprint arXiv:1511.07386.
  • Kokkinos, I. (2017). Ubernet: Training a universal convolutional neural network for low-, mid-, and high-level vision using diverse datasets and limited memory. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 6129-6138).
  • Lee, C. Y., Xie, S., Gallagher, P., Zhang, Z., & Tu, Z. (2015, February). Deeply-supervised nets. In Artificial intelligence and statistics (pp. 562-570). PMLR.
  • Maninis, K. K., Pont-Tuset, J., Arbeláez, P., & Van Gool, L. (2017). Convolutional oriented boundaries: From image segmentation to high-level tasks. IEEE transactions on pattern analysis and machine intelligence, 40(4), 819-833.
  • Marmanis, D., Schindler, K., Wegner, J. D., Galliani, S., Datcu, M., & Stilla, U. (2018). Classification with an edge: Improving semantic image segmentation with boundary detection. ISPRS Journal of Photogrammetry and Remote Sensing, 135, 158-172.
  • Rottensteiner, F., Sohn, G., Jung, J., Gerke, M., Baillard, C., Benitez, S., & Breitkopf, U. (2012). The ISPRS benchmark on urban object classification and 3D building reconstruction. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences I-3 (2012), Nr. 1, 1(1), 293-298.
  • Shen, W., Wang, X., Wang, Y., Bai, X., & Zhang, Z. (2015). Deepcontour: A deep convolutional feature learned by positive-sharing loss for contour detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3982-3991).
  • Shorten, C., & Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6(1), 1-48.
  • Xie, S., & Tu, Z. (2015). Holistically-nested edge detection. Proceedings of the IEEE international conference on computer vision (pp. 1395-1403).
  • Yu, Z., Feng, C., Liu, M. Y., & Ramalingam, S. (2017). Casenet: Deep category-aware semantic edge detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5964-5973).
Year 2021, , 358 - 368, 25.09.2021
https://doi.org/10.28979/jarnas.911130

Abstract

Project Number

119Y363

References

  • Altınoluk E., Akçay Ö., Kınacı A. C., Avşar Ö., Polat A. B. & Aydar U. (2020). Effects of Orthophoto Band Combinations on Semantic Segmentation. Intercontinental Geoinformation Days (IGD), (pp. 9-12).
  • Chen, L. C., Papandreou, G., Kokkinos, I., Murphy, K., & Yuille, A. L. (2014). Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv preprint arXiv:1412.7062.
  • Chen, L. C., Barron, J. T., Papandreou, G., Murphy, K., & Yuille, A. L. (2016). Semantic image segmentation with task-specific edge detection using cnns and a discriminatively trained domain transform. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4545-4554).
  • Cheng, G., & Han, J. (2016). A survey on object detection in optical remote sensing images. ISPRS Journal of Photogrammetry and Remote Sensing, 117, 11-28.
  • Dai, J., He, K., & Sun, J. (2016). Instance-aware semantic segmentation via multi-task network cascades. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3150-3158).
  • Guo, Y., Liu, Y., Georgiou, T., & Lew, M. S. (2018). A review of semantic segmentation using deep neural networks. International journal of multimedia information retrieval, 7(2), 87-93.
  • Han, X., Zhong, Y., & Zhang, L. (2017). An efficient and robust integrated geospatial object detection framework for high spatial resolution remote sensing imagery. Remote Sensing, 9(7), 666.
  • Hariharan, B., Arbelaez, P., Bourdev, L., Maji, S., & Malik, J. (2011). Semantic Contours from Inverse Detectors, International conference on computer vision (pp. 991-998). IEEE.
  • Kinzie, J., & Kuh, G. D. (2004). Going DEEP: Learning from campuses that share responsibility for student success. About Campus, 9(5), 2-8.
  • Kokkinos, I. (2015). Pushing the boundaries of boundary detection using deep learning. arXiv preprint arXiv:1511.07386.
  • Kokkinos, I. (2017). Ubernet: Training a universal convolutional neural network for low-, mid-, and high-level vision using diverse datasets and limited memory. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 6129-6138).
  • Lee, C. Y., Xie, S., Gallagher, P., Zhang, Z., & Tu, Z. (2015, February). Deeply-supervised nets. In Artificial intelligence and statistics (pp. 562-570). PMLR.
  • Maninis, K. K., Pont-Tuset, J., Arbeláez, P., & Van Gool, L. (2017). Convolutional oriented boundaries: From image segmentation to high-level tasks. IEEE transactions on pattern analysis and machine intelligence, 40(4), 819-833.
  • Marmanis, D., Schindler, K., Wegner, J. D., Galliani, S., Datcu, M., & Stilla, U. (2018). Classification with an edge: Improving semantic image segmentation with boundary detection. ISPRS Journal of Photogrammetry and Remote Sensing, 135, 158-172.
  • Rottensteiner, F., Sohn, G., Jung, J., Gerke, M., Baillard, C., Benitez, S., & Breitkopf, U. (2012). The ISPRS benchmark on urban object classification and 3D building reconstruction. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences I-3 (2012), Nr. 1, 1(1), 293-298.
  • Shen, W., Wang, X., Wang, Y., Bai, X., & Zhang, Z. (2015). Deepcontour: A deep convolutional feature learned by positive-sharing loss for contour detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3982-3991).
  • Shorten, C., & Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6(1), 1-48.
  • Xie, S., & Tu, Z. (2015). Holistically-nested edge detection. Proceedings of the IEEE international conference on computer vision (pp. 1395-1403).
  • Yu, Z., Feng, C., Liu, M. Y., & Ramalingam, S. (2017). Casenet: Deep category-aware semantic edge detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5964-5973).
There are 19 citations in total.

Details

Primary Language English
Subjects Artificial Intelligence, Engineering
Journal Section Research Article
Authors

Özgün Akçay 0000-0003-0474-7518

A. Cumhur Kınacı 0000-0002-8832-5453

Emin Özgür Avşar 0000-0002-3804-1209

Umut Aydar 0000-0002-3987-6435

Project Number 119Y363
Publication Date September 25, 2021
Submission Date April 7, 2021
Published in Issue Year 2021

Cite

APA Akçay, Ö., Kınacı, A. C., Avşar, E. Ö., Aydar, U. (2021). Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images. Journal of Advanced Research in Natural and Applied Sciences, 7(3), 358-368. https://doi.org/10.28979/jarnas.911130
AMA Akçay Ö, Kınacı AC, Avşar EÖ, Aydar U. Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images. JARNAS. September 2021;7(3):358-368. doi:10.28979/jarnas.911130
Chicago Akçay, Özgün, A. Cumhur Kınacı, Emin Özgür Avşar, and Umut Aydar. “Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images”. Journal of Advanced Research in Natural and Applied Sciences 7, no. 3 (September 2021): 358-68. https://doi.org/10.28979/jarnas.911130.
EndNote Akçay Ö, Kınacı AC, Avşar EÖ, Aydar U (September 1, 2021) Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images. Journal of Advanced Research in Natural and Applied Sciences 7 3 358–368.
IEEE Ö. Akçay, A. C. Kınacı, E. Ö. Avşar, and U. Aydar, “Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images”, JARNAS, vol. 7, no. 3, pp. 358–368, 2021, doi: 10.28979/jarnas.911130.
ISNAD Akçay, Özgün et al. “Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images”. Journal of Advanced Research in Natural and Applied Sciences 7/3 (September 2021), 358-368. https://doi.org/10.28979/jarnas.911130.
JAMA Akçay Ö, Kınacı AC, Avşar EÖ, Aydar U. Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images. JARNAS. 2021;7:358–368.
MLA Akçay, Özgün et al. “Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images”. Journal of Advanced Research in Natural and Applied Sciences, vol. 7, no. 3, 2021, pp. 358-6, doi:10.28979/jarnas.911130.
Vancouver Akçay Ö, Kınacı AC, Avşar EÖ, Aydar U. Boundary Extraction Based on Dual Stream Deep Learning Model in High Resolution Remote Sensing Images. JARNAS. 2021;7(3):358-6.


TR Dizin 20466




Academindex 30370    

SOBİAD 20460               

Scilit 30371                            

29804 As of 2024, JARNAS is licensed under a Creative Commons Attribution-NonCommercial 4.0 International Licence (CC BY-NC).