TY - JOUR
T1 - SEMANTIC SEGMENTATION of AERIAL IMAGERY VIA MULTI-SCALE SHUFFLING CONVOLUTIONAL NEURAL NETWORKS with DEEP SUPERVISION
AU - Chen, K.
AU - Weinmann, M.
AU - Sun, X.
AU - Yan, M.
AU - Hinz, S.
AU - Jutzi, B.
AU - Weinmann, M.
N1 - Publisher Copyright:
© Authors 2018.
PY - 2018/9/23
Y1 - 2018/9/23
N2 - In this paper, we address the semantic segmentation of aerial imagery based on the use of multi-modal data given in the form of true orthophotos and the corresponding Digital Surface Models (DSMs). We present the Deeply-supervised Shuffling Convolutional Neural Network (DSCNN) representing a multi-scale extension of the Shuffling Convolutional Neural Network (SCNN) with deep supervision. Thereby, we take the advantage of the SCNN involving the shuffling operator to effectively upsample feature maps and then fuse multiscale features derived from the intermediate layers of the SCNN, which results in the Multi-scale Shuffling Convolutional Neural Network (MSCNN). Based on the MSCNN, we derive the DSCNN by introducing additional losses into the intermediate layers of the MSCNN. In addition, we investigate the impact of using different sets of hand-crafted radiometric and geometric features derived from the true orthophotos and the DSMs on the semantic segmentation task. For performance evaluation, we use a commonly used benchmark dataset. The achieved results reveal that both multi-scale fusion and deep supervision contribute to an improvement in performance. Furthermore, the use of a diversity of hand-crafted radiometric and geometric features as input for the DSCNN does not provide the best numerical results, but smoother and improved detections for several objects.
AB - In this paper, we address the semantic segmentation of aerial imagery based on the use of multi-modal data given in the form of true orthophotos and the corresponding Digital Surface Models (DSMs). We present the Deeply-supervised Shuffling Convolutional Neural Network (DSCNN) representing a multi-scale extension of the Shuffling Convolutional Neural Network (SCNN) with deep supervision. Thereby, we take the advantage of the SCNN involving the shuffling operator to effectively upsample feature maps and then fuse multiscale features derived from the intermediate layers of the SCNN, which results in the Multi-scale Shuffling Convolutional Neural Network (MSCNN). Based on the MSCNN, we derive the DSCNN by introducing additional losses into the intermediate layers of the MSCNN. In addition, we investigate the impact of using different sets of hand-crafted radiometric and geometric features derived from the true orthophotos and the DSMs on the semantic segmentation task. For performance evaluation, we use a commonly used benchmark dataset. The achieved results reveal that both multi-scale fusion and deep supervision contribute to an improvement in performance. Furthermore, the use of a diversity of hand-crafted radiometric and geometric features as input for the DSCNN does not provide the best numerical results, but smoother and improved detections for several objects.
KW - Aerial Imagery
KW - CNN
KW - Deep Supervision
KW - Multi-Modal Data
KW - Multi-Scale
KW - Semantic Segmentation
UR - http://www.scopus.com/inward/record.url?scp=85056355458&partnerID=8YFLogxK
U2 - 10.5194/isprs-annals-IV-1-29-2018
DO - 10.5194/isprs-annals-IV-1-29-2018
M3 - Conference article
AN - SCOPUS:85056355458
SN - 2194-9042
VL - 4
SP - 29
EP - 36
JO - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
JF - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
IS - 1
T2 - 2018 ISPRS Technical Commission I Midterm Symposium on Innovative Sensing - From Sensors to Methods and Applications
Y2 - 10 October 2018 through 12 October 2018
ER -