TY - GEN
T1 - Deep convolutional segmentation of remote sensing imagery
T2 - 38th Annual IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2018
AU - Huang, Bohao
AU - Collins, Leslie M.
AU - Bradbury, Kyle
AU - Malof, Jordan M.
N1 - Publisher Copyright:
© 2018 IEEE
PY - 2018/10/31
Y1 - 2018/10/31
N2 - In this work we consider the application of convolutional neural networks (CNNs) for the semantic segmentation of remote sensing imagery (e.g., aerial color or hyperspectral imagery). In segmentation the goal is to provide a dense pixel-wise labeling of the input imagery. However, remote sensing imagery is usually stored in the form of very large images, called “tiles”, which are too large to be segmented directly using most CNNs and their associated hardware. During label inference (i.e., obtaining labels for a new large tile) smaller sub-images, called “patches”, are extracted uniformly over a tile and the resulting label maps are “stitched” (or concatenated) to create a tile-sized label map. This approach suffers from computational inefficiency and risks of discontinuities at the boundaries between the output of individual patches. In this work we propose a simple alternative approach in which the input size of the CNN is dramatically increased only during label inference. We evaluate the performance of the proposed approach against a standard stitching approach using two popular segmentation CNN models on the INRIA building labeling dataset. The results suggest that the proposed approach substantially reduces label inference time, while also yielding modest overall label accuracy increases. This approach also contributed to our winning entry (overall performance) in the INRIA building labeling competition.
AB - In this work we consider the application of convolutional neural networks (CNNs) for the semantic segmentation of remote sensing imagery (e.g., aerial color or hyperspectral imagery). In segmentation the goal is to provide a dense pixel-wise labeling of the input imagery. However, remote sensing imagery is usually stored in the form of very large images, called “tiles”, which are too large to be segmented directly using most CNNs and their associated hardware. During label inference (i.e., obtaining labels for a new large tile) smaller sub-images, called “patches”, are extracted uniformly over a tile and the resulting label maps are “stitched” (or concatenated) to create a tile-sized label map. This approach suffers from computational inefficiency and risks of discontinuities at the boundaries between the output of individual patches. In this work we propose a simple alternative approach in which the input size of the CNN is dramatically increased only during label inference. We evaluate the performance of the proposed approach against a standard stitching approach using two popular segmentation CNN models on the INRIA building labeling dataset. The results suggest that the proposed approach substantially reduces label inference time, while also yielding modest overall label accuracy increases. This approach also contributed to our winning entry (overall performance) in the INRIA building labeling competition.
KW - Aerial imagery
KW - Building detection
KW - Convolutional neural networks
KW - Deep learning
KW - Semantic segmentation
UR - http://www.scopus.com/inward/record.url?scp=85064154866&partnerID=8YFLogxK
U2 - 10.1109/IGARSS.2018.8518701
DO - 10.1109/IGARSS.2018.8518701
M3 - Conference contribution
AN - SCOPUS:85064154866
T3 - International Geoscience and Remote Sensing Symposium (IGARSS)
SP - 6899
EP - 6902
BT - 2018 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2018 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 22 July 2018 through 27 July 2018
ER -