TY - GEN
T1 - On the extraction of training imagery from very large remote sensing datasets for deep convolutional segmenatation networks
AU - Huang, Bohao
AU - Reichman, Daniel
AU - Collins, Leslie M.
AU - Bradbury, Kyle
AU - Malof, Jordan M.
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/10/31
Y1 - 2018/10/31
N2 - In this work, we investigate strategies for training convolutional neural networks (CNNs) to perform recognition on remote sensing imagery. In particular we consider the particular problem of semantic segmentation in which the goal is to obtain a dense pixel-wise labeling of the input imagery. Remote sensing imagery is usually stored in the form of very large images, called "tiles", which are too big to be segmented directly using most CNNs and their associated hardware. Therefore smaller sub-images, called "patches", must be extracted from the available tiles. A popular strategy in the literature is to randomly sample patches from the tiles. However, in this work we demonstrate experimentally that extracting patches randomly from a uniform, non-overlapping spatial grid, leads to more accurate models. Our findings suggest the performance improvements are the result of reducing redundancy within the training dataset. We also find that sampling mini-batches of patches (for stochastic gradient descent) using constraints that maximizes the diversity of images within each batch leads to more accurate models. For example, in this work we constrained patches to come from varying tiles, or cities. These simple strategies contributed to our winning entry (in terms of overall performance) in the first year of the INRIA Building Labeling Challenge.
AB - In this work, we investigate strategies for training convolutional neural networks (CNNs) to perform recognition on remote sensing imagery. In particular we consider the particular problem of semantic segmentation in which the goal is to obtain a dense pixel-wise labeling of the input imagery. Remote sensing imagery is usually stored in the form of very large images, called "tiles", which are too big to be segmented directly using most CNNs and their associated hardware. Therefore smaller sub-images, called "patches", must be extracted from the available tiles. A popular strategy in the literature is to randomly sample patches from the tiles. However, in this work we demonstrate experimentally that extracting patches randomly from a uniform, non-overlapping spatial grid, leads to more accurate models. Our findings suggest the performance improvements are the result of reducing redundancy within the training dataset. We also find that sampling mini-batches of patches (for stochastic gradient descent) using constraints that maximizes the diversity of images within each batch leads to more accurate models. For example, in this work we constrained patches to come from varying tiles, or cities. These simple strategies contributed to our winning entry (in terms of overall performance) in the first year of the INRIA Building Labeling Challenge.
KW - Aerial imagery
KW - Building detection
KW - Convolutional neural networks
KW - Remote sensing data
KW - Semantic segmentation
UR - http://www.scopus.com/inward/record.url?scp=85064265951&partnerID=8YFLogxK
U2 - 10.1109/IGARSS.2018.8519523
DO - 10.1109/IGARSS.2018.8519523
M3 - Conference contribution
AN - SCOPUS:85064265951
T3 - International Geoscience and Remote Sensing Symposium (IGARSS)
SP - 6895
EP - 6898
BT - 2018 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2018 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 38th Annual IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2018
Y2 - 22 July 2018 through 27 July 2018
ER -