TY - JOUR
T1 - SSL4EO-S12
T2 - A large-scale multimodal, multitemporal dataset for self-supervised learning in Earth observation [Software and Data Sets]
AU - Wang, Yi
AU - Braham, Nassim Ait Ali
AU - Xiong, Zhitong
AU - Liu, Chenying
AU - Albrecht, Conrad M.
AU - Zhu, Xiao Xiang
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023/9/1
Y1 - 2023/9/1
N2 - Self-supervised pretraining bears the potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pretraining in the field is based on ImageNet or medium-sized, labeled remote sensing (RS) datasets. In this article, we share an unlabeled dataset Self-Supervised Learning for Earth Observation-Sentinel-1/2 (SSL4EO-S12) to assemble a large-scale, global, multimodal, and multiseasonal corpus of satellite imagery. We demonstrate SSL4EO-S12 to succeed in self-supervised pretraining for a set of representative methods: momentum contrast (MoCo), self-distillation with no labels (DINO), masked autoencoders (MAE), and data2vec, and multiple downstream applications, including scene classification, semantic segmentation, and change detection. Our benchmark results prove the effectiveness of SSL4EO-S12 compared to existing datasets. The dataset, related source code, and pretrained models are available at https://github.com/zhu-xlab/SSL4EO-S12.
AB - Self-supervised pretraining bears the potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pretraining in the field is based on ImageNet or medium-sized, labeled remote sensing (RS) datasets. In this article, we share an unlabeled dataset Self-Supervised Learning for Earth Observation-Sentinel-1/2 (SSL4EO-S12) to assemble a large-scale, global, multimodal, and multiseasonal corpus of satellite imagery. We demonstrate SSL4EO-S12 to succeed in self-supervised pretraining for a set of representative methods: momentum contrast (MoCo), self-distillation with no labels (DINO), masked autoencoders (MAE), and data2vec, and multiple downstream applications, including scene classification, semantic segmentation, and change detection. Our benchmark results prove the effectiveness of SSL4EO-S12 compared to existing datasets. The dataset, related source code, and pretrained models are available at https://github.com/zhu-xlab/SSL4EO-S12.
UR - http://www.scopus.com/inward/record.url?scp=85174545416&partnerID=8YFLogxK
U2 - 10.1109/MGRS.2023.3281651
DO - 10.1109/MGRS.2023.3281651
M3 - Article
AN - SCOPUS:85174545416
SN - 2473-2397
VL - 11
SP - 98
EP - 106
JO - IEEE Geoscience and Remote Sensing Magazine
JF - IEEE Geoscience and Remote Sensing Magazine
IS - 3
ER -