REFERRING IMAGE SEGMENTATION FOR REMOTE SENSING DATA

Zhenghang Yuan, Lichao Mou, Yuansheng Hua, Xiao Xiang Zhu

Research output: Contribution to conferencePaperpeer-review

Abstract

In this paper, we present a new task: referring image segmentation for remote sensing data, which targets segmenting out specific objects referred to by natural language. Due to the absence of a dataset for this task, we construct a dataset based on the SkyScapes dataset. Our dataset is designed with linguistically structured expressions that focus on object categories, attributes, and spatial relationships, enabling the generation of binary masks from semantic segmentation maps. To benchmark this task, we evaluate and compare the performance of three different convolutional neural network (CNN)-based methods and a Transformer-based method. Experimental results provide valuable insights into the adaptability of these methods to remote sensing data, highlighting the potential of our dataset as a resource for the remote sensing community to further explore vision-language tasks.

Original languageEnglish
Pages946-949
Number of pages4
DOIs
StatePublished - 2024
Event2024 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2024 - Athens, Greece
Duration: 7 Jul 202412 Jul 2024

Conference

Conference2024 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2024
Country/TerritoryGreece
CityAthens
Period7/07/2412/07/24

Keywords

  • Referring image segmentation
  • remote sensing
  • vision-language task

Fingerprint

Dive into the research topics of 'REFERRING IMAGE SEGMENTATION FOR REMOTE SENSING DATA'. Together they form a unique fingerprint.

Cite this