Change-Aware Visual Question Answering

Zhenghang Yuan, Lichao Mou, Xiao Xiang Zhu

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Scopus citations

Abstract

Change detection has been a hot research topic in the field of remote sensing, and it can provide information on observing changes of Earth's surface. However, segmentation-based change results are not very friendly to end users. Thus, in order to improve user experience and offer them high-level semantic information on change detection, we introduce a new task: change-aware visual question answering (VQA) on multi-temporal aerial images. Specifically, given a pair of multi-temporal aerial images and questions, this task aims to automatically provide natural language answers. By doing so, end users have better access to easy-to-understand change information through natural language. Besides, we also create a dataset made of multi-temporal image-question-answer triplets and a baseline method for this task. Experimental results offer valuable insights for the further research on this task.

Original languageEnglish
Title of host publicationIGARSS 2022 - 2022 IEEE International Geoscience and Remote Sensing Symposium
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages227-230
Number of pages4
ISBN (Electronic)9781665427920
DOIs
StatePublished - 2022
Event2022 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2022 - Kuala Lumpur, Malaysia
Duration: 17 Jul 202222 Jul 2022

Publication series

NameInternational Geoscience and Remote Sensing Symposium (IGARSS)
Volume2022-July

Conference

Conference2022 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2022
Country/TerritoryMalaysia
CityKuala Lumpur
Period17/07/2222/07/22

Keywords

  • aerial images
  • change detection
  • deep learning
  • natural language
  • visual question answering (VQA)

Fingerprint

Dive into the research topics of 'Change-Aware Visual Question Answering'. Together they form a unique fingerprint.

Cite this