TEMPORAL RELATIONS MATTER: A TWO-PATHWAY NETWORK FOR AERIAL VIDEO RECOGNITION

Pu Jin, Lichao Mou, Yuansheng Hua, Gui Song Xia, Xiao Xiang Zhu

Research output: Contribution to conferencePaperpeer-review

3 Scopus citations

Abstract

With the increasing volume of aerial videos, the demand for automatically parsing these videos is surging. To achieve this, current researches mainly focus on extracting a holistic feature with convolutions along both spatial and temporal dimensions. However, these methods are limited by small temporal receptive fields and cannot adequately capture long-term temporal dependencies which are important for describing complicated dynamics. In this paper, we propose a novel two-pathway network to model not only holistic features, but also temporal relations for aerial video classification. More specially, our model employs a two-pathway architecture: (1) a holistic representation pathway to learn a general feature of frame appearances and short-term temporal variations and (2) a temporal relation pathway to capture multi-scale temporal relations across arbitrary frames, providing long-term temporal dependencies. Our model is evaluated on event recognition dataset, ERA, and achieves the state-of-the-art results. This demonstrates its effectiveness and good generalization capacity.

Original languageEnglish
Pages8221-8224
Number of pages4
DOIs
StatePublished - 2021
Event2021 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2021 - Brussels, Belgium
Duration: 12 Jul 202116 Jul 2021

Conference

Conference2021 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2021
Country/TerritoryBelgium
CityBrussels
Period12/07/2116/07/21

Keywords

  • Aerial video classification
  • convolutional neural networks (CNNs)
  • holistic features
  • temporal relations
  • two-pathway
  • unmanned aerial vehicles (UAVs)

Fingerprint

Dive into the research topics of 'TEMPORAL RELATIONS MATTER: A TWO-PATHWAY NETWORK FOR AERIAL VIDEO RECOGNITION'. Together they form a unique fingerprint.

Cite this