Contrastive Learning Based Modality-Invariant Feature Acquisition for Robust Multimodal Emotion Recognition With Missing Modalities

Rui Liu, Haolin Zuo, Zheng Lian, Bjorn W. Schuller, Haizhou Li

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

Multimodal emotion recognition (MER) aims to understand the way that humans express their emotions by exploring complementary information across modalities. However, it is hard to guarantee that full-modality data is always available in real-world scenarios. To deal with missing modalities, researchers focused on meaningful joint multimodal representation learning during cross-modal missing modality imagination. However, the cross-modal imagination mechanism is highly susceptible to errors due to the "modality gap"issue, which affects the imagination accuracy, thus, the final recognition performance. To this end, we introduce the concept of a modality-invariant feature into the missing modality imagination network, which contains two key modules: 1) a novel contrastive learning-based module to extract modality-invariant features under full modalities and 2) a robust imagination module based on imagined invariant features to reconstruct missing information under missing conditions. Finally, we incorporate imagined and available modalities for emotion recognition. Experimental results on benchmark datasets demonstrate that our proposed method outperforms existing state-of-the-art strategies. Compared with our previous work, our extended version is more effective on multimodal emotion recognition with missing modalities.

Original languageEnglish
Pages (from-to)1856-1873
Number of pages18
JournalIEEE Transactions on Affective Computing
Volume15
Issue number4
DOIs
StatePublished - 2024
Externally publishedYes

Keywords

  • Contrastive learning
  • invariant feature
  • missing modality imagination
  • modality gap
  • multimodal emotion recognition

Fingerprint

Dive into the research topics of 'Contrastive Learning Based Modality-Invariant Feature Acquisition for Robust Multimodal Emotion Recognition With Missing Modalities'. Together they form a unique fingerprint.

Cite this