Real-time tracking of speakers' emotions, states, and traits on mobile platforms

Erik Marchi, Florian Eyben, Gerhard Hagerer, Björn Schuller

Research output: Contribution to journalConference articlepeer-review

13 Scopus citations

Abstract

We demonstrate audEERING's sensAI technology running natively on low-resource mobile devices applied to emotion analytics and speaker characterisation tasks. A show-case application for the Android platform is provided, where audEERING's highly noise robust voice activity detection based on LSTM-RNN is combined with our core emotion recognition and speaker characterisation engine natively on the mobile device. This eliminates the need for network connectivity and allows to perform robust speaker state and trait recognition efficiently in real-time without network transmission lags. Realtime factors are benchmarked for a popular mobile device to demonstrate the efficiency, and average response times are compared to a server based approach. The output of the emotion analysis is visualized graphically in the arousal and valence space alongside the emotion category and further speaker characteristics.

Original languageEnglish
Pages (from-to)1182-1183
Number of pages2
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume08-12-September-2016
StatePublished - 2016
Externally publishedYes
Event17th Annual Conference of the International Speech Communication Association, INTERSPEECH 2016 - San Francisco, United States
Duration: 8 Sep 201616 Sep 2016

Keywords

  • Computational paralinguistics
  • Emotion recognition
  • Mobile application

Fingerprint

Dive into the research topics of 'Real-time tracking of speakers' emotions, states, and traits on mobile platforms'. Together they form a unique fingerprint.

Cite this