Being bored? Recognising natural interest by extensive audiovisual integration for real-life application

Björn Schuller, Ronald Müller, Florian Eyben, Jürgen Gast, Benedikt Hörnler, Martin Wöllmer, Gerhard Rigoll, Anja Höthker, Hitoshi Konosu

Publikation: Beitrag in FachzeitschriftArtikelBegutachtung

141 Zitate (Scopus)

Abstract

Automatic detection of the level of human interest is of high relevance for many technical applications, such as automatic customer care or tutoring systems. However, the recognition of spontaneous interest in natural conversations independently of the subject remains a challenge. Identification of human affective states relying on single modalities only is often impossible, even for humans, since different modalities contain partially disjunctive cues. Multimodal approaches to human affect recognition generally are shown to boost recognition performance, yet are evaluated in restrictive laboratory settings only. Herein we introduce a fully automatic processing combination of Active-Appearance-Model-based facial expression, vision-based eye-activity estimation, acoustic features, linguistic analysis, non-linguistic vocalisations, and temporal context information in an early feature fusion process. We provide detailed subject-independent results for classification and regression of the Level of Interest using Support-Vector Machines on an audiovisual interest corpus (AVIC) consisting of spontaneous, conversational speech demonstrating "theoretical" effectiveness of the approach. Further, to evaluate the approach with regards to real-life usability a user-study is conducted for proof of "practical" effectiveness.

OriginalspracheEnglisch
Seiten (von - bis)1760-1774
Seitenumfang15
FachzeitschriftImage and Vision Computing
Jahrgang27
Ausgabenummer12
DOIs
PublikationsstatusVeröffentlicht - Nov. 2009

Fingerprint

Untersuchen Sie die Forschungsthemen von „Being bored? Recognising natural interest by extensive audiovisual integration for real-life application“. Zusammen bilden sie einen einzigartigen Fingerprint.

Dieses zitieren