Multi-modal activity and dominance detection in smart meeting rooms

Benedikt Hörnler, Gerhard Rigoll

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

6 Scopus citations

Abstract

In this paper a new approach for activity and dominance modeling in meetings is presented. For this purpose low level acoustic and visual features are extracted from audio and video capture devices. Hidden Markov Models (HMM) are used for the segmentation and classification of activity levels for each participant. Additionally, more semantic features are applied in a two-layer HMM approach. The experiments show that the acoustic feature is the most important one. The early fusion of acoustic and global-motion features achieves nearly as good results as the acoustic feature alone. All the other early fusion approaches are out-performed by the acoustic feature. More over, the two-layer model could not achieve the results of the acoustic features.

Original languageEnglish
Title of host publication2009 IEEE International Conference on Acoustics, Speech, and Signal Processing - Proceedings, ICASSP 2009
Pages1777-1780
Number of pages4
DOIs
StatePublished - 2009
Event2009 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2009 - Taipei, Taiwan, Province of China
Duration: 19 Apr 200924 Apr 2009

Publication series

NameICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
ISSN (Print)1520-6149

Conference

Conference2009 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2009
Country/TerritoryTaiwan, Province of China
CityTaipei
Period19/04/0924/04/09

Keywords

  • Activity detection
  • Human-machine interaction
  • Machine learning
  • Meeting analysis
  • Multi-modal low level features

Fingerprint

Dive into the research topics of 'Multi-modal activity and dominance detection in smart meeting rooms'. Together they form a unique fingerprint.

Cite this