A multi-modal mixed-state dynamic Bayesian network for robust meeting event recognition from disturbed data

Marc Al-Hames, Gerhard Rigoll

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

14 Scopus citations

Abstract

In this work we present a novel multi-modal mixed-state dynamic Bayesian network (DBN) for robust meeting event classification. The model uses information from lapel microphones, a microphone array and visual information to structure meetings into segments. Within the DBN a multistream hidden Markov model (HMM) is coupled with a linear dynamical system (LDS) to compensate disturbances in the data. Thereby the HMM is used as driving input for the LDS. The model can handle noise and occlusions in all channels. Experimental results on real meeting data show that the new model is highly preferable to all single-stream approaches. Compared to a baseline multi-modal early fusion HMM, the new DBN is more than 2.5%, respectively 1.5% better for clear and disturbed data, this corresponds to a relative error reduction of 17%, respectively 9%.

Original languageEnglish
Title of host publicationIEEE International Conference on Multimedia and Expo, ICME 2005
Pages45-48
Number of pages4
DOIs
StatePublished - 2005
EventIEEE International Conference on Multimedia and Expo, ICME 2005 - Amsterdam, Netherlands
Duration: 6 Jul 20058 Jul 2005

Publication series

NameIEEE International Conference on Multimedia and Expo, ICME 2005
Volume2005

Conference

ConferenceIEEE International Conference on Multimedia and Expo, ICME 2005
Country/TerritoryNetherlands
CityAmsterdam
Period6/07/058/07/05

Fingerprint

Dive into the research topics of 'A multi-modal mixed-state dynamic Bayesian network for robust meeting event recognition from disturbed data'. Together they form a unique fingerprint.

Cite this