TY - GEN
T1 - Using audio, visual, and lexical features in a multi-modal virtual meeting director
AU - Al-Hames, Marc
AU - Hörnler, Benedikt
AU - Scheuermann, Christoph
AU - Rigoll, Gerhard
PY - 2006
Y1 - 2006
N2 - Multi-modal recordings of meetings provide the basis for meeting browsing and for remote meetings. However it is often not useful to store or transmit all visual channels. In this work we show how a virtual meeting director selects one of seven possible video modes. We then present several audio, visual, and lexical features for a virtual director. In an experimental section we evaluate the features, their influence on the camera selection, and the properties of the generated video stream. The chosen features all allow a real- or near real-time processing and can therefore not only be applied to offline browsing, but also for a remote meeting assistant.
AB - Multi-modal recordings of meetings provide the basis for meeting browsing and for remote meetings. However it is often not useful to store or transmit all visual channels. In this work we show how a virtual meeting director selects one of seven possible video modes. We then present several audio, visual, and lexical features for a virtual director. In an experimental section we evaluate the features, their influence on the camera selection, and the properties of the generated video stream. The chosen features all allow a real- or near real-time processing and can therefore not only be applied to offline browsing, but also for a remote meeting assistant.
UR - http://www.scopus.com/inward/record.url?scp=77249169836&partnerID=8YFLogxK
U2 - 10.1007/11965152_6
DO - 10.1007/11965152_6
M3 - Conference contribution
AN - SCOPUS:77249169836
SN - 3540692673
SN - 9783540692676
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 63
EP - 74
BT - Machine Learning for Multimodal Interaction - Third International Workshop, MLMI 2006, Revised Selected Papers
T2 - 3rd International Workshop on Machine Learning for Multimodal Interaction, MLMI 2006
Y2 - 1 May 2006 through 4 May 2006
ER -