TY - GEN
T1 - Robust video-based recognition of dynamic head gestures in various domains - comparing a rule-based and a stochastic approach
AU - McGlaun, Gregor
AU - Althoff, Frank
AU - Lang, Manfred
AU - Rigoll, Gerhard
N1 - Publisher Copyright:
© 2004 Springer-Verlag Berlin Heidelberg.
PY - 2004
Y1 - 2004
N2 - This work describes two video-based approaches for detecting and classifying dynamic head-gestures. We compare a simple, fast, and efficient rule-based algorithm with a powerful, robust, and flexible stochastic implementation. In both realizations, the head is localized via a combination of color-And shape-based segmentation. For a continuous feature extraction, the rule-based approach uses a template-matching of the nose bridge. In addition, the stochastic algorithm applies features derived from the optical flow, and classifies them by a set of discrete Hidden Markov Models. The rule-based implementation evaluates the key-feature in a finite state machine. We extensively tested the systems in two different application domains (VR desktop scenario vs. automotive environment). Six different gestures can be classified with an overall recognition rate of 93.7% (rule-based) and 97.3% (stochastic) in the VR (92.6% and 95.5% in the automotive environment, respectively). Both approaches work independently from the image background. Concerning the stochastic concept, further gesture types can easily be implemented.
AB - This work describes two video-based approaches for detecting and classifying dynamic head-gestures. We compare a simple, fast, and efficient rule-based algorithm with a powerful, robust, and flexible stochastic implementation. In both realizations, the head is localized via a combination of color-And shape-based segmentation. For a continuous feature extraction, the rule-based approach uses a template-matching of the nose bridge. In addition, the stochastic algorithm applies features derived from the optical flow, and classifies them by a set of discrete Hidden Markov Models. The rule-based implementation evaluates the key-feature in a finite state machine. We extensively tested the systems in two different application domains (VR desktop scenario vs. automotive environment). Six different gestures can be classified with an overall recognition rate of 93.7% (rule-based) and 97.3% (stochastic) in the VR (92.6% and 95.5% in the automotive environment, respectively). Both approaches work independently from the image background. Concerning the stochastic concept, further gesture types can easily be implemented.
UR - http://www.scopus.com/inward/record.url?scp=7444239210&partnerID=8YFLogxK
U2 - 10.1007/978-3-540-24598-8_18
DO - 10.1007/978-3-540-24598-8_18
M3 - Conference contribution
AN - SCOPUS:7444239210
SN - 3540210725
SN - 9783540210726
T3 - Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science)
SP - 180
EP - 197
BT - Gesture-Based Communication in Human-Computer Interaction
A2 - Camurri, Antonio
A2 - Volpe, Gualtiero
PB - Springer Verlag
T2 - 5th International GestureWorkshop, GW 2003
Y2 - 15 April 2003 through 17 April 2003
ER -