TY - GEN
T1 - Real-time acoustic source localization in noisy environments for human-robot multimodal interaction
AU - Trifa, Vlad M.
AU - Koene, Ansgar
AU - Morén, Jan
AU - Cheng, Gordon
PY - 2007
Y1 - 2007
N2 - Interaction between humans involves a plethora of sensory information, both in the form of explicit communication as well as more subtle unconsciously perceived signals. In order to enable natural human-robot interaction, robots will have to acquire the skills to detect and meaningfully integrate information from multiple modalities. In this article, we focus on sound localization in the context of a multi-sensory humanoid robot that combines audio and video information to yield natural and intuitive responses to human behavior, such as directed eye-head movements towards natural stimuli. We highlight four common sound source localization algorithms and compare their performance and advantages for real-time interaction. We also briefly introduce an integrated distributed control framework called DVC, where additional modalities such as speech recognition, visual tracking, or object recognition can easily be integrated. We further describe the way the sound localization module has been integrated in our humanoid robot, CB.
AB - Interaction between humans involves a plethora of sensory information, both in the form of explicit communication as well as more subtle unconsciously perceived signals. In order to enable natural human-robot interaction, robots will have to acquire the skills to detect and meaningfully integrate information from multiple modalities. In this article, we focus on sound localization in the context of a multi-sensory humanoid robot that combines audio and video information to yield natural and intuitive responses to human behavior, such as directed eye-head movements towards natural stimuli. We highlight four common sound source localization algorithms and compare their performance and advantages for real-time interaction. We also briefly introduce an integrated distributed control framework called DVC, where additional modalities such as speech recognition, visual tracking, or object recognition can easily be integrated. We further describe the way the sound localization module has been integrated in our humanoid robot, CB.
UR - http://www.scopus.com/inward/record.url?scp=48749105042&partnerID=8YFLogxK
U2 - 10.1109/ROMAN.2007.4415116
DO - 10.1109/ROMAN.2007.4415116
M3 - Conference contribution
AN - SCOPUS:48749105042
SN - 1424416345
SN - 9781424416349
T3 - Proceedings - IEEE International Workshop on Robot and Human Interactive Communication
SP - 393
EP - 398
BT - 16th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN
T2 - 16th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN
Y2 - 26 August 2007 through 29 August 2007
ER -