TY - GEN
T1 - Robotic gaze control using reinforcement learning
AU - Rothbucher, Martin
AU - Denk, Christian
AU - Diepold, Klaus
PY - 2012
Y1 - 2012
N2 - This work examines how adaptive control can learn to point a camera at the active speaker in a conversation by using a Reinforcement Learning approach with audio and video data. A motivating scenario for this problem is a robotic platform that interacts with people around its environment. Using Reinforcement Learning, the task is specified with an observable objective referred to as the reward signal. Specifying this task with a reward signal enables an adaptive controller to improve its performance with experience. The reward for this task is generated by a visual feedback from the conversation participants that is detected by the robot's camera system. Multiple experiments have been performed on a robot system with audiovisual data to examine the feasibility and potential of this approach. Our experimental results demonstrate that the system learns very fast to identify the active speakers. Furthermore, our approach inherently learns how to deal with egonoise that originates from the robot's motor or background noise from the environment.
AB - This work examines how adaptive control can learn to point a camera at the active speaker in a conversation by using a Reinforcement Learning approach with audio and video data. A motivating scenario for this problem is a robotic platform that interacts with people around its environment. Using Reinforcement Learning, the task is specified with an observable objective referred to as the reward signal. Specifying this task with a reward signal enables an adaptive controller to improve its performance with experience. The reward for this task is generated by a visual feedback from the conversation participants that is detected by the robot's camera system. Multiple experiments have been performed on a robot system with audiovisual data to examine the feasibility and potential of this approach. Our experimental results demonstrate that the system learns very fast to identify the active speakers. Furthermore, our approach inherently learns how to deal with egonoise that originates from the robot's motor or background noise from the environment.
UR - http://www.scopus.com/inward/record.url?scp=84871963547&partnerID=8YFLogxK
U2 - 10.1109/HAVE.2012.6374444
DO - 10.1109/HAVE.2012.6374444
M3 - Conference contribution
AN - SCOPUS:84871963547
SN - 9781467315661
T3 - Proceedings - 2012 IEEE Symposium on Haptic Audio-Visual Environments and Games, HAVE 2012
SP - 83
EP - 88
BT - Proceedings - 2012 IEEE Symposium on Haptic Audio-Visual Environments and Games, HAVE 2012
T2 - 11th International Symposium on Haptic Audio-Visual Environments and Games, HAVE 2012
Y2 - 8 October 2012 through 9 October 2012
ER -