Abstract
This paper presents an implementation of visual learning by appearance in conjunction with an adaptive, non-linear controller for fine-positioning a manipulator onto a grasping position. We use principal component analysis to reduce the dimension of raw camera images (about 10,000 pixels) to lower-dimension vectors that can be used as inputs of our neuro-fuzzy controllers. It is shown that this approach leads to a very robust system that is stable under variable environment conditions. The approach needs no camera calibration and is applied to tasks of three degrees of freedom, e.g. translating the gripper in the x-y-plane and rotating it about the z-axis.
Original language | English |
---|---|
Pages (from-to) | 1164-1169 |
Number of pages | 6 |
Journal | Proceedings - IEEE International Conference on Robotics and Automation |
Volume | 2 |
State | Published - 1999 |
Externally published | Yes |
Event | Proceedings of the 1999 IEEE International Conference on Robotics and Automation, ICRA99 - Detroit, MI, USA Duration: 10 May 1999 → 15 May 1999 |