Appearance-based visual learning in a neuro-fuzzy model for fine-positioning of manipulators

Jianwei Zhang, Ralf Schmidt, Alois Knoll

Research output: Contribution to journalConference articlepeer-review

11 Scopus citations

Abstract

This paper presents an implementation of visual learning by appearance in conjunction with an adaptive, non-linear controller for fine-positioning a manipulator onto a grasping position. We use principal component analysis to reduce the dimension of raw camera images (about 10,000 pixels) to lower-dimension vectors that can be used as inputs of our neuro-fuzzy controllers. It is shown that this approach leads to a very robust system that is stable under variable environment conditions. The approach needs no camera calibration and is applied to tasks of three degrees of freedom, e.g. translating the gripper in the x-y-plane and rotating it about the z-axis.

Original languageEnglish
Pages (from-to)1164-1169
Number of pages6
JournalProceedings - IEEE International Conference on Robotics and Automation
Volume2
StatePublished - 1999
Externally publishedYes
EventProceedings of the 1999 IEEE International Conference on Robotics and Automation, ICRA99 - Detroit, MI, USA
Duration: 10 May 199915 May 1999

Fingerprint

Dive into the research topics of 'Appearance-based visual learning in a neuro-fuzzy model for fine-positioning of manipulators'. Together they form a unique fingerprint.

Cite this