Abstract
Robust object manipulation is still a hard problem in robotics, even more so in high degree-of-freedom (DOF) humanoid robots. To improve performance a closer integration of visual and motor systems is needed. We herein present a novel method for a robot to learn robust detection of its own hands and fingers enabling sensorimotor coordination. It does so solely using its own camera images and does not require any external systems or markers. Our system based on Cartesian Genetic Programming (CGP) allows to evolve programs to perform this image segmentation task in real-time on the real hardware. We show results for a Nao and an iCub humanoid each detecting its own hands and fingers. © 2013 IEEE.
Original language | English (US) |
---|---|
Title of host publication | 2013 IEEE Congress on Evolutionary Computation, CEC 2013 |
Pages | 1411-1418 |
Number of pages | 8 |
DOIs | |
State | Published - Aug 21 2013 |
Externally published | Yes |