Zugang zum Dokument
Towards Spatial Perception: Learning to Locate Objects From Vision
Leitner, Jürgen ; Harding, Simon ; Frank, Mikhail ; Förster, Alexander ; Schmidhuber, Jürgen
Our humanoid robot learns to provide position estimates of objects placed on a table, even while the robot is moving its torso, head and eyes (cm range accuracy). These estimates are provided by trained artificial neural networks (ANN) and a genetic programming (GP) method, based solely on the inputs from the two cameras and the joint encoder positions. No prior camera calibration and kinematic model is used. We find that ANN and GP are both able to localise objects robustly regardless of the robot's pose and without an explicit kinematic model or camera calibration. These approaches yield an accuracy comparable to current techniques used on the iCub.
||Forschungsinstitut für Kognition und Robotik (CoR-Lab)
Jürgen Leitner, Simon Harding, Mikhail Frank, Alexander Förster, Jürgen Schmidhuber,
Towards Spatial Perception: Learning to Locate Objects From Vision.
Proceedings of the Post-Graduate Conference on Robotics and Development of Cognition, J. Szufnarowska, Ed., September 2012