MIN Faculty
Department of Informatics
Knowledge Technology

Warning!

This webpage is outdated and has moved. Please find the official Knowledge Technology page at:

https://www.inf.uni-hamburg.de/en/inst/ab/wtm/


NeuralGrasp: Neural Network-based Robotic Grasping via Curiosity-driven Reinforcement Learning



PIs: Prof. Dr. Stefan Wermter, Dr. Cornelius Weber
Associates: Burhan Hafez

Summary

Reinforcement Learning (RL) has been successfully applied to solve a broad range of robotic learning problems, including navigation, grasping and object manipulation. Recent works on learning deep neural network representations have also further extended the capabilities of RL in complex domains.

A common limitation of standard RL approaches is that they strongly rely on external rewards received by the agent through interacting with its environment. Fully autonomous agents often live in environments lacking such external rewards or where these rewards are very sparse. This raises a challenge of how to optimally direct the agent choice of gathering experience data to learn from in the absence of external feedback, especially for grasping tasks where high-dimensional visual data and multiple degrees of freedom are involved.

We propose to address this problem in RL by looking into exploration strategies inspired by infant cognition and sensorimotor development. When infants start interacting with the world, they engage in activities for their own sake and not for achieving an extrinsic goal, mapping sensorimotor patterns to actions and rapidly shifting from motor babbling to organized exploratory behavior. This behavior is, according to studies from psychology and neuroscience, believed to be in response to an intrinsic motivation to reduce the uncertainty in action outcomes and not randomly selected. Such intrinsic motivation as curiosity plays an important role in child development and we believe it has similar impact on robot learning of grasping skills.

Project objectives

  1. Investigate neural network-based mechanisms for generating artificial curiosity signals.
  2. Evaluate exploration strategies for collecting useful experiences.
  3. Develop a curiosity-driven RL approach to the grasping problem in continuous state-and-action space.