Visuomotor Mechanical Search: Learning to Retrieve Target Objects in Clutter

Andrey Kurenkov, Joseph Taglic, Rohun Kulkarni, Marcus Dominguez-Kuhne, Animesh Garg, Roberto Martín-Martín, Silvio Savarese

IROS 2020


When searching for objects in cluttered environments, it is often necessary to perform complex interactions in order to move occluding objects out of the way and fully reveal the object of interest and make it graspable. Due to the complexity of the physics involved and the lack of accurate models of the clutter, planning and controlling precise predefined interactions with accurate outcome is extremely hard, when not impossible. In problems where accurate (forward) models are lacking, Deep Reinforcement Learning (RL) has shown to be a viable solution to map observations (e.g. images) to good interactions in the form of close-loop visuomotor policies. However, Deep RL is sample inefficient and fails when applied directly to the problem of unoccluding objects based on images. In this work we present a novel Deep RL procedure that combines i) teacher-aided exploration, ii) a critic with privileged information, and iii) mid-level representations, resulting in sample efficient and effective learning for the problem of uncovering a target object occluded by a heap of unknown objects. Our experiments show that our approach trains faster and converges to more efficient uncovering solutions than baselines and ablations, and that our uncovering policies lead to an average improvement in the graspability of the target object, facilitating downstream retrieval applications.


Arxiv paper can be found here.

If you find this work useful in your research, please cite as follows:

title={Visuomotor Mechanical Search: Learning to Retrieve Target Objects in Clutter},
author={Kurenkov, Andrey and Taglic, Joseph and Kulkarni, Rohun and Dominguez-Kuhne, Marcus and Mart{\'\i}n-Mart{\'\i}n, Roberto and Garg, Animesh and Savarese, Silvio},
booktitle={IEEE/RSJ Int. Conference. on Intelligent Robots and Systems (IROS)},