Deep Reinforcement Learning Based on Local GNN for Goal-conditioned Deformable Object Rearranging

IROS 2022

1Tsinghua University; 2Tencent Robotics X Lab;

Abstract

Object rearranging is one of the most common deformable manipulation tasks, where the robot needs to rearrange a deformable object into a goal configuration. Previous studies focus on designing an expert system for each specific task by model-based or data-driven approaches and the application scenarios are therefore limited.Some research has been attempting to design a general framework to obtain more advanced manipulation capabilities for deformable rearranging tasks, with lots of progress achieved in simulation. However, transferring from simulation to reality is difficult due to the limitation of the end-to-end CNN architecture.To address these challenges, we design a local GNN (graph neural network) based learning method, which utilizes two representation graphs to encode keypoints detected from images. Self-attention is applied for graph updating and cross-attention is applied for generating manipulation actions. Extensive experiments have been conducted to demonstrate that our framework is effective in multiple 1-D (rope, rope ring) and 2-D (cloth) rearranging tasks in simulation and can be easily transferred to a real robot by fine-tuning a keypoint detector.

Solution framework

Our RL agent encodes the representation vectors of keypoints by self-attention layers firstly and gets 2 local dynamic graphs to represent the current and goal configurations. Cross-attention layers are used to map the 2 graphs to the q-value matrix. Coordinates of the picking point are the keypoints in the current image and coordinates of the placing point are the keypoints in the goal image.




Simulation experiments

We first conduct experiments to evaluate the performance of our framework on multiple rearranging tasks. The robot is given random goal configurations by only visual input, and the robot needs to rearrange the deformable object to the goal configurations without any sub-goal input. We define that the robot completes a rearranging task within 30 picking and placing actions as a success, and the rest is a failure.


Real experiments

We test our proposed method in a physical environment. The rope is placed on the platform, and a UR5 robotic manipulator with a suction cup is placed in front of the platform for picking and placing. Images are captured with a Realsence camera, which is fixed on the top of the platform. The experimental results demonstrate that the skills our framework learned in the simulation are also effective in reality. Our model can learn multiple goal-conditioned deformable object rearranging skills from a large quantity of data in simulation and these skills can be used in reality with only keypoint detector fine-tune.


N shape

V shape

Straight line

BibTeX

@inproceedings{deng2022def,
        title={Deep reinforcement learning based on local GNN for goal-conditioned deformable object rearranging},
        author={Deng, Yuhong and Xia, Chongkun and Wang, Xueqian and Chen, Lipeng},
        booktitle={2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
        pages={1131--1138},
        year={2022},
        organization={IEEE}
      }