Antipodal Robotic Grasping using Generative Residual Convolutional Neural Network
Sulabh Kumra, Shirin Joshi, Ferat Sahin
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/skumra/robotic-graspingOfficialIn paperpytorch★ 0
- github.com/skumra/baxter-pnpOfficialIn papernone★ 0
- github.com/SteveHao74/shahao_GR-ConvNetpytorch★ 0
- github.com/qingchenkanlu/new_grasppytorch★ 0
Abstract
In this paper, we present a modular robotic system to tackle the problem of generating and performing antipodal robotic grasps for unknown objects from n-channel image of the scene. We propose a novel Generative Residual Convolutional Neural Network (GR-ConvNet) model that can generate robust antipodal grasps from n-channel input at real-time speeds (~20ms). We evaluate the proposed model architecture on standard datasets and a diverse set of household objects. We achieved state-of-the-art accuracy of 97.7% and 94.6% on Cornell and Jacquard grasping datasets respectively. We also demonstrate a grasp success rate of 95.4% and 93% on household and adversarial objects respectively using a 7 DoF robotic arm.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| Cornell Grasp Dataset | GR-ConvNet | 5 fold cross validation | 97.7 | — | Unverified |
| Jacquard dataset | GR-ConvNet | Accuracy (%) | 94.6 | — | Unverified |