Enhancing robotic grasping detection accuracy with the R2CNN algorithm and force-closure
Hsien-I Lin, Muhammad Shodiq, Hong-Qi Chu- Industrial and Manufacturing Engineering
- Computer Graphics and Computer-Aided Design
- Computer Science Applications
- Software
Abstract
This study aims to use an improved rotational region convolutional neural network (R2CNN) algorithm to detect the grasping bounding box for the robotic arm that reaches supermarket goods. This algorithm can calculate the final predicted grasping bounding box without any additional architecture, which greatly improves the speed of grasp inferences. In this study, we added the force-closure condition, so that the final grasping bounding box could achieve grasping stability in a physical sense. We experimentally demonstrated that the deep model treated object detection and grasping detection are the same tasks. We used transfer learning to improve the prediction accuracy of the grasping bounding box. In particular, the ResNet-101 network weights, which were originally used in object detection, were used to continue training with the Cornell dataset. In terms of grasping detection, we used the trained model weights that were originally used in object detection as the features of the to-be-grasped objects and fed them to the network for continuous training. For 2,828 test images, this method achieved nearly 98% accuracy and a speed of 14–17 frames per second.