Object Recognition, Localization and Grasp Detection Using a Unified Deep Convolutional Neural Network with Multi-task Loss
Citations Over Time
Abstract
Recognize an object and detect a good grasp in unstructured scenes is still a challenge. In this paper, the problem of detecting robotic grasps is expressed by a two-point representation in an unstructured scene with an RGB-D camera. A deep Convolutional Neural Network is designed to predict good grasps in real-time on GTX1080, with using region proposal techniques. A contribution of this work is our proposed network framework can perform classification, location and grasp detection simultaneously so that in a single step, it not only recognizes the category and bounding-box of the object, but also finds a good grasp line. Besides, in training process, we minimize a multi-task loss objective function of object classification, location and grasp detection in order to train the network end-to-end. Our experimental evaluation on a real robotic manipulator demonstrates that the robotic manipulator can fulfill the grasping task effectively.
Related Papers
- → AttentionNet: Aggregating Weak Directions for Accurate Object Detection(2015)181 cited
- → Multi-Task Self-Supervised Object Detection via Recycling of Bounding Box Annotations(2019)51 cited
- → Design and Implementation of an Object Detection System Using Faster R-CNN(2019)20 cited
- → Small Object Detection Method based on Improved YOLOv5(2022)7 cited
- → Fast and Accurate Object Detection Based on Fusion of YOLOv2 and R-CNN Predicted Result for Autonomous Driving(2022)1 cited