@Article{cmes.2022.021383, AUTHOR = {Xinheng Yuan, Hao Yu, Houlin Zhang, Li Zheng, Erbao Dong, Heng’an Wu}, TITLE = {A Multi-Scale Grasp Detector Based on Fully Matching Model}, JOURNAL = {Computer Modeling in Engineering \& Sciences}, VOLUME = {133}, YEAR = {2022}, NUMBER = {2}, PAGES = {281--301}, URL = {http://www.techscience.com/CMES/v133n2/48964}, ISSN = {1526-1506}, ABSTRACT = {Robotic grasping is an essential problem at both the household and industrial levels, and unstructured objects have always been difficult for grippers. Parallel-plate grippers and algorithms, focusing on partial information of objects, are one of the widely used approaches. However, most works predict single-size grasp rectangles for fixed cameras and gripper sizes. In this paper, a multi-scale grasp detector is proposed to predict grasp rectangles with different sizes on RGB-D or RGB images in real-time for hand-eye cameras and various parallel-plate grippers. The detector extracts feature maps of multiple scales and conducts predictions on each scale independently. To guarantee independence between scales and efficiency, fully matching model and background classifier are applied in the network. Based on analysis of the Cornell Grasp Dataset, the fully matching model can match all labeled grasp rectangles. Furthermore, background classification, along with angle classification and box regression, functions as hard negative mining and background predictor. The detector is trained and tested on the augmented dataset, which includes images of 320 × 320 pixels and grasp rectangles ranging from 20 to more than 320 pixels. It performs up to 98.87% accuracy on image-wise dataset and 97.83% on object-wise split dataset at a speed of more than 22 frames per second. In addition, the detector, which is trained on a single-object dataset, can predict grasps on multiple objects.}, DOI = {10.32604/cmes.2022.021383} }