Object Recognition and Grasping for Collaborative Robots Based on Vision

This study introduces a parallel YOLO–GG deep learning network for collaborative robot target recognition and grasping to enhance the efficiency and precision of visual classification and grasping for collaborative robots. First, the paper outlines the target classification and detection task, the g...

Full description

Bibliographic Details
Main Authors: Ruohuai Sun, Chengdong Wu, Xue Zhao, Bin Zhao, Yang Jiang
Format: Article
Language:English
Published: MDPI AG 2023-12-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/24/1/195
_version_ 1797358149366185984
author Ruohuai Sun
Chengdong Wu
Xue Zhao
Bin Zhao
Yang Jiang
author_facet Ruohuai Sun
Chengdong Wu
Xue Zhao
Bin Zhao
Yang Jiang
author_sort Ruohuai Sun
collection DOAJ
description This study introduces a parallel YOLO–GG deep learning network for collaborative robot target recognition and grasping to enhance the efficiency and precision of visual classification and grasping for collaborative robots. First, the paper outlines the target classification and detection task, the grasping system of the robotic arm, and the dataset preprocessing method. The real-time recognition and grasping network can identify a diverse spectrum of unidentified objects and determine the target type and appropriate capture box. Secondly, we propose a parallel YOLO–GG deep vision network based on YOLO and GG-CNN. Thirdly, the YOLOv3 network, pre-trained with the COCO dataset, identifies the object category and position, while the GG-CNN network, trained using the Cornell Grasping dataset, predicts the grasping pose and scale. This study presents the processes for generating a target’s grasping frame and recognition type using GG-CNN and YOLO networks, respectively. This completes the investigation of parallel networks for target recognition and grasping in collaborative robots. Finally, the experimental results are evaluated on the self-constructed NEU-COCO dataset for target recognition and positional grasping. The speed of detection has improved by 14.1%, with an accuracy of 94%. This accuracy is 4.0% greater than that of YOLOv3. Experimental proof was obtained through a robot grasping actual objects.
first_indexed 2024-03-08T14:57:41Z
format Article
id doaj.art-b78f48cdb39e4f1b87b5b4a592e72733
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-08T14:57:41Z
publishDate 2023-12-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-b78f48cdb39e4f1b87b5b4a592e727332024-01-10T15:08:59ZengMDPI AGSensors1424-82202023-12-0124119510.3390/s24010195Object Recognition and Grasping for Collaborative Robots Based on VisionRuohuai Sun0Chengdong Wu1Xue Zhao2Bin Zhao3Yang Jiang4College of Information Science and Engineering, Northeastern University, Shenyang 110819, ChinaCollege of Information Science and Engineering, Northeastern University, Shenyang 110819, ChinaDaniel L. Goodwin College of Business, Benedict University, Chicago, IL 60601, USACollege of Information Science and Engineering, Northeastern University, Shenyang 110819, ChinaFaculty of Robot Science and Engineering, Northeastern University, Shenyang 110169, ChinaThis study introduces a parallel YOLO–GG deep learning network for collaborative robot target recognition and grasping to enhance the efficiency and precision of visual classification and grasping for collaborative robots. First, the paper outlines the target classification and detection task, the grasping system of the robotic arm, and the dataset preprocessing method. The real-time recognition and grasping network can identify a diverse spectrum of unidentified objects and determine the target type and appropriate capture box. Secondly, we propose a parallel YOLO–GG deep vision network based on YOLO and GG-CNN. Thirdly, the YOLOv3 network, pre-trained with the COCO dataset, identifies the object category and position, while the GG-CNN network, trained using the Cornell Grasping dataset, predicts the grasping pose and scale. This study presents the processes for generating a target’s grasping frame and recognition type using GG-CNN and YOLO networks, respectively. This completes the investigation of parallel networks for target recognition and grasping in collaborative robots. Finally, the experimental results are evaluated on the self-constructed NEU-COCO dataset for target recognition and positional grasping. The speed of detection has improved by 14.1%, with an accuracy of 94%. This accuracy is 4.0% greater than that of YOLOv3. Experimental proof was obtained through a robot grasping actual objects.https://www.mdpi.com/1424-8220/24/1/195deep learningcollaborative robotsparallel networksgrasping detection
spellingShingle Ruohuai Sun
Chengdong Wu
Xue Zhao
Bin Zhao
Yang Jiang
Object Recognition and Grasping for Collaborative Robots Based on Vision
Sensors
deep learning
collaborative robots
parallel networks
grasping detection
title Object Recognition and Grasping for Collaborative Robots Based on Vision
title_full Object Recognition and Grasping for Collaborative Robots Based on Vision
title_fullStr Object Recognition and Grasping for Collaborative Robots Based on Vision
title_full_unstemmed Object Recognition and Grasping for Collaborative Robots Based on Vision
title_short Object Recognition and Grasping for Collaborative Robots Based on Vision
title_sort object recognition and grasping for collaborative robots based on vision
topic deep learning
collaborative robots
parallel networks
grasping detection
url https://www.mdpi.com/1424-8220/24/1/195
work_keys_str_mv AT ruohuaisun objectrecognitionandgraspingforcollaborativerobotsbasedonvision
AT chengdongwu objectrecognitionandgraspingforcollaborativerobotsbasedonvision
AT xuezhao objectrecognitionandgraspingforcollaborativerobotsbasedonvision
AT binzhao objectrecognitionandgraspingforcollaborativerobotsbasedonvision
AT yangjiang objectrecognitionandgraspingforcollaborativerobotsbasedonvision