Sim2Real Transfer for Vision-Based Grasp Verification
Sim2Real Transfer for Vision-Based Grasp Verification
The verification of successful grasps is a crucial aspect of robot manipulation, particularly when handling deformable objects. Traditional methods relying on force and tactile sensors often struggle with deformable and non-rigid objects. In this work, we present a vision-based approach for grasp verification to determine whether the robotic gripper has successfully grasped an object. Our method employs a two-stage architecture; first YOLO-based object detection model to detect and locate the robot's gripper and then a ResNet-based classifier determines the presence of an object. To address the limitations of real-world data capture, we introduce HSR-GraspSynth, a synthetic dataset designed to simulate diverse grasping scenarios. Furthermore, we explore the use of Visual Question Answering capabilities as a zero-shot baseline to which we compare our model. Experimental results demonstrate that our approach achieves high accuracy in real-world environments, with potential for integration into grasping pipelines. Code and datasets are publicly available at https://github.com/pauamargant/HSR-GraspSynth .
Pau Amargant、Peter H?nig、Markus Vincze
计算技术、计算机技术
Pau Amargant,Peter H?nig,Markus Vincze.Sim2Real Transfer for Vision-Based Grasp Verification[EB/OL].(2025-05-05)[2025-05-22].https://arxiv.org/abs/2505.03046.点此复制
评论