Where shall I touch? Vision-guided tactile poking for transparent object grasping

Jiaqi Jiang, Guanqun Cao, Aaron Butterworth, Thanh-Toan Do, Shan Luo

Research output: Contribution to journalArticleResearchpeer-review

4 Citations (Scopus)

Abstract

Picking up transparent objects is still a challenging task for robots. The visual properties of transparent objects such as reflection and refraction make the current grasping methods that rely on camera sensing fail to detect and localize them. However, humans can handle the transparent object well by first observing its coarse profile and then poking an area of interest to get a fine profile for grasping. Inspired by this, we propose a novel framework of vision-guided tactile poking for transparent objects grasping. In the proposed framework, a segmentation network is first used to predict the horizontal upper regions named as poking regions, where the robot can poke the object to obtain a good tactile reading, while leading to minimal disturbance to the object's state. A poke is then performed with a high-resolution GelSight tactile sensor. Given the local profiles improved with the tactile reading, a heuristic grasp is planned for grasping the transparent object. To mitigate the limitations of real-world data collection and labeling for transparent objects, a large-scale realistic synthetic dataset was constructed. Extensive experiments demonstrate that our proposed segmentation network can predict the potential poking region with a high mean average precision (mAP) of 0.360, and the vision-guided tactile poking can enhance the grasping success rate significantly from 38.9% to 85.2%. Thanks to its simplicity, our proposed approach could also be adopted by other force or tactile sensors and could be used for grasping of other challenging objects.

Original languageEnglish
Pages (from-to)233-244
Number of pages12
JournalIEEE/ASME Transactions on Mechatronics
Volume28
Issue number1
DOIs
Publication statusPublished - 1 Feb 2023

Keywords

  • Multimodal sensing
  • object segmentation
  • robot grasping and manipulation
  • tactile sensing
  • transparent objects
  • visual perception

Cite this