跳到主要导航 跳到搜索 跳到主要内容

Multimodal zero-shot learning for tactile texture recognition

  • Guanqun Cao
  • , Jiaqi Jiang
  • , Danushka Bollegala
  • , Min Li
  • , Shan Luo
  • University of Liverpool
  • King's College London

科研成果: 期刊稿件文章同行评审

21 引用 (Scopus)

摘要

Tactile sensing plays an irreplaceable role in robotic material recognition. It enables robots to distinguish material properties such as their local geometry and textures, especially for materials like textiles. However, most tactile recognition methods can only classify known materials that have been touched and trained with tactile data, yet cannot classify unknown materials that are not trained with tactile data. To solve this problem, we propose a tactile Zero-Shot Learning framework to recognise materials when they are touched for the first time, using their visual and semantic information, without requiring tactile training samples. The biggest challenge in tactile Zero-Shot Learning is to recognise disjoint classes between training and test materials, i.e., the test materials that are not among the training ones. To bridge this gap, the visual modality, providing tactile cues from sight, and semantic attributes, giving high-level characteristics, are combined together and act as a link to expose the model to these disjoint classes. Specifically, a generative model is learnt to synthesise tactile features according to corresponding visual images and semantic embeddings, and then a classifier can be trained using the synthesised tactile features for zero-shot recognition. Extensive experiments demonstrate that our proposed multimodal generative model can achieve a high recognition accuracy of 83.06% in classifying materials that were not touched before. The robotic experiment demo and the FabricVST dataset are available at https://sites.google.com/view/multimodalzsl.

源语言英语
文章编号104688
期刊Robotics and Autonomous Systems
176
DOI
出版状态已出版 - 6月 2024

学术指纹

探究 'Multimodal zero-shot learning for tactile texture recognition' 的科研主题。它们共同构成独一无二的指纹。

引用此