Adversarial Subgraph Contrastive Learning for Predicting Grasp Stability of Robotic Hands With Multimodal Signals

  • Junjie Liao
  • , Pengwen Xiong
  • , Meng Chu Zhou
  • , Peter X. Liu
  • , Aiguo Song

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Accurate prediction of grasp stability is crucial for reliable and precise operations with multi-fingered robotic hands. Traditional methods tend to oversimplify tactile information and pay equal attention to all regions of the data. This can obscure subtle yet critical variations and introduce noise, increasing the risk of stability assessment errors. To address these challenges, a novel self-supervised method, Adversarial Subgraph Contrastive Learning (ASCL), is proposed. It constructs an instance graph from the spatial distribution and features of perceptual nodes. It employs a bi-level adversarial strategy to enhance latent data representations by maximizing the mutual information between the instance graph and its semantic subgraphs, while minimizing it with its noisy subgraphs. To prevent trivial solutions and continuous relaxation of semantic subgraphs, node confidence and edge connection terms are incorporated to ensure stabilization. From an information-theoretic perspective, ASCL exhibits notable advantages on unlabeled or sparsely labeled data, well outperforming existing methods in empirical tests with robotic hands.

Original languageEnglish (US)
Pages (from-to)17720-17733
Number of pages14
JournalIEEE Transactions on Automation Science and Engineering
Volume22
DOIs
StatePublished - 2025

All Science Journal Classification (ASJC) codes

  • Control and Systems Engineering
  • Electrical and Electronic Engineering

Keywords

  • graph contrastive learning
  • graph information bottleneck
  • Grasp stability prediction
  • multimodal perception
  • mutual information
  • self-supervised learning

Fingerprint

Dive into the research topics of 'Adversarial Subgraph Contrastive Learning for Predicting Grasp Stability of Robotic Hands With Multimodal Signals'. Together they form a unique fingerprint.

Cite this