Song Li;Wei Sun;Qiaokang Liang;Jian Sun;Chongpei Liu
{"title":"基于时空注意机制和自适应门融合的抓握稳定性评价","authors":"Song Li;Wei Sun;Qiaokang Liang;Jian Sun;Chongpei Liu","doi":"10.1109/JSEN.2024.3493117","DOIUrl":null,"url":null,"abstract":"In the field of robotic grasping and manipulation, accurately assessing the stability of handheld objects plays a critical role in achieving proficient manipulation. Methods relying solely on visual or tactile information for slip detection often have limited applicability across different scenarios. Relevant studies have shown that combining visual and tactile sensing can significantly improve grasping performance. This study proposes a novel deep neural network architecture, specifically adopting a spatiotemporal attention mechanism to fuse multilevel spatiotemporal features, effectively integrating deep high-level features with shallow low-level features. It extracts important slip features across temporal and spatial dimensions from both visual RGB image sequences and tactile image sequences, thereby facilitating stability prediction. The gating mechanism builds a resilient network architecture that adaptively fuses features with appropriate weights, maintaining highly accurate and robust predictive performance even when sensor signal quality degrades. Validation results from both public and custom datasets demonstrate that the proposed model is highly effective in accurately predicting grasp stability, even in the presence of missing, occluded, noisy, or corrupted visual or tactile sensor signals. The practicality of this approach extends to various downstream applications in robotics, including grasp force control, generation of grasping strategies, and proficient manipulation in challenging scenarios.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 1","pages":"1872-1884"},"PeriodicalIF":4.3000,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Grasp Stability Assessment Through Spatio-Temporal Attention Mechanism and Adaptive Gate Fusion\",\"authors\":\"Song Li;Wei Sun;Qiaokang Liang;Jian Sun;Chongpei Liu\",\"doi\":\"10.1109/JSEN.2024.3493117\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the field of robotic grasping and manipulation, accurately assessing the stability of handheld objects plays a critical role in achieving proficient manipulation. Methods relying solely on visual or tactile information for slip detection often have limited applicability across different scenarios. Relevant studies have shown that combining visual and tactile sensing can significantly improve grasping performance. This study proposes a novel deep neural network architecture, specifically adopting a spatiotemporal attention mechanism to fuse multilevel spatiotemporal features, effectively integrating deep high-level features with shallow low-level features. It extracts important slip features across temporal and spatial dimensions from both visual RGB image sequences and tactile image sequences, thereby facilitating stability prediction. The gating mechanism builds a resilient network architecture that adaptively fuses features with appropriate weights, maintaining highly accurate and robust predictive performance even when sensor signal quality degrades. Validation results from both public and custom datasets demonstrate that the proposed model is highly effective in accurately predicting grasp stability, even in the presence of missing, occluded, noisy, or corrupted visual or tactile sensor signals. The practicality of this approach extends to various downstream applications in robotics, including grasp force control, generation of grasping strategies, and proficient manipulation in challenging scenarios.\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 1\",\"pages\":\"1872-1884\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2024-11-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10752890/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/10752890/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Grasp Stability Assessment Through Spatio-Temporal Attention Mechanism and Adaptive Gate Fusion
In the field of robotic grasping and manipulation, accurately assessing the stability of handheld objects plays a critical role in achieving proficient manipulation. Methods relying solely on visual or tactile information for slip detection often have limited applicability across different scenarios. Relevant studies have shown that combining visual and tactile sensing can significantly improve grasping performance. This study proposes a novel deep neural network architecture, specifically adopting a spatiotemporal attention mechanism to fuse multilevel spatiotemporal features, effectively integrating deep high-level features with shallow low-level features. It extracts important slip features across temporal and spatial dimensions from both visual RGB image sequences and tactile image sequences, thereby facilitating stability prediction. The gating mechanism builds a resilient network architecture that adaptively fuses features with appropriate weights, maintaining highly accurate and robust predictive performance even when sensor signal quality degrades. Validation results from both public and custom datasets demonstrate that the proposed model is highly effective in accurately predicting grasp stability, even in the presence of missing, occluded, noisy, or corrupted visual or tactile sensor signals. The practicality of this approach extends to various downstream applications in robotics, including grasp force control, generation of grasping strategies, and proficient manipulation in challenging scenarios.
期刊介绍:
The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following:
-Sensor Phenomenology, Modelling, and Evaluation
-Sensor Materials, Processing, and Fabrication
-Chemical and Gas Sensors
-Microfluidics and Biosensors
-Optical Sensors
-Physical Sensors: Temperature, Mechanical, Magnetic, and others
-Acoustic and Ultrasonic Sensors
-Sensor Packaging
-Sensor Networks
-Sensor Applications
-Sensor Systems: Signals, Processing, and Interfaces
-Actuators and Sensor Power Systems
-Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting
-Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data)
-Sensors in Industrial Practice