See, Hear, and Feel: Smart Sensory Fusion for Robotic Manipulation

Modalities Stimulus modality Sensory Substitution
DOI: 10.48550/arxiv.2212.03858 Publication Date: 2022-01-01
ABSTRACT
Humans use all of their senses to accomplish different tasks in everyday activities. In contrast, existing work on robotic manipulation mostly relies one, or occasionally two modalities, such as vision and touch. this work, we systematically study how visual, auditory, tactile perception can jointly help robots solve complex tasks. We build a robot system that see with camera, hear contact microphone, feel vision-based sensor, three sensory modalities fused self-attention model. Results challenging tasks, dense packing pouring, demonstrate the necessity power multisensory for manipulation: displays global status but often suffer from occlusion, audio provides immediate feedback key moments are even invisible, touch offers precise local geometry decision making. Leveraging our significantly outperforms prior methods.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....