SonicSense: Object Perception from In-Hand Acoustic Vibration
Jiaxun Liu, Boyuan Chen
TL;DR
SonicSense introduces a cost-effective, four-finger robot hand equipped with contact microphones to sense in-hand acoustic vibrations. Through a heuristic interaction policy and end-to-end models, it achieves material classification, 3D shape reconstruction, and object re-identification across 83 real-world objects, including complex geometries and heterogeneous materials. The work demonstrates robustness to ambient noise, leverages synthetic data augmentation for shape learning, and shows strong task performance with dedicated datasets and evaluation. This holistic approach advances tactile perception in robotics by moving beyond small, controlled object sets to diverse, real-world scenarios, enabling richer object understanding from acoustic cues.
Abstract
We introduce SonicSense, a holistic design of hardware and software to enable rich robot object perception through in-hand acoustic vibration sensing. While previous studies have shown promising results with acoustic sensing for object perception, current solutions are constrained to a handful of objects with simple geometries and homogeneous materials, single-finger sensing, and mixing training and testing on the same objects. SonicSense enables container inventory status differentiation, heterogeneous material prediction, 3D shape reconstruction, and object re-identification from a diverse set of 83 real-world objects. Our system employs a simple but effective heuristic exploration policy to interact with the objects as well as end-to-end learning-based algorithms to fuse vibration signals to infer object properties. Our framework underscores the significance of in-hand acoustic vibration sensing in advancing robot tactile perception.
