Structured Bayesian Meta-Learning for Data-Efficient Visual-Tactile Model Estimation

Published: 05 Sept 2024, Last Modified: 08 Nov 2024CoRL 2024EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Multimodal perception, tactile sensing, few-shot learning
TL;DR: Our SBML framework enables data-efficient visual-tactile model estimation by learning a prior of visual-tactile model from diverse real world objects.
Abstract: Estimating visual-tactile models of deformable objects is challenging because vision suffers from occlusion, while touch data is sparse and noisy. We propose a novel data-efficient method for dense heterogeneous model estimation by leveraging experience from diverse training objects. The method is based on Bayesian Meta-Learning (BML), which can mitigate overfitting high-capacity visual-tactile models by meta-learning an informed prior and naturally achieves few-shot online estimation via posterior estimation. However, BML requires a shared parametric model across tasks but visual-tactile models for diverse objects have different parameter spaces. To address this issue, we introduce Structured Bayesian Meta-Learning (SBML) that incorporates heterogeneous physics models, enabling learning from training objects with varying appearances and geometries. SBML performs zero-shot vision-only prediction of deformable model parameters and few-shot adaptation after a handful of touches. Experiments show that in two classes of heterogeneous objects, namely plants and shoes, SBML outperforms existing approaches in force and torque prediction accuracy in zero- and few-shot settings.
Supplementary Material: zip
Spotlight Video: mp4
Website: https://shaoxiongyao.github.io/SBML
Publication Agreement: pdf
Student Paper: yes
Submission Number: 262
Loading