TY - JOUR
T1 - Structured Bayesian Meta-Learning for Data-Efficient Visual-Tactile Model Estimation
AU - Yao, Shaoxiong
AU - Zhu, Yifan
AU - Hauser, Kris
N1 - This work has partially been funded by USDA/NIFA Grant #2020-67021-32799, and we thank Toyota Research Institute for a loan of the Punyo tactile sensor. We thank Joao M. C. Marques, Jing-Chen Peng, Mengchao Zhang, Patrick Naughton, and Sicong Pan for their valuable comments and suggestions. We are grateful to Yidi Hua for segmenting object images. We also thank Simon Kato, Haonan Chen, Yuan Shen, Zhen Zhu, Shuijing Liu, Hameed Abdul-Rashid, Yiqiu Sun, Mengchao Zhang, Yangfei Dai, Yixuan Wang, Kaiwen Hong, Shuhong Zheng, and Yuxiang Liu for providing objects used in the experiments.
PY - 2024
Y1 - 2024
N2 - Estimating visual-tactile models of deformable objects is challenging because vision suffers from occlusion, while touch data is sparse and noisy. We propose a novel data-efficient method for dense heterogeneous model estimation by leveraging experience from diverse training objects. The method is based on Bayesian Meta-Learning (BML), which can mitigate overfitting high-capacity visual-tactile models by meta-learning an informed prior and naturally achieves few-shot online estimation via posterior estimation. However, BML requires a shared parametric model across tasks but visual-tactile models for diverse objects have different parameter spaces. To address this issue, we introduce Structured Bayesian Meta-Learning (SBML) that incorporates heterogeneous physics models, enabling learning from training objects with varying appearances and geometries. SBML performs zero-shot vision-only prediction of deformable model parameters and few-shot adaptation after a handful of touches. Experiments show that in two classes of heterogeneous objects, namely plants and shoes, SBML outperforms existing approaches in force and torque prediction accuracy in zero- and few-shot settings. Website: https://shaoxiongyao.github.io/SBML.
AB - Estimating visual-tactile models of deformable objects is challenging because vision suffers from occlusion, while touch data is sparse and noisy. We propose a novel data-efficient method for dense heterogeneous model estimation by leveraging experience from diverse training objects. The method is based on Bayesian Meta-Learning (BML), which can mitigate overfitting high-capacity visual-tactile models by meta-learning an informed prior and naturally achieves few-shot online estimation via posterior estimation. However, BML requires a shared parametric model across tasks but visual-tactile models for diverse objects have different parameter spaces. To address this issue, we introduce Structured Bayesian Meta-Learning (SBML) that incorporates heterogeneous physics models, enabling learning from training objects with varying appearances and geometries. SBML performs zero-shot vision-only prediction of deformable model parameters and few-shot adaptation after a handful of touches. Experiments show that in two classes of heterogeneous objects, namely plants and shoes, SBML outperforms existing approaches in force and torque prediction accuracy in zero- and few-shot settings. Website: https://shaoxiongyao.github.io/SBML.
KW - Multimodal perception
KW - few-shot learning
KW - tactile sensing
UR - http://www.scopus.com/inward/record.url?scp=86000764962&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=86000764962&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:86000764962
SN - 2640-3498
VL - 270
SP - 3072
EP - 3093
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 8th Conference on Robot Learning, CoRL 2024
Y2 - 6 November 2024 through 9 November 2024
ER -