Sim2Surf: A Sim2Real Surface Classifier for Vision-Based Tactile Sensors With a Bilevel Adaptation Pipeline

Gabriele M. Caddeo, Andrea Maracani, Paolo D. Alfano, Nicola A. Piga, Lorenzo Rosasco, Lorenzo Natale

Published: 01 Mar 2025, Last Modified: 26 Feb 2026IEEE Sensors JournalEveryoneRevisionsCC BY-SA 4.0
Abstract: In this work, we tackle the simulated to real (Sim2Real) gap in vision-based tactile sensors for surface classification. Specifically, we target four surface types: flat, curved, edge, and corner. We first train a diffusion model (DM) with a small set of real-world, unlabeled images gathered from everyday objects using the DIGIT sensor. Next, we employ the TACTO simulator to generate images by uniformly sampling object surfaces from the YCB model set. The images are transformed into the real domain using the DM and automatically labeled, allowing us to train a surface classifier without the need of manual annotations. To further align features across the real and simulated domains, we use an adversarial approach during training. The evaluation on tactile images from 15 3D-printed YCB objects shows an accuracy of 83.2%, significantly higher than the 35.9% achieved by training only on simulated images, confirming the effectiveness of our method. Testing on data acquired with different DIGIT sensors yields 81.6% accuracy on average, outperforming training with labeled real data. Moreover, our method demonstrates robustness also when applied to a different vision-based tactile sensor, GelSight Mini, obtaining a remarkable 83.3% accuracy on a balanced dataset. Lastly, we validate our approach with a 6-D object pose estimation task using tactile data.
Loading