Touch100k: A large-scale touch-language-vision dataset for touch-centric multimodal representation

Published: 01 Jan 2025, Last Modified: 18 Jul 2025Inf. Fusion 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•The first 100k tactile dataset with multi-granular descriptions.•A well-designed process for building multimodal datasets.•A method that efficiently captures the links between touch, language, and vision.•Top performance in tactile-related tasks across various benchmarks and settings.
Loading