October 31, 2024
In this work, we introduce general purpose touch representations for the increasingly accessible class of vision-based tactile sensors. Such sensors have led to many recent advances in robot manipulation as they markedly complement vision, yet solutions today often rely on task and sensor specific handcrafted perception models. Collecting real data at scale with task centric ground truth labels, like contact forces and slip, is a challenge further compounded by sensors of various form factor differing in aspects like lighting and gel markings. To tackle this we turn to self-supervised learning (SSL) that has demonstrated remarkable performance in computer vision. We present Sparsh, a family of SSL models that can support various vision-based tactile sensors, alleviating the need for custom labels through pre-training on 460k+ tactile images with masking and self-distillation in pixel and latent spaces. We also build TacBench, to facilitate standardized benchmarking across sensors and models, comprising of six tasks ranging from comprehending tactile properties to enabling physical perception and manipulation planning. In evaluations, we find that SSL pre-training for touch representation outperforms task and sensor-specific end-to-end training by 95.1% on average over TacBench, and Sparsh (DINO) and Sparsh (IJEPA) are the most competitive, indicating the merits of learning in latent space for tactile images.
Written by
Patrick Lancaster
Michael Kaess
Byron Boots
Mustafa Mukadam
Akash Sharma
Carolina Higuera
Chaithanya Bodduluri
Mike Lambeta
Mrinal Kalakrishnan
Taosha Fan
Tingfan Wu
Publisher
CoRL
Research Topics
Robotics
June 11, 2025
Aaron Foss, Ammar Rizvi, Chloe Evans, Justine T. Kao, Koustuv Sinha, Sasha Mitts
June 11, 2025
June 11, 2025
Mojtaba Komeili, Sarath Chandar, Abha Gejji, Ada Martin, Adrien Bardes, Ammar Rizvi, Artem Zholus, Claire Roberts, Daniel Dugas, David Fan, Francisco Massa, Francois Robert Hogan, Franziska Meier, Kapil Krishnakumar, Koustuv Sinha, Marc Szafraniec, Matthew Muckley, Mido Assran, Michael Rabbat, Nicolas Ballas, Patrick Labatut, Piotr Bojanowski, Quentin Garrido, Russell Howes, Sergio Arnaud, Vasil Khalidov, Xiaodong Ma, Yann LeCun, Yong Li
June 11, 2025
April 17, 2025
Ruslan Partsey, Ayush Jain, Ang Cao, Ishita Prasad, Aravind Rajeswaran, Abha Gejji, Ada Martin, Arjun Majumdar, Daniel Dugas, Franziska Meier, Krishna Murthy Jatavallabhula, Mido Assran, Mikael Henaff, Mike Rabbat, Mrinal Kalakrishnan, Nicolas Ballas, Oleksandr Maksymets, Paul McVay, Phillip Thomas, Alexander Sax, Sergio Arnaud, Vincent-Pierre Berges
April 17, 2025
October 31, 2024
Nolan Black, Romeo Mercado, Norb Tydingco, Gregg Kammerer, Ricardo Chavira, Eric Sanchez, Yitian Ding, Roberto Calandra, Mike Lambeta, Alexander Sohn, Ali Sengül, Byron Taylor, Dave Stroud, Haozhi Qi, Jake Khatha, Jitendra Malik, Kevin Sawyer, Kurt Jenkins, Kyle Most, Neal Stein, Thomas Craven-Bartle, Tingfan Wu, Victoria Rose Most
October 31, 2024

Our approach
Latest news
Foundational models