Imaginative and prescient-language-action fashions are the present cutting-edge in robotic manipulation. They nonetheless can’t choose up a potato chip with out crushing it.
That’s the consequence printed earlier this 12 months by the workforce behind the Video Tactile Motion Mannequin (VTAM). On a potato chip pick-and-place activity — a activity that calls for high-fidelity drive consciousness, the place imaginative and prescient alone can’t distinguish a crushing grasp from a holding one — VTAM outperformed the π0.5 baseline by 80%. Throughout the broader contact-rich benchmark suite, VTAM held a 90% common success charge.¹
The chip is an adversarial instance, and that’s exactly why it’s the proper check. On the level of grasp, solely contact dynamics carry helpful alerts. Strain, vibration, and drive/torque inform the coverage what is occurring, correcting the visible estimation errors that vision-only fashions can’t detect on their very own. A digital camera, nevertheless excessive its decision, can’t do this work.
Tactile isn’t plug-and-play
Tactile sensors don’t enhance mannequin efficiency on their very own. Most studying pipelines at the moment are constructed round imaginative and prescient and language; the 2 modalities with the most important datasets and probably the most mature architectures behind them. When tactile alerts are appended to a vision-first pipeline with out intentional design, they have a tendency to get downweighted, drowned out, or misplaced in coaching. VTAM works as a result of the structure forces the mannequin to forecast imaginative and prescient and tactile dynamics collectively, so the tactile sign immediately shapes the realized coverage fairly than getting absorbed into imaginative and prescient and language. Tactile knowledge solely delivers its worth when it’s intelligently included.
The sample is now constant throughout the literature
The chip is one finish of the spectrum, a case the place imaginative and prescient fails outright and tactile carries the sign alone. Most real-world duties sit additional alongside that spectrum, the place imaginative and prescient and tactile every contribute and the synergy between them is what drives coaching effectivity. The sample is now constant throughout the literature.
VTAM isn’t alone. The ManiSkill-ViTac 2025 benchmark formalises tactile-augmented analysis throughout insertion, software use, and precision meeting duties. Impartial analysis on tactile sensor configurations and grasp studying efficiency² reveals the identical raise. Insurance policies that mix imaginative and prescient with tactile suggestions persistently outperform vision-only equivalents on contact-rich duties, and have a tendency to achieve the identical success threshold from fewer demonstrations.
Failure detection is the second prize
A tactile-conditioned coverage registers incipient slip as a vibration signature tens to a whole lot of milliseconds earlier than the thing really strikes. That window is the distinction between re-grasping and a full restart — between 95% and 99% uptime on the identical line. Throughout a fleet, the operational case turns into arduous to disregard.
Failure detection is one case of a bigger functionality: producing correct, high-resolution labels for what really occurred in the course of the grasp. A binary success/fail label collapses data that the coaching pipeline might use. Did the grasp succeed cleanly, or did it succeed with inner slippage that the controller recovered from? Did the thing settle stably, or did it shift throughout transport? Tactile sensing can distinguish these circumstances, and embedded contact notion can label them on-device, turning each episode right into a extra informative coaching instance, not simply the failed ones.

Determine 1. VTAM combines a language mannequin, a predictive vision-tactile world mannequin, and a diffusion-based motion coverage. From simply 10 minutes of teleoperation per activity, it learns to foretell future actions, states, and forces — enabling contact-rich duties equivalent to chip pick-and-place, dynamic wiping, and secure peeling. Supply: arXiv:2603.23481.
What this implies for builders
Tactile sensing has moved from helpful addition to defensible requirement for any workforce aiming at production-grade contact-rich manipulation. The query is not whether or not to instrument. It’s whether or not to instrument now, or pay later in rebuilt datasets and recalibrated fashions.
VTAM put an actual quantity on the case and different current work retains pointing in the identical course. The subsequent era of basis fashions can be constructed on knowledge that captures contact fairly than vision-only.
Able to take the following step?
Speak to our technical workforce about tactile integration in your manipulation pipeline and study extra about how Robotiq can allow your utility.
¹ Video Tactile Motion Mannequin (VTAM), arXiv:2603.23481.
² Consultant findings embody Tactile Robotics: An Outlook (arXiv) and printed work on the affect of tactile sensor configurations on grasp studying effectivity.
