Imaginative and prescient-language-action fashions are the present cutting-edge in robotic manipulation. They nonetheless can not decide up a potato chip with out crushing it.
That’s the consequence printed earlier this 12 months by the staff behind the Video Tactile Motion Mannequin (VTAM). On a potato chip pick-and-place process — a process that calls for high-fidelity power consciousness, the place imaginative and prescient alone can not distinguish a crushing grasp from a holding one — VTAM outperformed the π0.5 baseline by 80%. Throughout the broader contact-rich benchmark suite, VTAM held a 90% common success price.¹
The chip is an adversarial instance, and that’s exactly why it’s the proper check. On the level of grasp, solely contact dynamics carry helpful indicators. Strain, vibration, and power/torque inform the coverage what is going on, correcting the visible estimation errors that vision-only fashions can not detect on their very own. A digital camera, nevertheless excessive its decision, can not try this work.
Tactile isn’t plug-and-play
Tactile sensors don’t enhance mannequin efficiency on their very own. Most studying pipelines immediately are constructed round imaginative and prescient and language; the 2 modalities with the most important datasets and probably the most mature architectures behind them. When tactile indicators are appended to a vision-first pipeline with out intentional design, they have an inclination to get downweighted, drowned out, or misplaced in coaching. VTAM works as a result of the structure forces the mannequin to forecast imaginative and prescient and tactile dynamics collectively, so the tactile sign immediately shapes the realized coverage moderately than getting absorbed into imaginative and prescient and language. Tactile information solely delivers its worth when it’s intelligently included.
The sample is now constant throughout the literature
The chip is one finish of the spectrum, a case the place imaginative and prescient fails outright and tactile carries the sign alone. Most real-world duties sit additional alongside that spectrum, the place imaginative and prescient and tactile every contribute and the synergy between them is what drives coaching effectivity. The sample is now constant throughout the literature.
VTAM isn’t alone. The ManiSkill-ViTac 2025 benchmark formalises tactile-augmented analysis throughout insertion, device use, and precision meeting duties. Impartial analysis on tactile sensor configurations and grasp studying efficiency² reveals the identical carry. Insurance policies that mix imaginative and prescient with tactile suggestions constantly outperform vision-only equivalents on contact-rich duties, and have a tendency to achieve the identical success threshold from fewer demonstrations.
Failure detection is the second prize
A tactile-conditioned coverage registers incipient slip as a vibration signature tens to a whole lot of milliseconds earlier than the thing really strikes. That window is the distinction between re-grasping and a full restart — between 95% and 99% uptime on the identical line. Throughout a fleet, the operational case turns into onerous to disregard.
Failure detection is one case of a bigger functionality: producing correct, high-resolution labels for what really occurred throughout the grasp. A binary success/fail label collapses data that the coaching pipeline might use. Did the grasp succeed cleanly, or did it succeed with inner slippage that the controller recovered from? Did the thing settle stably, or did it shift throughout transport? Tactile sensing can distinguish these instances, and embedded contact notion can label them on-device, turning each episode right into a extra informative coaching instance, not simply the failed ones.

Determine 1. VTAM combines a language mannequin, a predictive vision-tactile world mannequin, and a diffusion-based motion coverage. From simply 10 minutes of teleoperation per process, it learns to foretell future actions, states, and forces — enabling contact-rich duties reminiscent of chip pick-and-place, dynamic wiping, and steady peeling. Supply: arXiv:2603.23481.
What this implies for builders
Tactile sensing has moved from helpful addition to defensible requirement for any staff aiming at production-grade contact-rich manipulation. The query is not whether or not to instrument. It’s whether or not to instrument now, or pay later in rebuilt datasets and recalibrated fashions.
VTAM put an actual quantity on the case and different current work retains pointing in the identical route. The following era of basis fashions will probably be constructed on information that captures contact moderately than vision-only.
Able to take the subsequent step?
Discuss to our technical staff about tactile integration to your manipulation pipeline and be taught extra about how Robotiq can allow your software.
¹ Video Tactile Motion Mannequin (VTAM), arXiv:2603.23481.
² Consultant findings embody Tactile Robotics: An Outlook (arXiv) and printed work on the affect of tactile sensor configurations on grasp studying effectivity.

