AI & ML Practical Magic

If a robot touches an object in just a few spots, it can 'hallucinate' the rest of the shape so accurately it's like it has X-ray vision.

April 13, 2026

Original Paper

TouchAnything: Diffusion-Guided 3D Reconstruction from Sparse Robot Touches

Langzhe Gu, Hung-Jui Huang, Mohamad Qadri, Michael Kaess, Wenzhen Yuan

arXiv · 2604.08945

The Takeaway

This research allows robots to reconstruct objects they can't see by transferring visual knowledge from diffusion models to the sense of touch. It's a major step toward robots that can navigate and manipulate objects in the dark or in cluttered spaces.

From the abstract

Accurate object geometry estimation is essential for many downstream tasks, including robotic manipulation and physical interaction. Although vision is the dominant modality for shape perception, it becomes unreliable under occlusions or challenging lighting conditions. In such scenarios, tactile sensing provides direct geometric information through physical contact. However, reconstructing global 3D geometry from sparse local touches alone is fundamentally underconstrained. We present TouchAnyt