AI & ML Efficiency Breakthrough

Achieves 100x speedup in robotic action generation by distilling iterative flow/diffusion models into a one-step policy without a pre-trained teacher.

March 16, 2026

Original Paper

One-Step Flow Policy: Self-Distillation for Fast Visuomotor Policies

Shaolong Li, Lichao Sun, Yongchao Chen

arXiv · 2603.12480

The Takeaway

Iterative sampling in diffusion policies creates high latency, making them impractical for real-time, high-frequency robot control. This self-distillation framework enables single-step inference with state-of-the-art accuracy, removing the primary bottleneck for deploying generative policies in robotics.

From the abstract

Generative flow and diffusion models provide the continuous, multimodal action distributions needed for high-precision robotic policies. However, their reliance on iterative sampling introduces severe inference latency, degrading control frequency and harming performance in time-sensitive manipulation. To address this problem, we propose the One-Step Flow Policy (OFP), a from-scratch self-distillation framework for high-fidelity, single-step action generation without a pre-trained teacher. OFP u