Streaming Flow Policy: Simplifying diffusion/flow-matching policies by treating action trajectories as flow trajectories

Sunshine Jiang, Xiaolin Fang, Nicholas Roy, Tomás Lozano-Pérez, Leslie Pack Kaelbling, Siddharth Ancha
Proceedings of The 9th Conference on Robot Learning, PMLR 305:238-257, 2025.

Abstract

Recent advances in diffusion$/$flow-matching policies have enabled imitation learning of complex, multi-modal action trajectories. However, they are computationally expensive because they sample a *trajectory of trajectories*—a diffusion$/$flow trajectory of action trajectories. They discard intermediate action trajectories, and must wait for the sampling process to complete before any actions can be executed on the robot. We simplify diffusion$/$flow policies by *treating action trajectories as flow trajectories*. Instead of starting from pure noise, our algorithm samples from a narrow Gaussian around the last action. Then, it incrementally integrates a velocity field learned via flow matching to produce a sequence of actions that constitute a *single* trajectory. This enables actions to be streamed to the robot on-the-fly *during* the flow sampling process, and is well-suited for receding horizon policy execution. Despite streaming, our method retains the ability to model multi-modal behavior. We train flows that *stabilize* around demonstration trajectories to reduce distribution shift and improve imitation learning performance. Streaming flow policy outperforms prior methods while enabling faster policy execution and tighter sensorimotor loops for learning-based robot control.

Cite this Paper


BibTeX
@InProceedings{pmlr-v305-jiang25a, title = {Streaming Flow Policy: Simplifying diffusion/flow-matching policies by treating action trajectories as flow trajectories}, author = {Jiang, Sunshine and Fang, Xiaolin and Roy, Nicholas and Lozano-P\'{e}rez, Tom\'{a}s and Kaelbling, Leslie Pack and Ancha, Siddharth}, booktitle = {Proceedings of The 9th Conference on Robot Learning}, pages = {238--257}, year = {2025}, editor = {Lim, Joseph and Song, Shuran and Park, Hae-Won}, volume = {305}, series = {Proceedings of Machine Learning Research}, month = {27--30 Sep}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v305/main/assets/jiang25a/jiang25a.pdf}, url = {https://proceedings.mlr.press/v305/jiang25a.html}, abstract = {Recent advances in diffusion$/$flow-matching policies have enabled imitation learning of complex, multi-modal action trajectories. However, they are computationally expensive because they sample a *trajectory of trajectories*—a diffusion$/$flow trajectory of action trajectories. They discard intermediate action trajectories, and must wait for the sampling process to complete before any actions can be executed on the robot. We simplify diffusion$/$flow policies by *treating action trajectories as flow trajectories*. Instead of starting from pure noise, our algorithm samples from a narrow Gaussian around the last action. Then, it incrementally integrates a velocity field learned via flow matching to produce a sequence of actions that constitute a *single* trajectory. This enables actions to be streamed to the robot on-the-fly *during* the flow sampling process, and is well-suited for receding horizon policy execution. Despite streaming, our method retains the ability to model multi-modal behavior. We train flows that *stabilize* around demonstration trajectories to reduce distribution shift and improve imitation learning performance. Streaming flow policy outperforms prior methods while enabling faster policy execution and tighter sensorimotor loops for learning-based robot control.} }
Endnote
%0 Conference Paper %T Streaming Flow Policy: Simplifying diffusion/flow-matching policies by treating action trajectories as flow trajectories %A Sunshine Jiang %A Xiaolin Fang %A Nicholas Roy %A Tomás Lozano-Pérez %A Leslie Pack Kaelbling %A Siddharth Ancha %B Proceedings of The 9th Conference on Robot Learning %C Proceedings of Machine Learning Research %D 2025 %E Joseph Lim %E Shuran Song %E Hae-Won Park %F pmlr-v305-jiang25a %I PMLR %P 238--257 %U https://proceedings.mlr.press/v305/jiang25a.html %V 305 %X Recent advances in diffusion$/$flow-matching policies have enabled imitation learning of complex, multi-modal action trajectories. However, they are computationally expensive because they sample a *trajectory of trajectories*—a diffusion$/$flow trajectory of action trajectories. They discard intermediate action trajectories, and must wait for the sampling process to complete before any actions can be executed on the robot. We simplify diffusion$/$flow policies by *treating action trajectories as flow trajectories*. Instead of starting from pure noise, our algorithm samples from a narrow Gaussian around the last action. Then, it incrementally integrates a velocity field learned via flow matching to produce a sequence of actions that constitute a *single* trajectory. This enables actions to be streamed to the robot on-the-fly *during* the flow sampling process, and is well-suited for receding horizon policy execution. Despite streaming, our method retains the ability to model multi-modal behavior. We train flows that *stabilize* around demonstration trajectories to reduce distribution shift and improve imitation learning performance. Streaming flow policy outperforms prior methods while enabling faster policy execution and tighter sensorimotor loops for learning-based robot control.
APA
Jiang, S., Fang, X., Roy, N., Lozano-Pérez, T., Kaelbling, L.P. & Ancha, S.. (2025). Streaming Flow Policy: Simplifying diffusion/flow-matching policies by treating action trajectories as flow trajectories. Proceedings of The 9th Conference on Robot Learning, in Proceedings of Machine Learning Research 305:238-257 Available from https://proceedings.mlr.press/v305/jiang25a.html.

Related Material