Dense Policy: Bidirectional Autoregressive Learning of Actions

Abstract

Mainstream visuomotor policies predominantly rely on generative models for holistic action prediction, while current autoregressive policies, predicting the next token or chunk, have shown suboptimal results. This motivates a search for more effective learning methods to unleash the potential of autoregressive policies for robotic manipulation. This paper introduces a bidirectionally expanded learning approach, termed Dense Policy, to establish a new paradigm for autoregressive policies in action prediction. It employs a lightweight encoder-only architecture to iteratively unfold the action sequence from an initial single frame into the target sequence in a coarse-to-fine manner with logarithmic-time inference. Extensive experiments validate that our dense policy has superior autoregressive learning capabilities and can surpass existing holistic generative policies. Our policy, example data, and training code will be publicly available upon publication.

Cite

Text

Su et al. "Dense Policy: Bidirectional Autoregressive Learning of Actions." International Conference on Computer Vision, 2025.

Markdown

[Su et al. "Dense Policy: Bidirectional Autoregressive Learning of Actions." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/su2025iccv-dense/)

BibTeX

@inproceedings{su2025iccv-dense,
  title     = {{Dense Policy: Bidirectional Autoregressive Learning of Actions}},
  author    = {Su, Yue and Zhan, Xinyu and Fang, Hongjie and Xue, Han and Fang, Hao-Shu and Li, Yong-Lu and Lu, Cewu and Yang, Lixin},
  booktitle = {International Conference on Computer Vision},
  year      = {2025},
  pages     = {14486-14495},
  url       = {https://mlanthology.org/iccv/2025/su2025iccv-dense/}
}