Quantization-Free Autoregressive Action Transformer

Abstract

Current transformer-based imitation learning approaches introduce discrete action representations and train an autoregressive transformer decoder on the resulting latent code. However, the initial quantization breaks the continuous structure of the action space thereby limiting the capabilities of the generative model. We propose a quantization-free method instead that leverages Generative Infinite-Vocabulary Transformers (GIVT) as a direct, continuous policy parametrization for autoregressive transformers. This simplifies the imitation learning pipeline while achieving state-of-the-art performance on a variety of popular simulated robotics tasks. We enhance our policy roll-outs by carefully studying sampling algorithms, further improving the results.

Cite

Text

Sheebaelhamd et al. "Quantization-Free Autoregressive Action Transformer." Advances in Neural Information Processing Systems, 2025.

Markdown

[Sheebaelhamd et al. "Quantization-Free Autoregressive Action Transformer." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/sheebaelhamd2025neurips-quantizationfree/)

BibTeX

@inproceedings{sheebaelhamd2025neurips-quantizationfree,
  title     = {{Quantization-Free Autoregressive Action Transformer}},
  author    = {Sheebaelhamd, Ziyad and Tschannen, Michael and Muehlebach, Michael and Vernade, Claire},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/sheebaelhamd2025neurips-quantizationfree/}
}