Modular Quantization-Aware Training for 6d Object Pose Estimation

Abstract

Edge applications, such as collaborative robotics and spacecraft rendezvous, demand efficient 6D object pose estimation on resource-constrained embedded platforms. Existing 6D object pose estimation networks are often too large for such deployments, necessitating compression while maintaining reliable performance. To address this challenge, we introduce Modular Quantization-Aware Training (MQAT), an adaptive and mixed-precision quantization-aware training strategy that exploits the modular structure of modern 6D object pose estimation architectures. MQAT guides a systematic gradated modular quantization sequence and determines module-specific bit precisions, leading to quantized models that outperform those produced by state-of-the-art uniform and mixed-precision quantization techniques. Our experiments showcase the generality of MQAT across datasets, architectures, and quantization algorithms. Additionally, we observe that MQAT quantized models can achieve an accuracy boost (>7% ADI-0.1d) over the baseline full-precision network while reducing model size by a factor of 4x or more. Project Page: https://saqibjaved1.github.io/MQAT_

Cite

Text

Javed et al. "Modular Quantization-Aware Training for 6d Object Pose Estimation." Transactions on Machine Learning Research, 2024.

Markdown

[Javed et al. "Modular Quantization-Aware Training for 6d Object Pose Estimation." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/javed2024tmlr-modular/)

BibTeX

@article{javed2024tmlr-modular,
  title     = {{Modular Quantization-Aware Training for 6d Object Pose Estimation}},
  author    = {Javed, Saqib and Li, Chengkun and Price, Andrew Lawrence and Hu, Yinlin and Salzmann, Mathieu},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/javed2024tmlr-modular/}
}