RTFormer: Efficient Design for Real-Time Semantic Segmentation with Transformer

Abstract

Recently, transformer-based networks have shown impressive results in semantic segmentation. Yet for real-time semantic segmentation, pure CNN-based approaches still dominate in this field, due to the time-consuming computation mechanism of transformer. We propose RTFormer, an efficient dual-resolution transformer for real-time semantic segmenation, which achieves better trade-off between performance and efficiency than CNN-based models. To achieve high inference efficiency on GPU-like devices, our RTFormer leverages GPU-Friendly Attention with linear complexity and discards the multi-head mechanism. Besides, we find that cross-resolution attention is more efficient to gather global context information for high-resolution branch by spreading the high level knowledge learned from low-resolution branch. Extensive experiments on mainstream benchmarks demonstrate the effectiveness of our proposed RTFormer, it achieves state-of-the-art on Cityscapes, CamVid and COCOStuff, and shows promising results on ADE20K.

Cite

Text

Wang et al. "RTFormer: Efficient Design for Real-Time Semantic Segmentation with Transformer." Neural Information Processing Systems, 2022.

Markdown

[Wang et al. "RTFormer: Efficient Design for Real-Time Semantic Segmentation with Transformer." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/wang2022neurips-rtformer/)

BibTeX

@inproceedings{wang2022neurips-rtformer,
  title     = {{RTFormer: Efficient Design for Real-Time Semantic Segmentation with Transformer}},
  author    = {Wang, Jian and Gou, Chenhui and Wu, Qiman and Feng, Haocheng and Han, Junyu and Ding, Errui and Wang, Jingdong},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/wang2022neurips-rtformer/}
}