TrOCR: Transformer-Based Optical Character Recognition with Pre-Trained Models

Abstract

Text recognition is a long-standing research problem for document digitalization. Existing approaches are usually built based on CNN for image understanding and RNN for char-level text generation. In addition, another language model is usually needed to improve the overall accuracy as a post-processing step. In this paper, we propose an end-to-end text recognition approach with pre-trained image Transformer and text Transformer models, namely TrOCR, which leverages the Transformer architecture for both image understanding and wordpiece-level text generation. The TrOCR model is simple but effective, and can be pre-trained with large-scale synthetic data and fine-tuned with human-labeled datasets. Experiments show that the TrOCR model outperforms the current state-of-the-art models on the printed, handwritten and scene text recognition tasks. The TrOCR models and code are publicly available at https://aka.ms/trocr.

Cite

Text

Li et al. "TrOCR: Transformer-Based Optical Character Recognition with Pre-Trained Models." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I11.26538

Markdown

[Li et al. "TrOCR: Transformer-Based Optical Character Recognition with Pre-Trained Models." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/li2023aaai-trocr/) doi:10.1609/AAAI.V37I11.26538

BibTeX

@inproceedings{li2023aaai-trocr,
  title     = {{TrOCR: Transformer-Based Optical Character Recognition with Pre-Trained Models}},
  author    = {Li, Minghao and Lv, Tengchao and Chen, Jingye and Cui, Lei and Lu, Yijuan and Florêncio, Dinei A. F. and Zhang, Cha and Li, Zhoujun and Wei, Furu},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {13094-13102},
  doi       = {10.1609/AAAI.V37I11.26538},
  url       = {https://mlanthology.org/aaai/2023/li2023aaai-trocr/}
}