SVGformer: Representation Learning for Continuous Vector Graphics Using Transformers

Abstract

Advances in representation learning have led to great success in understanding and generating data in various domains. However, in modeling vector graphics data, the pure data-driven approach often yields unsatisfactory results in downstream tasks as existing deep learning methods often require the quantization of SVG parameters and cannot exploit the geometric properties explicitly. In this paper, we propose a transformer-based representation learning model (SVGformer) that directly operates on continuous input values and manipulates the geometric information of SVG to encode outline details and long-distance dependencies. SVGfomer can be used for various downstream tasks: reconstruction, classification, interpolation, retrieval, etc. We have conducted extensive experiments on vector font and icon datasets to show that our model can capture high-quality representation information and outperform the previous state-of-the-art on downstream tasks significantly.

Cite

Text

Cao et al. "SVGformer: Representation Learning for Continuous Vector Graphics Using Transformers." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00973

Markdown

[Cao et al. "SVGformer: Representation Learning for Continuous Vector Graphics Using Transformers." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/cao2023cvpr-svgformer/) doi:10.1109/CVPR52729.2023.00973

BibTeX

@inproceedings{cao2023cvpr-svgformer,
  title     = {{SVGformer: Representation Learning for Continuous Vector Graphics Using Transformers}},
  author    = {Cao, Defu and Wang, Zhaowen and Echevarria, Jose and Liu, Yan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {10093-10102},
  doi       = {10.1109/CVPR52729.2023.00973},
  url       = {https://mlanthology.org/cvpr/2023/cao2023cvpr-svgformer/}
}