Tradformer: A Transformer Model of Traditional Music Transcriptions

Abstract

We explore the transformer neural network architecture for modeling music, specifically Irish and Swedish traditional dance music. Given the repetitive structures of these kinds of music, the transformer should be as successful with fewer parameters and complexity as the hitherto most successful model, a vanilla long short-term memory network. We find that achieving good performance with the transformer is not straightforward, and careful consideration is needed for the sampling strategy, evaluating intermediate outputs in relation to engineering choices, and finally analyzing what the model learns. We discuss these points with several illustrations, providing reusable insights for engineering other music generation systems. We also report the high performance of our final transformer model in a competition of music generation systems focused on a type of Swedish dance.

Cite

Text

Casini and Sturm. "Tradformer: A Transformer Model of Traditional Music Transcriptions." International Joint Conference on Artificial Intelligence, 2022. doi:10.24963/IJCAI.2022/681

Markdown

[Casini and Sturm. "Tradformer: A Transformer Model of Traditional Music Transcriptions." International Joint Conference on Artificial Intelligence, 2022.](https://mlanthology.org/ijcai/2022/casini2022ijcai-tradformer/) doi:10.24963/IJCAI.2022/681

BibTeX

@inproceedings{casini2022ijcai-tradformer,
  title     = {{Tradformer: A Transformer Model of Traditional Music Transcriptions}},
  author    = {Casini, Luca and Sturm, Bob L. T.},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2022},
  pages     = {4915-4920},
  doi       = {10.24963/IJCAI.2022/681},
  url       = {https://mlanthology.org/ijcai/2022/casini2022ijcai-tradformer/}
}