Tradformer: A Transformer Model of Traditional Music Transcriptions
Abstract
We explore the transformer neural network architecture for modeling music, specifically Irish and Swedish traditional dance music. Given the repetitive structures of these kinds of music, the transformer should be as successful with fewer parameters and complexity as the hitherto most successful model, a vanilla long short-term memory network. We find that achieving good performance with the transformer is not straightforward, and careful consideration is needed for the sampling strategy, evaluating intermediate outputs in relation to engineering choices, and finally analyzing what the model learns. We discuss these points with several illustrations, providing reusable insights for engineering other music generation systems. We also report the high performance of our final transformer model in a competition of music generation systems focused on a type of Swedish dance.
Cite
Text
Casini and Sturm. "Tradformer: A Transformer Model of Traditional Music Transcriptions." International Joint Conference on Artificial Intelligence, 2022. doi:10.24963/IJCAI.2022/681Markdown
[Casini and Sturm. "Tradformer: A Transformer Model of Traditional Music Transcriptions." International Joint Conference on Artificial Intelligence, 2022.](https://mlanthology.org/ijcai/2022/casini2022ijcai-tradformer/) doi:10.24963/IJCAI.2022/681BibTeX
@inproceedings{casini2022ijcai-tradformer,
title = {{Tradformer: A Transformer Model of Traditional Music Transcriptions}},
author = {Casini, Luca and Sturm, Bob L. T.},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2022},
pages = {4915-4920},
doi = {10.24963/IJCAI.2022/681},
url = {https://mlanthology.org/ijcai/2022/casini2022ijcai-tradformer/}
}