MuPT: A Generative Symbolic Music Pretrained Transformer
Abstract
In this paper, we explore the application of Large Language Models (LLMs) to the pre-training of music. While the prevalent use of MIDI in music modeling is well-established, our findings suggest that LLMs are inherently more compatible with ABC Notation, which aligns more closely with their design and strengths, thereby enhancing the model's performance in musical composition. To address the challenges associated with misaligned measures from different tracks during generation, we propose the development of a $\underline{S}$ynchronized $\underline{M}$ulti-$\underline{T}$rack ABC Notation ($\textbf{SMT-ABC Notation}$), which aims to preserve coherence across multiple musical tracks. Our contributions include a series of models capable of handling up to 8192 tokens, covering 90\% of the symbolic music data in our training set. Furthermore, we explore the implications of the $\underline{S}$ymbolic $\underline{M}$usic $\underline{S}$caling Law ($\textbf{SMS Law}$) on model performance. The results indicate a promising research direction in music generation, offering extensive resources for further research through our open-source contributions.
Cite
Text
Qu et al. "MuPT: A Generative Symbolic Music Pretrained Transformer." International Conference on Learning Representations, 2025.Markdown
[Qu et al. "MuPT: A Generative Symbolic Music Pretrained Transformer." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/qu2025iclr-mupt/)BibTeX
@inproceedings{qu2025iclr-mupt,
title = {{MuPT: A Generative Symbolic Music Pretrained Transformer}},
author = {Qu, Xingwei and Bai, Yuelin and Ma, Yinghao and Zhou, Ziya and Lo, Ka Man and Liu, Jiaheng and Yuan, Ruibin and Min, Lejun and Liu, Xueling and Zhang, Tianyu and Du, Xeron and Guo, Shuyue and Liang, Yiming and Li, Yizhi and Wu, Shangda and Zhou, Junting and Zheng, Tianyu and Ma, Ziyang and Han, Fengze and Xue, Wei and Xia, Gus and Benetos, Emmanouil and Yue, Xiang and Lin, Chenghua and Tan, Xu and Huang, Wenhao and Fu, Jie and Zhang, Ge},
booktitle = {International Conference on Learning Representations},
year = {2025},
url = {https://mlanthology.org/iclr/2025/qu2025iclr-mupt/}
}