Processing math: 100%
Skip to yearly menu bar Skip to main content


Poster

MuPT: A Generative Symbolic Music Pretrained Transformer

Xingwei Qu · yuelin bai · Yinghao MA · Ziya Zhou · Ka Man Lo · Jiaheng Liu · Ruibin Yuan · Lejun Min · Xueling Liu · Tianyu Zhang · Xeron Du · Shuyue Guo · Yiming Liang · Yizhi Li · Shangda Wu · Junting Zhou · Tianyu Zheng · Ziyang Ma · Fengze Han · Wei Xue · Gus Xia · Emmanouil Benetos · Xiang Yue · Chenghua Lin · Xu Tan · Wenhao Huang · Jie Fu · Ge Zhang

Hall 3 + Hall 2B #43
[ ] [ Project Page ]
Thu 24 Apr midnight PDT — 2:30 a.m. PDT

Abstract: In this paper, we explore the application of Large Language Models (LLMs) to the pre-training of music. While the prevalent use of MIDI in music modeling is well-established, our findings suggest that LLMs are inherently more compatible with ABC Notation, which aligns more closely with their design and strengths, thereby enhancing the model's performance in musical composition.To address the challenges associated with misaligned measures from different tracks during generation, we propose the development of a S_ynchronized M_ulti-T_rack ABC Notation (SMT-ABC Notation), which aims to preserve coherence across multiple musical tracks. Our contributions include a series of models capable of handling up to 8192 tokens, covering 90\% of the symbolic music data in our training set. Furthermore, we explore the implications of the S_ymbolic M_usic S_caling Law (SMS Law) on model performance. The results indicate a promising research direction in music generation, offering extensive resources for further research through our open-source contributions.

Live content is unavailable. Log in and register to view live content