Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Symphony Generation with Permutation Invariant Language Model (2205.05448v2)

Published 10 May 2022 in cs.SD, cs.AI, cs.LG, and eess.AS

Abstract: In this work, we propose a permutation invariant LLM, SymphonyNet, as a solution for symbolic symphony music generation. We propose a novel Multi-track Multi-instrument Repeatable (MMR) representation for symphonic music and model the music sequence using a Transformer-based auto-regressive LLM with specific 3-D positional embedding. To overcome length overflow when modeling extra-long symphony tokens, we also propose a modified Byte Pair Encoding algorithm (Music BPE) for music tokens and introduce a novel linear transformer decoder architecture as a backbone. Meanwhile, we train the decoder to learn automatic orchestration as a joint task by masking instrument information from the input. We also introduce a large-scale symbolic symphony dataset for the advance of symphony generation research. Empirical results show that the proposed approach can generate coherent, novel, complex and harmonious symphony as a pioneer solution for multi-track multi-instrument symbolic music generation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jiafeng Liu (9 papers)
  2. Yuanliang Dong (3 papers)
  3. Zehua Cheng (10 papers)
  4. Xinran Zhang (28 papers)
  5. Xiaobing Li (27 papers)
  6. Feng Yu (58 papers)
  7. Maosong Sun (337 papers)
Citations (33)

Summary

We haven't generated a summary for this paper yet.