Papers
Topics
Authors
Recent
2000 character limit reached

Structure-informed Positional Encoding for Music Generation

Published 20 Feb 2024 in cs.SD, cs.AI, and eess.AS | (2402.13301v2)

Abstract: Music generated by deep learning methods often suffers from a lack of coherence and long-term organization. Yet, multi-scale hierarchical structure is a distinctive feature of music signals. To leverage this information, we propose a structure-informed positional encoding framework for music generation with Transformers. We design three variants in terms of absolute, relative and non-stationary positional information. We comprehensively test them on two symbolic music generation tasks: next-timestep prediction and accompaniment generation. As a comparison, we choose multiple baselines from the literature and demonstrate the merits of our methods using several musically-motivated evaluation metrics. In particular, our methods improve the melodic and structural consistency of the generated pieces.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Louis Bigo, Modeling Musical Scores Languages, HDR dissertation, UniversitƩ de Lille, 2023.
  2. ā€œAttention Is All You Need,ā€ NeurIPS, 2017.
  3. ā€œPop Music Transformer: Beat-based Modeling and Generation of Expressive Pop Piano Compositions,ā€ ACM International Conference on Multimedia (MM), 2020.
  4. ā€œPopMAG: Pop Music Accompaniment Generation,ā€ MM, 2020.
  5. ā€œA Survey on Deep Learning for Symbolic Music Generation: Representations, Algorithms, Evaluations, and Challenges,ā€ ACM Comput. Surv., vol. 56, no. 1, 2023.
  6. ā€œThe Jazz Transformer on the Front Line: Exploring the Shortcomings of AI-composed Music through Quantitative Measures,ā€ ISMIR, 2020.
  7. ā€œCompound Word Transformer: Learning to Compose Full-Song Music over Dynamic Directed Hypergraphs,ā€ AAAI, 2021.
  8. ā€œHow Large a Vocabulary Does Text Classification Need? A Variational Approach to Vocabulary Selection,ā€ NAACL, 2019.
  9. ā€œAllocating Large Vocabulary Capacity for Cross-Lingual Language Model Pre-Training,ā€ NAACL, 2021.
  10. ā€œWhich transformer architecture fits my data? A vocabulary bottleneck in self-attention,ā€ ICML, 2021.
  11. ā€œSymphony Generation with Permutation Invariant Language Model,ā€ ISMIR, 2022.
  12. ā€œA Domain-Knowledge-Inspired Music Embedding Space and a Novel Attention Mechanism for Symbolic Music Modeling,ā€ AAAI, 2023.
  13. ā€œPOP909: A Pop-song Dataset for Music Arrangement Generation,ā€ ISMIR, 2020.
  14. ā€œSelf-Attention with Relative Position Representations,ā€ ACL, 2018.
  15. ā€œEfficient Transformers: A Survey,ā€ ACM Comput. Surv., 2022.
  16. ā€œRelative Positional Encoding for Transformers with Linear Complexity,ā€ ICML, 2021.
  17. ā€œLearning a Fourier Transform for Linear Relative Positional Encodings in Transformers,ā€ 2023, preprint arXiv:2302.01925.
  18. MarcĀ G Genton, ā€œClasses of Kernels for Machine Learning: A Statistics Perspective,ā€ JMLR, vol. 2, 2001.
  19. ā€œCurriculum learning,ā€ ICML, 2009.
  20. ā€œAutomatic Analysis and Influence of Hierarchical Structure on Melody, Rhythm and Harmony in Popular Music,ā€ Joint Conference on AI Music Creativity (AIMC), 2020.
  21. ā€œMusic Transformer: Generating Music with Long-Term Structure,ā€ ICLR, 2019.
  22. ā€œMuseMorphose: Full-Song and Fine-Grained Music Style Transfer with Just One Transformer VAE,ā€ TASLP, 2022.
  23. ā€œReal-time drum accompaniment using transformer architecture,ā€ AIMC, 2022.
  24. ā€œTransformer Dissection: An Unified Understanding for Transformer’s Attention via the Lens of Kernel,ā€ EMNLP-IJCNLP, 2019.
  25. ā€œTransformer Language Models without Positional Encodings Still Learn Positional Information,ā€ EMNLP, 2022.
  26. ā€œThe Impact of Positional Encoding on Length Generalization in Transformers,ā€ 2023, preprint arXiv:2305.19466.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 3 likes about this paper.