Papers
Topics
Authors
Recent
Search
2000 character limit reached

Structure-informed Positional Encoding for Music Generation

Published 20 Feb 2024 in cs.SD, cs.AI, and eess.AS | (2402.13301v2)

Abstract: Music generated by deep learning methods often suffers from a lack of coherence and long-term organization. Yet, multi-scale hierarchical structure is a distinctive feature of music signals. To leverage this information, we propose a structure-informed positional encoding framework for music generation with Transformers. We design three variants in terms of absolute, relative and non-stationary positional information. We comprehensively test them on two symbolic music generation tasks: next-timestep prediction and accompaniment generation. As a comparison, we choose multiple baselines from the literature and demonstrate the merits of our methods using several musically-motivated evaluation metrics. In particular, our methods improve the melodic and structural consistency of the generated pieces.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Louis Bigo, Modeling Musical Scores Languages, HDR dissertation, Université de Lille, 2023.
  2. “Attention Is All You Need,” NeurIPS, 2017.
  3. “Pop Music Transformer: Beat-based Modeling and Generation of Expressive Pop Piano Compositions,” ACM International Conference on Multimedia (MM), 2020.
  4. “PopMAG: Pop Music Accompaniment Generation,” MM, 2020.
  5. “A Survey on Deep Learning for Symbolic Music Generation: Representations, Algorithms, Evaluations, and Challenges,” ACM Comput. Surv., vol. 56, no. 1, 2023.
  6. “The Jazz Transformer on the Front Line: Exploring the Shortcomings of AI-composed Music through Quantitative Measures,” ISMIR, 2020.
  7. “Compound Word Transformer: Learning to Compose Full-Song Music over Dynamic Directed Hypergraphs,” AAAI, 2021.
  8. “How Large a Vocabulary Does Text Classification Need? A Variational Approach to Vocabulary Selection,” NAACL, 2019.
  9. “Allocating Large Vocabulary Capacity for Cross-Lingual Language Model Pre-Training,” NAACL, 2021.
  10. “Which transformer architecture fits my data? A vocabulary bottleneck in self-attention,” ICML, 2021.
  11. “Symphony Generation with Permutation Invariant Language Model,” ISMIR, 2022.
  12. “A Domain-Knowledge-Inspired Music Embedding Space and a Novel Attention Mechanism for Symbolic Music Modeling,” AAAI, 2023.
  13. “POP909: A Pop-song Dataset for Music Arrangement Generation,” ISMIR, 2020.
  14. “Self-Attention with Relative Position Representations,” ACL, 2018.
  15. “Efficient Transformers: A Survey,” ACM Comput. Surv., 2022.
  16. “Relative Positional Encoding for Transformers with Linear Complexity,” ICML, 2021.
  17. “Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers,” 2023, preprint arXiv:2302.01925.
  18. Marc G Genton, “Classes of Kernels for Machine Learning: A Statistics Perspective,” JMLR, vol. 2, 2001.
  19. “Curriculum learning,” ICML, 2009.
  20. “Automatic Analysis and Influence of Hierarchical Structure on Melody, Rhythm and Harmony in Popular Music,” Joint Conference on AI Music Creativity (AIMC), 2020.
  21. “Music Transformer: Generating Music with Long-Term Structure,” ICLR, 2019.
  22. “MuseMorphose: Full-Song and Fine-Grained Music Style Transfer with Just One Transformer VAE,” TASLP, 2022.
  23. “Real-time drum accompaniment using transformer architecture,” AIMC, 2022.
  24. “Transformer Dissection: An Unified Understanding for Transformer’s Attention via the Lens of Kernel,” EMNLP-IJCNLP, 2019.
  25. “Transformer Language Models without Positional Encodings Still Learn Positional Information,” EMNLP, 2022.
  26. “The Impact of Positional Encoding on Length Generalization in Transformers,” 2023, preprint arXiv:2305.19466.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 3 likes about this paper.