Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The ACCompanion v0.1: An Expressive Accompaniment System (1711.02427v1)

Published 7 Nov 2017 in cs.SD, cs.HC, and eess.AS

Abstract: In this paper we present a preliminary version of the ACCompanion, an expressive accompaniment system for MIDI input. The system uses a probabilistic monophonic score follower to track the position of the soloist in the score, and a linear Gaussian model to compute tempo updates. The expressiveness of the system is powered by the Basis-Mixer, a state-of-the-art computational model of expressive music performance. The system allows for expressive dynamics, timing and articulation.

Citations (6)

Summary

  • The paper introduces an innovative accompaniment system that accurately tracks a soloist using a Hidden Markov Model and a switching Kalman filter.
  • It employs the Basis-Mixer framework with neural networks to predict expressive performance attributes such as timing micro-deviations and loudness trends.
  • The system features a user-friendly GUI that visualizes performance dynamics in real-time, offering immediate feedback on note accuracy and expressiveness.

An Overview of The ACCompanion v0.1: An Expressive Accompaniment System

The paper introduces the ACCompanion v0.1, a development in the domain of computational accompaniment systems, focusing on expressiveness for MIDI input. This expressive accompaniment system integrates a probabilistic score-following mechanism and the Basis-Mixer framework to achieve dynamic, articulated musical accompaniment that interacts with a soloist performance in real-time.

Score Following

The ACCompanion employs a Hidden Markov Model (HMM) based system for monophonic score following. The model's observable variables include performed MIDI pitches and inter-onset intervals (IOIs), while the hidden variable signifies the position within the score. Additionally, the system incorporates a switching Kalman filter to model tempo as a linear Gaussian process. This architecture enables the system to track the soloist's performance accurately, thereby aligning the accompaniment with the soloist's expressive nuances.

The Basis-Mixer: Model for Expressive Performance

Central to the ACCompanion's expressiveness is the Basis-Mixer (BM) framework, which leverages neural networks to emulate the subtleties of human musical expression. The BM framework encodes elements of a musical score into basis functions, predicting performance attributes for each note, such as loudness trends, beat period ratios, loudness deviation, timing micro-deviations, and articulation. These metrics, predicted by bidirectional recurrent and feedforward neural networks, enable nuanced control over the accompaniment dynamics, filling the void left by traditional computational models that often lack expressiveness.

Implementation and Interface

The implementation of the ACCompanion in Python translates the complex backend processes into a user-friendly interface. A graphical user interface (GUI) illustrates the performance in real-time through a visual representation, where the loudness and accuracy of notes are color-coded, providing immediate feedback to the user. This high degree of visual and interactive engagement facilitates experimentation with expressive musical elements.

Implications and Future Developments

While the paper discusses a preliminary model, the foundational aspects of the ACCompanion point towards significant implications in both performance and pedagogy. The ability to produce an expressive accompaniment opens new research avenues in collaborative music-making involving AI, the development of enhanced musical tutors, and real-time performance augmentation.

However, the paper also acknowledges the intrinsic limitations of the current version, pointing out the need for systematic evaluation and the incorporation of more advanced models, particularly regarding polyphonic score following and richer datasets for training the BM framework. The envisioned trajectory for the ACCompanion includes interacting with real ensemble data and utilizing sophisticated model variants, guiding the system towards more authentic and contextually adaptive musical output.

Conclusion

This document outlines a significant step towards addressing the limitations of non-expressive robotic accompaniment systems. The ACCompanion, though initially in its development phase, presents a compelling framework that integrates cutting-edge AI methodologies to accommodate the multifaceted nature of musical performance. Future iterations are expected to yield a more robust and expansive system, potentially setting a new standard in computational music accompaniment technology.

Youtube Logo Streamline Icon: https://streamlinehq.com