Notochord is a novel deep probabilistic model developed for real-time MIDI performance. It allows interpretable interventions at a sub-event level, facilitating diverse interactive musical functions like generation, harmonization, and improvisation. The model can generate polyphonic and multi-track MIDI with latency below ten milliseconds. Notochord's design emphasizes low-latency processing to ensure imperceptible delays between user actions and system responses. By training on the Lakh MIDI dataset, Notochord offers open-source software for exploring creative AI in real-time music performance settings.
The paper discusses the challenges of integrating deep learning models into performance settings where instantaneous feedback is crucial. It highlights the importance of studying creative AI in musical domains to understand unique interactions between users and intelligent systems. Notochord's design aims to bridge the gap between very low latency requirements in music applications and the diversity of potential use cases it enables.
The content delves into the technical aspects of Notochord, including its autoregressive factorization approach, sub-event order modeling, sub-event distributions using mixture models, and neural network architecture implementation. The paper also presents results from training Notochord on the Lakh MIDI dataset and provides insights into its efficacy in generating musical sequences with varying levels of information available.
Overall, Notochord represents a significant advancement in leveraging deep learning for real-time music performance applications by offering a flexible and interpretable model that responds instantaneously to user inputs.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Victor Shepa... at arxiv.org 03-19-2024
https://arxiv.org/pdf/2403.12000.pdfDeeper Inquiries