Document Type
Article
Publication Title
arXiv
Abstract
Music contains hierarchical structures beyond beats and measures. While hierarchical structure annotations are helpful for music information retrieval and computer musicology, such annotations are scarce in current digital music databases. In this paper, we explore a data-driven approach to automatically extract hierarchical metrical structures from scores. We propose a new model with a Temporal Convolutional Network-Conditional Random Field (TCN-CRF) architecture. Given a symbolic music score, our model takes in an arbitrary number of voices in a beat-quantized form, and predicts a 4-level hierarchical metrical structure from downbeat-level to section-level. We also annotate a dataset using RWC-POP MIDI files to facilitate training and evaluation. We show by experiments that the proposed method performs better than the rule-based approach under different orchestration settings. We also perform some simple musicological analysis on the model predictions. All demos, datasets and pre-trained models are publicly available on Github 1 © 2022, CC BY.
DOI
10.48550/arXiv.2209.10259
Publication Date
9-21-2022
Keywords
Audio and Speech Processing (eess.AS), Machine Learning (cs.LG), Sound (cs.SD)
Recommended Citation
J. Jiang, D. Chin, Y. Zhang, and G. Xia, "Learning Hierarchical Metrical Structure Beyond Measures", 2022, doi: 10.48550/arXiv.2209.10259
Comments
Preprint: arXiv
Archived with thanks to arXiv
Preprint License: CC by 4.0
Uploaded 31 October 2022